249
Systems Research Institute Polish Academy of Sciences Victor B. Taylor, MSc Ph.D. Thesis HeBIS: A Biologically Inspired Data Classification System Supervisor: Prof. Dr. hab. Inż. Janusz Kacprzyk Afiliacja Systems Research Institute Polish Academy of Sciences Warsaw, September 2010

 · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

  • Upload
    others

  • View
    3

  • Download
    0

Embed Size (px)

Citation preview

Page 1:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Systems Research Institute Polish Academy of Sciences

Victor B. Taylor, MSc

Ph.D. Thesis

HeBIS: A Biologically Inspired Data Classification System

Supervisor: Prof. Dr. hab. Inż. Janusz Kacprzyk Afiliacja

Systems Research Institute

Polish Academy of Sciences

Warsaw, September 2010

Page 2:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 2 of 249

09/28/10

Page 3:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 3 of 249

09/28/10

Page 4:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 4 of 249

09/28/10

ABSTRACT

This research develops and investigates an algorithm for the self-organized development of a

classification network. This idea for this classification network, known as HeBIS (Heterogeneous

Biologically Inspired System), is based on a heterogeneous mixture of intelligent yet simple

processing units (cells) that can potentially consist of several types of machine learning constructs. A

list of these constructs can include self-organizing feature maps (SOFM), artificial neural networks

(ANN), and support vector machines (SVM) that communicate with each other via the diffusion of

artificial proteins. The context for the self-organization of the network and the communication

between the processing cells is that of an artificial genetic regulatory network (GRN). An evolved

GRN based on an artificial chemistry of simulated proteins is used as the controller. Artificial genes

within each processing cell are essentially excitatory and inhibitory switches that control the

concentration and diffusion of artificial proteins throughout a simulated environmental lattice. This

GRN controls both the growth of the classification network and the specific behaviors of the

individual processing cells. These controls also use artificial chemistry analogs of problem

descriptors such as second-order statistics.

Self-organization and evolution of the network occur on several levels: the high-level topology of the

network as well as parameters and behaviors that affect the internal organization of each processing

cell. The artificial proteins used for communications and the transfer of regulatory information

between and within the processing elements are also evolved as are the environmental proteins used

to represent the input feature vector for the set of training and test exemplars. An evolutionary

process incorporating particle swarm optimization is used to define an artificial genome that defines

these elements as well as the classification information that the network presents to the user.

Behaviors such as input/output signal conditioning, machine learning processing, and

environmental/regulatory communications are evolved and are used as the genome’s input to the

high-level evolutionary process.

The HeBIS algorithm architecture is discussed in detail and extensions for future research are

proposed. Performance of a classification network based on this novel technique with a single type of

cellular machine learning element, a SOFM, is examined. This performance is compared with that of

a baseline standalone SOFM. For this case study, the problem considered is how well HeBIS learns

the empirical algorithm for cloud/no-cloud pixel detection that is used by the National Aeronautics

Page 5:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 5 of 249

09/28/10

and Space Administration’s (NASA) for its multispectral optical datasets acquired from the Moderate

Resolution Imaging Spectroradiometer (MODIS) sensor on the earth-orbiting Aqua satellite.

Page 6:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 6 of 249

09/28/10

ACKNOWLEDGMENTS

Page 7:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 7 of 249

09/28/10

DECLARATION

Page 8:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 8 of 249

09/28/10

Table of Contents

ABSTRACT .......................................................................................................................................................................... 4

ACKNOWLEDGMENTS .................................................................................................................................................... 6

DECLARATION .................................................................................................................................................................. 7

TABLE OF CONTENTS ..................................................................................................................................................... 8

FIGURES ............................................................................................................................................................................ 11

TABLES .............................................................................................................................................................................. 16

TERMS AND ACRONYMS .............................................................................................................................................. 18

1. INTRODUCTION ...................................................................................................................................................... 19

1.1. PROBLEM STATEMENT ........................................................................................................................................... 20 1.2. DELIMITATIONS OF THE RESEARCH........................................................................................................................ 21 1.3. KEY CONTRIBUTIONS ............................................................................................................................................ 21 1.4. ORGANIZATION OF THIS THESIS ............................................................................................................................. 21

2. LITERATURE REVIEW .......................................................................................................................................... 24

2.1. MACHINE LEARNING AND SELF-ORGANIZATION .................................................................................................... 25 2.1.1. Classification overview ................................................................................................................................. 25 2.1.2. Artificial neural networks ............................................................................................................................. 28 2.1.3. Self-organizing feature maps ........................................................................................................................ 34

2.2. EVOLUTIONARY COMPUTATION............................................................................................................................. 40 2.2.1. Particle swarm optimization ......................................................................................................................... 40

2.3. BIOLOGICAL AND ARTIFICIAL EVOLUTIONARY DEVELOPMENT .............................................................................. 43 2.4. SUMMARY ............................................................................................................................................................. 51

3. HETEROGENEOUS BIOLOGICALLY INSPIRED SYSTEM (HEBIS) ............................................................ 54

3.1. OVERVIEW ............................................................................................................................................................ 54 3.2. FUNDAMENTALS .................................................................................................................................................... 55

3.2.1. Processing cell .............................................................................................................................................. 55 3.2.2. Environment.................................................................................................................................................. 57

3.2.2.1. 1-D environment ........................................................................................................................................................ 58 3.1.1.1. 2-D environment ........................................................................................................................................................ 58 3.1.1.2. 3-D environment ........................................................................................................................................................ 59

3.1.2. Genetic regulatory network .......................................................................................................................... 59 3.1.2.1. Gene coding ............................................................................................................................................................... 60 3.1.2.2. Protein communications ............................................................................................................................................. 64

3.1.3. Basic cell processing .................................................................................................................................... 66 3.1.3.1. Cell genome ............................................................................................................................................................... 66 3.1.3.2. Intrinsic behaviors ...................................................................................................................................................... 67

3.1.3.2.1. NumberProteinsInCell ......................................................................................................................................... 67 3.1.3.2.2. NumberProteinsInLocalEnviro ............................................................................................................................ 67 3.1.3.2.3. ConcentrationStandardDeviationLocalEnviro ..................................................................................................... 67 3.1.3.2.4. ConcentrationMeanLocalEnviro .......................................................................................................................... 68 3.1.3.2.5. ConcentrationMaxLocalEnviro ............................................................................................................................ 68 3.1.3.2.6. ConcentrationMinLocalEnviro ............................................................................................................................ 68 3.1.3.2.7. KillSelf ................................................................................................................................................................ 68 3.1.3.2.8. NumberFeatures ................................................................................................................................................... 68

3.1.3.3. Learned behaviors ...................................................................................................................................................... 69 3.1.3.3.1. AddCell................................................................................................................................................................ 69 3.1.3.3.2. PruneSelf ............................................................................................................................................................. 69 3.1.3.3.3. ChangeToSOFMAndTrain .................................................................................................................................. 69 3.1.3.3.4. Classify ................................................................................................................................................................ 70

3.1.3.4. Cell types ................................................................................................................................................................... 70 3.1.3.4.1. SOFM .................................................................................................................................................................. 70 3.1.3.4.2. Pass-Thru ............................................................................................................................................................. 70

3.2. INPUT FEATURE VECTOR REPRESENTATIONS.......................................................................................................... 71

Page 9:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 9 of 249

09/28/10

3.2.1. Direct feature-to-protein mapping ................................................................................................................ 71 3.3. PATTERN TRAINING FOR CLASSIFICATION .............................................................................................................. 72

3.3.1. Self-organization principles .......................................................................................................................... 73 3.3.1.1. Protein analogs of statistical features ......................................................................................................................... 74 3.3.1.2. Cellular fission and death ........................................................................................................................................... 74

3.3.2. Particle swarm optimization ......................................................................................................................... 75 3.3.3. Training algorithm ....................................................................................................................................... 75

3.3.3.1. Training Algorithm: Presentation of training vectors and classes to the system ........................................................ 75 3.4. OUTPUT CODING .................................................................................................................................................... 77 3.5. POST PROCESSING OF CLASSIFICATION RESULTS .................................................................................................... 77

4. SIMULATIONS AND ANALYSES .......................................................................................................................... 78

4.1. SIMULATION LIMITS .............................................................................................................................................. 78 4.2. GENERAL METHODOLOGY ..................................................................................................................................... 80

4.2.1. Remote sensing cloud/no-cloud problem ...................................................................................................... 80 4.2.1.1. Description ................................................................................................................................................................. 80 4.2.1.2. Sensor and datasets .................................................................................................................................................... 80

4.3. THE CONSTRUCTION OF SIMPLE GENETIC REGULATORY NETWORKS ...................................................................... 95 4.3.1. Introduction/Methodology ............................................................................................................................ 95 4.3.2. Experiments 1 and 2 – Proteins .................................................................................................................... 96

4.3.2.1. Setup .......................................................................................................................................................................... 96 4.3.2.2. Experiments 1 and 2 results and discussion ............................................................................................................... 97 4.3.2.3. Experiments 1 and 2 conclusions ............................................................................................................................. 100

4.3.3. Experiments 3 and 4 - Protein Chemistry ................................................................................................... 101 4.3.3.1. Setup ........................................................................................................................................................................ 101 4.3.3.2. Experiments 3 and 4 results and discussion ............................................................................................................. 102 4.3.3.3. Experiments 3 and 4 conclusions ............................................................................................................................. 113

4.3.4. Experiment 5 - Gene activation .................................................................................................................. 114 4.3.4.1. Setup ........................................................................................................................................................................ 114 4.3.4.2. Experiment 5 results and discussion ......................................................................................................................... 115 4.3.4.3. Experiment 5 conclusions ........................................................................................................................................ 127

4.4. SELF-ORGANIZATION IN THE HEBIS ENVIRONMENT ............................................................................................ 128 4.4.1. Introduction/Methodology .......................................................................................................................... 128 4.4.2. Fitness function description ........................................................................................................................ 129 4.4.3. Experiment 6 - Swarm fitness characterization .......................................................................................... 132

4.4.3.1. Setup ........................................................................................................................................................................ 132 4.4.3.2. Experiment 6 results and discussion ......................................................................................................................... 132 4.4.3.3. Experiment 6 conclusions ........................................................................................................................................ 135

4.4.4. Experiment 7 - Initial location of processing cells ..................................................................................... 136 4.4.4.1. Setup ........................................................................................................................................................................ 136 4.4.4.2. Experiment 7 results and discussion ......................................................................................................................... 136 4.4.4.3. Experiment 7 conclusions ........................................................................................................................................ 137

4.4.5. Experiment 8 - Cellular actions .................................................................................................................. 138 4.4.5.1. Setup ........................................................................................................................................................................ 138 4.4.5.2. Experiment 8 results and discussion ......................................................................................................................... 139 4.4.5.3. Experiment 8 conclusions ........................................................................................................................................ 141

4.4.6. Experiment 9 - Protein statistical analogs.................................................................................................. 141 4.4.6.1. Setup ........................................................................................................................................................................ 141 4.4.6.2. Experiment 9 results and discussion ......................................................................................................................... 142 4.4.6.3. Experiment 9 conclusions ........................................................................................................................................ 143

4.4.7. Experiment 10 - Output protein comparison .............................................................................................. 143 4.4.7.1. Setup ........................................................................................................................................................................ 144 4.4.7.2. Experiment 10 results and discussion ....................................................................................................................... 144 4.4.7.3. Experiment 10 conclusions ...................................................................................................................................... 145

4.5. CLASSIFICATION ACCURACY ............................................................................................................................... 146 4.5.1. Introduction ................................................................................................................................................ 146 4.5.2. Training algorithm parameters description ................................................................................................ 146 4.5.3. Fully-engaged HeBIS.................................................................................................................................. 147

4.5.3.1. Experiment 11 - Size of geographic processing environment .................................................................................. 147 4.5.3.1.1. Setup .................................................................................................................................................................. 148 4.5.3.1.2. Experiment 11 results and discussion ................................................................................................................ 149 4.5.3.1.3. Experiment 11 conclusions ................................................................................................................................ 153

4.5.3.2. Experiment 12 - Size of intracellular SOFM kernel ................................................................................................. 153 4.5.3.2.1. Setup .................................................................................................................................................................. 153

Page 10:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 10 of 249

09/28/10

4.5.3.2.2. Experiment 12 results and discussion ................................................................................................................ 154 4.5.3.2.3. Experiment 12 conclusions ................................................................................................................................ 160

4.5.3.3. Experiment 13 - Protein chemistry reaction probability ........................................................................................... 161 4.5.3.3.1. Setup .................................................................................................................................................................. 161 4.5.3.3.2. Experiment 13 results and discussion ................................................................................................................ 162 4.5.3.3.3. Experiment 13 conclusions ................................................................................................................................ 167

4.5.3.4. Experiment 14 - Shotgun .......................................................................................................................................... 168 4.5.3.4.1. Setup .................................................................................................................................................................. 168 4.5.3.4.2. Experiment 14 results and discussion ................................................................................................................ 171 4.5.3.4.3. Experiment 14 conclusions ................................................................................................................................ 194

4.6. CLASSIFICATION ROBUSTNESS ............................................................................................................................. 196 4.6.1. Introduction/Methodology .......................................................................................................................... 196 4.6.2. Experiment 15 - Noise ................................................................................................................................ 196

4.6.2.1. Setup ........................................................................................................................................................................ 196 4.6.2.2. Experiment 15 results and discussion ....................................................................................................................... 198 4.6.2.3. Experiment 15 conclusions ...................................................................................................................................... 199

4.6.3. Experiment 16 - Missing features ............................................................................................................... 200 4.6.3.1. Setup ........................................................................................................................................................................ 200 4.6.3.2. Experiment 16 results and discussion ....................................................................................................................... 200 4.6.3.3. Experiment 16 conclusions ...................................................................................................................................... 203

5. SUMMARY AND CONCLUSIONS ....................................................................................................................... 204

6. FUTURE RESEARCH ............................................................................................................................................ 209

6.1. FITNESS FUNCTIONS ............................................................................................................................................ 209 6.2. ADDITIONAL MACHINE LEARNING KERNELS ........................................................................................................ 209 6.3. MODULARITY AND LEARNED FUNCTIONALITY .................................................................................................... 209 6.4. ADDITIONAL CELLULAR ACTIONS ........................................................................................................................ 210

6.4.1. Scale-free networks ..................................................................................................................................... 211 6.4.2. Mutual information ..................................................................................................................................... 211 6.4.3. Cellular instantiation .................................................................................................................................. 212

6.5. GRAPHICS PROCESSING UNIT .............................................................................................................................. 213 6.6. PROTEIN-BASED COMMUNICATIONS FOR ARTIFICIAL DEVICES IN A BIOLOGICAL SYSTEM .................................... 214

7. APPENDICES .......................................................................................................................................................... 215

7.1. DATA ................................................................................................................................................................... 216 7.2. GENOME MAPPINGS FOR SHOTGUN DATA ............................................................................................................ 234 7.3. PROTEIN DIFFUSION EXAMPLE ............................................................................................................................. 235 7.4. HEBIS FITNESS FUNCTION DETAILS ..................................................................................................................... 238 7.5. HEBIS SHOTGUN EXPERIMENT CORRELATION MAPS AND PDF .......................................................................... 240 7.6. HEBIS TRAINING CYCLE DETAILS ........................................................................................................................ 241

BIBLIOGRAPHY ............................................................................................................................................................ 242

Page 11:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 11 of 249

09/28/10

Figures

FIGURE 1. PERFORMANCE OF EMPIRICAL LEARNING SYSTEMS. ............................................................................................. 26 FIGURE 2. AN ARTIFICIAL NEURON. ...................................................................................................................................... 29 FIGURE 3. ARTIFICIAL FEEDFORWARD NEURAL NETWORK. ................................................................................................. 29 FIGURE 4. SELF-ORGANIZING FEATURE MAP. ...................................................................................................................... 36 FIGURE 5. SELF-ORGANIZING FEATURE MAP TOPOLOGY PRESERVATION EXAMPLE [61]. ................................................... 37 FIGURE 6. PARTICLE SWARM OPTIMIZATION ALGORITHM [90]. ............................................................................................ 42 FIGURE 7. ENVIRONMENTAL LATTICE AND PROCESSING CELL OVERVIEW. ........................................................................... 55 FIGURE 8. MAJOR BLOCKS OF CELL FUNCTIONALITY. ........................................................................................................... 56 FIGURE 9. LINEAR GRID NUMBERING SCHEME. ..................................................................................................................... 58 FIGURE 10. PLANAR GRID NUMBERING SCHEME. .................................................................................................................. 59 FIGURE 11. THREE-DIMENSIONAL LATTICE NUMBERING SCHEME. ........................................................................................ 59 FIGURE 12. GENOME/GENE PROTEIN HIERARCHY. ................................................................................................................ 61 FIGURE 13. STANDARD REGULATORY/ENVIRONMENTAL AND SWITCH PROTEIN DESCRIPTIONS. ........................................... 61 FIGURE 14. DIRECT FEATURE-TO-PROTEIN MAPPING. ........................................................................................................... 72 FIGURE 15. SCHEMATIC OF THE TRAINING ALGORITHM. ....................................................................................................... 76 FIGURE 16. TRAINING/TESTING PIXEL AND ITS RELATIONSHIP TO ITS SURROUNDING GEOGRAPHIC PIXELS. ......................... 78 FIGURE 17. TRAINING/TESTING PIXEL AND THE SURROUNDING MULTISPECTRAL INFORMATION. ......................................... 79 FIGURE 18. PSEUDOCOLOR IMAGE FOR A2002193183000 DATASET. GREY AND WHITE COLORS CORRESPOND TO CLOUD

PIXELS, BLACK CORRESPONDS TO WATER, AND GREEN AND BROWN REFER TO LAND PIXELS. ........................................ 84 FIGURE 19. GROUND TRUTH (CLOUD/NO-CLOUD) FOR A2002193183000 DATASET. RED CORRESPONDS TO LAND PIXELS

WHEREAS WHITE PIXELS REFERENCE CLOUDS AND BLACK CORRESPONDS TO WATER. ................................................... 85 FIGURE 20. LAND MASK (CLOUD/NO-CLOUD) FOR A2002193183000 DATASET. RED CORRESPONDS TO LAND PIXELS AND

BLACK CORRESPONDS TO WATER PIXELS. ...................................................................................................................... 86 FIGURE 21. CLOUD/NO-CLOUD CLASS BREAKDOWN ACCORDING TO SPECIFIC WAVELENGTH-BAND FEATURE. ..................... 87 FIGURE 22. 412 NM CLOUD/NO-CLOUD SCATTER PLOT WITH CLOUD (C0) PIXELS REPRESENTED AS 1 AND NO-CLOUD (C0)

PIXELS REPRESENTED AS -1 ON THE ABSCISSA. MAGNITUDES ARE LOG-NORMALIZED AND BIASED AND SCALED TO FALL

WITHIN THE RANGE [0, 1]. ............................................................................................................................................. 88 FIGURE 23. 443 NM CLOUD/NO-CLOUD SCATTER PLOT WITH CLOUD (C0) PIXELS REPRESENTED AS 1 AND NO-CLOUD (C0)

PIXELS REPRESENTED AS -1 ON THE ABSCISSA. MAGNITUDES ARE LOG-NORMALIZED AND BIASED AND SCALED TO FALL

WITHIN THE RANGE [0, 1]. ............................................................................................................................................. 88 FIGURE 24. 469 NM CLOUD/NO-CLOUD SCATTER PLOT WITH CLOUD (C0) PIXELS REPRESENTED AS 1 AND NO-CLOUD (C0)

PIXELS REPRESENTED AS -1 ON THE ABSCISSA. MAGNITUDES ARE LOG-NORMALIZED AND BIASED AND SCALED TO FALL

WITHIN THE RANGE [0, 1]. ............................................................................................................................................. 89 FIGURE 25. 488 NM CLOUD/NO-CLOUD SCATTER PLOT WITH CLOUD (C0) PIXELS REPRESENTED AS 1 AND NO-CLOUD (C0)

PIXELS REPRESENTED AS -1 ON THE ABSCISSA. MAGNITUDES ARE LOG-NORMALIZED AND BIASED AND SCALED TO FALL

WITHIN THE RANGE [0, 1]. ............................................................................................................................................. 89 FIGURE 26. 531 NM CLOUD/NO-CLOUD SCATTER PLOT WITH CLOUD (C0) PIXELS REPRESENTED AS 1 AND NO-CLOUD (C0)

PIXELS REPRESENTED AS -1 ON THE ABSCISSA. MAGNITUDES ARE LOG-NORMALIZED AND BIASED AND SCALED TO FALL

WITHIN THE RANGE [0, 1]. ............................................................................................................................................. 89 FIGURE 27. 551 NM CLOUD/NO-CLOUD SCATTER PLOT WITH CLOUD (C0) PIXELS REPRESENTED AS 1 AND NO-CLOUD (C0)

PIXELS REPRESENTED AS -1 ON THE ABSCISSA. MAGNITUDES ARE LOG-NORMALIZED AND BIASED AND SCALED TO FALL

WITHIN THE RANGE [0, 1]. ............................................................................................................................................. 90 FIGURE 28. 555 NM CLOUD/NO-CLOUD SCATTER PLOT WITH CLOUD (C0) PIXELS REPRESENTED AS 1 AND NO-CLOUD (C0)

PIXELS REPRESENTED AS -1 ON THE ABSCISSA. MAGNITUDES ARE LOG-NORMALIZED AND BIASED AND SCALED TO FALL

WITHIN THE RANGE [0, 1]. ............................................................................................................................................. 90 FIGURE 29. 645 NM CLOUD/NO-CLOUD SCATTER PLOT WITH CLOUD (C0) PIXELS REPRESENTED AS 1 AND NO-CLOUD (C0)

PIXELS REPRESENTED AS -1 ON THE ABSCISSA. MAGNITUDES ARE LOG-NORMALIZED AND BIASED AND SCALED TO FALL

WITHIN THE RANGE [0, 1]. ............................................................................................................................................. 91 FIGURE 30. 667 NM CLOUD/NO-CLOUD SCATTER PLOT WITH CLOUD (C0) PIXELS REPRESENTED AS 1 AND NO-CLOUD (C0)

PIXELS REPRESENTED AS -1 ON THE ABSCISSA. MAGNITUDES ARE LOG-NORMALIZED AND BIASED AND SCALED TO FALL

WITHIN THE RANGE [0, 1]. ............................................................................................................................................. 91 FIGURE 31. 678 NM CLOUD/NO-CLOUD SCATTER PLOT WITH CLOUD (C0) PIXELS REPRESENTED AS 1 AND NO-CLOUD (C0)

PIXELS REPRESENTED AS -1 ON THE ABSCISSA. MAGNITUDES ARE LOG-NORMALIZED AND BIASED AND SCALED TO FALL

WITHIN THE RANGE [0, 1]. ............................................................................................................................................. 91

Page 12:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 12 of 249

09/28/10

FIGURE 32. 748 NM CLOUD/NO-CLOUD SCATTER PLOT WITH CLOUD (C0) PIXELS REPRESENTED AS 1 AND NO-CLOUD (C0)

PIXELS REPRESENTED AS -1 ON THE ABSCISSA. MAGNITUDES ARE LOG-NORMALIZED AND BIASED AND SCALED TO FALL

WITHIN THE RANGE [0, 1]. ............................................................................................................................................. 92 FIGURE 33. 859 NM CLOUD/NO-CLOUD SCATTER PLOT WITH CLOUD (C0) PIXELS REPRESENTED AS 1 AND NO-CLOUD (C0)

PIXELS REPRESENTED AS -1 ON THE ABSCISSA. MAGNITUDES ARE LOG-NORMALIZED AND BIASED AND SCALED TO FALL

WITHIN THE RANGE [0, 1]. ............................................................................................................................................. 92 FIGURE 34. 869 NM CLOUD/NO-CLOUD SCATTER PLOT WITH CLOUD (C0) PIXELS REPRESENTED AS 1 AND NO-CLOUD (C0)

PIXELS REPRESENTED AS -1 ON THE ABSCISSA. MAGNITUDES ARE LOG-NORMALIZED AND BIASED AND SCALED TO FALL

WITHIN THE RANGE [0, 1]. ............................................................................................................................................. 93 FIGURE 35. 1240 NM CLOUD/NO-CLOUD SCATTER PLOT WITH CLOUD (C0) PIXELS REPRESENTED AS 1 AND NO-CLOUD (C0)

PIXELS REPRESENTED AS -1 ON THE ABSCISSA. MAGNITUDES ARE LOG-NORMALIZED AND BIASED AND SCALED TO FALL

WITHIN THE RANGE [0, 1]. ............................................................................................................................................. 93 FIGURE 36. 2130 NM CLOUD/NO-CLOUD SCATTER PLOT WITH CLOUD (C0) PIXELS REPRESENTED AS 1 AND NO-CLOUD (C0)

PIXELS REPRESENTED AS -1 ON THE ABSCISSA. MAGNITUDES ARE LOG-NORMALIZED AND BIASED AND SCALED TO FALL

WITHIN THE RANGE [0, 1]. ............................................................................................................................................. 94 FIGURE 37. BASELINE NUMBER OF PROTEINS IN ENVIRONMENTAL LATTICE FOR ZERO-LENGTH GENOME IN EXPERIMENT 1.

VERTICAL BARS CORRESPOND TO THE STANDARD DEVIATION OF THE SAMPLE MEAN. .................................................. 98 FIGURE 38. BASELINE NUMBER OF PROTEINS IN ENVIRONMENTAL LATTICE FOR THE 3-GENE GENOME IN EXPERIMENT 2.

VERTICAL BARS CORRESPOND TO THE STANDARD DEVIATION OF THE SAMPLE MEAN. .................................................. 98 FIGURE 39. NUMBER OF PROTEINS IN ENVIRONMENT COMPARED BETWEEN THE BASELINE GENOME FROM EXPERIMENT 1

AND THE MULTI-GENE GENOME FROM EXPERIMENT 2. VERTICAL BARS CORRESPOND TO THE STANDARD DEVIATION OF

THE SAMPLE MEAN. ..................................................................................................................................................... 100 FIGURE 40. NUMBER OF PROTEINS IN ENVIRONMENTAL LATTICE FOR A 0-GENE GENOME VS. CELLULAR ITERATION FOR

REACTION PROBABILITIES OF 0 %, 0.1 %, 1 %, AND 10 % WITH ERROR BARS REMOVED FOR CLARITY. ....................... 103 FIGURE 41. NUMBER OF PROTEINS IN ENVIRONMENTAL LATTICE FOR A 3-GENE GENOME VS. CELLULAR ITERATION FOR

REACTION PROBABILITIES OF 0 %, 0.1 %, 1 %, AND 10 % WITH ERROR BARS REMOVED FOR CLARITY. ....................... 104 FIGURE 42. NUMBER OF PROTEINS IN ENVIRONMENTAL LATTICE FOR A 0-GENE GENOME VS. CELLULAR ITERATION FOR A

REACTION PROBABILITY OF 0.1 %. VERTICAL BARS CORRESPOND TO THE STANDARD DEVIATION OF THE SAMPLE

MEAN........................................................................................................................................................................... 105 FIGURE 43. NUMBER OF PROTEINS IN ENVIRONMENTAL LATTICE FOR A 0-GENE GENOME VS. CELLULAR ITERATION FOR A

REACTION PROBABILITY OF 1.0 %. VERTICAL BARS CORRESPOND TO THE STANDARD DEVIATION OF THE SAMPLE

MEAN........................................................................................................................................................................... 105 FIGURE 44. NUMBER OF PROTEINS IN ENVIRONMENTAL LATTICE FOR A 0-GENE GENOME VS. CELLULAR ITERATION FOR A

REACTION PROBABILITY OF 10 %. VERTICAL BARS CORRESPOND TO THE STANDARD DEVIATION OF THE SAMPLE MEAN. .................................................................................................................................................................................... 106

FIGURE 45. NUMBER OF PROTEINS IN ENVIRONMENTAL LATTICE FOR A 3-GENE GENOME VS. CELLULAR ITERATION FOR A

REACTION PROBABILITY OF 0.1 %. VERTICAL BARS CORRESPOND TO THE STANDARD DEVIATION OF THE SAMPLE

MEAN........................................................................................................................................................................... 106 FIGURE 46. NUMBER OF PROTEINS IN ENVIRONMENTAL LATTICE FOR A 3-GENE GENOME VS. CELLULAR ITERATION FOR A

REACTION PROBABILITY OF 1.0 %. VERTICAL BARS CORRESPOND TO THE STANDARD DEVIATION OF THE SAMPLE

MEAN........................................................................................................................................................................... 107 FIGURE 47. NUMBER OF PROTEINS IN ENVIRONMENTAL LATTICE FOR A 3-GENE GENOME VS. CELLULAR ITERATION FOR A

REACTION PROBABILITY OF 10 %. VERTICAL BARS CORRESPOND TO THE STANDARD DEVIATION OF THE SAMPLE MEAN. .................................................................................................................................................................................... 107

FIGURE 48. FITTING AND STATISTICAL INFORMATION FOR A 0-GENE GENOME IN AN ENVIRONMENT WITH A REACTION

PROBABILITY OF 0.1 %. VERTICAL BARS ON THE UPPER CHART CORRESPOND TO THE STANDARD DEVIATION OF THE

SAMPLE MEAN. ............................................................................................................................................................ 109 FIGURE 49. FITTING AND STATISTICAL INFORMATION FOR A 0-GENE GENOME IN AN ENVIRONMENT WITH A REACTION

PROBABILITY OF 1.0 %. VERTICAL BARS ON THE UPPER CHART CORRESPOND TO THE STANDARD DEVIATION OF THE

SAMPLE MEAN. ............................................................................................................................................................ 110 FIGURE 50. FITTING AND STATISTICAL INFORMATION FOR A 0-GENE GENOME IN AN ENVIRONMENT WITH A REACTION

PROBABILITY OF 10%. VERTICAL BARS ON THE UPPER CHART CORRESPOND TO THE STANDARD DEVIATION OF THE

SAMPLE MEAN. ............................................................................................................................................................ 110 FIGURE 51. FITTING AND STATISTICAL INFORMATION FOR A 3-GENE GENOME IN AN ENVIRONMENT WITH A REACTION

PROBABILITY OF 0.1 %. VERTICAL BARS ON THE UPPER CHART CORRESPOND TO THE STANDARD DEVIATION OF THE

SAMPLE MEAN. ............................................................................................................................................................ 111 FIGURE 52. FITTING AND STATISTICAL INFORMATION FOR A 3-GENE GENOME IN AN ENVIRONMENT WITH A REACTION

PROBABILITY OF 1.0 %. VERTICAL BARS ON THE UPPER CHART CORRESPOND TO THE STANDARD DEVIATION OF THE

SAMPLE MEAN. ............................................................................................................................................................ 112

Page 13:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 13 of 249

09/28/10

FIGURE 53. FITTING AND STATISTICAL INFORMATION FOR A 3-GENE GENOME IN AN ENVIRONMENT WITH A REACTION

PROBABILITY OF 10 %. VERTICAL BARS ON THE UPPER CHART CORRESPOND TO THE STANDARD DEVIATION OF THE

SAMPLE MEAN. ............................................................................................................................................................ 112 FIGURE 54. COMPARISON BETWEEN 0-GENE AND 3-GENE GENOMES FOR VARYING ENVIRONMENTAL REACTION

PROBABILITIES. ERROR BARS REMOVED FOR CLARITY. ............................................................................................... 113 FIGURE 55. EXAMPLE GENE ACTIVATION MAP. ................................................................................................................... 116 FIGURE 56. 3-GENE GENOME ACTIVATION VS. ITERATION FOR C0 (178) USING GENOMES 93 AND 146 IN (A) AND (B),

RESPECTIVELY. ............................................................................................................................................................ 117 FIGURE 57. 3-GENE GENOME ACTIVATION VS. ITERATION FOR C0 (142) USING GENOMES 9 AND 147, RESPECTIVELY, IN (A)

AND (B). ...................................................................................................................................................................... 117 FIGURE 58. 3-GENE GENOME ACTIVATION VS. CELLULAR ITERATION FOR C0 (48) USING GENOME 178. ............................. 118 FIGURE 59. 10-GENE GENOME ACTIVATION VS. CELLULAR ITERATION FOR ORIGINAL AND TWO CLONED CELLS WITHIN THE

ENVIRONMENTAL MATRIX- C0 (42) TEST PIXEL FOR GENOME 117. THE ORIGINAL GENOME IS PRESENTED IN (A) WHILE

THE TWO CLONED GENOMES ARE PRESENTED IN (B) AND (C). ...................................................................................... 119 FIGURE 60. 10-GENE GENOME ACTIVATION VS. CELLULAR ITERATION FOR ORIGINAL AND TWO CLONED CELLS WITHIN THE

ENVIRONMENTAL MATRIX- C0 (50) TEST PIXEL USING GENOME 117. THE ACTIVATION FOR THE ORIGINAL CELL IS

SHOWN IN (A) AND THE ACTIVATIONS FOR THE CLONED CELLS ARE SHOWN IN (B) AND (C). ........................................ 120 FIGURE 61. 10-GENE GENOME ACTIVATION VS. CELLULAR ITERATION FOR ORIGINAL AND TWO CLONED CELLS WITHIN THE

ENVIRONMENTAL MATRIX- C1 (51) TEST PIXEL USING GENOME 117. THE ORIGINAL CELL ACTIVATION IS PRESENTED IN

(A) AND THE ACTIVATIONS FOR THE CLONED CELLS ARE LISTED IN (B) AND (C). ......................................................... 120 FIGURE 62. 10-GENE GENOME ACTIVATION VS. CELLULAR ITERATION FOR THE SAME TEST PIXEL C0 (42). GENOMES 91,

123, AND 135 ARE DISPLAYED IN (A), (B), AND (C), RESPECTIVELY. GENOMES 123 AND 135 SHOW ORIGINAL AND TWO

CLONED CELL ACTIVATIONS FOR EACH OF THESE GENOMES. ....................................................................................... 122 FIGURE 63. 40-GENE GENOME ACTIVATION FOR GENOME 30 FOR TEST PIXELS C0 (26) IN THE TOP IMAGE (A) AND C1 (25) IN

THE BOTTOM IMAGE (B). THE GENOME SHOWS MULTI-GENE ACTIVATION FOR C0 AND SINGLE GENE ACTIVATION FOR

C1 WITH DIFFERING RESPONSES. ................................................................................................................................. 123 FIGURE 64. 40-GENE GENOME ACTIVATION FOR GENOME 66. THE TOP (A) AND MIDDLE(B) IMAGES ARE THE MULTI-GENE

ACTIVATION PROFILES FOR C0 (30) WITH AN ORIGINAL AND CLONED CELL. THE BOTTOM (C) IMAGE SHOW THE

ACTIVATION FOR THE SAME GENOME, BUT FOR A C1 (31). .......................................................................................... 124 FIGURE 65. 40-GENE GENOME ACTIVATION OF GENOME 90 FOR TEST PIXELS C0 (30), C1 (31), C0 (32), C0 (34), C0 (54),

AND C1 (55), RESPECTIVELY FROM TOP TO BOTTOM, IN FIGURES (A) – (F). .................................................................. 126 FIGURE 66. DECISION REGION MAPPING AND BOUNDARY BASED ON THE VALUE OF Corr

C0 max. ....................................... 130

FIGURE 67. AVERAGE BEST GENOME FITNESS VS. BREED # FOR 1-PARTICLE PSO SWARM. VERTICAL BARS CORRESPOND TO

THE STANDARD DEVIATION OF THE SAMPLE MEAN. ..................................................................................................... 133 FIGURE 68. AVERAGE BEST GENOME FITNESS VS. BREED # FOR 100-PARTICLE PSO SWARM. VERTICAL BARS CORRESPOND

TO THE STANDARD DEVIATION OF THE SAMPLE MEAN. ................................................................................................ 133 FIGURE 69. AVERAGE BEST GENOME FITNESS VS. BREED # FOR 250-PARTICLE PSO SWARM. VERTICAL BARS CORRESPOND

TO THE STANDARD DEVIATION OF THE SAMPLE MEAN. ................................................................................................ 134 FIGURE 70. AVERAGE BEST GENOME FITNESS VS. BREED # FOR 500-PARTICLE PSO SWARM. VERTICAL BARS CORRESPOND

TO THE STANDARD DEVIATION OF THE SAMPLE MEAN. ................................................................................................ 134 FIGURE 71. BEST AVERAGE PEAK GENOME FITNESS VS. THE NUMBER OF PARTICLES IN THE PSO SWARM. VERTICAL BARS

CORRESPOND TO THE STANDARD DEVIATION OF THE SAMPLE MEAN OF THE PEAK FITNESS FOR EACH SWARM TESTED. .................................................................................................................................................................................... 135

FIGURE 72. CV AVERAGE FITNESS VS. INITIAL CELL LOCATION FROM BEST BRED GENOME. VERTICAL BARS CORRESPOND

TO THE STANDARD DEVIATION OF THE SAMPLE MEAN. ................................................................................................ 137 FIGURE 73. CV AVERAGE FITNESS VS. ACTIVATED CELLULAR ACTION. VERTICAL BARS CORRESPOND TO THE STANDARD

DEVIATION OF THE SAMPLE MEAN. .............................................................................................................................. 139 FIGURE 74. CV AVERAGE FITNESS VS. ACTIVITY LEVEL OF CELLULAR PROTEIN STATISTICS. VERTICAL BARS CORRESPOND

TO THE STANDARD DEVIATION OF THE SAMPLE MEAN. ................................................................................................ 142 FIGURE 75. CV AVERAGE FITNESS VS. STATIC OR PSO-EVOLVED SETTING OF THE OUTPUT C0/C1 PROTEIN. VERTICAL BARS

CORRESPOND TO THE STANDARD DEVIATION OF THE SAMPLE MEAN. .......................................................................... 145 FIGURE 76. MULTI-SPECTRAL DATA CUBE FOR A 5X5 GEOGRAPHIC REGION WITH 15 BANDS OF MULTISPECTRAL DATA. ... 149 FIGURE 77. FULL-IMAGE AVERAGE CLASSIFICATION ACCURACY VS. SIZE OF GEOGRAPHIC REGIONS SURROUNDING TEST

PIXEL. VERTICAL BARS CORRESPOND TO THE STANDARD DEVIATION OF THE SAMPLE MEAN. ..................................... 150 FIGURE 78. FULL-IMAGE CLASSIFICATION ACCURACY VS. FITNESS FOR A 3X3 GEOGRAPHIC REGION SURROUNDING TEXT

PIXEL. .......................................................................................................................................................................... 151 FIGURE 79. FULL-IMAGE CLASSIFICATION ACCURACY VS. FITNESS FOR A 5X5 GEOGRAPHIC REGION SURROUNDING TEXT

PIXEL. .......................................................................................................................................................................... 152 FIGURE 80. CLASSIFICATION ACCURACY VS. FITNESS FOR CASE WITH 0 NEURONS IN INTRACELLULAR SOFM. .................. 155 FIGURE 81. CLASSIFICATION ACCURACY VS. FITNESS FOR CASE WITH 1 NEURON IN INTRACELLULAR SOFM. ................... 156

Page 14:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 14 of 249

09/28/10

FIGURE 82. CLASSIFICATION ACCURACY VS. FITNESS FOR CASE WITH 4 NEURONS IN INTRACELLULAR SOFM. ................. 157 FIGURE 83. CLASSIFICATION ACCURACY VS. FITNESS FOR CASE WITH 9 NEURONS IN INTRACELLULAR SOFM. ................. 159 FIGURE 84. CLASSIFICATION ACCURACY VS. FITNESS FOR CASE WITH 81 NEURONS IN INTRACELLULAR SOFM. ............... 160 FIGURE 85. FULL-IMAGE AVERAGE CLASSIFICATION ACCURACY VS. THE NUMBER OF NEURONS IN THE INTRACELLULAR

SOFM. VERTICAL BARS CORRESPOND TO THE STANDARD DEVIATION OF THE SAMPLE MEAN OF THE CLASSIFICATION

ACCURACY. ................................................................................................................................................................. 160 FIGURE 86. CLASSIFICATION ACCURACY VS. FITNESS WITH 0.0 REACTION PROBABILITY. .................................................. 163 FIGURE 87. CLASSIFICATION ACCURACY VS. FITNESS WITH 0.001 REACTION PROBABILITY. .............................................. 164 FIGURE 88. CLASSIFICATION ACCURACY VS. FITNESS WITH 0.01 REACTION PROBABILITY. ................................................ 165 FIGURE 89. CLASSIFICATION ACCURACY VS. FITNESS WITH 0.1 REACTION PROBABILITY. .................................................. 166 FIGURE 90. FULL-IMAGE CLASSIFICATION ACCURACY VS. THE PROBABILITY OF PROTEIN REACTION IN THE

ENVIRONMENTAL LATTICE. VERTICAL BARS CORRESPOND TO THE STANDARD DEVIATION OF THE SAMPLE MEAN. .... 167 FIGURE 91. SAMPLE ROC CURVE WITH FALSE-POSITIVE RATE ALONG THE ABSCISSA AND TRUE-POSITIVE RATE AS THE

ORDINATE. ................................................................................................................................................................... 170 FIGURE 92. HEBIS CLASSIFICATION IMAGERY AND ROC FOR 2010_05_28_07_58_58_61 TEST. THIS IS HEBIS SELECTED

TRIAL # 0. .................................................................................................................................................................... 175 FIGURE 93. HEBIS CLASSIFICATION IMAGERY AND ROC FOR 2010_06_11_01_18_20_77 TEST. THIS IS HEBIS SELECTED

TRIAL # 1. .................................................................................................................................................................... 176 FIGURE 94. HEBIS CLASSIFICATION IMAGERY AND ROC FOR 2010_06_10_08_03_09_38 TEST. THIS IS HEBIS SELECTED

TRIAL # 2. .................................................................................................................................................................... 177 FIGURE 95. HEBIS CLASSIFICATION IMAGERY AND ROC FOR 2010_06_09_16_37_07_17 TEST. THIS IS HEBIS SELECTED

TRIAL # 3. .................................................................................................................................................................... 178 FIGURE 96. HEBIS CLASSIFICATION IMAGERY AND ROC FOR 2010_05_28_12_49_29_76 TEST. THIS IS HEBIS SELECTED

TRIAL # 4. .................................................................................................................................................................... 179 FIGURE 97. HEBIS CLASSIFICATION IMAGERY AND ROC FOR 2010_06_03_10_39_23_85 TEST. THIS IS HEBIS SELECTED

TRIAL # 5. .................................................................................................................................................................... 180 FIGURE 98. HEBIS CLASSIFICATION IMAGERY AND ROC FOR 2010_05_28_00_32_27_50 TEST. THIS IS HEBIS SELECTED

TRIAL # 6. .................................................................................................................................................................... 181 FIGURE 99. HEBIS CLASSIFICATION IMAGERY AND ROC FOR 2010_06_09_17_43_13_23 TEST. THIS IS HEBIS SELECTED

TRIAL # 7. .................................................................................................................................................................... 182 FIGURE 100. HEBIS CLASSIFICATION IMAGERY AND ROC FOR 2010_05_28_12_42_02_75 TEST. THIS IS HEBIS SELECTED

TRIAL # 8. .................................................................................................................................................................... 183 FIGURE 101. HEBIS CLASSIFICATION IMAGERY AND ROC FOR 2010_06_01_233_35_56_3 TEST. THIS IS HEBIS SELECTED

TRIAL # 9. .................................................................................................................................................................... 184 FIGURE 102. HEBIS CLASSIFICATION IMAGERY AND ROC FOR 2010_06_10_07_35_57_34 TEST. THIS IS HEBIS SELECTED

TRIAL # 10. .................................................................................................................................................................. 185 FIGURE 103. COMPARISON PLOT OF HEBIS CLASSIFICATION ACCURACY VS. FITNESS OF THE GENOME FOR 200 TRIALS.

PROTEIN CHEMISTRY IS DEACTIVATED. ....................................................................................................................... 189 FIGURE 104. COMPARISON PLOT OF HEBIS CLASSIFICATION ACCURACY VS. FITNESS OF THE GENOME FOR 79 TRIALS WITH

PROTEIN CHEMISTRY ACTIVATED. ............................................................................................................................... 189 FIGURE 105. HEBIS CLASSIFICATION ACCURACY VS. REACTION PROBABILITY FOR 79 TRIALS. PROTEIN CHEMISTRY IS

ACTIVATED. ................................................................................................................................................................. 190 FIGURE 106. HEBIS CLASSIFICATION ACCURACY VS. REACTION PROBABILITY AND FITNESS FOR 79 TRIALS. PROTEIN

CHEMISTRY IS ACTIVATED. .......................................................................................................................................... 191 FIGURE 107 HEBIS CLASSIFICATION ACCURACY VS. REACTION PROBABILITY AND MINIMUM PROTEIN CORRELATION FOR 79

TRIALS. PROTEIN CHEMISTRY IS ACTIVATED. .............................................................................................................. 192 FIGURE 108. CLASSIFICATION ACCURACY VS. ENVIRONMENTAL DIFFUSION RATE FOR 200 SHOTGUN TRIALS. PROTEIN

CHEMISTRY IS DEACTIVATED. ...................................................................................................................................... 193 FIGURE 109. CLASSIFICATION ACCURACY VS. GENOME FITNESS AND ENVIRONMENTAL DIFFUSION RATE FOR 200 SHOTGUN

TRIALS. PROTEIN CHEMISTRY IS DEACTIVATED. ......................................................................................................... 193 FIGURE 110. NOISE COMPARISON PLOT FOR CLASSIFICATION ACCURACY VS. NOISE STANDARD DEVIATION FOR BOTH HEBIS

AND SOFM TRIALS. VERTICAL BARS CORRESPOND TO THE STANDARD DEVIATION OF THE SAMPLE MEAN OF

CLASSIFICATION ACCURACY. ....................................................................................................................................... 199 FIGURE 111. COMPARISON OF BEFORE AND AFTER CLASSIFICATION ACCURACY FOR THE MODIS BAND 16 KNOCKOUT. .. 201 FIGURE 112. COMPARISON OF BEFORE AND AFTER CLASSIFICATION ACCURACY FOR THE MODIS BAND 7 KNOCKOUT. .... 202 FIGURE 113. ONE METHOD OF PRESENTING TRAINING/TEST DATA TO HEBIS. EACH BEHAVIOR IS TRAINED SEPARATELY,

CANDIDATE GENOMES ARE CREATED, AND THE CANDIDATES THEN UNDERGO EVOLUTIONARY OPTIMIZATION IN A FINAL

CV/GA LOOP. ............................................................................................................................................................. 210 FIGURE 114. PROCESSING MODEL FOR RESEARCH INFRASTRUCTURE. ................................................................................ 214

Page 15:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 15 of 249

09/28/10

FIGURE 115. PARAMETER ACTIVATION MAPS FOR GENOMES DISCOVERED DURING SHOTGUN EXPERIMENTS. THE COLOR BAR

FROM TABLE 81 IS APPLICABLE TO THE EVOLVED ELEMENTS OF THE GENOMES IN THIS FIGURE. ................................. 234 FIGURE 116. SINGLE PROTEIN DIFFUSION FROM FOUR SITES WITHIN AN 11X11X11 CUBIC ENVIRONMENT. THIS FRAME

SHOWS FOUR INITIAL SITES OF PROTEIN ACTIVATION AT THE BEGINNING OF THE SIMULATION .................................... 235 FIGURE 117. FRAME #2 IN THE SIMULATION THIS FRAME IS A SNAPSHOT OF ACTIVITY IN THE ENVIRONMENTAL LATTICE

AFTER ITERATION 2 OF THE DIFFUSION SIMULATION. THE TWO RED PROTEIN SITES ARE STILL ACTIVELY PRODUCING

PROTEINS WHEREAS THE TWO GREEN SITES ARE DECAYING. THE LIGHT BLUE COLOR REPRESENTS SITES WITHIN THE

LATTICE THAT HAVE THE LOWEST NON-ZERO PROTEIN CONCENTRATIONS AT THIS POINT IN THE SIMULATION. .......... 235 FIGURE 118. FRAME #6. THE SITES COLORED RED ARE STILL ACTIVELY PRODUCING WHEREAS THE LIGHT-BLUE-COLORED

AND DARK-BLUE-COLORED SITES POSSESS LOWER CONCENTRATIONS OF THE SIMULATED PROTEIN. THE DARKER BLUE

SITES CONTAIN LOWER CONCENTRATIONS OF PROTEIN THAN THE LIGHT-BLUE SITES. THIS IS THE SNAPSHOT FROM

ITERATION 6 OF THE SIMULATION. ............................................................................................................................... 236 FIGURE 119. FRAME #26. AFTER 26 ITERATIONS, THE ARTIFICIAL PROTEIN HAS DIFFUSED THROUGHOUT A LARGE PORTION

OF THE 11X11X11 ENVIRONMENTAL MATRIX. HOTTER COLORS (E.G. RED, YELLOW, GREEN) CORRESPOND TO HIGHER

CONCENTRATIONS OF THE PROTEINS WHEREAS COOLER COLORS (E.G. LIGHT BLUE, BLUE) CORRESPOND TO AREAS OF

RELATIVELY LOW CONCENTRATIONS. .......................................................................................................................... 236 FIGURE 120. FRAME #40. AT ITERATION 40, THE PROTEIN HAS DIFFUSED THROUGHOUT THE ENVIRONMENTAL LATTICE.

THE RED SITES ARE THE LOCATIONS OF THE ORIGINAL AND CONTINUING PROTEIN SOURCES. HOTTER COLORS

CORRESPOND TO HIGHER PROTEIN CONCENTRATIONS WHEREAS COOLER COLORS CORRESPOND TO LOWER

CONCENTRATIONS. ...................................................................................................................................................... 237 FIGURE 121. REGIONS OF EQUIVALENT Corr

C0 max FOR THE FITNESS FUNCTION. .............................................................. 238

FIGURE 122. Θcorr

2 PORTION OF FITNESS FUNCTION. ........................................................................................................ 238

FIGURE 123. Magcorr

2 PORTION OF FITNESS FUNCTION. ...................................................................................................... 239

FIGURE 124. CORRELATION COEFFICIENT GRID FOR PROCESSING PARAMETERS AND CLASSIFICATION RESULTS. PARAMETERS AND RESULTS ARE NUMBERED FROM 1 TO 26. ....................................................................................... 240

FIGURE 125. SIGNIFICANCE P-VALUE GRID FOR PROCESSING PARAMETERS AND CLASSIFICATION RESULTS OBTAINED WITH A

STUDENT’S T-TEST. PARAMETERS AND RESULTS ARE NUMBERED FROM 1 TO 26. ....................................................... 240 FIGURE 126. HEBIS CLASSIFICATION TRAINING CYCLE. ..................................................................................................... 241

Page 16:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 16 of 249

09/28/10

Tables

TABLE 1. BITWISE XOR FUNCTIONAL MAPPING ................................................................................................................... 65 TABLE 2. ORBITAL INFORMATION FOR NASA'S AQUA SATELLITE. ...................................................................................... 80 TABLE 3. NOMINAL RESOLUTIONS FOR THE MODIS SENSOR. ............................................................................................. 81 TABLE 4. 36 BANDS OF MULTISPECTRAL DATA FROM MODIS. ........................................................................................... 81 TABLE 5. NASA/CEOS DATASET LEVEL DEFINITION ......................................................................................................... 82 TABLE 6. 17 BANDS FROM MODIS FOR A2002193183000 LAC_X_NIR. ........................................................................... 82 TABLE 7. MULTISPECTRAL BANDS USED FROM MODIS [111]............................................................................................... 83 TABLE 8. SIMULATION PARAMETERS FOR EXPERIMENT 1. ................................................................................................... 96 TABLE 9. SIMULATION PARAMETERS FOR EXPERIMENT 2. .................................................................................................... 97 TABLE 10. STATISTICAL SUMMARY FOR EXPERIMENTS 1 AND 2. ......................................................................................... 99 TABLE 11. SIMULATION PARAMETERS FOR EXPERIMENT 3. ............................................................................................... 101 TABLE 12. SIMULATION PARAMETERS FOR EXPERIMENT 4. ............................................................................................... 102 TABLE 13. STATISTICAL SUMMARY FOR EXPERIMENT 3. ................................................................................................... 108 TABLE 14. STATISTICAL SUMMARY FOR EXPERIMENT 4. ................................................................................................... 108 TABLE 15. SIMULATION PARAMETERS FOR EXPERIMENT 5. ............................................................................................... 115 TABLE 16. SIMULATION PARAMETERS FOR SELF-ORGANIZATION EXPERIMENTS............................................................... 129 TABLE 17. EXPERIMENT 6 TRIAL DISTRIBUTION ................................................................................................................ 132 TABLE 18. TRIAL DISTRIBUTION FOR EXPERIMENT 7. ......................................................................................................... 136 TABLE 19. TRIAL DISTRIBUTION FOR EXPERIMENT 8 ......................................................................................................... 139 TABLE 20. EXPERIMENT 9 PARAMETERS ............................................................................................................................ 142 TABLE 21. EXPERIMENT 10 PARAMETERS .......................................................................................................................... 144 TABLE 22. RANGE OF PERTINENT HEBIS TRAINING PARAMETERS FOR CLASSIFICATION .................................................. 147 TABLE 23. PERTINENT SOFM TRAINING PARAMETERS FOR CLASSIFICATION.................................................................... 147 TABLE 24. TRIAL DISTRIBUTION ACROSS GEOGRAPHIC REGION SIZE FOR EXPERIMENT 11. .............................................. 148 TABLE 25. SIMULATION PARAMETERS FOR EXPERIMENT 11. ............................................................................................. 148 TABLE 26. CONFUSION MATRIX FOR 3X3 GEOGRAPHIC REGION ........................................................................................ 150 TABLE 27. CONFUSION MATRIX FOR 5X5 GEOGRAPHIC REGION ........................................................................................ 152 TABLE 28. HEBIS KERNEL SIZES FOR THE INTRACELLULAR SOFM IN EXPERIMENT 12. ................................................... 153 TABLE 29. SIMULATION PARAMETERS FOR EXPERIMENT 12. ............................................................................................. 154 TABLE 30. CONFUSION MATRIX FOR INTRACELLULAR SOFM WITH 0 NEURONS ............................................................... 154 TABLE 31. CONFUSION MATRIX FOR INTRACELLULAR SOFM WITH 1 NEURON ................................................................. 156 TABLE 32. CONFUSION MATRIX FOR INTRACELLULAR SOFM WITH 4 NEURONS ............................................................... 157 TABLE 33. CONFUSION MATRIX FOR INTRACELLULAR SOFM WITH 9 NEURONS ............................................................... 158 TABLE 34. CONFUSION MATRIX FOR INTRACELLULAR SOFM WITH 81 NEURONS ............................................................. 159 TABLE 35. PROTEIN REACTION PROBABILITY DISTRIBUTION FOR EXPERIMENT 13. ........................................................... 161 TABLE 36. SIMULATION PARAMETERS FOR EXPERIMENT 13. ............................................................................................. 162 TABLE 37. CONFUSION MATRIX FOR 0.0 PROTEIN REACTION PROBABILITY ...................................................................... 162 TABLE 38. CONFUSION MATRIX FOR 0.001 PROTEIN REACTION PROBABILITY .................................................................. 163 TABLE 39. CONFUSION MATRIX FOR 0.01 PROTEIN REACTION PROBABILITY .................................................................... 165 TABLE 40. CONFUSION MATRIX FOR 0.1 PROTEIN REACTION PROBABILITY ...................................................................... 166 TABLE 41. LIST OF HEBIS PARAMETERS TO RANDOMIZE FOR EXPERIMENT 14. ............................................................... 169 TABLE 42. LIST OF SOFM PARAMETERS TO RANDOMIZE FOR EXPERIMENT 14. ................................................................ 169 TABLE 43. DISTRIBUTION OF CLASS AND INFRASTRUCTURE PIXELS IN A2002193183000................................................. 171 TABLE 44. OPERATIONAL PARAMETERS FOR SELECTED HEBIS SHOTGUN EXPERIMENTS ................................................. 172 TABLE 45. OPERATIONAL PARAMETERS FOR SOFM EXPERIMENTS ................................................................................... 172 TABLE 46. CLASSIFICATION RESULTS FOR SELECTED HEBIS SHOTGUN EXPERIMENTS ..................................................... 173 TABLE 47. CLASSIFICATION RESULTS FOR SELECTED SOFM EXPERIMENTS ...................................................................... 173 TABLE 48. FEATURE AND RESULT INDICES FOR CORRELATION COEFFICIENT AND P-VALUE MATRICES ........................... 187 TABLE 49. BEST CLASSIFICATION ACCURACIES FOR THE SELECTED HEBIS AND SOFM EXAMPLES ................................. 195 TABLE 50. DATASET DEFINITIONS FOR EXPERIMENT 15. .................................................................................................... 197 TABLE 51. SIMULATION PARAMETERS FOR THE HEBIS "BEST" GENOME FOR EXPERIMENT 15. ....................................... 197 TABLE 52. SIMULATION PARAMETERS FOR THE 2X1 SOFM "BEST" CODEBOOK FOR EXPERIMENT 15.............................. 197 TABLE 53. SIMULATION PARAMETERS FOR THE 3X1 SOFM "BEST" CODEBOOK FOR EXPERIMENT 15.............................. 197 TABLE 54. MISSING FEATURE COMPARISON FOR CLASSIFICATION ACCURACY USING HEBIS AND SOFM ALGORITHMS

WITH MODIS BAND 16 KNOCKOUT ............................................................................................................................ 201

Page 17:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 17 of 249

09/28/10

TABLE 55. MISSING FEATURE COMPARISON FOR CLASSIFICATION ACCURACY USING HEBIS AND SOFM ALGORITHMS

WITH MODIS BAND 7 KNOCKOUT .............................................................................................................................. 202 TABLE 56. EXPERIMENT 6 AGGREGATE BREEDING DATA. ................................................................................................. 216 TABLE 57. EXPERIMENT 7 DATA. ....................................................................................................................................... 216 TABLE 58. EXPERIMENT 8 DATA. ....................................................................................................................................... 217 TABLE 59. EXPERIMENT 9 DATA ........................................................................................................................................ 217 TABLE 60. EXPERIMENT 10 DATA. ..................................................................................................................................... 217 TABLE 61. EXPERIMENT 11 DATA FOR 3X3 AND 5X5 GEOGRAPHIC REGION COMPARISON ................................................ 217 TABLE 62. EXPERIMENT 11 SCATTER DATA FOR 3X3 GEOGRAPHIC REGION ...................................................................... 218 TABLE 63. EXPERIMENT 11 SCATTER DATA FOR 5X5 GEOGRAPHIC REGION ...................................................................... 219 TABLE 64. EXPERIMENT 12 - INTRACELLULAR SOFM DATA ............................................................................................. 219 TABLE 65. EXPERIMENT 12 SCATTER DATA FOR 0X0 INTRACELLULAR SOFM .................................................................. 220 TABLE 66. EXPERIMENT 12 SCATTER DATA FOR 1X1 INTRACELLULAR SOFM .................................................................. 220 TABLE 67. EXPERIMENT 12 SCATTER DATA FOR 2X2 INTRACELLULAR SOFM .................................................................. 221 TABLE 68. EXPERIMENT 12 SCATTER DATA FOR 3X3 INTRACELLULAR SOFM .................................................................. 221 TABLE 69. EXPERIMENT 12 SCATTER DATA FOR 9X9 INTRACELLULAR SOFM .................................................................. 222 TABLE 70. EXPERIMENT 13 AGGREGATE CLASSIFICATION DATA ...................................................................................... 222 TABLE 71. EXPERIMENT 13 DATA FOR 0.0 REACTION PROBABILITY .................................................................................. 223 TABLE 72. EXPERIMENT 13 DATA FOR 0.001 REACTION PROBABILITY .............................................................................. 223 TABLE 73. EXPERIMENT 13 DATA FOR 0.01 REACTION PROBABILITY ................................................................................ 224 TABLE 74. EXPERIMENT 13 DATA FOR 0.1 REACTION PROBABILITY .................................................................................. 224 TABLE 75. EXPERIMENT 14 - STATISTICAL SUMMARY DATA FOR HEBIS CLASSIFICATION ACCURACY AND FITNESS

SCATTER DATA ........................................................................................................................................................... 225 TABLE 76. EXPERIMENT 14 HEBIS SCATTER DATA FOR FITNESS AND CLASSIFICATION ACCURACY ................................. 225 TABLE 77. EXPERIMENT 14 - AGGREGATE CLASSIFICATION ACCURACY RESULTS FOR SOFM. ......................................... 230 TABLE 78. SELECTED RESULTS FROM EXPERIMENT 14 ...................................................................................................... 231 TABLE 79. DATA FOR EXPERIMENT 15- CLASSIFICATION ACCURACY FOR 0.1 PROBABILITY NOISE INJECTION WITH

VARYING NOISE STANDARD DEVIATIONS ................................................................................................................... 232 TABLE 80. DATA FOR EXPERIMENT 16 – CLASSIFICATION ACCURACY WITH MODIS BAND 15 KNOCKED OUT ................ 232 TABLE 81. COMPARISON OF SELECTED GENOMES FOR SHOTGUN EXPERIMENTS ............................................................... 233

Page 18:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 18 of 249

09/28/10

Terms and Acronyms ANN: Artificial Neural Network

CDF: Cumulative Distribution Function CEOS: Committee on Earth Observation Satellites

CV: Cross Validation

EC: Evolutionary Computation ED: Evolutionary Development ESA: European Space Agency Evodevo: Evolutionary Development GA: Genetic Algorithm

Genetic Algorithm: GRN: Genetic Regulatory Network

ML: Machine Learning MODIS: Moderate Resolution Imaging Spectroradiometer NASA: National Aeronautics and Space Administration

PDF: Probability Density Function PSO: Particle Swarm Optimization

ROC: Receiver Operating Characteristic curve

SOFM: Self Organizing Feature Map SVM: Support Vector Machine SI: Swarm Intelligence TOA: Top of Atmosphere TBD: To Be Determined

Page 19:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 19 of 249

09/28/10

1. Introduction

The need currently exists on a worldwide basis for the timely extraction of knowledge for the

management of natural resources. Extraction of this knowledge for local, regional, and global

applications is being driven by the desire to more precisely assess the issues associated with both

anthropogenic and natural drivers in the environment. Data are being collected from a multitude of

earth-orbiting sensor platforms from the European Space Agency (ESA), the National Aeronautics

and Space Administration (NASA) and a host of other national agencies as well as private concerns.

This torrent of raw data, for example, NASA’s Moderate Resolution Imaging Spectroradiometer

(MODIS) sensors routinely collect a terabyte of earth imagery on a daily basis, brings with it the

issue of extraction of usable knowledge from these information-rich datasets for efficient

management of the earth’s natural resources.

State-of-the-art on-orbit optical sensors routinely collect earth resource data that are multispectral or

hyperspectral in nature. This increase in the number of available bands of information promises to

provide better discrimination of desired classes only if an appropriate level of class precision is

available [1].

Remote sensing classification problems are difficult to solve because of many issues. A sampling of

these pitfalls includes spectral and spatial noise in the geographic areas of interest, loss of usable data

due to cloud cover, noisy multi-temporal datasets, and noisy human and machine-generated class

labels. A multitude of complex regional characteristics around the globe, e.g., the particulates in

different coastal regions, terrain effects, and open ocean effects and other disturbances also affect the

analytical chain associated with knowledge acquisition and analysis techniques. These problems

contribute to the difficulty of solving computational classification problems in the area of satellite-

based optical remote sensing.

This research examines a method through which classification knowledge for potential use in remote

sensing applications may be acquired from multi-spectral datasets that are routinely used by

researchers and government policy leaders. This research is based on contemporary machine learning

techniques that have been combined to develop a novel system that is based on recent thought in

biological evolutionary development. In particular, it is based on the observation that biological

evolution has provided life as we know it with successful means of navigating the data-rich pitfalls

and rewards associated with day-to-day survival in a harsh environment. It is hoped that the

application of the ideas associated with biologically complex structures beyond the typical realm of

Page 20:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 20 of 249

09/28/10

neural networks can provide novel means for processing human-produced, information-rich

constructions in the future.

The idea behind this dissertation is the research and development of an algorithm that creates a self-

organizing classification network. This classification network can be based on a mixture of

heterogeneous machine learning constructs such as self-organizing feature maps (SOFM), artificial

neural networks (ANN), and support vector machines (SVM) [2]. The biological context for self-

organization and communications between the simple processing constructs, or cells, in the network

is that of genetic regulatory networks and evolutionary development [3,4,5]. Each processing cell

consists of an artificial genome that has excitatory and inhibitory switches that are controlled through

the communication of artificial proteins in a simulated environmental lattice in which the cells reside.

Within each cell, protein switches control the expression of particular proteins from each processing

cell. These proteins diffuse through the lattice and in turn are used for communications between the

processing elements. This communication is based on an artificial protein chemistry with

concentration levels.

Self-organization and evolution of the network occur on several levels. The high-level topology of

the network- both the number and types of the simple processing cells (SOFM, ANN, or SVM)- can

change as can the internal organization of each processing element. At this lower processing-element

level, this could include parameters such as the choice of kernel used for a particular SVM element.

Artificial proteins used for communications between processing elements as well as between the

“outside” environment (the input data patterns) and the classification topology also adapt to the

application domain. These communications and environmental proteins are released into the

classification lattice if their corresponding genes are switched on and expressed.

1.1. Problem statement

Claim: Biological inspiration is a powerful paradigm in classification and hybridization introduces

interesting and useful qualities. It can provide powerful tools for solving relevant satellite image

classification problems.

We will investigate in detail as to whether the combination of an artificial genetic regulatory network

(GRN) and a basic machine learning element in a rudimentary self-organizing network is effective

when applied to binary classification in a multi-dimensional space; i.e. using multispectral feature

Page 21:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 21 of 249

09/28/10

vectors acquired from an optical satellite image. This detailed case study will examine how our novel

hybridization idea works and also how to set it up for a class of practical classification problems.

1.2. Delimitations of the research

• Simple GRNs will be constructed o Simplified artificial protein representation o Protein interaction through a simple protein “chemistry” o Size and complexity are limited by the available computational resources

• Self-organization o Facilitated by evolutionary computation; i.e. particle swarm optimization; and a small

set of self-organizing rules inspired by biological evolutionary development and statistical analysis

o Rudimentary classification training algorithm is based on repeated presentation of samples to the classification network in the PSO framework.

• Accuracy for a real-world binary cloud/no-cloud classification problem will be addressed

• Robustness for benchmark problems will be addressed in the areas of datasets with noisy features and datasets that have missing features

• A single binary classification using remotely sensed multispectral optical data will be examined

• Comparisons are limited to simple implementations of SOFM machine learning kernels with research into SVM and ANN kernels left for future research.

• It is not the purpose of this research to delve deeply into the merits of one SOFM training algorithm versus another one.

1.3. Key contributions

• Novel application of a hybridized biological construct with machine learning to a practical computational classification problem

• Determination of the effectiveness of a simplified GRN applied to multi-dimensional classification.

• Artificial proteins communicate classification information and results to and from the cellular machine learning kernels.

• Training of a GRN via particle swarm optimization (PSO).

• Application of a GRN-based classification system to a real-world multispectral remote sensing problem domain, i.e. cloud detection in optical satellite imagery.

• Performance comparison of HeBIS and a SOFM-only classification algorithm on a remotely-sensed multispectral dataset with unadulterated features, noisy features, and deleted features.

1.4. Organization of this thesis

The thesis is organized as follows:

Page 22:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 22 of 249

09/28/10

Chapter 1 outlines the objectives and limitations of this research and lists the key contributions of this

dissertation.

Chapter 2 reviews the current state of the art for classification systems based on machine learning

techniques that are embedded in self-organizing structures- as is the case with HeBIS. As such, this

chapter provides direct linkage to the origins of the core ideas used in the HeBIS architecture.

Specifically, the areas examined are Artificial Neural Networks (ANN) and Self-Organizing Feature

Maps (SOFM). ANNs and associated research into their self-organization are presented as important

background in addition to being an introduction to the SOFM theory. Promising new developments

are discussed as well as the problems associated with each of these machine learning paradigms.

Attempts to alleviate these problems through the use of a “bare bones” Genetic Regulatory Network

(GRN) based on an artificial protein chemistry form the basis for the remainder of the dissertation

using the HeBIS self-organizing system. Information on this GRN is presented from a computational

development viewpoint and an overview of the current state-of-the art in this research domain is also

included. This information is presented in the context of the cell-to-cell and intra-cell

communications that are enabled by a computational environment based on artificial proteins. A

literature overview of GRNs applied to classification problems is also presented.

In Chapter 3, a detailed architecture of the Heterogeneous Biologically Inspired System (HeBIS) is

presented. This treatment includes information on the protein environment and lattice structure of the

system; the artificial proteins, their different types and encodings, and the associated simulated

protein chemistry; and the basic processing cells. Also examined are the inherent and learned

behaviors that each cell can acquire through the protein reactions within the simulated environment.

Finally, the classification training algorithm is examined. The underlying particle swarm optimizer

that is used for system optimization is also discussed and sample training and classification

processing data flows are given to solidify the presentation of this material.

In Chapter 4, the simulation results are presented. Through these simulations, HeBIS classifications

are examined and compared to classifications based on SOFMs. Simulation results are also presented

which characterize salient properties of various instantiations of the HeBIS architecture.

Chapter 5 discusses and summarizes the comparison results in detail.

Page 23:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 23 of 249

09/28/10

Finally, in Chapter 6, the dissertation concludes with an outline of potential avenues of future work in

this research area.

Page 24:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 24 of 249

09/28/10

2. Literature review Hybridization Background

Algorithm Discussion GRN Analyses

GRN Training with PSO GRN Action Analyses

Remote Sensing Background Remote Sensing Application – Analyses and Comparisons

Robustness Analyses

The list of research areas that this dissertation touches upon is quite extensive. This stems from the

fact that the idea that biology has something to teach engineers and computer scientists has become

more accepted by researchers over the last several years. During the last decade, basic biological

research has become cheaper to perform and its volume has increased. This is now being coupled

with an exponential increase in computing power and information processing. The intersection of

these disciplines has seen a fertile exchange in ideas between the biological researchers and the

“information processors”. This dissertation is itself a result of this exchange and it focuses on the

attempt to apply biological principles to knowledge extraction; specifically the area of automated

multi-class classification. To do this, an overview of work that is directly applicable to this research

is required.

This review is divided into four sections that concentrate on machine learning and evolutionary

development within the context of self-organization and classification. Section 2.1 introduces

machine learning as it is applied in the specific domains of general classification, artificial neural

networks and self-organizing feature maps. Section 2.2 covers evolutionary computation with

emphasis on particle swarm optimization. Section 2.3 presents biological and artificial evolutionary

development and provides an overview of the research and applications associated with pattern

classification and creation. Finally, Section 2.4 summarizes the advantages and limitations of the

outlined techniques and proposes that there may be improvements in pattern classification if ideas

from these different research areas are blended together through HeBIS, the Heterogeneous

Biologically Inspired System.

HeBIS research is based on simple SOFM pattern recognition kernels (cells) with a GRN-based

communications infrastructure wrapped around them. The emphasis in this research is to determine

whether a GRN can be successfully used to create a classification network which can be used as the

basis for further research, not to examine the relative advantages or disadvantages of different

subclasses of this or other simple processing kernels.

Page 25:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 25 of 249

09/28/10

This review only provides a concise overview of these topics. Appropriate references are included

for further detailed examination.

2.1. Machine learning and self-organization

2.1.1. Classification overview

Machine learning entails methods by which computers may be programmed to learn. The accuracy

and precision of a computerized classification system are the primary attributes of such a system.

Learning tasks can be classified into analytical and empirical techniques. With analytical learning,

no external experiences – data and environmental descriptions- are required whereas empirical

learning explicitly requires the use of external data and experience [6]. This research is primarily

concerned with empirical learning in both supervised and unsupervised learning environments for

classification applications.

Classification is the process through which an object is mapped to a specific class within a set of

classes that has been defined for the problem. In this research, an object and its associated definition

or class is referred to as a labeled example or exemplar. The set of labeled exemplars constitute the

training set of data in which the object is a feature vector of many descriptive numerical features that

is mapped to a specific class label. These training data are applied to a given learning algorithm and

the result is a specific instantiation of a classifier. In turn, this classifier is evaluated for its precision

and accuracy by applying it to a test data set that is composed of a separate set of labeled examples

that have been taken from the same underlying statistical distribution as the training set.

Classifiers should generalize well to datasets that they have not been directly trained on. In other

words, a good classifier is one that, once it has been trained on a small training set, may be used to

effectively classify larger sets of data. Classification rate is the statistic that is the percentage of test

examples which are correctly classified. The misclassification rate is the converse, those test

exemplars that have been misclassified by the classifier [7]. These statistics are further refined in the

cases where successive classification decisions are not independent and when the classification

decisions are not equally important [6]. The latter leads to the development of the Receiver

Operating Characteristic curve, ROC, that is used to gauge classification performance for ranking

classifiers [8]. Through an ROC, the performance of a classifier is examined by changing the

threshold that is used to decide between two classes. As this threshold changes, one can construct an

Page 26:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 26 of 249

09/28/10

ROC from the false-positive classification rates and the corresponding true-positive classification

rates.

Empirical learning systems, whether supervised or unsupervised, trade performance between three

factors: the complexity of the classifier, the amount of training data, and the generalization ability of

the system when it is applied to new, unseen exemplars. As the classifier’s complexity increases, its

generalization ability increases, peaks, and then decreases. As more training data become available to

the classifier, more detailed information becomes available about the problem’s statistical manifold.

However, as the complexity of the classifier increases, the system’s generalization accuracy will

increase and then decrease after a certain point is reached. These points are noted in Figure 1.

Figure 1. Performance of empirical learning systems.

Page 27:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 27 of 249

09/28/10

Low variance and high variance classifiers are defined respectively as systems that exhibit a small

degree or a high degree of change in classification performance as different (and noisy) exemplars are

presented and tested [7].

A high bias system is defined as one which exhibits high classification precision on the problem, but

with low recall and a low bias system is one which exhibits low precision with high recall, with

recall =t p

t p + fn

ffffffffffffffffffffff,

(1)

precision =t p

t p + fp

ffffffffffffffffffffff,

(2)

where t p is the true-positive rate, fp is the false-positive classification rate, and f

n is the false-

negative classification rate.

A low bias system essentially can represent almost any classifier whereas a high bias system is not

complex enough to represent the optimal classifier.

Many mechanisms exist through which classifier complexity is matched to the complexity of the

training data [9,10,11].

Unsupervised learning requires no exemplar-class training pairs because the interrelationships

between the examples’ features are automatically categorized and clustered by these types of

algorithms according to a set of rules that is defined before the feature vectors are presented to the

learning system [12, 13].

Typically, machine learning algorithms make weak or no assumptions about the training data.

Therefore, machine learning techniques generally require a large number of training data so that the

problem’s statistical manifold can be adequately sampled. However, if domain knowledge is applied,

the size of the required data set (for a given level of classifier precision) is typically much less than it

Page 28:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 28 of 249

09/28/10

is in the case in which no knowledge is used. This introduction of bias into the system is risky,

however, since the a priori knowledge must be correct or the added bias may preclude the discovery

of an accurate classifier.

2.1.2. Artificial neural networks

Historically, research into artificial neural networks (ANNs) has been motivated by the differences

between mammalian brains and human-engineered digital computers. Researchers have typically

focused their efforts in two ways. The first is an attempt to better understand how the brain works by

simulating its topology through models of varying complexity. Secondly, researchers have attempted

to mimic the brain’s operation in a quest to improve engineered information processing systems.

It is through this second camp of researchers that the modern variants of artificial neural networks

have been utilized for complex and nonlinear applications such as pattern recognition. The

processing and self-organizing abilities of even the smallest mammalian brain outstrip current

supercomputers given almost any applicable performance metric.

A small sampling of the early work in the field includes [14,15,16,17,18,19].

A generic artificial neural network is composed of a collection of simple processing elements that are

interconnected. This type of architecture is one that is extended in this current work with the HeBIS

network’s processing cells that are interconnected through a GRN. Each of the simple processing

elements in an artificial neural network is called an artificial neuron and is based on a simple

mathematical model of a biological neuron.

Figure 2 shows an artificial neuron that receives a set of numerical inputs, applies a multiplicative

weighting function to each input and then sums these results over all of the weighted inputs to the

neuron.

Page 29:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 29 of 249

09/28/10

Figure 2. An artificial neuron.

This individual weight is called a synaptic weight and it mimics the excitatory and inhibitory

responses at the input of a biological neuron. In the artificial case, a negative-valued weight acts to

inhibit that input whereas a positive-valued weight excites that input in the artificial neuron. The

summed result of these weights is then nonlinearly mapped through a normalizing activation function

to the neuron’s output. This activation function is typically chosen to be a scaled sigmoid function.

At this point, the output signal (a number) is either passed on as an input to another neuron or is the

output of a layered feedforward network as in Figure 3.

Figure 3. Artificial Feedforward Neural Network.

A feedforward neural network is typically composed of three layers: an input layer, a hidden layer

and an output layer. In its most general form, the neurons in these layers are interconnected within

the layers and also between the layers via synapses that are excitatory or inhibitory. These responses

Page 30:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 30 of 249

09/28/10

are controlled by the previously mentioned synaptic weights. A properly-sized feedforward network

composed of three layers can theoretically approximate any arbitrary function [20]. Similarly, the

idea is that HeBIS forms a layered network of interconnected processing cells, albeit one that may not

be as apparent as that of an ANN.

ANN architectures are mature and can be differentiated according to the following characteristics:

• Neuron Models,

• Synaptic interconnection (network) models), and

• Training paradigms.

Common artificial neuron models use simplified versions of the actual operating properties inherent

in the biological neuron. For example, biological neurons appear to actually process their inputs and

outputs according to a pulsing signal model. This neuron model has been mostly ignored by the

computational research community but interest has increased recently [21]. Besides the common

sigmoidal activation functions, other nonlinear functions have also been examined [20].

Synaptic interconnection models define how the neurons in the different layers of a network are

connected to each other. Types of architectures include feedforward networks in which the inputs to

the network are processed by the input layer and then “fed forward” to the hidden layer and then the

output layer. Recurrent neural networks are oscillatory and function by feeding information

backwards through the network or directly back to the originating neuron [20,22]. Another model is

the Self-Organizing Feature Map which is described in detail in a separate section [23].

Many paradigms for the training of artificial neural networks’ synaptic weights have been researched.

Backpropagation is a workhorse training technique for the neural network community [24]. Other

neural network training techniques include [25, 26, 27, 28, 29, 30].

Researchers have also applied classical mathematical tools to the training issue. Sequential Monte

Carlo methods are used to train the ANN as each new training example is presented to the network

[31]. Iterative training such as this is useful in instances when the training datasets are large, consist

of thousands (or more) of features, and when the dataset’s statistics are time-varying and/or non-

Gaussian. [31] uses Monte Carlo sampling to characterize the training set’s probability distribution

for such an iterative and time-varying process. This technique, HySIR, was shown in 2000 to have

Page 31:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 31 of 249

09/28/10

only a linear dependence on the number of sampling trajectories used and thereby has realizable

computational requirements.

Non-traditional training methods are also used to train artificial neural networks- as are also used with

HeBIS. Vieira in 2005 [32] used a genetic algorithm (GA) coupled with elitism to train the Iterative

Artificial Neural Network (IANN). The authors researched the problem of fitting the size of a neural

network to the complexity of the problem at hand- an issue that HeBIS addresses through self-

organization. They proceed on the basis that ANN training based on high-dimensional data is

difficult and that an overly complex network is probably over-fitted to the training data. Network

pruning or dimensionality reduction of the training features are common techniques used to match the

complexities of the system to the problem [20]. The IANN uses an iterative training process on a

layered ANN in which each layer contains the same number of neurons. The number of layers, the

number of neurons in each layer, and the connection weights are determined with a genetic algorithm

that minimizes the quadratic error of the network’s output. Vieira believes that this iterative

algorithm’s advantages are its robustness, ease of implementation, and the ability to use high-

dimensional feature vectors, even when large numbers of features are irrelevant [32].

Evolutionary development (development or evo-devo) is described in detail in Section 2.3. However,

it is briefly mentioned in this section on neural networks because it has been found to be useful when

applied to the construction of a simulated robot’s morphology and the creation of its artificial “brain”

[33,34]. Development facilitates the creation of a compact set of rules that can be used to create

complex structures. The individual rules (roughly equivalent to biological genes) in the derived rule

set are not mapped on a one-to-one basis to the actual realized constructions (morphologies) that are

created by the interpretation of the rules. Four improvements in abilities and topology generation for

artificial neural networks point to the importance of incorporating evolutionary development into the

HeBIS algorithm.

First, large and complex networks can be created through the process of development by using a

small number of genes. Second, evolutionary development appears to be the basis of the gradual

development of biological morphological patterns that are complex and fractal such as the

mammalian brain’s synaptic connections. The term, fractal, in this research describes a pattern or

system in which structures or patterns are replicated on different scales. Applying such biological

design principles into human-engineered systems has been shown to be a valid effort [35, 36].

Artificial grammar systems such as L- systems have been found to be capable of introducing fractal-

Page 32:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 32 of 249

09/28/10

like morphologies in artificial systems [37, 38, 39]. With grammar-based systems, strings of

characters represent the state of the neural network (i.e. neuron weight, bias, and network topology)

and are changed according to a set of string parsing rules. These parsing rules modify the strings and

in turn, the instantiation of the neural network is changed. Third, because the linkages between

genotype (the set of processing rules) and phenotype (the realized structure: neural network, physical

body, etc.) in a development-based system are not direct but are complex, nonlinear, and indirect,

this allows radical alterations in neurological network architectures to be realized through slight

changes in the artificial genes. Finally, development allows the formation of modular functionality

within the network [40].

These works point to the utility of basing HeBIS on evolutionary development principles, in

particular with respect to ANNs.

Jung researched a topographical method for development-based neural network creation and reported

on it in [41] in 2005. His method has been shown to create artificial neural architectures that mimic

the neuron architecture patterns that are found in the mammalian visual cortex. Importantly, these

ANNs are based on a compact genetic representation that recreates modular fractal neural modules

(patterns) that are present in mammalian brains.

The Self-Optimizing Neural Network (SONN) is an ANN training paradigm through which a neural

network topology is optimized by creating subnetworks of the various training classes in a problem.

These subnetworks are created by statistically determining the important input features and then

combining and mapping these features through separate networks to the output. It is another method

by which a neural network’s complexity (i.e. topology and weights) may be tuned to the complexity

of the data, in this case without explicitly using biological developmental principles [42,43]. HeBIS

also has behaviors that allow for statistical sampling of the classification manifold, but from an

artificial protein perspective. This is discussed in detail in Chapter 3. The SONN is reported to have

no requirements for a prior configuration of training parameters and is directly applicable to different

datasets. The authors indicate that the technique derives minimally complex networks while still

retaining expressive generalization. The study’s first author, Hizyk, does mention that SONN

construction “can be compared to the processes that take place during brain development” [42]. It is

still unclear as to what SONN’s performance is when compared against classical ANN architectures,

but it is interesting nonetheless.

Page 33:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 33 of 249

09/28/10

Kitano’s Neurogenetic Learning algorithm (NGL) [44] is an extension of his grammar-encoding work

for neural networks that are trained by GA [45]. Through these methods, the creation of a neural

network topology and weight tuning are performed simultaneously. In [44], he showed that an

initialization of the neuronal weights decreases training convergence time. This network initialization

is controlled by a GA and a grammar and is used in the gradient-descent backpropagation algorithm

that trains the neural network.

Neural network classifiers have routinely been applied to problems in remote sensing of the

environment to perform, for example, land cover, ocean phytoplankton, and atmospheric composition

classifications [46, 47, 48, 49, 50, 51]. Recently, Chin implemented the Statistical Self-Organizing

Learning System (SSOLS) [52]. It combines a functional-link neural network [53] with a self-

organized hidden layer that adapts to the training dataset’s statistics. The number of neurons in the

hidden layer is increased according to the results of a series of statistical hypothesis tests performed

on the data as they are presented to the network. A combination of cross-validation and a Student’s

T-test is used to minimize the effects of overfitting and to keep the size of the network to a minimum.

SSOLS is applied to a hyperspectral dataset that consists of 191 optical bands in the 0.4 – 2.4 mµ

portion of the visible and infrared spectrum. The data are classified on the pixel level into seven land

cover classes that include background features, corn, soybean, wheat, alfalfa, pasture, and a sensor-

distortion class. In a comparison of training and testing accuracies between the SSOLS, a radial basis

function (RBF) support vector machine (SVM), a matched filter, and other classification methods, the

SSOLS achieved the highest accuracies in conjunction with the SVM.

Another type of artificial neural network is the functional neural network. This is a flat feedforward

network that has no hidden layer [53]. Each feature of the input vector that is presented to the

network undergoes a linear transformation with scaling and added bias. The transformed feature is

then mapped through a function that is typically a neuronal tanh sigmoid. The mapped features are

then simply scaled and summed to create the output of the network

The HeBIS processing network can also be somewhat described as a functional link network since

each cell encapsulates a specific mapping function.

In summary, artificial neural networks posses the following advantages and disadvantages:

Page 34:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 34 of 249

09/28/10

Advantages

• Mature training methods exist

• Straightforward iterative training

• Domain specific knowledge not required Disadvantages

• Long training times for datasets with large number of features

• Difficult to tune with respect to the choice of the number of elements in the middle layer

• Difficult to analyze; nonlinear analysis of neural weights

2.1.3. Self-organizing feature maps

As part of self-organizing systems, Self-Organizing Feature Maps (SOFM), also known as Self-

Organizing Maps (SOM), are introduced here separately from ANNs because of their importance to

this work. The HeBIS network can also be thought of as a SOFM, but with individual processing

elements that are more complicated than the simple SOFM neurons. Several of the general ideas in

this section are incorporated into the HeBIS architecture.

SOFMs are biologically inspired processes that construct structural representations of high-

dimensional data. The SOFM described by Kohonen [23] is composed of simple neurons embedded

in a lattice that is typically of much lower dimension than the input data. Metric relationships defined

between these neurons are referred to as weight vectors. These weights are modified as the feature

vectors are presented to the SOFM on a one-by-one basis. Presentation of a multi-dimensional

feature vector to the SOFM consists of mapping the vector to the neuron with the weights that

matches the vector most “closely” according to a defined distance metric; with the Euclidean metric

being commonly used. This matching process is somewhat analogous to vector quantization, but

with the addition of a weight-updating process that is defined over a subset of the surrounding

neurons. This is known as the neighborhood updating function. This neighborhood function can take

many forms [23].

After iterative presentation of data vectors is complete and the SOFM has converged [54], it now

contains a structural representation of the input dataset that maintains a degree of topological parity

with the original data. The weight vectors of the converged SOFM are collectively referred to as the

feature map. Typically, a SOFM is defined in two or three dimensions and is used in a variety of

applications. These include, visualization of the “important” similarity relationships between the

Page 35:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 35 of 249

09/28/10

input, high-dimensional data vectors, pattern classification, and function approximation [55, 56].

Figure 4 shows a schematic of the input/output relationships and a sample of a map whose neurons

are being used to map a multi-class classification problem for remote sensing purposes. The input

vector is represented as X with the associated weights, wij, describing the links from the input vector

to the output cells in the SOFM N.

More formally, a basic SOFM may be defined as a set of neurons that is arranged in a lattice of

specific shape and dimensionality. The neuron’s position in the network can be denoted by the

vector, s , and the weight vector of this neuron by ws. The set of all ws

for all the neurons in the

network comprise the feature map. A pattern vector taken from the input data space is denoted as p .

The SOFM is trained by taking a feature vector, p , from the input data space and determining which

of the neurons in the lattice is “closest” to the input pattern as

W c@PN

N

N

N

N

N= minS

W s@PN

N

N

N

N

N

(3)

where the weight vector, wc, is associated with the neuron at sc

that is found to be the closest match

as in ( ). The weight vector of the neurons located in a defined neighborhood of this closest neuron

are then updated with the neighborhood updating function. In the original incremental SOM [12],

this neighborhood function is updated at discrete time intervals, t, and is defined as

W n t + 1` a

=W n t0

b c

+ αh C,S n

b c

P t` a

@W n t` a

B C

(4)

where NW marks the weights in the defined neighborhood of neurons, NS , and h() is the relaxation

function that depends on the neurons’ position in the neighborhood of the closest matched neuron.

Page 36:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor

The relaxation function, h(), is a smoothing kernel that is defined over the

neighborhood. Its spatial extent and behavior over time (for convergence h()

decreases as the distance metric increases) define the elasticity of the SOFM and how the mapped

lattice responds to input data vectors [23

as

Initialization of the SOFM is either random or may be chosen in some fashion if one has some

knowledge of the input data manifold [54

Figure 4. Self-Organizing Feature Map.

The relaxation function, h(), is a smoothing kernel that is defined over the

neighborhood. Its spatial extent and behavior over time (for convergence h()

decreases as the distance metric increases) define the elasticity of the SOFM and how the mapped

sponds to input data vectors [23]. The learning rate, alpha, is bundled in this h() parameter

h ` a= αh ` a .

M is either random or may be chosen in some fashion if one has some

e of the input data manifold [54].

36 of 249

The relaxation function, h(), is a smoothing kernel that is defined over the neuron’s local

neighborhood. Its spatial extent and behavior over time (for convergence h() Q 0 as t Q 1 and h()

decreases as the distance metric increases) define the elasticity of the SOFM and how the mapped

]. The learning rate, alpha, is bundled in this h() parameter

(5)

M is either random or may be chosen in some fashion if one has some

Page 37:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 37 of 249

09/28/10

Since its publication by Kohonen in 1982, many researchers have contributed to the theoretical

underpinnings of the SOFM. Many changes and enhancements to the basic algorithm have been

suggested [57, 58, 59].

One important feature of the SOFM is its ability to preserve topological relationships when high-

dimensional data are mapped to lower-dimensional or equi-dimensional structures. This preservation

of neighborhood has been studied extensively for regular lattices, but irregular lattices were also

examined [60,61]. In [61], the study’s authors, Neme and Miramontes, created many types or

irregular lattices, including several in which locality was defined as a small-world (1/f) network

where f is the distance between connected neurons in the network [62,63]. These SOFM lattices were

trained and topological preservation metrics were calculated for each. Mutual information between

the networks’ statistics and their degree of topological preservation was analyzed. Interestingly, the

report states that topological preservation was found to be better in non-regular lattices than in the

regular ones [61]. An example of topology preservation as training of the SOFM progresses is

presented in Figure 5.

Figure 5. Self-Organizing Feature Map Topology Preservation Example [61].

It is hoped that HeBIS will incorporate topological features of the data manifold into its self-

organized network of machine learning cells.

Vector quantization (VQ), SOFMs, and Learning Vector Quantization (LVQ) are closely related [23].

In [64], the Dynamic Time Warping algorithm (DTW) is proposed by which sequences of features

Page 38:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 38 of 249

09/28/10

may be applied to an SOFM for clustering and an LVQ for pattern recognition. DTW attempts to

compensate for and tolerate input sequence differences due to sequence lengths and spatial variation

by normalizing according to these sequence lengths.

Two limitations exist with the SOFM. The first of these limitations is the generic SOFM’s static map

architecture: the number of neurons and the orientation of the lattice are fixed before training begins.

With data that has unknown characteristics, it can be difficult to determine the SOFM’s architecture

that provides good training results. The second limitation is that the generic SOFM has no ability to

arrange the data according to their natural hierarchical relationships. Since the data from many

problem domains are hierarchical, this limitation precludes using SOFMs for satisfactory analyses in

such cases.

The Growing Hierarchical SOM (GHSOM) is one attempt to addresses these limitations [65]. It is an

outgrowth of prior SOFM enhancements [57,66] that attempted to also address these limitations.

GHSOM modulates both the size of the network at each layer and also decides which neuron will be

selected to seed a more detailed hierarchical map. These selections are dependent on data

representation quality measures for SOFMS such as the Mean Quantization Error (MQE) of a neuron

and the total Quantization Error (QE). With one of these measures and a threshold parameter as

another control, the depth of the hierarchy can be roughly set. The choice of the threshold parameter

can still be somewhat complicated.

The GHSOM and the Growing Neural Gas (GNG) [67] algorithms attempt to improve unsupervised

SOFM learning by dynamically adding neurons to the network. In the case of [65], [67] and others,

the network is only resized after a defined number of training iterations have occurred between

updates. Because of this, these networks may not respond well to data distributions that are non-

stationary.

Marsland, Shapiro, and Nehmzow have developed an unsupervised technique by which the size of the

network can change as the data statistics change [68]. Their Grown When Required (GWR)

algorithm responds to non-stationary aspects of the data by adding new neurons, but it stops its

growth when the data’s statistics have become stationary. The GWR was compared with the GNG

and the Restricted Coulomb Energy Classifier (RCE) [69] in Marsland’s study.

Page 39:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 39 of 249

09/28/10

The Generative Topographic Map (GTM) is an alternative to the SOM [70]. This algorithm, unlike

the SOFM, works by optimizing a specified objective function. Whereas training convergence is not

guaranteed for the SOFM, the authors state that it is guaranteed for the GTM. Also, the GTM

explicitly defines a probability density function (pdf) on the mapped data space, something which the

SOFM does not explicitly accomplish. Similarly, though, the GTM clusters input data, but

accomplishes this task through the modeling of a small number of latent variables.

Yet another extension to the SOFM algorithm is the SOM of SOMS, also known as SOM2 [71].

Whereas a SOFM operates by highlighting the differences and similarities between individual feature

vectors in a dataset, the SOM2 has a different focus. It attempts to examine the similarities and

differences directly between the distributions of classes (i.e. the different manifolds in a multi-

dimensional space) in the input data space.

SOFMs have also been used to cluster and aid in the analysis of temporal sequences of data [72,73].

One of the more recent, the Time Enhanced SOM (TESOM) is described in [74]. It requires the

tuning of many parameters for adequate training.

SOFMs have also been used in remote sensing applications. [75] discusses the application of an

SOFM to the classification of vegetative land cover with a combination of multispectral data from the

Airborne Thematic Mapper (ATM) and multifrequency/multipolarization data from the NASA

Airborne Synthetic Aperture Radar (AIRSAR). [76] compares classification results based on SOFM

and LVQ algorithms that are applied to a LANDSAT Thematic Mapper (TM) dataset and cloud cover

classification is discussed in [77] and [78].

In summary, self-organizing feature maps possess the following advantages and disadvantages:

Advantages

• Inter-feature relationships can be visualized

• Easy to implement

• Not computationally expensive

• Straightforward iterative training

• Simple, not very many training parameters to tune

• Domain specific knowledge is not required

• Relatively easy to analyze

Page 40:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 40 of 249

09/28/10

Disadvantages

• The size of the network is set before training and is not adaptable as the input statistics change

• Multi-dimensional -> lower-dimensional mapping(e.g. 3 dimension mapped to 2 dimensions) may gloss over critical feature relationships

• Convergence is not guaranteed

• Hierarchical data relationships are not mapped in the generic algorithm

2.2. Evolutionary computation

In the current work, Genetic algorithms (GA) and Particle Swarm Optimization (PSO) are introduced

primarily as tools to facilitate a broad search of potential HeBIS genomes. The evolved genome

denotes the proteins that are used in the HeBIS communications infrastructure, its Genetic Regulatory

Network (GRN). The genome also sets the evolved capabilities of each processing cell. This search

is necessary to provide a fast ability to examine a large set of innovative solutions. Both the GA and

PSO belong to the class of techniques in which populations of individuals are harnessed to find and

optimize problem solutions [79].

With a genetic algorithm, which is one of many optimization strategies in evolutionary computation

(EC), artificial evolution is used to search through the problem space [80]. A population of

chromosomes is mated and mutated using biological evolution as a guidebook. In Swarm

Intelligence (SI), PSO guides a population of potential problem solutions by communicating

information about the search space between members of the population [81]. Both methods have

been the subject of much research in the application and pure research communities in engineering as

well as the biological sciences. HeBIS uses a PSO as the basis of the training algorithm and hence

this will be discussed in further detail. For more details on genetic algorithms, please see

[82,83,84,85,80,86,87,88,89,90,91,92].

2.2.1. Particle swarm optimization

Particle Swarm Optimization is a variant of EC that is useful for multi-dimensional nonlinear

optimization [90]. The algorithm is traced back to Kennedy and Eberhart’s initial research in 1995

and is closely related to genetic algorithms and evolutionary programming [81]. It is also a

population-based optimization procedure, but the population’s members communicate information

Page 41:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 41 of 249

09/28/10

about the solution space to the entire population. The term swarm refers to the population’s members

(the particles) and is based on defining principles for swarms in the artificial life community [81].

PSO is different from the GA and evolutionary programming strategies in that no selection is

performed on the potential solutions in the population. Each particle is defined as a point in the

problem’s N-dimensional solution space and each particle flies (searches) through this solution space.

The velocity at which the particle flies through the space is dynamic and is adjusted according to the

particle’s best fitness value as well as the best-found fitness values of other particles within a defined

neighborhood. With the ith particle, Xi, a vector defined as

X i = x i1 ,x i2 , …,x iN

( )

(6)

in an N-dimensional space, the particle’s own best fitness value (at this time step in the iterative

process) is defined at the location

P i = pi1 , pi2 , … piN

( )

(7)

and the index over all P that represents the global best fitness value is denoted as g. Finally, with the

local velocity of Pi described as

V i = v i1 ,v i2 , …,v iN

( )

(8)

the update equations for the core PSO algorithm are

vin

= vin

+ C1 rand ` a pin@ x

in

` a

+ C 2 Rand ` a pgn@ xin

` a

(9)

and

xin

= xin

+ vin

(10)

Page 42:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 42 of 249

09/28/10

where C1 and C2 are constants and rand() and Rand() are random functions defined on [0,1].

Mutation in an EC context occurs in the PSO through the changes in velocity for each particle as

successive generations of particles are updated through (9) and (10).

PSO is competitive with GA on classical GA benchmarks [93]. Applications of particle swarm

optimization are diverse, ranging from recent research in multi-objective optimization [94] to remote

sensing applications such as ocean color reflectance inversion [95]. Research into hybrid

combination of PSO and GA is also ongoing [96].

Figure 6 shows a system of several swarm particles over 25 time iterations as they explore an

optimization problem that has a complicated three-dimensional manifold.

Figure 6. Particle Swarm Optimization Algorithm [90].

In summary, particle swarm optimization algorithms posses the following advantages and

disadvantages:

Advantages

• Easy to implement

• Simple algorithm can be applied to many different types of problem sets

• Domain specific knowledge is not required

• Fairly mature (real world and multi-objective problems)

Page 43:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 43 of 249

09/28/10

• Typically a faster optimization process than genetic algorithms

Disadvantages

• Accuracy tradeoffs required for global and local searching

• Not mature for multi-objective problems

2.3. Biological and artificial evolutionary development

Biological evolutionary development, commonly referred to as evo-devo, is the term for the processes

which shape the phenotypic development of an entity based on its genotype [97]. Although it is a

biological phenomenon, in this research it is directly applied to the creation of the HeBIS information

classification system. It provides the basis for self-organized control and communications in the

HeBIS classifier. Biological evolutionary development outlines the process through which a

fertilized human egg can grow from a single-cell entity to an entity that is composed of trillions of

independently functioning cells [98]. The fact that the instructions required to create a fully-

functioning and sentient human are encoded into such a compact structure is one of the bases of this

research. These compact structures, the cell’s DNA-based chromosomes, consist of a limited

alphabet of four distinct organic chemicals (complementary chemical base pairs consisting of

adenine, cytosine, thymine, and guanine) form a book of life by using only 4 billion characters.

These 4 billion characters encode for 20 000 to 30 000 genes within a collection of genes defined as a

genome. Each of these genes, when expressed, can form an organic chemical known as a protein.

From this relatively small number of genes that code for individual proteins, a complete and complex

organism is coded. These proteins constitute the control and communications infrastructure for

biological processes [98]. Evolutionary development attempts to explain the apparently huge

information density contained in DNA by placing this information in the context of a physical,

temporally changing, chemical environment. It is through this environment and the complex

signaling network which comprise it that the entity described by DNA may grow into a mature entity

composed of trillions of cells [98].

It is this physical environment which provides a rich level of chemical interaction, evolved complex

spatial and temporal protein-to-protein signaling (communications) networks, and a level of self-

organization that human engineers would like to exploit for their information processing systems.

The power of biological evolutionary development appears to stem from the constant reuse of

proteins within evolved signaling networks [99]. These evolved signaling networks are called genetic

regulatory networks (GRN). The term regulatory is used because it is through the GRN that a

physical body’s biological processes are regulated and controlled. It is the idea of a GRN, in an

Page 44:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 44 of 249

09/28/10

artificial form, that provides another of the bases for this research. The GRN and the expression of

proteins provide the control and signaling through which an actual organism is built This concept

may be used as the basis to build a physical organism such as evolved, fully functioning robots.

However, it can also be applied to design a control structure for information classification which is

relevant to this research

HeBIS’ self-organizing abilities are based on five key aspects of biological evolutionary development

as defined by Bentley and Kumar in [37]:

• Cleavage divisions- The process by which a zygote goes through a series of divisions which create more cells within the volume of the original cell.

• Pattern formation- The spatio-temporal process through which the layout and cellular activities of an organism are organized. The concept of positional information enables individual cells to be positioned within an organism according to prescribed chemical diffusion gradients.

• Morphogenesis- The process through which an organism grows and enables its 3-dimensional body plan.

• Cellular differentiation- This idea enables cells in the organism to acquire differing properties such as skin cells, kidney cells, or neurons. These properties are ultimately caused by the morphed changes in gene expression and signaling in the individual cell. They are caused by, among other factors, intercellular communication and asymmetric division during cell fission.

• Growth – The increase in mass of the organism through the pertinent examples of cell fission and proliferation.

These five fundamental aspects of evolutionary development utilize a cellular signaling mechanism

for both intra-cellular and inter-cellular communications. These signaling mechanisms consist of

both short-range and long-range responses and are collectively referred to as the cellular genetic

regulatory network (GRN) [98].

HeBIS incorporates an artificial GRN that is derived from biological GRN attributes. Because of

this, it is important that pertinent details about biological GRNs be discussed. What follows is an

overview of the salient biological concepts.

Cell signaling as enabled by the GRN, is predicated on the presence of the protein. A protein is a

linear organic polymer that is composed of 20 building blocks called amino acids. A protein is

composed of a sequence of these amino acids and, once created, acquires a distinct, complex three-

dimensional shape that defines how it will chemically react with other proteins. This research uses

Page 45:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 45 of 249

09/28/10

the construct of an artificial protein for communications of control information and data within the

HeBIS classification network.

A protein is ultimately produced within the cell by its ribosomes, but it is defined by a specific gene

located on the cell’s chromosomal DNA. Each DNA-encoded protein is defined by a sequence of

codons upon which each of the codons is defined as a 3-base pair sequence of DNA that codes for a

single amino acid. Artificial genes are used in an artificial genome in this research.

For a protein to be produced, the protein description as defined by its codons must be transcribed to

the ribosomes via an intermediary, messenger RNA or mRNA. However, before this occurs, the gene

that controls the protein must be activated or expressed [98]. This expression is controlled by a

region of DNA base pairs that is adjacent to the codons that code for the specific protein. This region

is referred to as a promoter region or the cis-regulatory region. This region responds to the presence

of particular proteins and acts as a switch that activates (turns on) or inhibits (turns off) protein

transcription by this gene. These “particular” genes are called regulatory proteins and are emitted by

other genes that have been expressed. This system provides a complex (and multi-tasked)

communications paradigm that can create control chains and negative and positive regulatory

feedback loops over long periods of time and across distant regions of the physical organism.

Through these GRN interactions that recursively control protein expression (and inhibition), an

organism undergoes development of its body plan, parts, and initial brain synaptic connections

(neurogenesis) [5]. Certain genes and their associated GRN have been found to have been conserved

across biological evolutionary time and across species. This set of genes, the hox genes, appears to

compose a compact and shared evolutionary toolkit that controls morphology [45]. Day-to-day

operation aspects of the organism, after it has developed fully, are also controlled by a GRN.

The idea of such a compactly coded, robust, and adaptive toolkit that is based on a GRN with

multiple, layered feedback and feedforward communications and control loops is very important to

this research. It is useful for artificial neurogenesis and also for the development of the HeBIS

artificial information processing structure.

The incorporation of a GRN into an evolutionary computation system allows a complex and

nonlinear mapping to occur between a genotype and the realized phenotype. This is important

because discovery and optimization systems based on evolutionary computation typically force the

genome to become more complex as the problem’s complexity increases. This is because the

Page 46:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 46 of 249

09/28/10

problem components are generally linearly mapped back to individual artificial genes. Therefore, as

the problem or design space becomes more complex, the number of required genes increases

prohibitively. This scaling issue is exacerbated by the issue that the human designer must also

explicitly inject her own knowledge into the situation with respect to the parameters that should be

searched or optimized to find a solution to the problem at hand. Quite often, lack of adequate

knowledge of the problem domain hinders the finding of a solution or biases solutions towards

human-realized alternatives. The EC workhorse, the genetic algorithm with one-to-one gene-to-

functionality mapping, is the epitome of this scaling problem.

Remote sensing classification is a complex problem. HeBIS research investigates the possibility of

melding the powerful nonlinear mapping ability of a GRN with evolutionary computation to create an

accurate and robust classifier.

Kumar in [37] has coined the term “computational development” to describe the application of

biological development principles to computer science problems. This term can be used to describe

that research and the resulting techniques which couple the EC-derived genotype to its physical or

neurological parameters. In [37], Kumar and Bentley define a set of advantages and disadvantages of

artificial systems that exploit biological development. In doing so, these systems exploit an important

commonality between biological development and evolutionary computation: the desire to

understand self-organized structure. A subset of these defined computational development

advantages and disadvantages includes:

Advantages

• Reduction of the size of the genotype

• Inherent emergence of complexity

• Complex phenotypes based on simple genotypes

• Hierarchical development with reuse of modular functionality

• Adaptability

• Noise robustness

• Regenerative capabilities of damaged functionality

• Regulatory capabilities

Disadvantages

• Computer-based evolution is difficult

• Analyses of the feedback and feedforward loops of the signaling networks within the evolved GRNs are difficult

• Computationally expensive (more than traditional EC methods)

Page 47:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 47 of 249

09/28/10

Much research remains in computational development, where biological development meets

evolutionary computation. Although the research areas in which biological principles are applied are

diverse and deep, the application of biological development is still limited to

• GRN studies and pattern creation,

• Evolved robotic morphology, and

• Artificial immunology.

Principles of biological development have been applied to the creation of artificial chemistries,

research into genetic regulatory networks and pattern creation, robotic morphology, control and

artificial immunology, and remote sensing.

Kitano in [100] was one of the first researchers to incorporate biological evolutionary development

into a computational model based on biological principles. At its core, this model was based on what

Kitano referred to as evolutionary large scale chaos. Through the Super Coupled Map (SCM),

individual cells reacted and communicated with each other on a two-dimensional lattice through

localized and global artificial enzyme reactions. Evolutionary large scale chaos was defined as the

application of a genetic algorithm to the evolution of a cellular metabolism. Kitano showed that

stable cell differentiation occurred as the simulations progressed. This differentiation was expressed

as the types of chemical activity present in the cells, specifically as the cell’s chromosomes

underwent expression and repression. HeBIS incorporates this idea of a lattice structure in which

communications between cells is based on interactions between artificial proteins.

Recently, in 2004, Madina, Ono, and Ikesani [101] studied the evolution of cellular processes that

were based on an artificial chemistry. They modeled their system using a 3-dimensional lattice of 643

(262 144) positions in which their artificial chemistry was based on the discrete diffusion of chemical

particles. While limited to 300 particles per position, the researchers were able to show the

spontaneous creation of membrane-like structures, or proto-cells, as they called them. Importantly,

they found that this formation of complex structures occurred from random initial conditions.

Although Kauffman pioneered the study of the genome as a complex network in 1993 [102] through

the use of random Boolean networks, the computing and biological communities took issue with his

modeling. The problem, it was believed, was that the random Boolean network was limited and did

not share many of the underlying processes of biological GRNs.

Page 48:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 48 of 249

09/28/10

However, in 1999, Torsen Reil at the University of Sussex expanded upon these initial GRN models

by incorporating a simplified but biologically plausible model [103]. In his model, he made use of

template matching and the concept of an artificial gene which was expressed or inhibited in an analog

of biological expression. Reil incorporated the idea that the artificial genes were switched on or off

through the manipulation of a unique switch region that was situated next to each artificial gene in

his artificial genomes. His simplified TATA box (the gene switch in the eukaryotic genome) which is

also known as a standard promoter used simple template matching and was activated if a matching

sequence in the genome was encountered.

Reil’s research is also notable because he discovered attractors of complex and dynamic gene

expression that ranged from chaotic behavior through complex behavior, and finally simple ordered

expression. The most interesting results were those complex instances in which many cycle-lengths

of expression were discovered to be inherent in the GRN. Furthermore, his research indicated that a

high degree of robustness is afforded by a GRN (both biological and artificial). It also indicated that

robust adaptation may occur when natural selection is allowed to act upon the raw properties inherent

in a GRN. The importance of scale-free or 1/f networks are also hinted at in Reil’s research results.

These are important results that are incorporated into HeBIS through the implementation of a simple

GRN.

An overview of formalisms in the study of genetic regulatory networks is contained in [104]. In this

2004 report, Geard outlined GRN modeling approaches and their various advantages and

disadvantages from the perspective of computational expense and biological accuracy.

Over the past few years, Bongard has specialized in the development of robotic morphologies and

neurogenesis for their control. His work is based on a genetic regulatory network in which simulated

genes produce transcription factors that directly affect phenotypic expression and the regulation of

other genes [4]. This GRN makes use of genes that possess promoter sites that are activated only if

the environmental concentration of the matching transcription factor is within an evolved range of

concentration values. Bongard’s research for his 2004 dissertation employed the concept of a two-

stage process of fitness determination. In the first stage, the simulated robot morphology (for a

simulated multi-articulated robot) is grown from an evolved GRN and in the second stage it is

evaluated in a virtual environment [33]. The GRNs were evolved with the use of a genetic algorithm

acting on a population of 300 genomes.

Page 49:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 49 of 249

09/28/10

Bongard discovered that the two-stage fitness determination process allowed evolution to experiment

with morphology and controller (neurogenesis) development separately and in a modular fashion.

The evolutionary success of the various morphologies was found to be due to the early evolution of

modular regulatory networks, the GRNs. Because of this, it was decided to use a two-stage fitness

process for testing HeBIS genomes.

Bentley has also used computational development as the basis for evolving robotic controllers, in this

case for a toy robot known as the wonderborg [3]. Locomotive actions for the robot are keyed

directly to action genes that are expressed through an evolved regulatory network. Information from

the physical world is represented in the environmental simulation as environmental proteins. This

research was the first to utilize a complex artificial chemistry based on the Mandelbrot fractal set. It

also highlighted the concept of a cell embedded in an environment that possesses internal and external

behaviors that are controlled by the evolved GRN. Through evolutionary computation, specifically a

genetic algorithm, GRN-based controllers were grown that successfully guided the robot through a

maze. The controllers are unique in the sense that they automatically created functional modules that

were used with variation for maze navigation. This concept of a module was not hard-wired into the

system, but emerged naturally as a consequence of using a GRN for signaling and regulation-

something which does not occur in GP (genetic programming) or parameterized L-systems [38].

Importantly, this research showed that evo-devo principles allow an information processing system to

efficiently build a system to find a solution [105].

Scaling issues in artificial neural networks have been examined in the context of rudimentary systems

that break the one-to-one genotype-phenotype mapping. In 1990, Hiroaki Kitano developed the

grammar encoding method to address the scalability issues of artificial neural networks that were

created with genetic algorithms [45]. Kitano's research addressed two issues with such a system

when scaled to large problems:

• ANN morphogenesis cannot be captured in a one-to-one mapping and,

• Complex problems lead to a corresponding increase in the complexity of the genome that the GA is optimizing.

The development of this grammar encoding method was based on Lindenmayer’s L-system [38] from

1968. The system was one of the first attempts to separate genotype development from phenotype

Page 50:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 50 of 249

09/28/10

evaluation. This also has been done recently in [33]. Kitano determined that this separation allows

his technique to generate more regular patterns than a direct-encoding method for the GA and that

subcircuits (modules) were formed and exploited by the GA. Kitano also discovered that this method

could be used to create a useful neural network using a shorter length of chromosome than the length

of chromosome that was required when using direct-encoded GA.

Evolutionary development combined with genetic programming has also been used to create simple

visual patterns. Pattern recognition and the ability to model patterns are essential to the operation of

an effective classifier. Therefore, this is why the creation of visual patterns is relevant to the HeBIS

classifier. It is possible that HeBIS will encapsulate or generate the pattern of the training data’s

manifold (or a statistical version of it) in the GRN. Designed to create French flag maps in a toy

environment, Miller and Banzhaf’s Cartesian Genetic Programming (CGP) was designed around

chemical diffusion that is much simpler than that used in HeBIS. Through the diffusion of chemicals

in a toy environment, a limited set of developmental rules was invoked in attempts to create the

desired pattern. The set of developmental rules was replicated in each artificial cell defined on a two-

dimensional Cartesian lattice in which each position in the lattice ultimately became part of an acyclic

graph that was self-organized to create the pattern. Although each cell ran the same program, the

local outcome of each was affected by the internal states of the neighboring cells. Cell differentiation

was displayed via the creation of a French flag pattern over the lattice. The research showed that

developmental principles can be successfully applied to genetic programming, albeit this was for a

rudimentary problem. Simple self-organization was demonstrated. Miller and Banzhaf suggested

that the idea of limiting the cells that can arbitrarily reproduce and differentiate (i.e. stem cells), could

be examined in future research. The authors believed that the cellular program was relatively small

for the level of complexity contained in the problem.

Artificial immunology has also been used as the basis for classification research. deCastro and Von

Zubin in 2001 [106] experimented with their artificial immune system model, aiNet in the context of

determining the distributions of test datasets and performing automated clustering of redundant data

samples. The network was designed more so to implement rudimentary learning capabilities instead

of being a completely biologically plausible model of the vertebrate immune system. The model’s

performance was analyzed for several benchmark problems that consisted of non-linearly separable

classes as well as the CHAINLINK problem which was formed from the point distribution of two

interwoven rings in 3-dimensional space. Results were comparable to a Kohonen SOFM trained on

the same problems, but the aiNet model is sensitive to a relatively larger number of tuning parameters

Page 51:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 51 of 249

09/28/10

than is the SOM. aiNet has high computational costs per training iteration that are O m2b c

where m

is the number of “memory” antibodies used by the system in a specific problem. Therefore, as the

complexity of the problem increases, the time required to train aiNet increases exponentially.

Unlike the artificial immunology classification research, HeBIS is more of a hybrid of biological and

ML principles. It does not attempt to map the ML constructs directly into biological analogs.

Not much research effort has been applied to the application of evolutionary development concepts to

classifying remote sensing imagery. One example uses the idea of an artificial immune system (AIS)

and applies it to this topic [107]. This presents an AIS that harnesses its various elements such as

antigens, antibodies, shape-space, and immune memory for the unsupervised classification of

multispectral optical satellite data. The algorithm, the Unsupervised Artificial Immune Classifier

(UAIC), classifies data on a pixel-by-pixel basis. The training data are mapped to the system’s

antigens and the classes/clusters are represented by UAIC’s antibodies. UAIC’s performance is

reported in comparisons with clustering algorithms such as K-Means, ISODATA, Fuzzy K-Means,

and a SOFM. The authors report that it outperforms these algorithms on the test dataset. However,

computation times for the algorithm are not adequately discussed. This has direct bearing on HeBIS

and indicates that the HeBIS algorithm may require significant computing resources for simulation.

2.4. Summary

It seems that biological development applied to classification problems may help by breaking the

one-to-one genotype-to-phenotype matching, but there are still problems [45]. One aspect of these

problems is the amount of computing power that is required to simulate large-scale GRN reactions in

an artificial chemistry [101]. Also, the evolutionary computation required to create a usable (and

complex) classification structure is clearly daunting with the computation power that is available to

the typical researcher.

The creation of a purely artificial-protein based classification system is difficult to realize, especially

one that can be as easily and as quickly trained as those produced by artificial neural networks, self-

organizing feature maps, and support vector machines. However, it is the combination of the SOFM

element with artificial biological development that is investigated in this work. It may mitigate the

previously outlined problems and produce a self-organizing classification network which possesses

Page 52:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 52 of 249

09/28/10

enough complexity and robustness to quickly and automatically adapt to different problems with high

precision. HeBIS is meant to examine a few of the basic issues associated with such a goal.

The modular building blocks for this architecture consist of evolved self-organizing feature maps. As

will be discussed in detail in later sections, communications between the processing cells are

conducted via artificial proteins that are diffused throughout the simulated environmental lattice.

Each cell in the artificial evolutionary development environment can have a SOFM machine learning

processing kernel. The communication among these cells will be defined by a fixed-length genome

that is evolved through evolutionary computation using the ideas of computational development. The

genes in each cell are associated with artificial proteins that will be activated or inhibited by evolved

genetic switches that are also contained within each processing cell’s genome [5,82].

A simple self-organizing algorithms will be examined for its potential contributions to this evolved

heterogeneous architecture that is based on computational development and artificial neurogenesis

[2,23,100]. General principles from the fields of computational development and genetic regulatory

networks will be used to guide the research towards developing complex learning structures that are

modular and robust [33]. HeBIS’ potential advantages and disadvantages are:

HeBIS Advantages

• Introduction of complexity via artificial protein chemistries and GRN signaling

• Potential for compact genomes because of the nonlinear mapping between genotype and the parameter (phenotype) space

• Graceful degradation of classification performance through the evolved GRN o Robustness o Noise

• Automatic modularization of functionality o Graceful degradation / robustness

� Missing data � Noisy data

o Adaptability � Damage-resistance- another module takes over processing

HeBIS Disadvantages

• Difficult to evolve (GRN)

• Difficult to analyze

• Computationally expensive to train and operate

Page 53:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 53 of 249

09/28/10

Page 54:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 54 of 249

09/28/10

3. Heterogeneous Biologically Inspired System (HeBIS) Hybridization Background

Algorithm Discussion Remote Sensing Background

GRN Analyses GRN Training with PSO

GRN Action Analyses Remote Sensing Application – Analyses and Comparisons

Robustness Analyses

3.1. Overview

HeBIS is a self-organizing classification system that is inspired by the ability of a live biological

organism to grow itself from the stage of a fertilized egg to the complete species’ phenotype through

the processing of the daughter genome that is obtained from the mating of its parents. The complex

development procedure is encoded through the total number of genes in the genome and the interplay

of these proteins in time and within the chemical environment in which they reside via a GRN. The

use of this GRN allows for a dense encoding of development information that is both self-organizing

and robust. It is also this network of protein-to-protein metabolic reactions that forms the basis and

feedback loops for the stable metabolic control of a human being. At its core, the HeBIS

environment for the classification of multispectral remote sensing data is inspired by the evolutionary

developmental power described and controlled by real-life biological GRNs.

HeBIS mimics portions of the mammalian protein signaling network from a GRN and embeds

machine learning elements in a simulated protein environment. An artificial GRN is evolved to

provide both the underlying communications infrastructure via protein-to-protein interactions as well

as to introduce a significant level of complexity into the system. This complexity provides the engine

for the self-organizing components of the classification network. Once trained, the evolved genome

is used to classify previously unseen examples from the problem at hand.

The HeBIS processing infrastructure is composed of the following functional areas, each of which

will be discussed in detail in this chapter:

• Simulated protein interaction environment

• Genetic regulatory network (GRN)

• Basic cell types, processing and behaviors

• Input feature vector and class representations

• Output class representation

• Classification training

• Evolutionary optimization

Page 55:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 55 of 249

09/28/10

Figure 7 provides an overview of the HeBIS environmental lattice in relation to its cells, genomes,

and proteins.

Outp

ut

ProcessingKernel

(ANN,SOM,or SVM)

Pro

tein

Hol

ogra

m

Environmental Protein

Regulatory Protein

Cell Receptor Protein

Pha

sed

Inpu

t Det

ecto

r

Environmental Protein

Cell Membrane

Control

Genome

Figure 7. Environmental lattice and processing cell overview.

In this chapter, the processing cell will be introduced as well as the simulated environment in which

artificial protein signaling occurs as well as the metabolic interactions between these proteins.

Proceeding from there, the artificial proteins will be described with an emphasis on the core

processing cell types and their intrinsic and learned behaviors. Input and output feature vector and

class representations will be described as well as the training method examined in this research.

Finally, implementation issues will be discussed.

3.2. Fundamentals

3.2.1. Processing cell

At this time, it is necessary to introduce information about the workhorse of the HeBIS environment,

the cell. Detailed functioning of the cell is included in later sections, but an overview is required to

provide the proper context for discussion of the HeBIS genetic regulatory network and its place in the

simulated environment. The major blocks of cell functionality to be discussed are portrayed in Figure

8.

Page 56:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 56 of 249

09/28/10

Figure 8. Major blocks of cell functionality.

The processing cell is an artificial entity which resides at a fixed spatial location within the

processing protein environment. The cell is delimited by a cell membrane that separates the cell’s

internal process from processes that are defined to exist only in the surrounding protein environment.

The interior volume delineated by the cell membrane is referred to as the cell’s cytoplasm. Major

structures contained within the cytoplasm are the cell receptor proteins, the control genome, the

output block, and the cell’s processing kernel. All these structures are purely artificial and mimic key

functions of their biological, real-world analogs.

The cell receptor proteins serve to map specific environmental proteins into the cell’s internal cell

control systems and the output system maps output results back out into the environment via

environmental proteins. Both the receptor and output blocks map numerical values to artificial

protein concentrations with appropriate scaling.

Each cell has a processing kernel which provides the machine learning functionality. This

functionality is currently limited to be a SOFM.

This architecture creates a genome within an individual cell. This genome has the ability to grow a

network of cells with processing machine learning kernels that will solve a specified classification

problem. The overlying structure provides the learning and analysis mechanism so that a genome

may be trained via the application of evolutionary computation, i.e. particle swarm optimization.

Page 57:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 57 of 249

09/28/10

The cell’s genome consists of genes which code for artificial proteins that are evolved by the HeBIS

training algorithm. It controls the execution of the cell’s innate and learned behaviors and it activates,

deactivates, and modifies these actions according to the concentrations of regulatory and recognized

environmental proteins. Through these interactions, a single cell self-organizes into a more complex

classification network.

The genome and its GRN that are created are optimized for a specific problem through the application

of a particle swarm optimizer to a specific training dataset taken from the problem domain.

These major functionalities are treated in more detail in the following sections of this dissertation.

3.2.2. Environment

The HeBIS protein environment is simulated and does not involve the use of actual proteins. The

environment is a discrete-time and discrete-space simulation that is based on a user-specified multi-

dimensional lattice structure. The environment comes in three flavors, a one-dimensional linear

array, a two-dimensional planar array, and a three-dimensional cubic lattice. However, only the

three-dimensional cubic lattice is examined in the current research. The discretized space coordinate

system that defines each structure is based on unit integers that are members of the set of counting

numbers, that is N such that n [0, ).

Simulation of the protein metabolic reactions and processing intervals occur in discrete unit time

steps such that t [0, ).

The environment uses an indexing arrangement in which the most fundamental processing element,

the cell, may be accessed and polled for its internal operating state. It also provides a simulated

medium through which artificial proteins diffuse, encounter each other, and react through the

simulated genetic regulatory network. 1-D, 2-D and 3-D lattices are now discussed in more detail.

Page 58:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 58 of 249

09/28/10

3.2.2.1. 1-D environment

The one-dimensional linear environment is pictured in Figure 9. At each integral unit along the

defined x-coordinate system, a processing cell may or may not exist. This is decided either through

the environment’s initialization or as the artificial GRN is processed. Therefore with the grid

numbering scheme defined in the figure, a maximum of 6 cells can exist in this example. Only six

cells can exist because boundary conditions are invoked in the environment so that a 1-unit buffer

zone always exists between the maximum extent of the cells and the edge of the environment. In this

case, the buffer positions are defined at the positions, x = 0 and x = 7. For purposes of protein

diffusion, the initial values at the beginning of the protein diffusion simulation are zero for the

boundary positions, x = 0 and x = 7. Protein concentrations for these boundary elements remain fixed

at zero through the simulation. Nearest-neighbor spatial relationships (to be defined later) between

cells on the linear grid are defined using a Euclidean L2 norm. For example, the nearest neighbors of

a cell located at position 3 are cells located at positions 2 and 4. Grid positions are numbered with

unit integers such that x 2 [0, Max -1] with Max defined as the maximum size of the linear grid.

Diffusion is defined only at the integral cell positions and not on the inter-cell regions, for example x

2 (0,1).

Figure 9. Linear grid numbering scheme.

3.1.1.1. 2-D environment

The 2-dimensional planar environment is shown in Figure 10. It is gridded and enumerated in a

fashion similar to the 1-dimensional linear environment. In the 2-dimensional case, cells may exist

only at discrete unit intervals along the imposed x-y coordinate system. Therefore, cell number 7 is

located at the x-y position (1,1). Boundary conditions are set in a fashion similar to that of the 1-

dimensional lattice: a one-cell wide buffer zone surrounds those positions which cells may reside and

in which protein diffusion may occur. For the example in Figure 3, with Max = 24, the buffer zones

are defined as the elements with x = 0, y = [0,3]; x = 5, y = [0,3]; x = [0,5], y = 0; and x = [0,5], y = 3.

This leaves a potentially active processing region composed of (for this example) 8 positions, P, such

that P 2 { 7, 8, 9, 10, 13,14, 15, 16}. Boundary conditions for the buffer positions are set to 0 for

protein diffusion.

Page 59:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 59 of 249

09/28/10

Figure 10. Planar grid numbering scheme.

3.1.1.2. 3-D environment

The third and final type of lattice is the 3-dimensional cubic lattice. It is an extension of the schemes

that have been outlined for the 1-dimensional and 2-dimensional environments. Presented in Figure

11, the same theory that regards boundary conditions and cell numbering in the lower dimensional

environments, applies to the 3-dimensional cubic lattice.

Figure 11. Three-dimensional lattice numbering scheme.

3.1.2. Genetic regulatory network

Biological development uses genes to define proteins that are in turn used for multiple overlaid

purposes [3]. These purposes include the activation of other genes, the suppression of other genes,

Page 60:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 60 of 249

09/28/10

and inter-cell and intra-cell signaling and communications. This GRN is a modular system that

recurrently uses the same gene for different purposes and for the creation of many different types of

proteins that have different and complex effects. These protein and metabolic interactions are

harnessed for complex and recurrent communications between the simple processing building blocks

in the learning system [108].

Artificial protein creation is activated or suppressed by other artificial proteins that are in turn

controlled by genetic switches present in the evolved processing cell’s fixed-length genome. These

proteins diffuse through the simulated environment which contains the network of modular

heterogeneous processing cells. The diffusion occurs at discrete time intervals and each protein’s

concentration decreases over a specified time interval [109]. Concentration is defined as a

mass/volume ratio and for HeBIS’ artificial proteins its unit is [generic-mass-unit/generic-unit-

volume]. Communications between processing elements is inherent in this system as the proteins and

their corresponding activation and inhibition genetic switches evolve during training [3].

3.1.2.1. Gene coding

Within each HeBIS cell is a fixed-length genome that consists of at minimum, one gene. The

ultimate number of genes that may be contained in the genome is unrestrained although it is fixed for

each research simulation. The hierarchy between the cell’s genome, its genes, and potential

expressed proteins is shown in Figure 12. This is a much simplified version of what occurs in a true

biological regulatory environment.

Page 61:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 61 of 249

09/28/10

Figure 12. Genome/Gene protein hierarchy.

Each gene portrayed in the figure contains several functional regions. These regions include the

switch protein, the gene type, the action that may be associated with the gene, and one or more

protein descriptions. The switch protein region is the description of the perfectly-matched protein

that activates the specific action or actions that have been associated with the gene. Biologically, the

switch protein (see Figure 13) defines the protein match for which proteins that are matched against

this template, must have a high affinity or else the action will not occur. Affinity is defined as the

degree to which a protein matches the switch protein template.

Figure 13. Standard regulatory/environmental and switch protein descriptions.

Next, the gene type region defines this gene as a combination of one or more of the following

attributes in an OR’d fashion: Action Gene || Regulatory Protein || Environmental Protein || Cell

Page 62:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 62 of 249

09/28/10

Receptor Protein. Once the gene is excited, the gene type region determines whether the cell starts or

stops production of regulatory and/or environmental proteins and also determines whether any of the

learned cell behaviors is executed.

The action region defines a diverse set of potential cellularactions. This list includes the following

cell behaviors that will be discussed in detail in a later section.

• AddCell

• PruneSelf

• ChangeToSOFMAndTrain

• Classify The protein description regions (regulatory, environmental, and cell receptor) within the gene

describe the proteins that are produced and released into either the cytoplasm or environment. The

regulatory and cell receptor proteins are only for intracellular release whereas the coded

environmental proteins are released directly into the environmental lattice in which the cells reside.

Two broad classes of protein codings exist: switch proteins and standard proteins. Switch codings

only describe switch proteins and standard codings describe environmental, regulatory, and cell

receptor proteins. Figure 13 portrays these codings and their important regions.

Both codings describe a protein as a vector of four integers, < p0, p1, p2, p3> . This vector defines a

point in a [1, 254] four-dimensional cube.

The difference in the two codings lies with the concentration region. In the case of a standard protein,

the single concentration value is a real number on the interval [0,1] that describes a current real-

valued concentration. For these regulatory, environmental, and cell receptor proteins, this number is

the value of the protein’s concentration as it is first produced by the gene at each excited timestep.

For example, when a specific environmental protein is produced, it is initially produced in the first

time step with the initial concentration indicated in the concentration mask. In subsequent timesteps,

if the gene remains excited then more of the protein will be created with this concentration level at

this point in the environmental lattice. As this protein diffuses through the cytoplasm or the

environment, concentration levels of this protein at all points in the lattice are updated.

Page 63:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 63 of 249

09/28/10

However, the switch protein requires minimum and maximum activation thresholds. If a test protein

matches the protein description template, <p0, p1, p2, p3> it still must have a local concentration at the

cell that is within the prescribed thresholds of the switch. If it does not, then the associated gene will

not express itself.

This similarity measure is currently implemented as a two-part function. More formally, the

determination of the switch’s activation is identical to the question of whether two proteins, P1 and

P2, are matches for each other. In this case, an exact match between the proteins is defined if and

only if

P1 ,P2

* +

P1

N

N

N

N

N

N P2

N

N

N

N

N

N

fffffffffffffffffffffffffffffff= cos θ

` a

= 1

(11)

and

P1

N

N

N

N

N

N

P2

N

N

N

N

N

N

fffffffffffffff= 1

,

(12)

where < , > is an inner product defined on the two vectors, || || is the L2 norm, and is the inner

angle between the two proteins, P1 and P2, when they are referenced as vectors.

The L2 norm is defined only over the domain of descriptive protein parameters; p0, p1, p2, and p3.

This norm is

P1

N

N

N

N

N

N= p02 + p1

2 + p22 + p3

2qwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwww

.

(13) A partial match between the proteins and in turn between a switch and a test protein is enabled

through a constraint relaxation such that a gene switch is considered to be activated if

Page 64:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 64 of 249

09/28/10

1@∆1 ≤P0 ,Switch* +

P0

N

N

N

N

N

N SwitchN

N

N

N

fffffffffffffffffffffffffffffffffffffffffffff≤ 1

,

(14)

and

1@∆2 ≤P0

N

N

N

N

N

N

SwitchN

N

N

N

ffffffffffffffffffffffffffff≤ 1

.

(15)

The empirical effects of the choices of ∆1 and ∆2 are examined in the Simulations and Analyses

chapter.

Each of the two protein codings also has a protein ID region. This region is used purely for

bookkeeping purposes by the simulation’s infrastructure. It has no implied biological meaning or

utility.

3.1.2.2. Protein communications

Protein communications between the processing cells is founded on three points: protein diffusion

through the environment and within the cytoplasm, a defined artificial chemistry that allows the

proteins to react with each other and to create new proteins, and protein template matching.

• Diffusion of an artificial protein occurs in the 3-D environmental lattice. The protein

diffusion equation is simply the discretized scalar Laplacian. In a 3-dimensional

Cartesian coordinate system, this is defined as

(16)

where h = 2 and ψ x,y,z` a

is the value of the protein concentration at (x,y,z) in the

discrete spatial environment.

Page 65:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 65 of 249

09/28/10

• The HeBIS artificial protein chemistry is a simple one that bears little resemblance to its

biological model. The chemistry is defined and occurs with normal Gaussian probability

at a location in the environment where two or more proteins are present such that

p t` a

reaction=

1

2πpwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwffffffffffffffff

e@

t 2

2

ffffffff .

(17)

Note that the protein chemistry is not currently defined for intracellular reactions. If two

or more proteins exist at a specific location, a chemical reaction will occur between them

with a discrete Gaussian probability. The resultant new protein vector, <pnew0, pnew1,

pnew2, pnew3, ϕ new> is calculated by a simple exclusive-OR (XOR) operation on the bits of

each of the integer-valued parameters, p0… p3. The XOR operation is tabulated in Table

1.

Table 1. Bitwise XOR Functional Mapping

Bit0

Bit1

XOR

0 0 0

0 1 1

1 0 1

1 1 0

This protein chemistry is defined so that the chemical reaction depletes the protein

reactants that are available at the environmental location. This depletion assumes that the

proteins that are undergoing the reaction are reduced in concentration and that

“concentration” is conserved. As an example, assume that at location (2,3,2) in a cubic

environmental lattice that three proteins are being metabolized. With these proteins

defined as P0, P1, and P2 these proteins have concentrations of 1, 2, and 0.5 units

respectively. If the chemical reaction occurs, then it proceeds according to

1P 0 + 2P1 + 0.5P 2u 3.5P 3 .

(18)

Page 66:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 66 of 249

09/28/10

The concentration of P3, the newly formed protein, has an initial concentration of 3.5

units- the summation of all the reactants that are available for the metabolic function at

this location. In turn, the concentrations of P0, P1, and P2 at location (2,3,2) are reduced

to 0 before the next time step of the simulation occurs. Hence, concentration is conserved

through the metabolic reaction and a negative feedback loop for concentration levels is

imposed on the global environment.

• Protein matching, the final pillar of HeBIS communications, has been described in detail

in the previous section. Its purpose is to mimic the protein-based excitatory/inhibitory

functions that occur on a genetic level as well as the shape-based recognition that

biological proteins use.

3.1.3. Basic cell processing

The core of HeBIS is the basic processing cell. Embedded in the artificial environment, it contains

the system’s distributed intelligence for its interaction with the network’s GRN. The system’s self-

organizing features that allow it to learn are implemented at this level. The cell has many behaviors

that are controlled by an evolved fixed-length genome. These behaviors can be automatically

modified as processing continues. Some of these behaviors are immutable and are designed into the

system from the start and others are learned through interaction with the GRN and the proteins.

This section continues the discussion of HeBIS with details about the cell’s control genome, the

different cell types, intrinsic and learned behaviors, and available machine learning kernels. Initially,

the cell genome is discussed.

3.1.3.1. Cell genome

The role of the cell genome is to control the individual cell such that its local behaviors add structure

and ability to the entire classification network of cells. The classification network is built and

operates through the self-organized construction of complex behaviors. These behaviors are built

from a “dictionary” of simpler actions that are executed and controlled within each cell. Importantly,

the simple cellular actions can become more complex as cells interact with other cells via the artificial

GRN. With its individualized behaviors, the cell genome functions in two modes of operation. These

two modes, learning and processing (or production), are separate but not distinct from each other.

In the learning mode an initial cell genome with its variable number of genes is evolved to classify a

set of training examples via evolutionary computation. This evolved genome allows the network to

Page 67:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 67 of 249

09/28/10

grow (if necessary) from its original single cell beginnings, to a network that is sized and matched in

various aspects to the statistics of the problem. Secondly, in the processing/production operational

mode, the cell genome uses its evolved “control” program and GRN to classify unseen patterns from

the same problem domain. Thus, the evolved behaviors for a single cell interact in a complex fashion

with other cells that are autonomously added (or deleted) from the network.

The various behaviors and cell types are now examined in more detail.

3.1.3.2. Intrinsic behaviors

The cell’s intrinsic behaviors are actions which are performed at the beginning of each discrete time

step for each cell- after the proteins have chemically reacted with each other and they have diffused

further into the environment. These behaviors provide statistical information to the genome about the

proteins that are present both in the local environment as well as inside the cell.

The nomenclature for these intrinsic and learned behaviors is designed such that the names directly

referenced in the following sections are at the same time, function names programmed in the code.

The individual English words are allowed to run into each other to form a new, composite algorithm

name and topic heading.

3.1.3.2.1. NumberProteinsInCell

This behavior counts the proteins that are present in the cytoplasm. These proteins include all

internal regulatory proteins, cell receptor proteins, and environmental proteins that have been passed

into the cytoplasm. The number of these proteins is made available to the cell’s controller and to its

genome as a distinct cytoplasm protein.

3.1.3.2.2. NumberProteinsInLocalEnviro

This behavior counts the proteins that are present only in the local environment outside of the cell.

By definition, this includes only environmental proteins. The sampled proteins are only those that are

present in sufficient concentration at the cell’s location within the environment. The number of these

proteins is made available to the cell’s controller and its genome as a distinct cytoplasm protein.

3.1.3.2.3. ConcentrationStandardDeviationLocalEnviro

This behavior determines the concentrations of all the proteins that are present in the indexed

locations in the environment surrounding the location of this cell. A standard deviation is then

calculated from these concentrations via

Page 68:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 68 of 249

09/28/10

σconcentration local environment

=1

N@1ffffffffffffffffffX

n = 0

N@ 1

Conc n` a

@µlocal environment

b c2

h

j

i

k

0.5

,

(19)

and provided to the cell’s controller and its genome as a distinct cytoplasm protein.

3.1.3.2.4. ConcentrationMeanLocalEnviro

This behavior determines the mean of the concentrations of all the proteins that are present in the

environment at the cell’s location. This information is calculated with

µlocal environment

=1

N

fffffffXn = 0

N@ 1

Conc n` a

,

(20)

and provided to the cell’s controller and its genome as a distinct cytoplasm protein.

3.1.3.2.5. ConcentrationMaxLocalEnviro

This behavior determines the maximum concentration of all the proteins that are present in the

environment local to the individual cell. This information is provided to the cell’s controller and its

genome as a distinct cytoplasm protein.

3.1.3.2.6. ConcentrationMinLocalEnviro

This behavior determines the minimum concentration of all the proteins that are present in the

environment local to the individual cell. This information is provided to the cell’s controller and its

genome as a distinct cytoplasm protein.

3.1.3.2.7. KillSelf

This action allows the cell to commit suicide if a set number of discrete simulation time intervals

have occurred. The default behavior is that this feature is disabled.

3.1.3.2.8. NumberFeatures

This behavior determines the number of features that are being used within this time step by the cell’s

processing kernel. This information is provided to the cell’s controller and its genome as a distinct

cytoplasm protein.

Page 69:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 69 of 249

09/28/10

3.1.3.3. Learned behaviors

The cell’s learned behaviors are those actions which occur as a result of the cell’s interaction as part

of the genetic regulatory network. Each of these actions is a behavior that is expressed because a

specific gene (or gene sequence) is activated because the requisite proteins are present in the cell’s

cytoplasm in appropriate (evolved) concentrations. These learned behaviors may use information that

has been encoded as proteins by the cell’s intrinsic behaviors. These simple learned behaviors are

combined into more complex behaviors through the nonlinear GRN mapping. These behaviors

implement the communications and control for the classification network and they make use of

information provided by the intrinsic behaviors and the environmental GRN. Inhibition of a behavior

occurs when the excitatory proteins are no longer present in sufficient concentrations. These

behaviors take place after protein chemistry, protein diffusion, and the innate behaviors have

executed.

3.1.3.3.1. AddCell

The cell that executes this behavior adds a copy of itself to the environmental lattice immediately next

to itself. The exact location is one unit away from the initiating cell’s position and in the direction of

the average concentration gradient.

3.1.3.3.2. PruneSelf

The cell that executes this behavior deletes itself from the environmental lattice and ceases all

processing and GRN activities. Proteins that are being produced by this cell both within the

cytoplasm and in the environment cease production beginning at the next simulation time step.

3.1.3.3.3. ChangeToSOFMAndTrain

This behavior causes the cell to convert its processing kernel to a self-organizing feature map.

Training of the SOFM then proceeds and uses the protein holograms that are in the phased input

detector. If the cell is already an SOFM, the SOFM’s parameters are reset and training commences

during the current time interval.

Default SOFM parameters:

� Rectangular lattice � Lattice contains 9 elements in a 3x3 orientation � Simple neighborhood update function

Page 70:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 70 of 249

09/28/10

In subsequent test/production runs of this behavior, a feature that is not present in the protein-based

feature vector (exemplar) is replaced with an equivalent value of 0.0.

3.1.3.3.4. Classify

This behavior takes the cell’s input present in its phased input detector and uses that as input data as it

executes the processing kernel (SOFM). The resulting output is released into the environment as an

artificial protein.

3.1.3.4. Cell types

Two types of cells are used in HeBIS and they are defined by the type of machine learning processing

kernel that is active: the SOFM and Pass-Thru. By default, each position in the environmental lattice

is initialized as a Pass-Thru node. The Pass-Thru node exists to simply allow protein diffusion and

communications to occur in an uninhibited fashion through a position in the environmental lattice.

As the network’s structure organizes, pass-thru nodes can be converted to processing cells according

to the interactions of the genetic regulatory network.

3.1.3.4.1. SOFM

The SOFM cell type has a SOFM as its processing kernel. The kernel defaults to the following:

� Rectangular lattice � Lattice contains 9 elements in a 3x3 orientation � Simple neighborhood update function � Conservative learning rate

3.1.3.4.2. Pass-Thru

This is a cell which has no processing kernel. No GRN actions are performed by this cell. In

essence, it is invisible to the other cell types in the environmental network.

Page 71:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 71 of 249

09/28/10

3.2. Input feature vector representations

A single method is currently used to map test and training data from benchmark and real-world

datasets into artificial proteins that are used in the HeBIS environment. This is direct feature-to-

protein mapping. With this method, it is assumed that the data to be converted consists of N

exemplars. Each exemplar is defined by a feature vector which consists of M real-values and finite

entries (features). These features describe specific characteristics of the exemplar. With this

representation, the exemplar is referenced as a point in an M-dimensional feature space. The feature

vector may or may not be sparsely defined.

Given a set of exemplars, a modified set is created in a preprocessing stage that uses the collective

statistics of the feature vectors from the problem domain. This preprocessing stage normalizes the

range of each of the M features independently of the other M-1 features. This global normalization

shifts and scales each feature so that its normalized range now resides within [0,1] and has an

appropriately modified probability density function for each feature. The parameters used to

normalize the features are saved during preprocessing for later denormalization. Preprocessing

occurs for training, test, and operational data.

In this preprocessing stage, each feature is assigned to a unique protein description, <p0,p1,p2,p3>.

The class label is also assigned to a unique protein description.

3.2.1. Direct feature-to-protein mapping

This data mapping operates on a single training or test exemplar. Each feature and class label (if

available) of the normalized exemplar is mapped to the unique protein that has been defined to

represent that feature during preprocessing. The normalized value of the feature is directly mapped to

the concentration of the protein and the phase is set to 0.

The exemplar’s training class label is also normalized so that it is a discrete point in the interval [0,1].

For example, in a two-class problem with class labels C0 and C1, the normalized values of C0 and C1

are chosen as C0 = 0.25 and C1 = 0.75. These values are the centroids of two non-overlapping

subintervals of [0,1]. These two subintervals are [0,0.5) and [0.5,10.]. Multi-class labels are handled

by mapping the labels to the centroids of multiple non-overlapping intervals on [0,1]. Currently, the

subintervals are equally sized but it is foreseeable that the lengths of these subintervals could be set

according to some statistical measure of the training data. This issue will not be examined in this

research.

Page 72:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 72 of 249

09/28/10

Since the centroids of the subintervals are used, in some sense this class mapping strategy can also be

examined in the future as a fuzzy interpretation of the class label.

A similar process is applicable for multi-class problems where the number of classes is greater than

two.

This procedure for direct feature-to-protein mapping is outlined in Figure 14. The specific

feature/class protein mapping is arbitrarily established in advance during preprocessing. The only

constraint is that the feature/class protein mapping must be unique for each feature. Therefore, a

particular protein is always associated uniquely with a specific feature/class during the course of the

simulation.

Figure 14. Direct feature-to-protein mapping.

3.3. Pattern training for classification

In past sections of this chapter, the fundamental elements of the HeBIS architecture have been

introduced. This section aims to unify these elements under the umbrella of the network training

mechanism. These training mechanisms are based on the self-organization principles that control

each cell, the development of the GRN, and the overlying PSO that tunes the cell’s genome. Once

tuned, a single cell can potentially spawn a network of cells that in unison can perform data

classifications to varying degrees of success.

Page 73:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 73 of 249

09/28/10

This research examines the applicability of the HeBIS algorithms to the classification of sets of

features. HeBIS is oriented towards using biological principles to determine whether self-

organization ideas from biology may be used to create successful supervised classification networks

based on machine learning constructs. In this research, a supervised learning environment controlled

by a PSO is used. This optimizer tunes the gene and protein descriptions of the cell’s fixed-length

genome. It is this genome that is used to control the individual behavior and network organizational

abilities of the cell. These behaviors and abilities are the building blocks for complex behaviors that

may be used for classification as the cells in a network interact with each other.

By classification, it is meant that given a set of L exemplars, each of which consists of M real-valued

features, each of the exemplars can be mapped to a finite number of elements within a chosen set of

classes, C. Each of the L elements of this set, C, has a label associated with it, hence the term class

label.

This research determines if it is possible to self-organize a network of machine learning elements that

can map multidimensional feature vectors, exemplars, from the feature space to their correct class

labels. The mapping is determined in the network training phase using supervised learning, but the

network’s cellular topology, and the details of the ML intercommunications via the GRN are chosen

in an unsupervised manner through the optimization of the cell’s genome.

3.3.1. Self-organization principles

There are two important parts of the low-level HeBIS algorithm. The first is the development of a

genetic regulatory network. This network allows communications between the processing cells to be

encoded by an artificial protein chemistry (metabolism) with concentrations that are diffused

throughout an environmental lattice. These protein concentrations are both time and space

dependent. This communication is primarily defined by the cell’s genome.

The second part, which is discussed here, is the small set of self-organizing principles that are

investigated in this research. This small set of principles is not designed to be all inclusive, but is

meant to form the basis of this research which will determine if useful classification can occur based

on

Page 74:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 74 of 249

09/28/10

• a small set of rules

• biologically-inspired communications

• simple machine learning elements.

This set of self-organizing principles provides feedback to the cell for the local optimization of the

cell’s classification activities and defines the interactions between cells in the network. This set

includes

• Protein analogs of mathematical statistical features

• Cellular fission

• Cellular death

These self-organizing principles are now discussed.

3.3.1.1. Protein analogs of statistical features

These statistical analogs are the building blocks of the principles that follow. These analogs have

already been described but are enumerated again for convenience:

• NumberProteinsInCell – The number of distinct proteins that are present in the cell’s cytoplasm.

• NumberProteinsInLocalEnviro – The number of distinct proteins that are present in the cell’s external environment.

• ConcentrationStandardDeviationLocalEnviro – The standard deviation of the concentrations of the proteins that are present in the local environment.

• ConcentrationMeanLocalEnviro – The average concentration of all the proteins that are located in the cell’s external environment at a specific location.

• ConcentrationMaxLocalEnviro – The maximum concentration of all the proteins that are located in the cell’s external environmental lattice at the location of the cell.

• ConcentrationMinLocalEnviro – The minimum concentration of all the proteins that are located in the cell’s external environmental lattice at the location of the cell.

3.3.1.2. Cellular fission and death

Next are the principles of cellular fission and cellular death. These are expressed by the network’s

GRN and the evolved gene-protein process. There is also an inherent lifetime to each of the cells.

Once a cell is born through either cellular creation or cellular fission (a cloning procedure), the cell

has a finite number of simulation intervals during which it may live and remain active. Once it

reaches that limit it dies, unless it has already initiated its own death because of genes that have been

expressed through the GRN.

Page 75:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 75 of 249

09/28/10

Death allows inefficient or unused ML elements to be pruned from the classification network.

Fission allows a useful ML element (with its trained knowledge) to be utilized as a complete module

in another location in the classification topology.

3.3.2. Particle swarm optimization

A particle swarm algorithm is used to optimize the entire cellular genome so that it can grow a self-

organizing classification network. The HeBIS PSO optimizes the genes in the cellular genome at the

protein level. That is, the four parameters that describe each protein each comprise a dimension of

the support of the multi-dimensional space for the proteins in the genome as well as the proteins that

define the features of the training and testing exemplars. Protein output concentrations as well as

switch protein concentration ranges also add to the dimensionality of the PSO. More information on

the implementation is included in the Appendices.

3.3.3. Training algorithm

A single training algorithm is examined in this research for simplicity. In this case, the network’s

training is controlled by a PSO. This improves the network’s performance on unseen examples by

providing a better estimate of classification generalization performance.

3.3.3.1. Training Algorithm: Presentation of training vectors and classes to the system

The training algorithm is diagrammed in Figure 15 and its flow is as follows. For purposes of

illustration, this example is based on a pool of 100 training exemplars. Scaling to datasets that have a

smaller or larger number of points is straightforward.

Training Algorithm

1. Choose Q, the size of the particle swarm and N, the number of network training iterations. Set the particle average fitness values to 0.

2. Create a PSO population of Q particles. Initialize the parameters of each particle randomly with regards to initial values and velocities. 3. Initialize a HeBIS environment for each of the Q particles. Each environment will consist of a single processing cell with a single genome. The location of the processing cell and the genome’s parameters are taken from the particle parameters from the PSO. 4. Choose an exemplar from the training pool and present it to each of the environments that is controlled by one of the Q particles in the PSO.

Page 76:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 76 of 249

09/28/10

5. Begin the protein diffusion and self-organization process for each of the environments. Allow each simulation to continue for a fixed number of iterations or until all protein concentrations equal zero. 6. Determine the fitness of the genome associated with the ith PSO particle and record this. This fitness is determined by comparing the desired label of the exemplar to the label produced by the network. 7. Repeat steps 3, 4,5, and 6 until all the exemplars have been presented for training to the HeBIS network. 8. Average the fitness values of all the training data for each genome across all the training points presented to the network during this iteration of the training algorithm. 9. Update each of the Q PSO particles with its respective, averaged fitness. 10. Repeat Steps 2-9 for N iterations. 11. Save the best PSO-evolved genome.

Figure 15. Schematic of the training algorithm.

Page 77:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 77 of 249

09/28/10

Cross-validation (CV) can be used in conjunction with the training procedure to improve the

modeling HeBIS’ performance by providing a better estimate of the classifier’s generalization ability

[110].

3.4. Output coding

The classification result for a particular test case is a complex behavior that is evolved by the training

algorithm. The result is constrained to be a unique protein which maps class membership to a class

protein’s normalized concentration value 2 [0,1]. The concentration mapping is defined in the same

manner as for the input class mapping. To reiterate, class membership is defined through the class

u subinterval definition 2 [0,1]. The difference between the output and input class mapping is that

the input mapping is manually constrained to sub-intervals of [0,1], but the output mapping is a result

of the evolved network and its GRN. This mapping is evaluated and validated when the network’s

classification is tested during the evaluation phase of the training procedure.

3.5. Post processing of classification results

The detected output protein provides information about the classification of the test example.

Denormalization of the output protein concentration is accomplished simply by mapping the peak

concentration to the closest class centroid that has been previously defined for the input

normalization. For example, in a 2-class problem, the class centroids are defined as C0 = 0.25 and

C1 = 0.75. If the peak output protein concentration level is C = 0.20, then | C – C0 | < | C – C1 | and

the output class is found to be C0.

Since this is, at its core, a ranking classifier, it is possible to use data from the cross-validation

procedure to form a Receiver Operating Characteristic curve (ROC) that could be used to actively set

the output threshold so as to improve classification accuracy. This output threshold is the number in

the interval, [0,1], that can be used to set the classifier’s operating point- the estimate of the expected

number of correctly classified exemplars given the user-defined level of acceptable false positives.

Page 78:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 78 of 249

09/28/10

4. Simulations and analyses

4.1. Simulation limits

HeBIS with its diffusion of artificial proteins in a discrete time/space simulation, is currently a

processing-intensive application. Since the purpose of this research is to begin the exploration of this

classification paradigm, the number of testing and training exemplars was limited so as to provide

reasonable simulation times. Even so, with 20 or 40 multispectral center pixels and their surrounding

geographic samples used for training and cross-validation testing, simulation times were found to run

between 1.5 and 10 days on a modern multiprocessor computer. This depended on such parameters

as the number of particles in the PSO, the dimensionality of the tested HeBIS implementation (i.e.,

number of genes, number of features), and the size of the geographic region included in the

classification training process. Figure 16 and Figure 17 provide a strawman diagram for the

surrounding data cube that is associated with a specific pixel.

5.

Figure 16. Training/Testing pixel and its relationship to its surrounding geographic pixels.

Page 79:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 79 of 249

09/28/10

Figure 17. Training/Testing pixel and the surrounding multispectral information.

Initially limiting the exploration of the system in this manner allowed the creation of a base of

experience which will be used in future and more detailed HeBIS research.

Page 80:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 80 of 249

09/28/10

4.2. General methodology Hybridization Background

Algorithm Discussion Remote Sensing Background

GRN Analyses GRN Training with PSO

GRN Action Analyses Remote Sensing Application – Analyses and Comparisons

Robustness Analyses

4.2.1. Remote sensing cloud/no-cloud problem

4.2.1.1. Description

A remote sensing problem is used as a simplified benchmark for comparison and experiments in this

dissertation. We have constructed a problem that consists of learning whether a pixel in a remotely

sensed image is a no-cloud pixel or if it is a cloudy pixel. The baseline determination of whether the

pixel is considered to be cloud (Class C0) or no-cloud (Class C1) is empirically determined from the

data after it is processed from the raw radiance data. Thus, HeBIS is evaluated on its ability to learn

this empirical mapping.

Background on this dataset and its source are now discussed.

4.2.1.2. Sensor and datasets

The data used in this dissertation were acquired from NASA’s Moderate Resolution Imaging

Spectroradiometer (MODIS) on the Aqua satellite platform. Aqua’s orbital information is listed in

Table 2.

Table 2. Orbital Information for NASA's Aqua Satellite.

Parameter Value

Orbit 705 km circular

Equator Crossing 1:30 pm local

Orbit Type Ascending

Inclination Near-polar

Synchronicity Sun-synchronous

The MODIS sensor is capable of capturing data over a ground swath of 2 330 km crosstrack with a 10

km along-track swath measured at earth nadir. It is a multispectral optical sensor with 36 spectral

bands available with the best ground resolutions listed in Table 3 for the spectral bands.

Page 81:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 81 of 249

09/28/10

Table 3. Nominal Resolutions for the MODIS Sensor.

Band Resolution

1-2 250 m

3-7 500 m

8-36 1000 m

The full list of the available multispectral data available from MODIS is available in Table 4.

Table 4. 36 Bands of Multispectral Data from MODIS.

Band # Bandwidth [nm] Spectral Radiance [W]

1 620-670 21.8

2 841-876 24.7

3 459-479 35.3

4 545-565 29.0

5 1230-1250 5.4

6 1628-1652 7.3

7 2105-2155 1.0

8 405-420 44.9

9 438-448 41.9

10 483-493 32.1

11 526-536 27.9

12 546-556 21.0

13 662-672 9.5

14 673-683 8.7

15 743-753 10.2

16 862-877 6.2

17 890-920 10.0

18 931-941 3.6

19 915-965 15.0

20 3660-3840 0.45 (300K)

21 3929-3989 2.38 (335 K)

22 3929-3989 0.67 (300K)

23 4020-4080 0.78 (300 K)

24 4433-4498 0.17 (250 K)

25 4482-4549 0.59 (275 K)

26 1360-1390 6.0

27 6535-6895 1.16 (240 K)

28 7175-7475 2.18 (250 K)

29 8400-8700 9.58 (300 K)

30 9580-9880 3.69 (250 K)

31 10780-11280 9.55 (300 K)

32 11770-12270 8.94 (300 K)

33 13185-13485 4.52 (260 K)

34 13485-13785 3.76 (250 K)

35 13785-14085 3.11 (240 K)

36 14085-14385 2.08 (220 K)

Page 82:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 82 of 249

09/28/10

The HeBIS research dataset was created from top-of-the-atmosphere (TOA) radiances that were

derived from modified ocean Level 2 data products from MODIS [116]. Level 2 datasets consist of

calibrated radiances with geolocation references in addition to ancillary information about data

quality and error estimates. A definition of these NASA/CEOS data levels is contained in Table 5.

Table 5. NASA/CEOS Dataset Level Definition

Level Description

Level 1A Raw radiance counts from all bands Spacecraft and instrument telemetry for geolocation, calibration, and data processing

Level 1B Radiances that are calibrated and geolocated from the position of the sensor’s aperture Calibration data, quality flags, error estimates, Generated from Level 1A data

Level 2 Geophysical values for each pixel Values derived via Level 1A raw sensor information, atmospheric corrections, and bio-optical algorithms

Level 3 Generated from multiple Level 2 datasets All data from the following temporal periods:

1 day 8 days 1 month 1 year

Data are binned into non-overlapping bins across the earth’s ocean and land surfaces 9 km x 9 km bins 4 km x 4km bins

Since these data are nominally stored in hierarchical data format (HDF) files, a procedure was

generated to convert these files to a format accessible by HeBIS. This data conversion consisted of

acquiring the data from the NASA data archive, extracting the Level 2 data from the 36 bands, and

then converting the data for each pixel and band to floating point numbers. These 4-byte floating

points are then stored in flat files for easier customized access than the standard HDF format.

Although 36 multispectral bands are available, we only use those bands that are typically associated

with ocean remote sensing. These 17 bands are listed in Table 6.

Table 6. 17 Bands from MODIS for A2002193183000 LAC_x_NIR.

Band Name

Cloud Albedo

Lt 412

Lt 443

Lt 469

Lt 488

Lt 531

Lt 551

L5 555

L5 645

Page 83:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 83 of 249

09/28/10

Lt 667

Lt 678

Lt 748

Lt 859

Lt 869

Lt 1240

Lt 1640

Lt 2130

These intermediate data are used to create a cloud/no-cloud mask and then the final dataset of 15

channels is used for ingestion into HeBIS. The 15 final bands are included in Table 7.

Table 7. Multispectral bands used from MODIS [111]. HeBIS Band Index Primary Use MODIS Band Bandwidth [nm]

1 Land/Cloud/Aerosols Boundaries

1 620-670

2 2 841-876

3 Land/Cloud/Aerosols Properties 3 459-479

4 4 545-565

5 5 1230-1250

6 7 2105-2155

7 Ocean Color/Phytoplankton/Biogeochemistry 8 405-420

8 9 438-448

9 10 483-493

10 11 526-536

11 12 546-556

12 13 662-672

13 14 673-783

14 15 743-753

15 16 862-877

The HeBIS dataset is a 15-band dataset from MODIS that also includes a cloud/no-cloud pixel

classification created during Level-2 TOA processing. These bands consist of optical and short-wave

infrared radiances. The dataset is A2002193183000_cloud_albedo. It was acquired in 2002 on the

193rd day of the year over the Chesapeake Bay. The Bay is located in the Mid-Atlantic region of the

east coast of the United States. It consists of a processed swath that is 234 pixels wide and 430 pixels

long. Figure 18 shows the pseudocolor image for the visible bands, Figure 19 shows the cloud/no-

cloud ground truth, and Figure 20 shows the land mask for the dataset. A land mask was applied over

the region such that only observations over water (specifically the Bay) were included in the dataset.

Page 84:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor

Figure 18. Pseudocolor image for A2002193183000 dataset.

black corresponds to water, and green and brown re

image for A2002193183000 dataset. Grey and white colors correspond to cloud pixels,

black corresponds to water, and green and brown refer to land pixels.

84 of 249

Grey and white colors correspond to cloud pixels,

Page 85:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor

Figure 19. Ground truth (cloud/no

white pixels reference clouds and black corresponds to water.

. Ground truth (cloud/no-cloud) for A2002193183000 dataset. Red corresponds to land pixels whereas

white pixels reference clouds and black corresponds to water.

85 of 249

Red corresponds to land pixels whereas

Page 86:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor

Figure 20. Land mask (cloud/no

corresponds to water pixels.

A cloud/no-cloud mask was used which

corrected for molecular atmospheric

is classified as a cloudy pixel if

Otherwise, if the approximated surface reflectance is less than or equal to 0.027, then the pixel is

classified as a no-cloud pixel.

to be 2105 nm; ρW

λ` a

is the approximated surface reflectance;

atmosphere reflectance; t g

surface to sensor; T is the direct transmittance of the atmosphere through Rayleigh and aerosols

scattering, from the surface to the sensor;

from the ocean’s surface; ρ

(cloud/no-cloud) for A2002193183000 dataset. Red corresponds to land pixels and black

was used which is based on MODIS band 16 reflectance

corrected for molecular atmospheric scattering and other scattering components

pixel if the approximated surface reflectance is greater than 0.027

ρW

λ` a

tλ` a

t g λ` a

fffffffffffffffffff@T Aρ

gλ` a

@ρr

λ` a

h

j

i

k

1

td

λ` a

ffffffffffffffffff .

roximated surface reflectance is less than or equal to 0.027, then the pixel is

cloud pixel. In Equation (21), for MODIS data, λ is the wav

is the approximated surface reflectance; ρt

λ` a

λ` a

is the transmittance of the atmospheric gases, sun to surface and

is the direct transmittance of the atmosphere through Rayleigh and aerosols

scattering, from the surface to the sensor; ρg

λ` a

is the glint reflectance due to specular sun reflection

ρr

λ` a

is reflectance originated from molecular (Rayleigh) scattering; and

86 of 249

Red corresponds to land pixels and black

band 16 reflectance that has been

and other scattering components. A pixel over water

the approximated surface reflectance is greater than 0.027 [112] as in

(21)

roximated surface reflectance is less than or equal to 0.027, then the pixel is

is the wavelength and is defined a

is the total top-of-the-

is the transmittance of the atmospheric gases, sun to surface and

is the direct transmittance of the atmosphere through Rayleigh and aerosols

is the glint reflectance due to specular sun reflection

is reflectance originated from molecular (Rayleigh) scattering; and

Page 87:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 87 of 249

09/28/10

td

λ` a

is the total transmittance of the atmosphere through Rayleigh scattering from the surface to the

sensor.

Because the band data are exponentially distributed, log-normalization is used within HeBIS so that

the data may be more easily processed. This normalization process consists of applying the natural

log function to the data within each band and determining the minimum and maximum values of each

of these transformed bands. These band minima and maxima are then used to scale the range of the

normalized values for each band such that the resulting range 2 [0.0, 1.0].

Figure 21 shows a cloud/no-cloud class breakdown according to wavelength in the 15 bands that are

included in the dataset. Figure 22 through Figure 36 are detailed zooms of this plot. These are

scatter plots in which the cloud class (C0) is represented as a “1” on the abscissa and the no-cloud

class (C1) is plotted as a “-1”. The ordinate values are the log-normalized values of band magnitudes

for each of the pixels in the dataset.

Figure 21. Cloud/no-cloud class breakdown according to specific wavelength-band feature.

Page 88:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 88 of 249

09/28/10

Figure 22. 412 nm cloud/no-cloud scatter plot with cloud (C0) pixels represented as 1 and no-cloud (C0) pixels

represented as -1 on the abscissa. Magnitudes are log-normalized and biased and scaled to fall within the range

[0, 1].

Figure 23. 443 nm cloud/no-cloud scatter plot with cloud (C0) pixels represented as 1 and no-cloud (C0) pixels

represented as -1 on the abscissa. Magnitudes are log-normalized and biased and scaled to fall within the range

[0, 1].

Page 89:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 89 of 249

09/28/10

Figure 24. 469 nm cloud/no-cloud scatter plot with cloud (C0) pixels represented as 1 and no-cloud (C0) pixels

represented as -1 on the abscissa. Magnitudes are log-normalized and biased and scaled to fall within the range

[0, 1].

Figure 25. 488 nm cloud/no-cloud scatter plot with cloud (C0) pixels represented as 1 and no-cloud (C0) pixels

represented as -1 on the abscissa. Magnitudes are log-normalized and biased and scaled to fall within the range

[0, 1].

Figure 26. 531 nm cloud/no-cloud scatter plot with cloud (C0) pixels represented as 1 and no-cloud (C0) pixels

represented as -1 on the abscissa. Magnitudes are log-normalized and biased and scaled to fall within the range

[0, 1].

Page 90:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 90 of 249

09/28/10

Figure 27. 551 nm cloud/no-cloud scatter plot with cloud (C0) pixels represented as 1 and no-cloud (C0) pixels

represented as -1 on the abscissa. Magnitudes are log-normalized and biased and scaled to fall within the range

[0, 1].

Figure 28. 555 nm cloud/no-cloud scatter plot with cloud (C0) pixels represented as 1 and no-cloud (C0) pixels

represented as -1 on the abscissa. Magnitudes are log-normalized and biased and scaled to fall within the range

[0, 1].

Page 91:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 91 of 249

09/28/10

Figure 29. 645 nm cloud/no-cloud scatter plot with cloud (C0) pixels represented as 1 and no-cloud (C0) pixels

represented as -1 on the abscissa. Magnitudes are log-normalized and biased and scaled to fall within the range

[0, 1].

Figure 30. 667 nm cloud/no-cloud scatter plot with cloud (C0) pixels represented as 1 and no-cloud (C0) pixels

represented as -1 on the abscissa. Magnitudes are log-normalized and biased and scaled to fall within the range

[0, 1].

Figure 31. 678 nm cloud/no-cloud scatter plot with cloud (C0) pixels represented as 1 and no-cloud (C0) pixels

represented as -1 on the abscissa. Magnitudes are log-normalized and biased and scaled to fall within the range

[0, 1].

Page 92:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 92 of 249

09/28/10

Figure 32. 748 nm cloud/no-cloud scatter plot with cloud (C0) pixels represented as 1 and no-cloud (C0) pixels

represented as -1 on the abscissa. Magnitudes are log-normalized and biased and scaled to fall within the range

[0, 1].

Figure 33. 859 nm cloud/no-cloud scatter plot with cloud (C0) pixels represented as 1 and no-cloud (C0) pixels

represented as -1 on the abscissa. Magnitudes are log-normalized and biased and scaled to fall within the range

[0, 1].

Page 93:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 93 of 249

09/28/10

Figure 34. 869 nm cloud/no-cloud scatter plot with cloud (C0) pixels represented as 1 and no-cloud (C0) pixels

represented as -1 on the abscissa. Magnitudes are log-normalized and biased and scaled to fall within the range

[0, 1].

Figure 35. 1240 nm cloud/no-cloud scatter plot with cloud (C0) pixels represented as 1 and no-cloud (C0) pixels

represented as -1 on the abscissa. Magnitudes are log-normalized and biased and scaled to fall within the range

[0, 1].

Page 94:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 94 of 249

09/28/10

Figure 36. 2130 nm cloud/no-cloud scatter plot with cloud (C0) pixels represented as 1 and no-cloud (C0) pixels

represented as -1 on the abscissa. Magnitudes are log-normalized and biased and scaled to fall within the range

[0, 1].

Page 95:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 95 of 249

09/28/10

4.3. The construction of simple genetic regulatory networks Hybridization Background

Algorithm Discussion Remote Sensing Background

GRN Analyses GRN Training with PSO

GRN Action Analyses Remote Sensing Application – Analyses and Comparisons

Robustness Analyses

Now, we will begin the research experiments by examining the construction and efficacy of simple

genetic regulatory networks.

4.3.1. Introduction/Methodology

This section deals with the construction of simple genetic regulatory networks. Its purpose is the

determination as to whether an artificial GRN is being created within an experimental setup which

isolates a portion of the HeBIS operational environment. This portion is dedicated to management of

the protein environment.

We will examine HeBIS in both the context of the underpinnings of a GRN as well as examine

simple GRN processes in more detail. The underpinnings include diffusion of an artificial protein

through an artificial matrix in addition to a study of activation and inhibition of proteins. These

proteins are controlled through the protein switches which are contained within the genes which can

code for regulatory, environmental and output proteins. Following this, simple GRN processes will

be examined by characterizing gene activation and inhibition of selected artificial genomes.

Within this experimental setup, only genome complexity is examined. We want to determine if

complexity can be introduced through appropriate manipulation of our logic/control processor of

choice, the artificial genome. For our purposes, complexity is defined as follows:

• The number of proteins in the environmental matrix changes and becomes greater than the number contained in the matrix when a zero-length genome is used.

• Patterns of gene activation exist within the genome over the course of HeBIS simulations for non-zero-length genomes.

These gene-level patterns of activation can be used potentially for the stages of classification which

will be studied in later sections.

Page 96:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 96 of 249

09/28/10

Our baselines is the zero-length genome, i.e. a genome which has no explicit genes that can be sued

for interaction with environmental proteins.

4.3.2. Experiments 1 and 2 – Proteins

The protein simulations will only examine the proteins in the base GRN and they will not deal with

any intracellular processes. Thus, for this analysis, the intracellular SOFM kernel remains

deactivated.

Two experiments are conducted:

• Experiment 1: Baseline protein simulations, and

• Experiment 2: Baseline multi-gene protein simulations.

4.3.2.1. Setup

Experiment 1

In this test, 15 proteins corresponding to normalized multispectral channel information from the

cloud/no-cloud dataset are injected into the protein environment. The number of genes in the genome

is set to 0 genes and we will determine the number of proteins that are present in the environmental

matrix. This variable will be sampled periodically as the protein diffusion simulation progresses.

The simulations parameters for this experiment are listed in Table 8.

Table 8. Simulation Parameters for Experiment 1.

Simulation Parameter Value

# Trials 100

# Genes 0

Protein diffusion rate 0.1

Minimum protein environmental concentration 0.01

Maximum number of network diffusion iterations 1000

PSO breedings 10

Topology and size of artificial protein environment 3x3x17

Protein correlation (affinity) 0.9

Reaction probability 0.0 %

# PSO particles 200

This baseline will aid in the determination as to whether a GRN and associated complexity can be

introduced into the system by an appropriate genome. The PSO Breeding parameter is set to a non-

zero value so as to maintain constancy throughout the follow-on simulations.

Page 97:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 97 of 249

09/28/10

Experiment 2

This experiment also injects 15 normalized proteins that correspond to cloud/no-cloud spectral

information. However, the activated genome is one which possesses three (3) genes instead of zero

(0) genes as in Experiment 1. The number of proteins in the environmental matrix is tracked

periodically during the diffusion simulation. Simulation parameters for this experiment are listed in

Table 9.

Table 9. Simulation Parameters for Experiment 2.

Simulation Parameter Value

# Trials 100

# Genes 3

Protein diffusion rate 0.1

Minimum protein environmental concentration 0.01

Maximum number of network diffusion iterations 1000

PSO breedings 10

Topology and size of artificial protein environment 3x3x17

Protein correlation (affinity) 0.9

Reaction probability 0.0 %

# PSO particles 200

Protein correlation refers to the affinity that two proteins have for each other as described in detail in

the algorithm section of this document. Higher values of correlation [0,1] imply that the proteins

only match if they are very similar to each other.

It will be determined in Experiment 2 if the HeBIS PSO interacting with a small multi-gene genome

can induce complexity through protein diffusion patterns and protein creation that are different from

the results obtained through the Experiment 1 baseline.

4.3.2.2. Experiments 1 and 2 results and discussion

The proteins that are used as input in these initial two experiments consist only of the artificial

proteins that have been mapped from the 15 spectral input bands for a test case of 18 pixels- 9 pairs

of C0 and C1 case exemplars. These input spectral proteins are assigned random protein parameters-

essentially the 4 bytes that describe the unique coding for the artificial protein. Since only 254 values

out of the possible 256 potential values for each description byte are used, this provides a potential

space of 64516 (2543) proteins which is somewhat less than 65536 that would be available if all 256

values for each byte were used. Each trial is conducted with a randomly chosen set of parameters

from this set.

Page 98:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 98 of 249

09/28/10

Figure 37 and Figure 38 show the average number of proteins and corresponding standard deviations

that are present in the environment for both the 0-gene (Experiment 1) and 3-gene (Experiment 2)

cases at the end of each iteration of the simulation.

Figure 37. Baseline number of proteins in environmental lattice for zero-length genome in Experiment 1. Vertical

bars correspond to the standard deviation of the sample mean.

Figure 38. Baseline number of proteins in environmental lattice for the 3-gene genome in Experiment 2. Vertical

bars correspond to the standard deviation of the sample mean.

Page 99:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 99 of 249

09/28/10

This information is derived by averaging the number of environmental proteins at each iteration

across all 200 particles of the particle swarm as well as across all 18 pixels that are used for the

specific trial. There is a small standard deviation among the trials after the end of iteration 0- the step

at which the 15 initial proteins have been injected into the environmental lattice. Higher standard

deviations are seen during the middle portion of the sequence of iterations as the proteins with lower

normalized concentrations are removed from the matrix through diffusion. The abrupt drop in the

number of proteins in the matrix at the end of the final iteration occurs when the protein

concentrations fall below the 0.01 minimum normalized concentration level. At that point, all the

proteins have concentrations less than this threshold and are subsequently removed from the matrix.

Table 10 summarizes the statistics that describe the diffusion characteristics for these initial two

experiments.

Table 10. Statistical Summary for Experiments 1 and 2.

Experiment 1 1 2 2

Data iteration # proteins iteration # proteins

Reaction 0 0 0 0

Max. 4.800e+001 1.497e+001 4.700e+001 1.495e+001 Mean 2.400e+001 9.844e+000 2.350e+001 9.898e+000 Median 2.400e+001 1.028e+001 2.350e+001 1.049e+001 Min. 0.000e+000 3.506e+000 0.000e+000 4.087e+000 Mode 0.000e+000 4.563e+000 0.000e+000 4.588e+000 Range 4.800e+001 1.146e+001 4.700e+001 1.086e+001 Standard deviation 1.4289e+001 4.290e+000 1.400e+001 4.204e+000

In Experiment 1, an average of 9.84 proteins remained active during the maximum 48 iterations that

occurred over the range of trials. Similarly, an average of 9.90 proteins had concentrations above the

minimum concentration threshold during a maximum of 47 iterations for the second experiment.

This slight difference is due to the fact that 200 different randomized 18-point training datasets were

used for the two experiments.

Page 100:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 100 of 249

09/28/10

A cubic equation was fitted to the averaged protein data for each of the two experiments. Norms of

the residuals from the fitted curves were 2.7599 and 2.6116, respectively. These residuals indicate

the “goodness of fit” of the curve and the average deviations of the data from this curve.

Another highlight is that the average number of proteins in the lattice as a function of iteration

number is dependent on the diffusion parameter from (16). 0.1 was used for this parameter and was

chosen to determine whether the system was working and not necessarily as an optimum. A diffusion

parameter greater than 0.1- the value used in these experiments- would maintain a number of

proteins in the environment lower than for these baselines. Conversely, a diffusion parameter with a

value less than the baseline would cause more proteins to remain in the environmental matrix for a

longer period of time. The effects on HeBIS operation by changing the diffusion parameter will be

examined in more detail later in this dissertation in Experiment 14.

4.3.2.3. Experiments 1 and 2 conclusions

It can be concluded that HeBIS’ protein diffusion works as it has been shown that artificial proteins

diffuse through the artificial environment. Figure 39 shows that the diffusion characteristics of the

protein environments are essentially the same for the 0-gene and 3-gene experiments.

Figure 39. Number of proteins in environment compared between the baseline genome from Experiment 1 and the

multi-gene genome from Experiment 2. Vertical bars correspond to the standard deviation of the sample mean.

Page 101:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 101 of 249

09/28/10

This was an expected result because the more complex parts of the GRN were deactivated for these

two experiments. An example of protein diffusion over time is presented in the Appendix in Figure

116 through Figure 120 , beginning on page 235.

4.3.3. Experiments 3 and 4 - Protein Chemistry

The next experimental phase examines the addition of complexity into the protein matrix and whether

this changes the complexity of the evolved genetic regulatory network. In this case, a straightforward

protein chemistry is introduced into the environment. This chemistry has been discussed prior to this

in 3.1.2.2 (page 64), the section that treats HeBIS protein communications in detail. The import

aspect of this chemistry is the probability of reaction and how it affects the creation of new proteins

both in a zero-gene instance and a multi-gene case. This probability of reaction can potentially be

thought of as a level of mutation in the GRN. This would roughly correspond to mutation in the

context of optimization based on genetic algorithms.

Two experiments are conducted:

• Experiment 3: Protein simulation with a zero-length genome and non-zero reaction probabilities.

• Experiment 4: Protein simulation with a multi-gene genome and non-zero reaction probabilities.

4.3.3.1. Setup

Experiment 3

This study is similar to Experiment 1 except for there being a non-zero protein reaction probability.

The number of proteins in the environmental matrix is also sampled periodically during the

simulation. 75 trials are split equally among three reaction probabilities: 0.1%, 1%, and 10%.

Simulation parameters for this experiment are listed in Table 11.

Table 11. Simulation Parameters for Experiment 3.

Simulation Parameter Value

# Trials 75

# Genes 0

Protein diffusion rate 0.1

Minimum protein environmental concentration 0.01

Maximum number of network diffusion iterations 1000

PSO breedings 10

Page 102:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 102 of 249

09/28/10

Topology and size of artificial protein environment 3x3x17

Protein correlation (affinity) 0.9

Reaction probability 0.1 %, 1 %, 10 %

# PSO particles 200

Protein creation and diffusion complexity are examined in light of the chosen reaction probabilities.

Experiment 4

This is similar to Experiment 2, however now we examine a non-zero reaction probability with a

multi-gene genome. As in Experiment 3, the number of environmental proteins is sampled

periodically and compared. Simulation parameters are listed in Table 12.

Table 12. Simulation Parameters for Experiment 4.

Simulation Parameter Value

# Trials 75

# Genes 3

Protein diffusion rate 0.1

Minimum protein environmental concentration 0.01

Maximum number of network diffusion iterations 1000

PSO breedings 10

Topology and size of artificial protein environment 3x3x17

Protein correlation (affinity) 0.9

Reaction probability 0.1 %, 1 %, 10 %

# PSO particles 200

With these data, we will examine whether the number of created proteins is different than those in the

Experiments in Section 4.3.2. Protein diffusion and creation patters are also briefly examined.

4.3.3.2. Experiments 3 and 4 results and discussion

Overall, there is a significant increase in the mean number of proteins in the environmental matrix for

experiments 3 and 4 when compared to that in experiments 1 and 2. The initial two experiments are

considered to be the baseline with 0% probability of protein chemistry occurring. As a reminder,

protein chemistry describes the potential creation and destruction of environmental proteins based

purely on “chemical” reactions between the proteins. This is different, but is also a component of the

creation and destruction of proteins based on an evolved GRN. Differences in protein

creation/destruction for experiments 3 and 4 are linked to the probability of a chemical reaction

Page 103:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 103 of 249

09/28/10

occurring over the simulation run. Cubic parameterized curves fitted to the average protein data is

also used as an additional proxy for the complexity in protein interactions that occurs in light of an

underlying probabilistic chemistry. The residuals associated with these fitted cubic equations and the

data indicate an overall error that is used as a type of texture to further describe the complexity

associated with protein interaction. The similar responses for the 0-gene and 3-gene trials for

experiments 3 and 4 were expected since an evolved GRN was not created or activated for these

trials.

Figure 40 and Figure 41 point out the averaged number of proteins in the environmental lattice for

the 0-gene and 3-gene genomes, respectively.

Figure 40. Number of proteins in environmental lattice for a 0-gene genome vs. cellular iteration for reaction

probabilities of 0 %, 0.1 %, 1 %, and 10 % with error bars removed for clarity.

Page 104:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 104 of 249

09/28/10

Figure 41. Number of proteins in environmental lattice for a 3-gene genome vs. cellular iteration for reaction

probabilities of 0 %, 0.1 %, 1 %, and 10 % with error bars removed for clarity.

Both show a significant change in the case in which the reaction probability is at its highest- 10%.

The mean number of active proteins for the baseline 0% (experiments 1 and 2), 0.1%, and 1.0%

reaction probabilities show a dependence on protein production (beyond the initial 15 multispectral

proteins injected at iteration 0) that increases slightly as the reaction probability increases.

The mean number of proteins in the lattice ranges from 9.8437 (0% probability case) to 14.4816

(10% probability case) for the 0-gene genome and 9.8983 (0% probability case) to 15.3183% (10%

probability case) for the 3-gene genome. Figure 42, Figure 43, Figure 44, Figure 45, Figure 46 and

Figure 47 highlight these results in more detail in addition to the high standard deviations relative to

the mean for the number of active environmental proteins.

Page 105:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 105 of 249

09/28/10

Figure 42. Number of proteins in environmental lattice for a 0-gene genome vs. cellular iteration for a reaction

probability of 0.1 %. Vertical bars correspond to the standard deviation of the sample mean.

Figure 43. Number of proteins in environmental lattice for a 0-gene genome vs. cellular iteration for a reaction

probability of 1.0 %. Vertical bars correspond to the standard deviation of the sample mean.

Page 106:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 106 of 249

09/28/10

Figure 44. Number of proteins in environmental lattice for a 0-gene genome vs. cellular iteration for a reaction

probability of 10 %. Vertical bars correspond to the standard deviation of the sample mean.

Figure 45. Number of proteins in environmental lattice for a 3-gene genome vs. cellular iteration for a reaction

probability of 0.1 %. Vertical bars correspond to the standard deviation of the sample mean.

Page 107:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 107 of 249

09/28/10

Figure 46. Number of proteins in environmental lattice for a 3-gene genome vs. cellular iteration for a reaction

probability of 1.0 %. Vertical bars correspond to the standard deviation of the sample mean.

Figure 47. Number of proteins in environmental lattice for a 3-gene genome vs. cellular iteration for a reaction

probability of 10 %. Vertical bars correspond to the standard deviation of the sample mean.

Page 108:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 108 of 249

09/28/10

The 0.1% and 1% reactions (Figure 42, Figure 43, Figure 45 and Figure 46) did not add a large number

of extra proteins into the lattice compared to the 0% baselines in Figure 37 and Figure 38. However,

the data for the 10% reaction probability show a large increase in the average number of active

proteins before the protein concentrations drop to the minimum supported during the simulation.

This is apparent relative to the 0% baseline as well as the 0.1 % and 1.0% cases for both genomes.

On average, the 0-gene instance at 10% probability experienced 70 iterations whereas the 3-gene

instance at the same probability of reaction went through 76 iterations. Table 13 and Table 14

summarize other differences in the mean number of experimental iterations in addition to the mean

numbers of proteins present in the environmental lattice.

Table 13. Statistical Summary for Experiment 3.

Experiment 3 3 3 3 3 3

Data Type iteration #proteins iteration #proteins iteration #proteins

Reaction 10 10 1 1 0.1 0.1

Max. 6.900e+001 2.456e+001 4.700e+001 1.534e+001 4.700e+001 1.496e+001

Mean 3.450e+001 1.448e+001 2.350e+001 1.035e+001 2.350e+001 1.041e+001

Median 3.450e+001 1.526e+001 2.3500000e+001 1.113e+001 2.350e+001 1.104e+001

Min. 0.000e+000 2.000e+000 0.000e+000 3.438e+000 0.000e+000 4.438e+000

Mode 0.000e+000 3.800e+000 0.000e+000 1.496e+001 0.000e+000 4.563e+000

Range 6.900e+001 2.256e+001 4.700e+001 1.190e+001 4.700e+001 1.053e+001

Standard deviation 2.035e+001 7.030e+000 1.400e+001 4.348e+000 1.400e+001 3.845e+000

Table 14. Statistical Summary for Experiment 4.

Experiment 4 4 4 4 4 4

Data Type iteration #proteins iteration #proteins iteration #proteins

Reaction 10 10 1 1 0.1 0.1

Max. 7.600e+001 2.582e+001 4.700e+001 1.506e+001 4.800e+001 1.498e+001

Mean 3.800e+001 1.532e+001 2.350e+001 1.024e+001 2.400e+001 1.005e+001

Median 3.800e+001 1.768e+001 2.350e+001 1.181e+001 2.400e+001 1.049e+001

Min. 0.000e+000 1.000e+000 0.000e+000 3.000e+000 0.000e+000 4.778e+000

Mode 0.000e+000 1.000e+000 0.000e+000 3.556e+000 0.000e+000 5.182e+000

Range 7.600e+001 2.482e+001 4.700e+001 1.206e+001 4.800e+001 1.020e+001

Standard deviation 2.237e+001 7.711e+000 1.400e+001 4.637e+000 1.429e+001 3.677e+000

Page 109:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 109 of 249

09/28/10

Residual norm (texture) for the fitted cubic equations are presented in Figure 48, Figure 49, Figure

50, Figure 51, Figure 52 and Figure 53.

Figure 48. Fitting and statistical information for a 0-gene genome in an environment with a reaction probability of

0.1 %. Vertical bars on the upper chart correspond to the standard deviation of the sample mean.

Page 110:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 110 of 249

09/28/10

Figure 49. Fitting and statistical information for a 0-gene genome in an environment with a reaction probability of

1.0 %. Vertical bars on the upper chart correspond to the standard deviation of the sample mean.

Figure 50. Fitting and statistical information for a 0-gene genome in an environment with a reaction probability of

10%. Vertical bars on the upper chart correspond to the standard deviation of the sample mean.

Page 111:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 111 of 249

09/28/10

Figure 51. Fitting and statistical information for a 3-gene genome in an environment with a reaction probability of

0.1 %. Vertical bars on the upper chart correspond to the standard deviation of the sample mean.

Page 112:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 112 of 249

09/28/10

Figure 52. Fitting and statistical information for a 3-gene genome in an environment with a reaction probability of

1.0 %. Vertical bars on the upper chart correspond to the standard deviation of the sample mean.

Figure 53. Fitting and statistical information for a 3-gene genome in an environment with a reaction probability of

10 %. Vertical bars on the upper chart correspond to the standard deviation of the sample mean.

Page 113:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 113 of 249

09/28/10

For the 0-gene genome, residuals increased from 2.618 through 3.635 and 13.9773 as the reaction

probability increased from 0.1 % to 10 %. This compares with 2.7599 for the 0% baseline from

experiment 1. Residuals for the 3-gene genome showed an increase from 3.0707 through 3.5035 and

14.9639 as the corresponding reaction probability increased from 0.1 % to 10%. The 10 % baseline

for 3 genes had a residual of 2.6116.

4.3.3.3. Experiments 3 and 4 conclusions

Comparing the deactivated protein chemistries in experiments 1 and 2 with the corresponding

experiments with the chemistries activated, it is seen that more proteins are available over time within

the lattice with nonzero reaction probabilities. The average number of iterations before the available

proteins fall below the minimum concentration threshold remained about the same for the 0 %, 0.1 %,

and 1.0 % cases, regardless of the number of genes present in the genome. However, with 10 %

reaction probability, the average number of lattice iterations increased significantly as can be seen in

Figure 54.

Figure 54. Comparison between 0-gene and 3-gene genomes for varying environmental reaction probabilities.

Error bars removed for clarity.

The mean number of active proteins as well as the maximum number of proteins tended to increase

for both genome cases as the probability of protein reaction increased. For the 0-gene instance (Table

10 and Table 13), the mean increased from 9.8437 (0 %) to 14.4818 (10 %) and the maximum

Page 114:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 114 of 249

09/28/10

number of proteins in the lattice increased from 14.965 ( 0%) to 24.56 (10 %). For the 3-gene

instance (Table 10 and Table 14), the mean increased from 9.8983 (0 %) to 15.3183 (10 %) and the

maximum number of proteins in the lattice increased from 14.945 (0 %) to 25.82 (10 %). Both

results show a significant change in environmental behavior when protein chemistry is activated.

Complexity of protein production also shows up as ripples in the averaged protein data in the

corresponding plots (Figure 42, Figure 43, Figure 44, Figure 45, Figure 46 and Figure 47) for the 0.1

%, 1.0 %, and 10 % activated protein chemistry experiments. Additionally, the levels of residual

norm increases as the reaction probabilities increase- as shown in the fitted cubic equations. This

indicate a complex change in the average number of proteins and a change in texture.

The additional proteins present in the matrix (compared to the 0 % baselines) potentially mean that

there are more possibilities for exploring the search space of solutions associated with the evolved

GRN. Additionally, the “mutations” produced by the protein reactions can potentially be used to

improve exploration within the protein parameter space for classification. These issues will be

explored in more detail in Experiment 14.

4.3.4. Experiment 5 - Gene activation

Genes in HeBIS can be thought of as being modules of logic and control that are modified by the

presence of evolved proteins that match the switch region in a specified gene. Investigation into the

workings of these genes in an intracellular context will be examined in more detail in a later section

of this dissertation. For now, it needs to be determined whether and to what extent gene activation

and inhibition can occur. Can gene activation/inhibition patterns occur within the HeBIS GRN?

For this examination, a single experiment will be conducted:

• Experiment 5: Gene Activation/Inhibition Complexity

4.3.4.1. Setup

Experiment 5

Periodic sampling of the genes’ expression in time occurs for a set of genomes. It is determined if

these genes have been activated or subsequently inhibited. These activations/inhibitions of the genes

are caused by the proteins that are instantiated in the protein diffusion matrix. The genes are allowed

Page 115:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 115 of 249

09/28/10

to activate and inhibit the functionality of evolved GRN proteins only for environmental protein

types. Regulatory proteins are not considered because they are constrained to intracellular

functionality that is not examined at this time. Cellular cloning is allowed at this stage. Other

simulation parameters for this experiment are listed in Table 15.

Table 15. Simulation Parameters for Experiment 5.

Simulation Parameter Value

# Trials 3

# Genes (environmental) 3, 10, 40

Protein diffusion rate 0.1

Minimum protein environmental concentration 0.01

Maximum number of network diffusion iterations 1000

PSO breedings 10

Topology and size of artificial protein environment 3x3x17

Protein correlation (affinity) 0.9

Reaction probability 0.0 %

# PSO particles 200

Several genomes are examined for each of the trials. They are selected to highlight certain

characteristics of these simple, evolved GRNs. The fitness function used in this case is the standard

HeBIS fitness function which is discussed in detail in 4.4.2 on page 129.

4.3.4.2. Experiment 5 results and discussion

Examples of complex GRN behaviors are highlighted in this section by using genome activation

maps that are taken directly from the PSO at specified cellular iterations. These genome activation

maps summarize the activity of the evolved GRN in a graphical format. The before-mentioned

fitness function is based on a single protein output. This fitness function is use to examine whether

activation/inhibition of genes within evolved GRNs works., but not necessarily how well these

actions work.

An example of a gene activation map is in Figure 55. With a gene activation map, the response of the

entire genome (collection of genes) within a cell is summarized according to environmental and

cellular stimuli. Examining this image in more detail, from left to right, the vertical columns that are

separated by a thin black line represent a particular gene number. From top to bottom, the rows that

are separated by the black lines represent the activity level of the genes at that particular time iteration

of time within the simulation. Genes are indexed beginning with 0, for example Gene0, on the left-

hand side of the image and they increment as the map is traversed towards the right-hand side to a

Page 116:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor

maximum of GeneN-1. Time indices begin at the

downwards. A white color in the intersection of gene number and

that that specific gene did not activate during that time inte

did activate during the interval. In this example, gene 1 (the second column) activated in iteration 0

and continued activating through iteration 20.

The first 9 genes are infrastructure genes within the genome and can also be expressed during this

experiment. For example, G

Gene1 is the output gene associated with Class

numbered beginning with Gene

genome refers to a genome that contains M environmental genes with a total of M+3 genes.

Genomes are numbered according to their

In Figure 56, Gene1 is activated for both imaged

and right-hand sides of the image. Activa

iteration 20 whereas activation of Gene1 in Genome

ime indices begin at the top with 0 and increment as the map is traversed

color in the intersection of gene number and cellular

that that specific gene did not activate during that time interval and a red color indicates that the gene

did activate during the interval. In this example, gene 1 (the second column) activated in iteration 0

and continued activating through iteration 20.

Figure 55. Example gene activation map.

The first 9 genes are infrastructure genes within the genome and can also be expressed during this

For example, Gene0, is the output gene associated with Class 0, C0, th

is the output gene associated with Class 1, C1, the no-cloud class. Environmental genes are

numbered beginning with Gene9. To simplify the nomenclature, for this experiment an M

genome refers to a genome that contains M environmental genes with a total of M+3 genes.

according to their particle index in the PSO used in this experiment.

is activated for both imaged 3-gene genomes, 93 and 146, respectively on the left

hand sides of the image. Activation of Genome93 begins in cellular iteration 0 and ends in

iteration 20 whereas activation of Gene1 in Genome146 begins in iteration 6 and ends in iteration 12.

116 of 249

top with 0 and increment as the map is traversed

cellular iteration index indicates

rval and a red color indicates that the gene

did activate during the interval. In this example, gene 1 (the second column) activated in iteration 0

The first 9 genes are infrastructure genes within the genome and can also be expressed during this

Class 0, C0, the cloud class and

cloud class. Environmental genes are

To simplify the nomenclature, for this experiment an M-gene

genome refers to a genome that contains M environmental genes with a total of M+3 genes.

index in the PSO used in this experiment.

genomes, 93 and 146, respectively on the left

begins in cellular iteration 0 and ends in

begins in iteration 6 and ends in iteration 12.

Page 117:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor

Figure 56. 3-gene genome activation

respectively.

Examining Genome93 it is seen that the protein concentrations which activate the switch for the gene

evolved such that they activate once the test datum is injected into the environmental ma

of artificial proteins. Since protein chemistry is inactive in this example, Genome

and switching on due to an intermediate range of concentrations for the valid protein in the matrix.

This highlights the fact that diffu

progresses through the simulation.

Figure 57 highlights differing genetic responses between

being applied to the genomes as an

(a)

Figure 57. 3-gene genome activation vs.

(a) (b)

activation vs. iteration for C0 (178) using genomes 93 and 146

it is seen that the protein concentrations which activate the switch for the gene

evolved such that they activate once the test datum is injected into the environmental ma

of artificial proteins. Since protein chemistry is inactive in this example, Genome

and switching on due to an intermediate range of concentrations for the valid protein in the matrix.

This highlights the fact that diffusion is constantly decreasing protein concentrations as time

progresses through the simulation.

highlights differing genetic responses between 3-gene genomes, 9 and 147

he genomes as an exemplar from C0.

(a) (b)

activation vs. iteration for C0 (142) using genomes 9 and 147, respectively

117 of 249

93 and 146 in (a) and (b),

it is seen that the protein concentrations which activate the switch for the gene

evolved such that they activate once the test datum is injected into the environmental matrix as a set

of artificial proteins. Since protein chemistry is inactive in this example, Genome146 is thus reacting

and switching on due to an intermediate range of concentrations for the valid protein in the matrix.

sion is constantly decreasing protein concentrations as time

9 and 147, with pixel 142

(142) using genomes 9 and 147, respectively, in (a) and (b).

Page 118:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor

The maps show that Gene1 activated with Genome

interesting point is that different genes activated and the delta time of the response for each of the

genomes was slightly different: Gene

after iteration 24 whereas Gene

A sustained gene expression response is exemplified in

genome, Gene2 is expressed for 30 cel

Figure 58. 3-gene genome activation vs. cellular iteration for C0 (48) using genome 178

The first example of cellular cloning is contained in

GRN that has been evolved with

which is from the C0 class. The left

cell whereas the middle and right

cloned cells that were created during the training of this genome.

activated with Genome9 and Gene5 activated with Genome

interesting point is that different genes activated and the delta time of the response for each of the

genomes was slightly different: Gene1 responded beginning at iteration 2 and ended its activation

Gene5 activated in iteration 0 and ended its activation after iteration 26.

A sustained gene expression response is exemplified in Figure 58 with 3-gene Genome

is expressed for 30 cellular iterations.

gene genome activation vs. cellular iteration for C0 (48) using genome 178

The first example of cellular cloning is contained in Figure 59. This gene activation map shows the

N that has been evolved with 10-gene Genome117. It produced 2 clones while training on pixel 42

which is from the C0 class. The left-hand portion of the figure is the activation map for the original

right-hand portions of the figure show the activation maps for the two

cloned cells that were created during the training of this genome.

118 of 249

ivated with Genome147. An

interesting point is that different genes activated and the delta time of the response for each of the

responded beginning at iteration 2 and ended its activation

activated in iteration 0 and ended its activation after iteration 26.

gene Genome178. Within the

gene genome activation vs. cellular iteration for C0 (48) using genome 178.

activation map shows the

. It produced 2 clones while training on pixel 42

hand portion of the figure is the activation map for the original

hand portions of the figure show the activation maps for the two

Page 119:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor

(a)

Figure 59. 10-gene genome activation vs. cellular iteration for origina

environmental matrix- C0 (42) test pixel for genome 117

cloned genomes are presented in (b) and (c).

In this case, the same genome reacted

environmental lattice. The clonings occurred at different baseline times and are

can be considered as time-encoded patterns of protein concentrations within the lattice.

Although each cell (original and 2

for each of these three cells. Gene

gene activated in Clone0 between 2

24.

Complex behavior is further exemplified in

show an evolved GRN for 10

and pixel 51 from C1. Cloning occurred in addition to a modulation of the expression that differed

between the originals and clones in addition to across the classes. Since clones are placed in the

environmental lattice next to the original cell, only a single cell can occupy a specific location with

the environment.

(b) (c)

gene genome activation vs. cellular iteration for original and two cloned cells within the

C0 (42) test pixel for genome 117. The original genome is presented in (a) while the two

cloned genomes are presented in (b) and (c).

ase, the same genome reacted to differing concentrations at different

environmental lattice. The clonings occurred at different baseline times and are

encoded patterns of protein concentrations within the lattice.

Although each cell (original and 2 clones) all shared the same genome, Gene

for each of these three cells. Gene15 activated in the original cell between iterations

between 2 and 26; and the gene in Clone1 activated between iter

Complex behavior is further exemplified in Figure 60 and Figure 61. Respectively, these figures

10-gene Genome117 as it responded to the proteins from pixel 50 from C0

and pixel 51 from C1. Cloning occurred in addition to a modulation of the expression that differed

between the originals and clones in addition to across the classes. Since clones are placed in the

ice next to the original cell, only a single cell can occupy a specific location with

119 of 249

l and two cloned cells within the

The original genome is presented in (a) while the two

ns at different points in the

environmental lattice. The clonings occurred at different baseline times and are responding to what

encoded patterns of protein concentrations within the lattice.

clones) all shared the same genome, Gene15 responded differently

activated in the original cell between iterations 0 and 33; the

activated between iteration 0 and

Respectively, these figures

to the proteins from pixel 50 from C0

and pixel 51 from C1. Cloning occurred in addition to a modulation of the expression that differed

between the originals and clones in addition to across the classes. Since clones are placed in the

ice next to the original cell, only a single cell can occupy a specific location with

Page 120:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor

(a)

Figure 60. 10-gene genome activation vs. cellular iteration

environmental matrix- C0 (50) test pixel using genome 117

and the activations for the cloned cells are shown in (b) and (c).

(a)

Figure 61. 10-gene genome a

environmental matrix- C1 (51) test pixel using genome 117

activations for the cloned cells are listed in (b) and (c).

The varying expression responses in

(or clone) within the matrix and the con

These 6 GRNs activate Gene

original cell expressed Gene

(b) (c)

activation vs. cellular iteration for original and two cloned cells within the

(50) test pixel using genome 117. The activation for the original cell is shown in (a)

and the activations for the cloned cells are shown in (b) and (c).

(b) (c)

gene genome activation vs. cellular iteration for original and two cloned cells within the

C1 (51) test pixel using genome 117. The original cell activation is presented in (a) and the

activations for the cloned cells are listed in (b) and (c).

The varying expression responses in Figure 60 and Figure 61 are dictated by the location of the cell

(or clone) within the matrix and the concentration of the evolved switch protein at that discrete point.

These 6 GRNs activate Gene15, but at different cellular iterations according

original cell expressed Gene15 between iterations 0 and 38, Clone0 expressed it between iterations 1

120 of 249

original and two cloned cells within the

The activation for the original cell is shown in (a)

ctivation vs. cellular iteration for original and two cloned cells within the

The original cell activation is presented in (a) and the

are dictated by the location of the cell

protein at that discrete point.

to the class. For C0, the

xpressed it between iterations 1

Page 121:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 121 of 249

09/28/10

and 32, and Clone1 expressed it between 0 and 31. With C1, the original cell expressed Gene15

between iterations 0 and 28, Clone0 expressed between iterations 3 and 20, and Clone1 expressed

between 0 and 18. As with the prior cloning example, a complex GRN has been created with a

temporal response that is different according to the class and also location within the lattice.

Figure 62 shows that further complexity is possible within the HeBIS GRN. This example consists of

the different expressions which occurred with the same training pixel, but with different 10-gene

genomes taken from the PSO. 10-gene Genomes 91, 123 and 135 were applied to pixel 42 from the

C0 (cloud) class. In this case, cloning has also occurred, but the expressions from the different

genomes are composed of different genes. The first map is for Genome91; the second, third and

fourth maps are for the original and two clones created by Genome123; and the fifth, sixth and seventh

maps are for the original and two clones created by Genome135.

Page 122:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor

(a)

(b)

(c)

Figure 62. 10-gene genome activation vs. cellular iteration for

135 are displayed in (a), (b), and (c), respectively

activations for each of these genomes.

Genome91 shows a single gene response from Gene

differing GRN responses from the originals and the clones due to environmental pr

concentration differences due to location and protein diffusion within the lattice.

The first multiple-gene expression example is summarized in the GRN activation map in

Genome30 consists of 40 environmental genes for which multiple genes were expressed as the GRN

reacted differently to data from different classes.

gene genome activation vs. cellular iteration for the same test pixel C0 (42)

in (a), (b), and (c), respectively. Genomes 123 and 135 show original and two cloned cell

activations for each of these genomes.

shows a single gene response from Gene5 and the remainder of the maps highlight

differing GRN responses from the originals and the clones due to environmental pr

concentration differences due to location and protein diffusion within the lattice.

gene expression example is summarized in the GRN activation map in

nvironmental genes for which multiple genes were expressed as the GRN

reacted differently to data from different classes.

122 of 249

same test pixel C0 (42). Genomes 91, 123, and

23 and 135 show original and two cloned cell

and the remainder of the maps highlight

differing GRN responses from the originals and the clones due to environmental protein

concentration differences due to location and protein diffusion within the lattice.

gene expression example is summarized in the GRN activation map in Figure 63.

nvironmental genes for which multiple genes were expressed as the GRN

Page 123:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor

(a)

(b)

Figure 63. 40-gene genome activation for genome 30 for test pixels C0 (26) in the top image

bottom image (b). The genome shows multi

responses.

The top image in the figure shows the map for pixel 26 which is from C0 and the bottom image

shows the map for pixel 25 which is from C1.

iterations 0 and 31 while also expressing Gene

expressed for the exemplar from C1 and it activated between iterations 0 and 24.

Further GRN complexity is shown with the example of multiple gene expression

cloning. This is exemplified in

Genome66 which is a 40-gene ge

gene genome activation for genome 30 for test pixels C0 (26) in the top image

The genome shows multi-gene activation for C0 and single gene activation for C1 with differing

The top image in the figure shows the map for pixel 26 which is from C0 and the bottom image

l 25 which is from C1. The exemplar from C0 expressed Gene

iterations 0 and 31 while also expressing Gene35 between iterations 2 and 23. Only a single gene was

expressed for the exemplar from C1 and it activated between iterations 0 and 24.

ther GRN complexity is shown with the example of multiple gene expression

cloning. This is exemplified in Figure 64 with complex temporal differences in expression for

gene genome.

123 of 249

gene genome activation for genome 30 for test pixels C0 (26) in the top image (a) and C1 (25) in the

gene activation for C0 and single gene activation for C1 with differing

The top image in the figure shows the map for pixel 26 which is from C0 and the bottom image

The exemplar from C0 expressed Gene8 between

between iterations 2 and 23. Only a single gene was

expressed for the exemplar from C1 and it activated between iterations 0 and 24.

ther GRN complexity is shown with the example of multiple gene expression occurring with

with complex temporal differences in expression for

Page 124:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor

(a)

(b)

(c)

Figure 64. 40-gene genome activation for genome 66

activation profiles for C0 (30) with an orig

same genome, but for a C1 (31).

The GRN causes a clone to be created with the injection of class C0 proteins, pixel 30, at the

beginning of the simulation. However, the GRN for this genome does not create a clone with

genome activation for genome 66. The top (a) and middle(b) images are the multi

activation profiles for C0 (30) with an original and cloned cell. The bottom (c) image show the activation

The GRN causes a clone to be created with the injection of class C0 proteins, pixel 30, at the

beginning of the simulation. However, the GRN for this genome does not create a clone with

124 of 249

images are the multi-gene

image show the activation for the

The GRN causes a clone to be created with the injection of class C0 proteins, pixel 30, at the

beginning of the simulation. However, the GRN for this genome does not create a clone with the

Page 125:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 125 of 249

09/28/10

injection of a class C1 exemplar, pixel 31. The GRN also expresses 2 genes in the original cell, but

only a single gene in the cloned cell.

In detail, for the C0 pixel, genes 3 and 27 are expressed in the original cell, but only Gene3 is

expressed in Clone0. For the C1 pixel, only Gene3 is expressed. Also, the temporal responses differ

for Gene3: in the original cell with the C0 class, Gene3 was expressed between iterations 0 and 31

with Gene27 expressed for only a single iteration beginning at iteration 0. The clone saw an

expression of Gene3 that occurred between iterations 2 and 22. Yet another variation is seen with the

C1 pixel as Gene3 only expressed between iterations 0 and 22.

The final example of complex responses within evolved HeBIS GRNs is displayed in Figure 65. This

highlights different responses given that different class exemplars are applied to the same 40-gene

genome, Genome90.

Page 126:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor

(a)

(b)

(c)

(d)

(e)

(f)

Figure 65. 40-gene genome activation

C1 (55), respectively from top to bottom

gene genome activation of genome 90 for test pixels C0 (30), C1 (31), C0 (32), C0 (34), C0 (54), and

top to bottom, in figures (a) – (f).

126 of 249

0 (32), C0 (34), C0 (54), and

Page 127:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 127 of 249

09/28/10

Six pixel in total were used in this instance, four from class C0 and 2 from class C1. These pixels

were pixel 30, 32, 34 and 54 for C0 and 31 and 55 for C1.

Class(pixel) C0(30), C1(31), C0(32) and C1(33) all cause the evolved GRN to express Gene25.

C0(54) and C1(55) express the GRN differently than the remaining four pixels. The GRN that results

from application of the first four pixels results in expression for a single iteration with a longer-term

expression for C1(55) at Gene36. The last two present GRNs that evolved one more generation

beyond the GRNs from the first four pixels. In the last two, Gene24 and Gene36 show a multiple-

gene expression for C0 whereas only Gene36 is expressed for the C1 exemplar.

4.3.4.3. Experiment 5 conclusions

The fitness function used in Experiment 5 provides for the evolution of a set of rudimentary GRNs

that express complex responses. However, the function is not necessarily geared towards

simultaneous multiple-gene expression. Routine multiple-gene expression could potentially be

accomplished with more complex fitness functions, however the flip-side of this is that the current

function quite possibly does not introduce unnecessary complexity into the structures required for

classification. This will be further discussed in the large-scale classification studies discussed in later

experiments within this thesis.

The PSO is providing a diversity of building-block GRN behaviors such as differing temporal

activations and cloning and follow-on experiments will determine whether these can be harnessed for

classification by adding more complex behaviors to the individual cell.

The observed cloning allows for a time-mediated response within an evolved GRN represented by a

specific genome. Figure 59, Figure 62 and Figure 65 show that the cloning behavior can provide

distinctly different responses for different class pixels.

Page 128:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 128 of 249

09/28/10

4.4. Self-organization in the HeBIS environment Hybridization Background

Algorithm Discussion Remote Sensing Background

GRN Analyses GRN Training with PSO

GRN Action Analyses Remote Sensing Application – Analyses and Comparisons

Robustness Analyses

We have now completed the examination of the genetic regulatory networks that HeBIS can create.

Now, we proceed to studying the presence and variability of self-organization with HeBIS. These

experiments examine this next layer of HeBIS’ functionality and builds from the GRN results. These

self-organization results will be used later as the basis for the detailed look at the utility of HeBIS for

a remote sensing classification application.

Within HeBIS, we examine its abilities to improve the fitness of the classifier genome as the genetic

regulatory network changes dynamically according to the different proteins that are present in the

system. We include the following list as a summation of HeBIS self-organization principles.

i. The number of particles in the particle swarm optimizer.

ii. Breed location of the initial processing cell within the protein diffusion matrix (environment).

iii. Cellular actions that are activated and inhibited according to the protein interactions that comprise the GRN.

iv. Cloning of a processing cell to another location in the environment. v. Protein statistical analogs that can potentially affect the GRN.

vi. Determination of output protein choice for the 2-class classification problems that we have chosen to study.

Although the size of the swarm in the PSO is not technically controlled from within HeBIS, it is

nonetheless the important basis that allows the creation of a genome with self-organization faculties.

Thus, we include this sizing parameter in our list and also examine it and its relation to the creation of

a genome which performs “good” classifications as defined by the fitness of the genome.

4.4.1. Introduction/Methodology

Self-organization is studied through an examination of evolved genomic fitness in the context of a

small-scale classification test based on cross-validation (CV). This CV is performed on a subset of

the training/test dataset and the results are compared when HeBIS self-organization training

Page 129:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 129 of 249

09/28/10

parameters are changed. A subset of the training data is used to facilitate a speed-up in the somewhat

lengthy training process. We compare the fitness of the best evolved genome given changes in these

parameters.

CV for all the experiment sin the self-organization section consist of 20 multispectral pixels sampled

uniformly from the optical cloud/no-cloud training dataset. A 10-fold CV is used which results in 20

test pixels. The base simulation parameters for these tests are listed in Table 16.

Table 16. Simulation Parameters for Self-Organization Experiments.

Simulation Parameter Value

# Trials 100, 25, or 20

# Environmental genes 10

Protein diffusion rate 0.1

Minimum protein environmental concentration 0.01

Maximum number of network diffusion iterations 1000

PSO breedings 10

Topology and size of artificial protein environment 3x3x17

Protein correlation (affinity) 0.9

Reaction probability 0.0 %

# PSO particles 200

# Folds in the cross-validation 10

Test dataset size 20

4.4.2. Fitness function description

The fitness function is used to optimize a genome to produce an accurate classification result during

the training and testing phases. HeBIS uses a single fitness function to judge the efficacy of a

particular genome. This efficacy is based on whether the desired output protein is present in the

environmental protein matrix at an appropriate concentration. The fitness function rewards the

existence of this protein relative to the maximum correlation with the proteins that are present in the

matrix during the training phase. The fitness function also considers the closeness of the match of the

various proteins in the matrix relative to the 4-tuple descriptor of the output protein.

The fitness of the genome consists of a scalar value that is derived from the genome’s response to a

series of pairs of training pixels. Each pair of pixels contains a cloud pixel from class C0 and a no-

cloud pixel from class C1. The fitness scalar, Fitnessvalue

, is defined by

Page 130:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 130 of 249

09/28/10

if (Labeldesired class

== C0)

Fitnessvalue

= 2(CorrC0 max

– 0.5)

else if ( Labeldesired class

== C1)

Fitnessvalue

= 2(0.5 – CorrC0 max

)

(22)

where CorrC0 max

and fitnessValue are contained within the interval, [0.0, 1.0].

This provides a decision boundary equal to 0.5. This decision is dependent on CorrC0 max

for the

desired single output protein that distinguishes between classes C0 and C1. Figure 66 shows this

mapping range for CorrC0 max

.

Figure 66. Decision region mapping and boundary based on the value of CorrC0 max

.

CorrC0 max

is defined as

CorrC0 max = 0.5 Mag

corr

2+ Θ

2b c

(23) with

Magcorr

=Proteintest

N

N

N

N

N

N

ProteinC0 Output

N

N

N

N

N

N

fffffffffffffffffffffffffffffffffffffffffffffffffff for

Proteintest

ProteinC0 Output

N

N

N

N

N

N

fffffffffffffffffffffffffffffffffffffffffffffffffff2 0.0, 1.0B C

(24)

and

Magcorr

=1

ProteinN

N

N

N

ProteinC0 Output

N

N

N

N

N

N

ffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffff for

Proteintest

ProteinC0 Output

N

N

N

N

N

N

fffffffffffffffffffffffffffffffffffffffffffffffffff2 1.0,1b c

(25)

Page 131:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 131 of 249

09/28/10

and

Θcorr =@1

πfffff

θinner angle

+ 1.

(26)

.

Plots corresponding to equations (23), (25) and (26), respectively are included in the Appendices in

Figure 121, Figure 122, and Figure 123, beginning on page 238.

Page 132:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 132 of 249

09/28/10

4.4.3. Experiment 6 - Swarm fitness characterization Hybridization Background

Algorithm Discussion Remote Sensing Background

GRN Analyses GRN Training with PSO

GRN Action Analyses Remote Sensing Application – Analyses and Comparisons

Robustness Analyses

The fitness of the swarm is characterized through a single experiment:

• Experiment 6 Swarm fitness characterization.

4.4.3.1. Setup

Experiment 6

This is a characterization of the particle swarm optimizer and gauges the utility of PSO for optimizing

the fitness function of the genome. All the other self-organization parameters are set to the default

static values- no PSO evolution modifies these parameters.

The best genome fitness is acquired over each CV-fold and this is averaged over the trial to produce a

new averaged CV fitness value. These trials are accumulated as described in Table 17.

Table 17. Experiment 6 Trial Distribution

# Particles in swarm # Trials

1 25

100 25

250 25

500 25

All the training parameters are as defined in section 4.4.

4.4.3.2. Experiment 6 results and discussion

Figure 67, Figure 68, Figure 69 and Figure 70 summarize the evolution of the best fitness achieved

within the particle swarm optimizer given a specific number of particles in the swarm. Ranging from

a minimum of 1 particle to a maximum of 500 particles, the plots show that there is a tendency for the

HeBIS PSO to perform better- that is to achieve higher fitness values- when there are more particles

being used by the optimizer. All four trial distributions indicate that the HeBIS PSO shows a

monotonic increase in average fitness from the first breeding through the end of the last breeding,

breeding 9.

Page 133:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 133 of 249

09/28/10

Figure 67. Average best genome fitness vs. breed # for 1-particle PSO swarm. Vertical bars correspond to the

standard deviation of the sample mean.

Figure 68. Average best genome fitness vs. breed # for 100-particle PSO swarm. Vertical bars correspond to the

standard deviation of the sample mean.

Page 134:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 134 of 249

09/28/10

Figure 69. Average best genome fitness vs. breed # for 250-particle PSO swarm. Vertical bars correspond to the

standard deviation of the sample mean.

Figure 70. Average best genome fitness vs. breed # for 500-particle PSO swarm. Vertical bars correspond to the

standard deviation of the sample mean.

For the 1-particle PSO, the swarm that has the worst optimized fitness, the breeding begins with an

average best fitness of 0.0 and by the end of the final breeding, it has risen to 0.0018308. In the case

of the 1-particle PSO, the fitness is changing due to the fact that several pixels are examined in CV

mode. This 1-particle fitness compares with the highest-achieved average best fitness of 0.0723 for

the 500-particle PSO. The standard deviations for the fitnesses of the genomes within the swarms are

also relatively high, thus indicating that the health of the swarms remains useful over the period of

breeding for the 100, 200 and 500-particle swarms. Standard deviations of the best average fitness

ranges from 0.0018 at the end of the second 1-particle PSO breeding to 0.0797 at the end of the tenth

Page 135:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 135 of 249

09/28/10

breeding of the 500-particle PSO. Table 56 on page 216 in the Appendices summarizes the breeding

data for the swarm in aggregate.

Figure 71 indicates the health of the HeBIS PSO. It shows that the best average peak fitness during

the averaged CV trials increases as the size of the swarm increases. The best average peak fitness

increases from 0.008308 in the case of the 1-particle swarm to 0.02386 for the 100-particle swarm ,

0.0503 for the 250-particle swarm, and finally to 0.0723 for the 500-particle swarm.

Figure 71. Best average peak genome fitness vs. the number of particles in the PSO swarm. Vertical bars

correspond to the standard deviation of the sample mean of the peak fitness for each swarm tested.

4.4.3.3. Experiment 6 conclusions

The HeBIS PSO works and is optimizing the limited portions of the fitness function that have been

activated for this experiment. The observed fitness values are low because of this, but nonetheless,

the optimizer appears to be working. Higher fitness values are seen in later experiments when more

complex cellular and GRN behaviors are activated. Thus, the PSO’s ability to search through a

complex search space for useful GRNs improves as the number of particles increases. This will be

discussed in further detail in the classification experiments.

Page 136:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 136 of 249

09/28/10

4.4.4. Experiment 7 - Initial location of processing cells

HeBIS allows the initial location of the first processing cell in the lattice to be randomly set and

evolved by the PSO. Subsequent breedings allow for this location to be optimized.

4.4.4.1. Setup

Experiment 7

This experiment determines whether any change in the genome’s fitness occurs over time as the PSO

iterates and the initial cell location is statically set to a specific position in the environmental lattice.

Only the initial location of the first (and only) processing cell is allowed to change via the PSO. The

best genome fitness information is acquired over each CV-fold and is then averaged. This new

averaged CV fitness represents the statistic for the trial.

300 trials are used where the initial locations can be varied over the 15 possible locations in a 3x3x17

environmental lattice for a single pixel. Table 18 shows the distribution of trials for this experiment.

Table 18. Trial distribution for Experiment 7.

# Trials Initial Cell Location (z-axis)

20 1

20 2

20 3

20 4

20 5

20 6

20 7

20 8

20 9

20 10

20 11

20 12

20 13

20 14

20 15

4.4.4.2. Experiment 7 results and discussion

The plot in Figure 72 summarizes the results for changing the location of the initial cell among the 15

possible cell locations in the default HeBIS 3x3x17 environmental lattice. Only 15 possible cell

Page 137:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 137 of 249

09/28/10

locations are used because of the 2-location buffer zone in the environmental matrix. In this plot, we

see that the mean fitness values are associated with relatively large standard deviations.

Figure 72. CV average fitness vs. initial cell location from best bred genome. Vertical bars correspond to the

standard deviation of the sample mean.

A peak does occur at location 9 which corresponds to HeBIS feature 9- a MODIS spectral bandwidth

of 483-493 nm. The mean peak fitness over 20 trials for this feature is 0.06287 with a standard

deviation of 0.02691.

4.4.4.3. Experiment 7 conclusions

As with Experiment 6, the mean fitnesses are low because the complex cellular and GRN behaviors

are not active. These trials do not show any significant differences in mean fitness between the initial

cell locations although the peak at HeBIS feature 9 is interesting.

Page 138:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 138 of 249

09/28/10

4.4.5. Experiment 8 - Cellular actions Hybridization Background

Algorithm Discussion Remote Sensing Background

GRN Analyses GRN Training with PSO

GRN Action Analyses Remote Sensing Application – Analyses and Comparisons

Robustness Analyses

Individual cellular actions form the backbone of the HeBIS algorithm. These actions are mediated by

the GRN as it responds to the proteins that are present in the environmental lattice and their

invocation is in turn evolved by the HeBIS PSO.

4.4.5.1. Setup

Experiment 8

The cellular actions experiment poses the question as to whether individual basic cellular actions can

produce changes in the corresponding fitness of a genome and its GRN. The genome is evolved by

application of the PSO and only a single cellular action is potentially allowed to activate during the

set of simulation trials for that particular action. This allows the determination as to whether these

individual cellular actions, which are used by the HeBIS self-organization building blocks, can

change the fitness of a simplified, bare-bones genome. Tied together by a fully activated GRN, we

will examine during the classification experiments whether these actions allow effective classification

of our cloud/no-cloud dataset.

The tested cellular actions include:

i. NOACTION ii. ADDCELL

iii. PRUNESELF iv. ACTIVATEENVIROPROTEIN v. INHIBITENVIROPROTEIN

vi. ACTIVATEREGPROTEIN vii. INHIBITREGPROTEIN

viii. CHANGETOSOFMANDTRAIN ix. CLASSIFY

The default simulation parameters listed in Table 16 on page 129 are used. The initial cell location

along the environment’s z-axis is set to be in the middle of the matrix at location 8. The best genome

fitness is acquired over each CV-fold and this is averaged over the 10 folds of the simulation. This

Page 139:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 139 of 249

09/28/10

new averaged CV fitness is calculated for each of the 25 trials for each tested cellular action listed in

Table 19.

Table 19. Trial Distribution for Experiment 8

Cellular Action Index Cellular Action # Trials

1 NOACTION 25

2 ADDCELL 25

3 PRUNESELF 25

4 ACTIVATEENVIROPROTEIN 25

5 INHIBITENVIROPROTEIN 25

6 ACTIVATEREGPROTEIN 25

7 INHIBITREGPROTEIN 25

8 CHANGETOSOFMANDTRAIN 25

9 CLASSIFY 25

4.4.5.2. Experiment 8 results and discussion

Figure 73 displays a composite plot for comparison of all the cellular actions that are tested in this

experiment. Mean fitness and standard deviations for each of the cellular actions listed in Table 19 is

presented in the composite plot. The cellular action index is mapped to the abscissa of the plot. The

fitness function is the default HeBIS single-protein function that is based on environmental proteins

that are measured to be close to the defined mappings for the C0 and C1 output classes.

Figure 73. CV average fitness vs. activated cellular action. Vertical bars correspond to the standard deviation of

the sample mean.

Page 140:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 140 of 249

09/28/10

The baseline performance for this experiment consists only of the corresponding averaged fitness for

the NOACTION cellular action. This mean fitness is 0.02097 with a corresponding standard

deviation of 0.0358.

Referencing the composite plot in Figure 73, only three additional actions will be discussed in detail

with reference to the baseline. These actions are ADDCELL, ACTIVATEENVIROPROTEIN, and

CLASSIFY.

The first of these, ADDCELL, showed a decrease in the mean fitness to 0.01288 with a standard

deviation of approximately 0.0312. Referencing the relatively low fitness magnitudes from the

experimental data, this is a 38.5 % decrease in mean fitness from the NOACTION baseline.

The second action, ACTIVATEENVIROPROTEIN, shows the largest relative increase in mean

fitness when compared to the mean fitnesses of the other actions. Its mean fitness of 0.06888

corresponds to a 228 % (3.28x) increase in fitness compared to the NOACTION baseline. The

corresponding standard deviation is comparable to that of the other actions in this portion of the study

at 0.03725. This increase is primarily because in most of the trial cases, the only proteins in the

lattice were those that were initially injected into the lattice as the spectral proteins from the specific

exemplar that HeBIS trained with. Thus, proteins beyond this initial set of spectral proteins are

allowed to be expressed since the ACTIVATEENVIROPROTEIN action is active. This allows a

greater chance for the evolved and tested GRN to more closely match the desired output proteins that

are being measured and rated by the HeBIS fitness function as the PSO evolves the crippled GRN.

The CLASSIFY action shows no significant change in mean fitness when compared to the baseline.

CLASSIFY’s mean fitness over the 25 trials is 0.02082 which compares with 0.02097 for

NOACTION. This occurs primarily because of one reason. The first is that the CLASSIFY action

relies on a SOFM that is internal to each cell within the environmental lattice. This SOFM relies on

an internal codebook that has been trained on environmental proteins that are present in the lattice at

the position of the particular cell. However, with these CLASSIFY trials, no SOFM was trained

because the action, CHANGETOSOFMANDTRAIN is inactive. Thus, the classification process that

is started by the invocation of the CLASSIFY action, does not complete.

Page 141:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 141 of 249

09/28/10

4.4.5.3. Experiment 8 conclusions

Examining the composite plot in Figure 73, it is seen that the mutually exclusive invocation of the

different cellular actions does change the measured mean fitness. This reinforces the belief that the

activation or inhibition of these cellular actions by the selected genome and its GRN can be

effectively applied to our classification problem. These constituent behaviors are adding complexity

on their own to the evolved GRN. The use of these cellular actions makes sense when they are used

in conjunction with each other. Hence, the classification experiments, 11-16, examine changes in

both classification and fitness with these cellular actions acting in concert with the evolved GRN

controlling the self-organization of the system. Table 58 presents the data for this experiment on

page 217 in the Appendices.

4.4.6. Experiment 9 - Protein statistical analogs

Statistics corresponding to protein concentration information that is local to each cell can be released

into the protein environmental matrix as the cell iterates through processing. The question as to

whether these statistical analogs are released into the environment is answered by the PSO. Inclusion

of protein statistics may or may not help in the creation of a GRN that is more fit for classification.

4.4.6.1. Setup

Experiment 9

The Protein Statistic Effect experiment varies the creation and use of the cell’s protein statistics that

can be released into the environmental protein matrix. All other cellular actions are deactivated for

this experiment except for the activation and inhibition functions of the regulatory and environmental

proteins. The location of the initial cell is also constrained to be at location 8 which is in the middle

of the default environmental lattice.

The default simulation parameters for this experiment are listed in Table 16. The best genome fitness

is measured over the CV-fold and this value is averaged over the 10 folds. This new averaged fitness

represents the aggregate fitness for this trial.

50 trials are run with protein statistics activated and with the protein descriptors bred by the PSO. A

further 50 trials are run with protein statistics still activated, but with the protein descriptors statically

set to default values. Results from these runs are compared to the baseline 25 NOACTION trials

from Experiment 8. Parameters that are specific to this experiment are listed in Table 20.

Page 142:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 142 of 249

09/28/10

Table 20. Experiment 9 Parameters

Protein Statistics Index Protein Statistics Action Number of Trials

10 STATS_ACTIVE_STATIC 50

11 STATS_ACTIVE_BRED 50

4.4.6.2. Experiment 9 results and discussion

The statistical protein analogs that area released by the individual cell into the environmental lattice

are those described in 3.1.3.2 and 3.3.1.1. Two cases exist for this experiment. In the first,

STATS_ACTIVE_STATIC, the four parameters that describe each of the statistical proteins remain

unmodified as the PSO iterates through successive breedings. Thus, in this case the proteins are

considered to be static. For the second, STATS_ACTIVE_BRED, the four parameters that describe

each of the statistical proteins are allowed to be modified and evolve as the PSO iterates through the

breedings, hence these are considered to be bred proteins. The NOACTION data from Experiment 8

are used as an appropriate baseline comparison since no protein statistics were active in Experiment

8. The results are plotted in Figure 74 with the protein statistics indices used as the abscissa.

Figure 74. CV average fitness vs. activity level of cellular protein statistics. Vertical bars correspond to the

standard deviation of the sample mean.

Page 143:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 143 of 249

09/28/10

The mean fitness for STATS_ACTIVE_STATIC is little changed from the NOACTION case from

Experiment 8. It is 0.02075 with a standard deviation of 0.02635 over the 50 trials whereas the

NOACTION trials resulted in a mean fitness of 0.02097. However, STATS_ACTIVE_BRED shows

a significant change in mean fitness compared to both the NOACTION baseline and the static

statistical protein trials. The mean fitness for STATS_ACTIVE_BRED is 0.30357 with a standard

deviation of 0.0483. This is a 1363 % increase (14.6x) in mean fitness over that for the static

statistical protein case.

4.4.6.3. Experiment 9 conclusions

The inclusion of bred statistical proteins into the environmental lattice significantly changes the mean

fitness of the best genome and its corresponding GRN for the training set of protein-converted pixels.

The addition of the statistical proteins into the environmental lattice adds more proteins that can be

used by the GRN as it is evolved by the PSO according to the fitness function. Allowing the set of

statistics proteins to breed along with the other proteins in the GRN increases the chances for protein

expression and inhibition to occur which is useful for significantly changing the fitness. It should be

noted that these statistics proteins are always released into the environmental lattice without regards

to whether their switch proteins have been activated by other proteins present in the environment. In

this sense, they are considered to be always “on” and no matching of the switch protein is required for

expression.

However, the presence of these statistics is not enough to improve the fitness even more given the

number of breedings in the test. Providing more breedings could potentially allow these proteins to

be subverted and to allow them to be actively have their functionality taken over for improvement in

classification.

4.4.7. Experiment 10 - Output protein comparison

In this study, we examine the effects of HeBIS output proteins that are evolved with the PSO. This

evolution occurs in the context of the GRN. Conceivably, the output proteins can change according

to the GRN proteins encountered during training. Hence, we have another level of self-organization.

Page 144:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 144 of 249

09/28/10

4.4.7.1. Setup

Experiment 10

All other self-organizing cellular actions are deactivated except for the activation and inhibition

functions for the environmental and regulatory proteins. Protein statistics are deactivated. The

output genes are set to be either evolved or locked down to static values. The best genome fitness is

measured over each CV-fold and this is averaged to calculate the aggregate fitness for the test.

Evolved output proteins from the output genes are limited to only the environmental CO output

protein. In this case, C0 is the only protein used within the fitness function to differentiate between

the two classes. Also, the location of the initial cell is constrained to be at location 8 which is the

middle environmental element in the default environmental lattice.

The switch gene in both the fixed and evolved cases is always evolved. 50 trials are run in static

mode and 50 trials are run in PSO-evolved mode. Parameters specific to Experiment 10 are listed in

Table 21.

Table 21. Experiment 10 Parameters

Protein Statistics Index Protein Statistics Action Number of Trials

12 OUTPUT_ACTIVE_STATIC 50

13 OUTPUT_ACTIVE_BRED 50

4.4.7.2. Experiment 10 results and discussion

Figure 75 plots the mean best fitness for the first case in which the output target protein for C0/C1

remains static during the PSO breedings and for the second case in which this C0/C1 target protein is

allowed to evolve. These output proteins are important because environmental proteins that closely

match the associated parameters for the C0/C1 protein are used directly by the HeBIS fitness function

to gauge classification success.

Page 145:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 145 of 249

09/28/10

Figure 75. CV average fitness vs. static or PSO-evolved setting of the output C0/C1 protein. Vertical bars

correspond to the standard deviation of the sample mean.

In both output protein cases compared to the NOACTION baseline in Figure 75, the mean value of

the best fitness is significantly different from the baseline fitness of 0.02097. For the

OUTPUT_ACTIVE_STATIC trials, the mean of the best fitness was 0.40678 and for the

OUTPUT_ACTIVE_BRED trials, the mean was 0.18473. Standard deviations were respectively

0.06797 and 0.0548.

4.4.7.3. Experiment 10 conclusions

It is interesting that there was a decrease in fitness with the bred output proteins when compared to

the trials in which static output proteins were used. This is the opposite outcome that occurred in

Experiment 9 in which the bred statistical proteins produced GRN fitnesses that were higher than

those from the static trials. This highlights the point that fixed C0 parameters are able to be utilized

more easily than the bred parameters because the fixed “target” more efficiently allows the PSO to

find higher fitness solutions given the low number of breedings in this controlled experiment: 10. An

important observation is that the trial GRNs are able to incorporate these fixed output protein

parameters more readily given the experimental constraints. In any case, complexity has been

introduced into the GRN with the addition of this particular constituent action.

Page 146:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 146 of 249

09/28/10

4.5. Classification accuracy Hybridization Background

Algorithm Discussion Remote Sensing Background

GRN Analyses GRN Training with PSO

GRN Action Analyses Remote Sensing Application – Analyses and Comparisons

Robustness Analyses

Classification accuracy of the HeBIS algorithm is now examined. In particular, comments will be

made with regards to this accuracy using our cloud/no-cloud dataset with respect to a few high-level

training parameters. In addition, we examine and compare the results from self-organizing feature

maps on the same dataset where applicable. These results are presented in the context of the HeBIS

algorithm being fully-engaged and active over a range of controllable simulation parameters.

These experiments consist of examples in which an entire test image is classified. Full-image

classification is used to mimic usage of the algorithm in the real world as well as to produce a larger

statistical base now that the core parts of HeBIS have been examined and shown to work individually.

In summary, we will accomplish the following in this section:

• Examine the performance of HeBIS at various points in the operating space of the algorithm,

• Compare HeBIS with a self-organizing feature map.

4.5.1. Introduction

4.5.2. Training algorithm parameters description

The pertinent training parameters for HeBIS and the SOFM simulations follow. Table 22 presents

the range of these parameters for HeBIS and similar information for the SOFM simulations is

contained in Table 23.

Page 147:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 147 of 249

09/28/10

Table 22. Range of Pertinent HeBIS Training Parameters for Classification

Simulation Parameter Description

Geographic size 3x3,5x5,7x7

Number of PSO breedings [0,20]

Number of PSO particles 400

Genome size [0,40]

Protein diffusion rate (cell) [0.0, 1.0]

Protein diffusion rate (environment) [0.0, 1.0]

Maximum number of protein diffusion iterations [0, 4000]

Minimum protein concentration (cell) 0.001, 0.01

Minimum protein concentration (environment) 0.001, 0.01

Topology and size of artificial protein environment 3x3x17, 5x5x17, 7x7,17

Protein chemical reaction probability [0.0, 1.0] %

Minimum protein correlation (affinity) [0.0, 0.9]

Active protein statistics ON/OFF

Protein magnitude [0,1] or [0,15]

SOFM kernel size 0,1,2,4,6,9

SOFM kernel topology 1x1, 1x2, 2x1, 2x2, 2x3, 3x2, 3x3

Full-size classification dataset size 100620 (234x430) pixels

Table 23. Pertinent SOFM Training Parameters for Classification

Simulation Parameter Description

Kernel size 1, 2, 6, 9

Number of coarse training iterations 100

Number of fine training iterations 1000

Coarse neighborhood size Limited by max distance within network topology

Fine neighborhood size 1

Neighborhood type square

4.5.3. Fully-engaged HeBIS

These experiments occur in the context of HeBIS being fully operational. In other words, all the self-

organization parameters are active and the underpinnings of the genetic regulatory network are also

active.

4.5.3.1. Experiment 11 - Size of geographic processing environment

In this experiment we examine the effects of choosing different sizes for the area that surrounds the

pixel that is classified. We examine whether surrounding pixels help to convey information to aid in

the HeBIS classification process.

Page 148:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 148 of 249

09/28/10

4.5.3.1.1. Setup

Experiment 11

Only HeBIS is considered in this experiment. 50 full-image trials are executed with the geographic

regions displayed in Table 24. Default C0/C1 descriptors are used and protein statistics are

deactivated.

Table 24. Trial Distribution across Geographic Region Size for Experiment 11.

Number of trials Geographic region size

25 3x3

25 5x5

The parameters listed in Table 25 are used for the training parameters for these trials.

Table 25. Simulation Parameters for Experiment 11.

Simulation Parameter Value

# Trials 50

# Environmental genes 3

Protein diffusion rate 0.1

Minimum protein environmental concentration 0.01

Maximum number of network diffusion iterations 1000

PSO breedings 10

Topology and size of artificial protein environment 3x3x17, 5x5x17

Protein correlation (affinity) 0.9

Reaction probability 0.0 %

# PSO particles 400

# Folds in the cross-validation 10

Test dataset size 100620 (234x430) pixels

Note that the intracellular SOFM kernel used in this experiment is deactivated by setting its number

of neurons to 0. At this point, only the utility of the geographic size parameter is being tested.

The resulting input data cubes for each pixel to be classified is exemplified in Figure 76 for a

geographic size of 5x5 (25 pixels) with 15 multispectral features.

Page 149:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 149 of 249

09/28/10

Figure 76. Multi-spectral data cube for a 5x5 geographic region with 15 bands of multispectral data.

4.5.3.1.2. Experiment 11 results and discussion

A 3x3 geographic region of multispectral data contains no surrounding pixels that are usable by

HeBIS- only the central column of the features associated with the centrally located test pixel is used

for environmental protein diffusion. The 5x5 geographic region contains the test pixel and those

pixels that are immediately adjacent to the central test column. Thus, for the 3x3 region, the lattice

consists of 15 (1x1x15) discrete locations in which diffusion occurs and for the 5x5 region, the lattice

contains 135 (3x3x15) discrete diffusion points. A 7x7 geographic region contains the adjacent

pixels and the pixels adjacent to those for 375 (5x5x15) discrete points. Regions larger than 5x5 were

not considered because of simulation times increase exponentially as the size of the geographic region

increases.

Figure 77 summarizes the overall classification accuracy of the test image given the size of the test

geographic region; either 3x3 or 5x5. Overall, there was a decrease in average classification accuracy

from 91.47 % to 83.96 % as the size of the region increased.

Page 150:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 150 of 249

09/28/10

Figure 77. Full-image average classification accuracy vs. size of geographic regions surrounding test pixel. Vertical

bars correspond to the standard deviation of the sample mean.

The confusion matrix for the 3x3 region is presented in Table 26. It indicates relatively good overall

classification rates for both the C0 and C1 classes with 93.92 % of C0 and 88.25 % of C1 being

classified correctly. The percentage of C0 pixels that are incorrectly classified as C1 pixels is 6.08 %

and the percentage of C1 pixels that are misclassified is higher at 11.75 %.

Table 26. Confusion Matrix for 3x3 Geographic Region

C0pred C1pred

C0actual 93.92 % 6.078 %

C1actual 11.75 % 88.25 %

In Figure 78, a scatter plot of the 3x3 classification accuracies is presented with reference to the

fitness of the associated genome/GRN that was used to perform the classification. There is a single

outlier at an approximate classification accuracy of 60 %. This is due to a failed genome/GRN for

which the cloud class (C0) is completely classified as being from the no-cloud, C1, class. This failure

mode occurs when no significant concentrations of the C0 class protein are present in the

environmental lattice. In this case, zero concentration of the output protein for a particular test pixel

causes the pixel to be interpreted being from the C1 class as described previously in the decision

region map in Figure 66 on page 133.

Page 151:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 151 of 249

09/28/10

Figure 78. Full-image classification accuracy vs. fitness for a 3x3 geographic region surrounding text pixel.

The correlation coefficient for a linear fit for the 3x3 data is 0.2631, but with a confidence estimate of

only 79.6 %- low and not in line with a desired confidence on the order of 95 % (Student’s t-test).

Table 27 lists the confusion matrix for the 5x5 region. Once again, it indicates a relatively good

overall classification rate for the C0 class with 91.86 % of the C0 pixels being classified correctly.

However, the misclassification rate for C1 pixels being incorrectly classified as C0 pixels has

significantly increased to 25.89 % from the 3x3 11.75 % misclassification rate. The overall

misclassification rate for C0 pixels being classified as members of the C1 class has only slightly

increased for the 5x5 case to 8.144 % from 6.08 % in the 3x3 case.

Page 152:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 152 of 249

09/28/10

Table 27. Confusion Matrix for 5x5 Geographic Region

C0pred C1pred

C0actual 91.86 % 8.144 %

C1actual 25.89 % 74.11 %

Figure 79 displays a scatter plot for the 5x5 classification accuracies that is also with reference to the

fitness of the GRN used in the classification. Four outliers at the failure accuracy of approximately

60 % exist in a fashion similar to the outlier that is present in the 3x3 data.

Figure 79. Full-image classification accuracy vs. fitness for a 5x5 geographic region surrounding text pixel.

The correlation coefficient for a linear fit for the 5x5 data is higher at 0.3761 when compared to the

3x3 data. This indicates that some correlation exists between the genome’s fitness and its

classification performance with a confidence of 93.6 % (Student’s t-test).

Page 153:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 153 of 249

09/28/10

4.5.3.1.3. Experiment 11 conclusions

The 3x3 geographic region has been shown to perform better in overall classification accuracy than

the 5x5 region. The cloud (C0) classifications are similar for both the 3x3 and 5x5 regions, however

the reduction in accuracy for the 5x5 region is due to misclassifications of no-cloud (C1) pixels as

cloud (C0) pixels. Thus, the surrounding pixels do not appear to aid the HeBIS classification process

while it is using the parameters set forth for this experiment. Potentially, the accuracy might be

improved by increasing the number of breedings beyond the 10 allowed in this experiment or by

using regions that are larger than the larger 5x5 region that was tested in this experiment. However,

these larger regions were not studied because of the exponential increase in processing time required

for these larger-region simulations.

Also, Experiment 11’s parameters do not constitute an optimal set of processing parameters. The

shotgun experiments in Experiment 14 will help to shed light on this tuning process.

4.5.3.2. Experiment 12 - Size of intracellular SOFM kernel

The effects of varying the size of the HeBIS intracellular SOFM kernel is examined. For our

purposes, this size is defined as the number of neurons in the feature map. The 2-D topology of these

neurons is not considered in this experiment and thus a square topology is used as the default.

4.5.3.2.1. Setup

Experiment 12

Comparisons are made for different sizes of the SOFM kernel as shown in Table 28.

Table 28. HeBIS Kernel Sizes for the Intracellular SOFM in Experiment 12.

Kernel Number of Neurons Number of Trials

0x0 0 20

1x1 1 20

2x2 4 20

3x3 9 18

9x9 81 20

Page 154:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 154 of 249

09/28/10

Table 28 also shows the distribution of the 100 trials among the 5 kernel mappings that are tested.

Other training parameters used for these trials are included in Table 29.

Table 29. Simulation Parameters for Experiment 12.

Simulation Parameter Value

# Trials 98

# Environmental genes 3

Protein diffusion rate 0.1

Minimum protein environmental concentration 0.01

Maximum number of network diffusion iterations 1000

PSO breedings 10

Topology and size of artificial protein environment 3x3x17

Protein correlation (affinity) 0.9

Reaction probability 0.0 %

# PSO particles 400

Test dataset size 100620 (234x430) pixels

Full-image testing is used to acquire classification accuracy information from each trial.

4.5.3.2.2. Experiment 12 results and discussion

The confusion matrix for the case in which 0 neurons are used in the intracellular SOFM is presented

in Table 30.

Table 30. Confusion Matrix for Intracellular SOFM with 0 Neurons

C0pred C1pred

C0actual 93.12 % 6.885 %

C1actual 15.57 % 83.43 %

Classification rates for the cloud and no-cloud classes are 93.12 % and 83.43 % respectively. The

ratio of misclassifications is defined as

(27)

Page 155:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 155 of 249

09/28/10

and it provides an indication of the degree of difference between the two types of misclassification.

The ratio for this case is 2.26 with 6.89 % of C0 pixels being incorrectly assigned to C1 and 15.57 %

of C1 pixels being misclassified as belonging to class C0. Overall mean classification accuracy on

the entire test image for the 0-neuron classifier is 88.94 %.

A scatter plot for classification accuracy vs. fitness with 0 neurons in the intracellular SOFM is

displayed in Figure 80. A modest correlation correlation of 0.3609 exists between the two parameters

with a confidence of 88.2 % (Student’s t-test).

Figure 80. Classification accuracy vs. fitness for case with 0 neurons in intracellular SOFM.

Adding a single neuron to the SOFM improves the detection of class C0 slightly to 95.79 % when

compared to the classification for the 0-neuron case. However, detection of C1 becomes worse and

falls to 70.39 % compared to 83.43 % for the 0-neuron case. C0->C1 misclassification falls to 4.21%

from 6.89 % for the 0-neuron case, but C1->C0 misclassification increases when compared to the 0-

neuron baseline. In this case, C1->C0 misclassification increases from 15.57 % to 29.61 %. The

spread in the misclassification ratio increases to 7.03 compared to 2.26 for the baseline. These results

Page 156:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 156 of 249

09/28/10

are summarized in Table 31. Mean classification accuracy for the 1-neuron classifier falls slightly to

84.84 % from the 88.94 % attained with the 0-neuron classifier.

Table 31. Confusion Matrix for Intracellular SOFM with 1 Neuron

C0pred C1pred

C0actual 95.79 % 4.214 %

C1actual 29.61 % 70.39 %

The classification-fitness scatter plot in Figure 81 highlights a high linear correlation coefficient

index of 0.7421 that has a corresponding significance of greater than 99 % (Student’s t-test).

Figure 81. Classification accuracy vs. fitness for case with 1 neuron in intracellular SOFM.

Although the confusion matrix in Table 32 for the 4-neuron SOFM shows better performance than for

the 1-neuron instance, it is not better than the 0-neuron case.

Page 157:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 157 of 249

09/28/10

Table 32. Confusion Matrix for Intracellular SOFM with 4 Neurons

C0pred C1pred

C0actual 94.29 % 5.715 %

C1actual 21.78 % 78.22 %

HeBIS with a 4-neuron SOFM classifies C0 pixels correctly 94.29 % of the time with a detection rate

of C1 pixels that, at 78.22 %, is significantly higher than what occurred for the 2-neuron case. The

misclassification ratio is still higher than for the baseline SOFM, but it has improved to 3.81 from the

2-neuron SOFM’s 7.03 ratio. The corresponding misclassification details are that class C0 was

misclassified 5.72 % of the time and class C1 was misclassified 21.78 % of the time. Improvements

in the overall classification accuracy stem from the decrease in the misclassification of no-cloud

pixels into cloud pixels. This overall mean classification accuracy is approximately 87.36 % which is

better than the 84.84 % mean classification accuracy determined for the 1-neuron trials.

A moderate amount of linear correlation exists between classification accuracy and fitness as is

evidenced by the 0.5105 correlation coefficient which is found from the 4-neuron results with a 98%

significance (Student’s t-test). A linear fit through the scatter data is shown in Figure 82.

Figure 82. Classification accuracy vs. fitness for case with 4 neurons in intracellular SOFM.

Page 158:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 158 of 249

09/28/10

The 9-neuron confusion matrix as listed in Table 33 shows that this classifier has the highest

aggregate classification rate on the test image of all the intracellular SOFMs examined. At 91.07%,

the classifier also has the lowest misclassification statistics. These are 4.81 % misclassification of C0

pixels as C1 pixels and 14.36 % for C1 pixels that have been incorrectly labeled as being members of

class C0. The misclassification ratio has dropped to 2.99 for this classifier which is better than for

both the 0-neuron and the 4-neuron classifiers.

Table 33. Confusion Matrix for Intracellular SOFM with 9 Neurons

C0pred C1pred

C0actual 95.1894 % 4.8106 %

C1actual 14.3611 % 85.6389 %

A low amount of linear correlation exists for the classification-fitness scatter data for the 9-neuron

classifier: 0.2576 with a low significance of approximately 69.8% (Student’s t-test). The plot of the

spread in data is presented in Figure 83 over the 20 trials for this portion of the experiment.

Page 159:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 159 of 249

09/28/10

Figure 83. Classification accuracy vs. fitness for case with 9 neurons in intracellular SOFM.

The 81-neuron classifier has the fourth best performance with regards to mean classification

accuracy. At 87.16 %, only the 1-neuron classifier performs worse in aggregate. In Table 34, the

confusion matrix shows the lowest misclassification ratio, 2.80. This is lower than that of the best

aggregate classifier, HeBIS with 9 intracellular SOFM neurons. Although the ratio of

misclassifications is lower than the 9-neuron case, the individual misclassification errors are higher.

C0->C1 misclassifications increase to 7.23 % and C1->C0 misclassifications increase to 20.24 %.

Table 34. Confusion Matrix for Intracellular SOFM with 81 Neurons

C0pred C1pred

C0actual 92.77 % 7.229 %

C1actual 20.24 % 79.76 %

The scatter data for the 81-neuron classifier support a correlation coefficient that is the highest of the

studied SOFM classifiers at 0.7259 with the highest significance of greater than 99.0 % (Student’s t-

test). The linear fit is presented in Figure 84.

Page 160:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 160 of 249

09/28/10

Figure 84. Classification accuracy vs. fitness for case with 81 neurons in intracellular SOFM.

4.5.3.2.3. Experiment 12 conclusions

Figure 85 summarizes the mean classification accuracies given the size of the intracellular SOFM

kernel used by the HeBIS algorithm.

Figure 85. Full-image average classification accuracy vs. the number of neurons in the intracellular SOFM.

Vertical bars correspond to the standard deviation of the sample mean of the classification accuracy.

Page 161:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 161 of 249

09/28/10

The best mean accuracy, 91.07 %, over the trials is obtained with the 9-neuron classifier and the

worst overall mean accuracy of 84.84 % was found with the 1-neuron classifier. From best mean

classification accuracy performance to the worst, the classifiers are ranked as follows: 9-neuron, 0-

neuron, 4-neuron, 81-neuron and 1-neuron.

Misclassifications of both types were found to be at a minimum with the 9-neuron classifier. It would

seem that the improvements in overall aggregate classification result from the decrease in C1->C0

misclassifications. The cloud class classifications (C0->C0) remain roughly the same at 93.12 %,

95.79 %, 94.29 %, 95.19 % and 92.77 % respectively for the 0-neuron, 1-neuron, 4-neuron, 9-neuron,

and 81-neuron classifiers. However, the standard deviations of mean classification accuracies are

high, most likely because of the relatively small number of trials conducted per SOFM kernel. The

number of trials used was due to time constraints for processing the simulations.

4.5.3.3. Experiment 13 - Protein chemistry reaction probability

The effects of protein chemistry on HeBIS classification accuracy is now examined.

4.5.3.3.1. Setup

Experiment 13

This experiment consists of fixing the primary HeBIS parameters while varying the probability that a

protein chemical reaction- beyond the genome-dictated interactions- can occur in the environmental

lattice. This is different from Experiment 2 in the sense that now the self-organization and GRN

modules are active. The trial distribution for the protein chemistry probabilities are listed in Table 35

and are similar to the trial distribution used in Experiment 4.

Table 35. Protein Reaction Probability Distribution for Experiment 13.

Number of trials Protein reaction probability

20 0.0 %

20 0.1 %

20 1.0 %

20 10 %

Page 162:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 162 of 249

09/28/10

The remainder of the HeBIS training parameters for this experiment are listed in Table 36.

Table 36. Simulation Parameters for Experiment 13.

Simulation Parameter Value

# Trials 80

# Genes 3

Intracellular SOFM size 2x1

Protein diffusion rate 0.1

Minimum protein environmental concentration 0.01

Maximum number of network diffusion iterations 1000

PSO breedings 10

Topology and size of artificial protein environment 3x3x17

Protein correlation (affinity) 0.9

Reaction probability 0.0, 0.1, 1.0, 10.0 %

# PSO particles 400

The size of the intracellular SOFM kernel is set at 2 neurons with a 2x1 topology. For this

experiment, training occurs over 18 pixels (9 pairs of C0/C1 pairs).

Classification accuracy data are acquired from testing a full image and these results are averaged over

similar trials. These averaged results are collected over the prescribed number of trials according to

the distribution in Table 35.

4.5.3.3.2. Experiment 13 results and discussion

The baseline for Experiment 13 is the case in which there is a 0.0 % probability of chemistry

occurring between proteins in the environmental matrix. The mean classification accuracy for this

baseline HeBIS classifier is 87.1 %. In the confusion matrix listed in Table 37 for this classifier, it is

seen that this mean classification accuracy translates into 91.954 % correct classification of class C0

pixels and 80.704% correct classification of C1 pixels. The misclassification ratio for this classifier is

2.398 and this is indicative of the high misclassification rate, 19.296 %, of C1 pixels into C0 pixels.

Table 37. Confusion Matrix for 0.0 Protein Reaction Probability

C0pred C1pred

C0actual 91.95 % 8.0460 %

C1actual 19.30 % 80.70 %

Page 163:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 163 of 249

09/28/10

The correlation coefficient for this classifier with regards to classification accuracy and the fitness of

the associated GRN is 0.63 (with a confidence of 99.3 %; Student’s t-test) and this is indicative of a

fit that possesses a moderate amount of linearity. A linear fit through the classification and fitness

data is presented in Figure 86.

Figure 86. Classification accuracy vs. fitness with 0.0 reaction probability.

The second classifier has a 0.1% chance of a protein chemical reaction occurring in the environmental

lattice. This classifier has a mean classification accuracy of 93.88 % which is 6.78% higher than for

the baseline 0% classifier. The confusion matrix for the 0.1% classifier is listed in Table 38. This

classifier has the highest classification accuracy, 92.926 %, for no-cloud pixels for the four protein

chemistry classifiers studied. The misclassification rate for this classifier decreased to 1.31 from

2.398 for the baseline classifier with a deactivated protein reaction chemistry.

Table 38. Confusion Matrix for 0.001 Protein Reaction Probability

C0pred C1pred

C0actual 94.60 % 5.398 %

C1actual 7.074 % 92.93 %

Page 164:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 164 of 249

09/28/10

The low misclassification ratio indicates that the misclassification rates are roughly comparable: a

C1->C0 misclassification occurred over 7.07% of the C1 pixels and a C0->C1 misclassification

occurred over 5.40% of the C0 pixels.

Very little linear correlation seems to exist between the classification accuracies and the fitness of the

evolved 0.1 % reaction probability GRNs. The correlation coefficient is 0.2308 with a corresponding

confidence of 67.3% (Student’s t-test) - a relatively low and uncertain result. The scatter plot of the

trials for the application of this classifier is in Figure 87.

Figure 87. Classification accuracy vs. fitness with 0.001 reaction probability.

The third classifier, allows for proteins to react with a probability of 1% during the classification

process. It performs with a mean classification accuracy of 85.47 % with the decidedly mixed

confusion matrix in Table 39. Classification accuracies for C0 and C1 are 84.24% and 87.09%

respectively. For C0, this classifier performs 7.71% worse than the baseline classifier, but it performs

6.39 % better than the baseline classifier on the C1 class. The misclassification rate for C1->C0 is

12.91% compared to the baseline 19.30% which is a 6.4 % decrease whereas the misclassification

rate for C0->C1 increases by 7.71% to 15.76. The corresponding misclassification ratio is 0.8189

which indicates that the C0->C1 misclassification rate is now slightly higher than the C1->C0

misclassification rate.

Page 165:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 165 of 249

09/28/10

Table 39. Confusion Matrix for 0.01 Protein Reaction Probability

C0pred C1pred

C0actual 84.24 % 15.76 %

C1actual 12.91 % 87.09 %

There exists a high confidence (98.2%; Student’s t-test) that the correlation coefficient of 0.5239 for

the classification accuracy and fitness data indicates a moderate amount of linear correlation between

these parameters. Figure 88 shows a scatter plot of the classification accuracy vs. fitness data in

addition to a fitted line through the data.

Figure 88. Classification accuracy vs. fitness with 0.01 reaction probability.

The final classifier which possesses a 10% protein reaction probability has the highest C0

classification accuracy of 96.57 %. Unfortunately, as Table 40 shows, it also has the lowest C1

Page 166:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 166 of 249

09/28/10

classification accuracy of 71.75 %. Its misclassifications are unbalanced as is indicated by a

misclassification ratio of 8.239. Overall, the mean classification accuracy across the entire test image

and the trials stands at 85.87 %.

Table 40. Confusion Matrix for 0.1 Protein Reaction Probability

C0pred C1pred

C0actual 96.57 % 3.430 %

C1actual 28.25 % 71.75 %

Low confidence (49.3%; Student’s t-test) in the 0.1577 correlation coefficient leads to the linear fit of

the classification and fitness data that is presented in Figure 89.

Figure 89. Classification accuracy vs. fitness with 0.1 reaction probability.

Page 167:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 167 of 249

09/28/10

The results for the four reaction probability classifiers are summarized in Figure 90 in a mean

classification accuracy vs. probability of reaction plot. Peak mean classification accuracy occurred

with the 0.1 % probability classifier and the minimum mean classification accuracy occurred with the

1 % probability classifier. There is a higher standard deviation for the 1 % probability classifier than

the 0.1% and 10 % classifiers.

Figure 90. Full-image classification accuracy vs. the probability of protein reaction in the environmental lattice.

Vertical bars correspond to the standard deviation of the sample mean.

4.5.3.3.3. Experiment 13 conclusions

Similar classification accuracies for the 1% and 10% probability classifiers are based on different

failure modes. In the 1% case, there is an increase in both misclassification rates, C0->C1 and C1-

>C0. However, in the 10% case, there is a large increase in C1->C0 misclassifications with a

relatively small C0->C1 misclassification rate.

As the protein chemistry probability increases, C0->C1 misclassification decreases from the baseline

8.046% (0.0 % classifier) to 5.398 % (0.1% classifier). C0->C1 misclassification then increases by a

factor of three (3x) to 15.75% with the 1% classifier before falling back to a low of 3.43% with the

10% classifier.

Page 168:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 168 of 249

09/28/10

Also, as the probability increases, C1->C0 misclassification decreases from the baseline 19.30 % with

the 1% classifier to 7.074 % with the 0.1% classifier. C1->C0 misclassification then increases to

12.91% with the 1% classifier and peaks at 28.25 % with the 10 % classifier.

The minima for both types of misclassification rates coincide with the highest overall classification

accuracy with the 0.1% classifier. It appears that too much “mutation” in the form of protein reaction

rate decreases the classification accuracy below that of the baseline classifier in which no protein

chemistry is active. “Too much” in this case occurs with the 1% probability of protein chemistry

reaction. Therefore it seems that a small amount of protein reaction probability improves the overall

classification accuracy of the HeBIS algorithm. A small amount, 0.1%, of reaction probability causes

the subsequent creation of new intermediate proteins during the HeBIS classification process. It

appears that class C1 (no-cloud) is more sensitive to the creation of these extra intermediate proteins.

4.5.3.4. Experiment 14 - Shotgun

We are now ready to examine HeBIS with its self-organization and genetic regulatory network bases

activated. Comparisons between HeBIS and a SOFM are made where applicable.

4.5.3.4.1. Setup

Experiment 14

All training parameters are considered and can be modified randomly (uniform distribution for each

selected parameter) to determine classification accuracies. For this experiment, these accuracies are

derived from full classifications of the test image. This test image comes from our standard cloud/no-

cloud dataset.

The list of HeBIS parameters that are varied in these randomized trials is presented in Table 41.

Page 169:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 169 of 249

09/28/10

Table 41. List of HeBIS Parameters to Randomize for Experiment 14.

Simulation Parameter Range

Number of particles in PSO swarm [1,1000]

Number of PSO breedings [1,100]

Number of environmental/regulatory genes [2, 1000]

Intracellular SOFM kernel size 1x2, 2x1, 3x3, 5x5

Minimum protein correlation (affinity) [0, 1.0]

Protein diffusion rate (cell) [0, 1.0]

Protein diffusion rate (environment) [0, 1.0]

Minimum protein concentration (cell) [0, 1.0]

Minimum protein concentration (environment) [0, 1.0]

Protein magnitude [0, 1.0], [0, 15]

Protein statistics active On/Off

Similarly, Table 42 shows the parameters that are varied for the SOFM trials.

Table 42. List of SOFM Parameters to Randomize for Experiment 14.

Simulation Parameter Range

Kernel size 3x3, 5x5, 7x7, 9x9

Number of trials for coarse initialization 100

Number of trials for fine initialization 1000

Size of coarse neighborhood Limited to max L2 distance within network topology

Size of fine neighborhood 1

In both the HeBIS and SOFM trials, the classification accuracies are examined. Additionally, we also

analyze Receiver Operating Characteristic (ROC) curves to analyze threshold operating points for

classification for HeBIS [117].

ROC curves are useful in estimating the performance of ranking classifiers. They are based on the

concepts of precision, recall, sensitivity, and specificity which are taken from a classifier’s confusion

matrix and are defined in the following equations:

, (28)

, (29)

, (30)

Page 170:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 170 of 249

09/28/10

, (31)

, (32)

and

. (33)

The ROC curve graphically estimate the effect that increasing the false positive rate for a specific

classifier will have on the true positive rate. An example of a simple ROC curve is show in Figure

91.

Figure 91. Sample ROC curve with false-positive rate along the abscissa and true-positive rate as the ordinate.

The area under the curve (AUC) is a single number which is the area under the tradeoff curve in the

ROC analysis. It ranges from a minimum of 0 to a maximum of 1.0. Higher AUC implies a better

classifier as the area approaches the maximum. In other words, it measures the discrimination power

of the classifier which is the ability of the classifier to correctly classify both positive and negative

class members. Generally, the AUC is not a perfect measure of classification performance, but it is

useful nonetheless.

For these experiments, the AUC is approximated via the construction of trapezoids under the curve.

Finally, two more useful error characterizations are

Page 171:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 171 of 249

09/28/10

(34)

and

. (35)

.

Further details on the operating theory behind ROC curves are available from [117] and [118].

4.5.3.4.2. Experiment 14 results and discussion

This experiment consists of both HeBIS and standalone SOFM trials on the test image from the

MODIS A2002193183000 dataset. The distribution of class pixels is listed in Table 1.

Table 43. Distribution of Class and Infrastructure Pixels in A2002193183000

Parameter Value

# C0 (cloud) pixels 19053

# C1 (no-cloud) pixels 14441

# Land pixels 67126

# Water pixels (C0 + C1) 33494

# Invalid pixels 0

The HeBIS trials in this experiment use a minimum environmental and intracellular protein

concentration of 0.01. HeBIS is trained on a total of 18 pixels- 9 pairs of C0/C1 pixels and the

standalone SOFMs are trained on 20 pixels that consist of 10 pairs of C0/C1 pixels. The previously

described 15 bands of spectral data are used as features for each of the class pixels. In the SOFM

case, each pixel consists of a single column of multispectral data without any information from the

surrounding pixels. This SOFM data format is used for training and testing as this is equivalent to a

3x3 HeBIS geographic region. Table 44 and Table 45 list more detailed operational parameters for

the selected HeBIS and standalone SOFM trials, respectively.

Page 172:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 172 of 249

09/28/10

Table 44. Operational Parameters for Selected HeBIS Shotgun Experiments

HeBISHeBISHeBISHeBIS

Sel.Sel.Sel.Sel. TrTrTrTrial #ial #ial #ial #

HeBISHeBISHeBISHeBIS

ExperimentExperimentExperimentExperiment LabelLabelLabelLabel

DeltaDeltaDeltaDelta

Sec.Sec.Sec.Sec.

BreedsBreedsBreedsBreeds MaxMaxMaxMax

Iter.Iter.Iter.Iter.

Part.Part.Part.Part. RRRReactioneactioneactioneaction

Prob.Prob.Prob.Prob.

OOOOutpututpututpututput

ProteinProteinProteinProtein ValueValueValueValue SwitchSwitchSwitchSwitch

MMMMinininin....

CorrCorrCorrCorr ValueValueValueValue

####

GenesGenesGenesGenes

SOFMSOFMSOFMSOFM

SizeXSizeXSizeXSizeX

SOFMSOFMSOFMSOFM

SizeYSizeYSizeYSizeY

Diff.Diff.Diff.Diff.

RateRateRateRate

Diff.Diff.Diff.Diff.

RateRateRateRate CellCellCellCell

StatStatStatStat

FlagFlagFlagFlag

0 05_28_07_58_58_61

1497 19 2770 197 0.000 0 0.9728 6 3 2 0.0432 0.7501 0

1 06_11_01_18_20_77

1417 19 2999 442 0.01841 1 0.9054 13 0 0 0.3654 0.5130 0

2 06_10_08_03_09_38

377 10 3431 120 0.02188 1 0.9806 14 1 1 0.4264 0.5133 1

3 06_09_16_37_07_17

554 10 1223 229 0.02685 0 0.9998 14 2 2 0.1645 0.5151 0

4 05_28_12_49_29_76

906 17 1184 110 0.000 1 0.9556 19 1 2 0.1226 0.4319 1

5 06_03_10_39_23_85

206 4 2943 270 0.000 0 0.9884 10 0 3 0.9508 0.9441 1

6 05_28_00_32_27_50

3310 16 348 426 0.000 0 0.9516 19 3 1 0.9432 0.8463 0

7 06_09_17_4

3_13_23

73 5 1264 24 0.002965 0 0.9849 15 1 0 0.9069 0.4742 1

8 05_28_12_42_02_75

447 15 779 104 0.000 1 0.9531 9 1 2 0.5592 0.0766 1

9 06_01_23_35_56_3

36881 17 591 436 0.000 0 0.9145 11 2 1 0.0152 0.2865 0

10 06_10_07_35_57_34

460 11 3859 209 0.02688 0 0.9756 4 1 3 0.8637 0.8795 1

Table 45. Operational Parameters for SOFM Experiments

SOFMSOFMSOFMSOFM TrialTrialTrialTrial ####

SOFMSOFMSOFMSOFM Experiment Experiment Experiment Experiment LabelLabelLabelLabel

SOFM SOFM SOFM SOFM SizeXSizeXSizeXSizeX

SOFM SOFM SOFM SOFM SizeYSizeYSizeYSizeY

1st stage 1st stage 1st stage 1st stage #iter.#iter.#iter.#iter.

1st stage 1st stage 1st stage 1st stage alphaalphaalphaalpha

1st stage 1st stage 1st stage 1st stage radiusradiusradiusradius

2nd stage 2nd stage 2nd stage 2nd stage #iter.#iter.#iter.#iter.

2nd stage 2nd stage 2nd stage 2nd stage alphaalphaalphaalpha

2nd stage 2nd stage 2nd stage 2nd stage radiusradiusradiusradius

0 2009.11.10 2 1 1000 0.05 2 100000 0.02 0.5

1 2009.11.11b 2 1 1000 0.05 2 100000 0.02 0.5

2 2009.06.17 2 1 100 0.05 2 1000 0.02 0.5

3 2009.11.11f 5 4 200 0.05 2 2000 0.02 0.5

4 2009.11.11e 3 2 200 0.05 2 2000 0.02 0.5

5 2009.11.11d 2 2 200 0.05 2 2000 0.02 0.5

6 2009.11.11c 3 1 200 0.05 2 2000 0.02 0.5

7 2009.11.11g 5 5 200 0.05 2 2000 0.02 0.5

8 2009.11.11h 8 7 200 0.05 2 2000 0.02 0.5

Page 173:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 173 of 249

09/28/10

Summaries of the classification results that are discussed in detail in this section are presented in the

following tables. Table 46 lists the results for the selected HeBIS experiments taken from the 279

that are available and Table 47 shows the SOFM experiments.

Table 46. Classification Results for Selected HeBIS Shotgun Experiments

HeBISHeBISHeBISHeBIS Sel.Sel.Sel.Sel. TrialTrialTrialTrial ####

HeBISHeBISHeBISHeBIS ExperimentExperimentExperimentExperiment LabelLabelLabelLabel

FitnessFitnessFitnessFitness % Correct% Correct% Correct% Correct AllAllAllAll VVVValidalidalidalid

% Incorrect% Incorrect% Incorrect% Incorrect AllAllAllAll VVVValidalidalidalid

%%%% C0C0C0C0 ValidValidValidValid CCCCorrectorrectorrectorrect

% C0% C0% C0% C0 VVVValid_alid_alid_alid_ IIIIncorrectncorrectncorrectncorrect

% C1% C1% C1% C1 VVVValid_alid_alid_alid_ CCCCorrectorrectorrectorrect

% C1% C1% C1% C1 VVVValidalidalidalid IIIIncorrectncorrectncorrectncorrect

0 05_28_07_58_58_61 0.000 43.12 56.88 0.000 100 100.0 0.000

1 06_11_01_18_20_77 0.1781 51.41 48.59 15.43 84.569 98.89 1.115

2 06_10_08_03_09_38 0.06596 56.88 43.12 100.0 0.000 0.000 100.0

3 06_09_16_37_07_17 0.2609 62.11 37.89 53.89 46.11 72.95 27.05

4 05_28_12_49_29_76 0.3987 88.98 11.02 81.64 18.36 98.67 1.330

5 06_03_10_39_23_85 0.6919 92.79 7.213 87.99 12.01 99.11 0.8864

6 05_28_00_32_27_50 0.7934 93.66 6.335 88.86 11.14 100.0 0.000

7 06_09_17_43_13_23 0.7637 94.28 5.723 92.27 7.726 96.92 3.082

8 05_28_12_42_02_75 0.5012 96.31 3.687 100.0 0.000 91.45 8.552

9 06_01_23_35_56_3 0.5382 98.38 1.618 99.61 0.3884 96.76 3.241

10 06_10_07_35_57_34 0.7135 99.15 0.8509 98.51 1.485 99.99 0.01385

Table 47. Classification Results for Selected SOFM Experiments

SOFMSOFMSOFMSOFM TrialTrialTrialTrial

####

SOSOSOSOFMFMFMFM Experiment LabelExperiment LabelExperiment LabelExperiment Label

%Correct Total%Correct Total%Correct Total%Correct Total %Incorrect Total%Incorrect Total%Incorrect Total%Incorrect Total %C0 Correct%C0 Correct%C0 Correct%C0 Correct %C0 Incorrect%C0 Incorrect%C0 Incorrect%C0 Incorrect %C1 Correct%C1 Correct%C1 Correct%C1 Correct %C1 Incorrect%C1 Incorrect%C1 Incorrect%C1 Incorrect

0 2009.11.10 63.408 36.59 35.67 64.33 100.0 0.000

1 2009.11.11b 63.408 36.59 35.67 64.33 100.0 0.000

2 2009.06.17 63.77 36.23 36.315 63.69 100.0 0.000

3 2009.11.11f 94.85 5.153 97.71 2.288 91.07 8.933

4 2009.11.11e 95.34 4.661 95.90 4.089 94.59 5.415

5 2009.11.11d 97.12 2.884 99.00 0.9970 94.63 5.374

6 2009.11.11c 97.18 2.821 98.73 1.275 95.14 4.861

7 2009.11.11g 97.20 2.669 95.49 4.509 99.45 0.2424

8 2009.11.11h 97.23 2.771 96.13 3.868 98.68 1.323

All the HeBIS trials used a classification discrimination threshold of 0.5 to classify C0/C1 according

to a single output protein. Note that the ROCs in the detailed examinations present the effects of

varying the threshold so as to tradeoff between classification recall and precision.

Page 174:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 174 of 249

09/28/10

Briefly examining Table 46 and Table 47, it is interesting to note that comparing both the HeBIS and

standalone SOFM trials with 2x1 SOFM kernels, it is seen that HeBIS outperforms the standalone

SOFM by approximately 33% in overall classification accuracy. This is with an intracellular SOFM

kernel which has the same topology as that of the standalone SOFM classifier, 2x1.

The following examination consists of several types of imagery for each of the listed trials. For each

listed HeBIS example, a figure is presented which contains four images. The classification image

shows those pixels in white that were classified as cloud (C0) pixels. Black pixels in this image

correspond to pixels that were classified as no-cloud (C1) pixels. Red pixels delineate land pixels

that are not considered during the classification. The second image is the difference image that

presents the misclassification errors. A blue pixel in this image shows a cloud pixel that was

misclassified as no-cloud (C0->C1) and a green pixel shows a no-cloud pixel that was misclassified

as a cloud pixel (C1->C0). The third image is the ground truth for the C0 and C1 class distribution.

White indicates cloud (C0) and black shows the no-cloud (C1) class with red as land. The fourth

image is the land mask that presents the distribution of water and land pixels with land pixels

represented as red and water pixels represented as black. There is also an ROC plot associated with

each trial directly underneath the four descriptive images.

Figure 92 shows the imagery for HeBIS selected trial # 0. Only 43.12 % of the pixels were correctly

classified and this was broken down as being 0% of the cloud pixels being classified correctly

although 100 % of the no-cloud pixels were correct. The difference image in this figure highlights the

C0->C1 misclassifications as blue pixels- the entire cloud field is blue in the difference image. The

AUC for the ROC is 0.5 which indicates that the classifier has no discrimination ability between

cloudy and no-cloud (clear-weather) pixels.

Page 175:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 175 of 249

09/28/10

2010_05_28_07_58_58_61

Classification Difference Cloud/No-Cloud

Ground Truth Land/Water Mask

Figure 92. HeBIS classification imagery and ROC for 2010_05_28_07_58_58_61 test. This is HeBIS selected trial

# 0.

Figure 93 summarizes the results for HeBIS trial # 1. In the classification image we see the

beginnings of an ability to classify clouds. In this trial, the overall classification rate was 51.4 %. of

the overall pixels being correctly classified as being cloudy- an increase of 8.28% from the 43.12 %

in trial #0. Of these correctly classified pixels, 15.43 % of the cloudy pixels were classified correctly

and 98.89 % of the clear-weather pixels were determined correctly. The AUC is 0.604 and the

resulting ROC curve shows a rise in recall at low precision. At the 1 % precision (FP) point, this

classifier possesses approximately 15 % of recall (TP).

Page 176:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 176 of 249

09/28/10

2010_06_11_01_18_77

Classification Difference Cloud/No-Cloud

Ground Truth Land/Water Mask

Figure 93. HeBIS classification imagery and ROC for 2010_06_11_01_18_20_77 test. This is HeBIS selected trial

# 1.

In Figure 94 we see that the AUC for HeBIS trial # 2 has risen to 0.854. This 0.25 increase in area

under the curve when compared to the AUC for trial #1 results in an improvement in recall to 6% at a

value of 1% precision, but this only results in an increase to 56.88 % for overall classification

performance. More telling is the fact that the breakdown of the misclassification results show that

although 100 % of C0 pixels were classified correctly, no C1 pixels were correct. This translates to

Page 177:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 177 of 249

09/28/10

the classification image being purely white over water regions and also to the no-cloud areas in the

difference image being filled with green as an indication of the C1->C0 misclassifications. This is an

example of a useless classifier in the sense that all pixels are considered to be cloudy, hence no

optical satellite data would be usable from this image based on the results of this classifier.

2010_06_10_08_03_09_38

Classification Difference Cloud/No-Cloud

Ground Truth Land/Water Mask

Figure 94. HeBIS classification imagery and ROC for 2010_06_10_08_03_09_38 test. This is HeBIS selected trial

# 2.

Page 178:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 178 of 249

09/28/10

The third selected HeBIS trial summarized in Figure 95, has a C0 classification rate of 53.89 % with

a C1 rate of 72.95 % which results in an overall classification rate of 62.11 %. It is noticeable in the

classification and difference images that there is a mix of C0->C1 and C1->C0 misclassifications.

The error field is dominated by the C0->C1 misclassifications that are highlighted by the speckled

blue field in the difference image. The AUC for this classifier is 0.757 and the ROC has a multi-step

response to varying FP and TP operating points. At a precision of 0.5%, the corresponding recall is

30 %, for a 20% precision, the recall rises to 50%, for a 28% precision, recall rises further to 54%. At

a 31% precision, the recall steps to 78 %.

2010_06_09_16_37_07_17

Classification Difference Cloud/No-Cloud

Ground Truth Land/Water Mask

Figure 95. HeBIS classification imagery and ROC for 2010_06_09_16_37_07_17 test. This is HeBIS selected trial

# 3.

Page 179:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 179 of 249

09/28/10

A rise to 88.98% overall classification accuracy is seen with the trial #4 classifier which is

summarized in Figure 96. This result consists of 81.64% of class C0 pixels being classified correctly

with a correspondingly large percentage of C1 pixels, 98.67%, being classified correctly as clear-

weather. The ROC curve shows a large increase in low-precision recall compared to the prior trials.

For example, the recall at the 1% precision level has risen to 82%. The green in the difference image

shows the C1->C0 misclassifications. In this image, this type of misclassification is observed to be

mostly in a few areas that are close to the coasts and also in a few areas in open water in the lower

section of the Chesapeake Bay. A rise in AUC to 0.902 corresponds to the improvement in

classification accuracies across both classes.

2010_05_28_12_49_29_76

Classification Difference Cloud/No-Cloud

Ground Truth Land/Water Mask

Figure 96. HeBIS classification imagery and ROC for 2010_05_28_12_49_29_76 test. This is HeBIS selected trial

# 4.

Page 180:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 180 of 249

09/28/10

Figure 97 shows that the classifier from trial # 5 has an overall classification accuracy of 92.79%.

This corresponds to 87.99 % of C0 pixels being classified correctly as cloudy and 99.11% of all C1

pixels being classified correctly as clear-weather pixels. The AUC of 0.984 correlates with what is so

far, the highest recall for the given lowest precision. For example, the classifier performs with 88%

recall with approximately 0% precision. The difference image indicates that more clouds are being

classified correctly in the confined water areas in the central north-south-oriented inlet and at the end

of the confined inlet in the northeast section of the image.

2010_06_03_10_39_23_85

Classification Difference Cloud/No-Cloud

Ground Truth Land/Water Mask

Figure 97. HeBIS classification imagery and ROC for 2010_06_03_10_39_23_85 test. This is HeBIS selected trial

# 5.

Page 181:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 181 of 249

09/28/10

HeBIS trial # 6 is summarized in Figure 98. An incremental change in overall classification

accuracy, 93.66%, is seen in comparison to the 92.79% of trial # 5. The 93.66% classification

accuracy is due to 88.86 % of the C0 class and 100% of the C1 class being classified correctly. The

AUC is 0.986 and subsequently, this classifier shows good performance at the low-precision end of

the ROC curve. At an approximate 0% precision, recall has increased by 0.5% to 88.5 % compared

to trial #5. Examining the difference image, it is noticed that the same regions that were classified

correctly for trial #5 were also classified correctly for this trial. However, more of the C0 pixels that

are present at the end of the northeastern water inlet are classified correctly as being cloudy. The

same occurs at the end of the lower-most southwestern water inlet.

2010_05_28_00_32_27_50

Classification Difference Cloud/No-Cloud

Ground Truth Land/Water Mask

Figure 98. HeBIS classification imagery and ROC for 2010_05_28_00_32_27_50 test. This is HeBIS selected trial

# 6.

Page 182:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 182 of 249

09/28/10

Trial # 7 shows an aggregate classification rate of 94.28 % with a split of 92.27 % of C0 pixels and

96.92 % of C1 pixels being classified correctly. The summary imagery for this trial are shown in

Figure 99. Out of all the trials listed so far, this one has the highest C1->C0 misclassification rate of

3.08 %. We also see the highest C0 classification rate of 92.27 % which compares to 88.86 % for

trial # 6 and 81.64 % for trial # 4. The 7.73 % misclassification rate for C0->C1 is presented in the

difference image as blue pixels. The subsequent 91% recall at an approximate 0 % precision level is

associated with this trial’s AUC of 0.986 which is essentially the same as for selected trial # 6.

2010_06_09_17_43_13_23

Classification Difference Cloud/No-Cloud

Ground Truth Land/Water Mask

Figure 99. HeBIS classification imagery and ROC for 2010_06_09_17_43_13_23 test. This is HeBIS selected trial

# 7.

Page 183:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 183 of 249

09/28/10

Figure 100 presents selected trial # 8 with an aggregate classification accuracy of 96.31%. 100% of

cloudy (C0) pixels and 91.45% of clear-weather (C1) pixels are classified correctly. This particular

example has the largest C1->C0 misclassification rate of 8.55% across all the selected trials. These

misclassified pixels show up as green in the difference image and they are largely limited to open-

water regions with a few clusters present in the coastal regions. Examining the ROC curve, we see

that 0% recall occurs at the 0 % precision level, but importantly, recall rises quickly to 100 % at the

low-end precision of 8.5%. The 0.957 AUC has decreased compared to that in trials # 6 and # 7.

2010_05_28_12_42_02_75

Classification Difference Cloud/No-Cloud

Ground Truth Land/Water Mask

Figure 100. HeBIS classification imagery and ROC for 2010_05_28_12_42_02_75 test. This is HeBIS selected trial

# 8.

Page 184:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 184 of 249

09/28/10

In Figure 101, the trial #9 data and imagery support an overall classification rate of 98.38 % with

99.62 % and 96.67 % of all C0 and C1 pixels, respectively, being classified correctly. This

corresponds to an overall misclassification rate of 1.62 % for which the misclassifications for C1->

C0 have decreased from the previous example’s 8.55 % to 3.24 %. These C1->C0 misclassifications

are clustered in the southwestern and northeastern water inlets along the coastal regions and near the

eastern shore inlets and small peninsulas. With an AUC that has increased significantly to 0.982

from trial # 8’s 0.957, there is a similarly approximate 0% recall at 0% precision. However, with a

modest shift in precision to the 4% level, recall jumps dramatically to the 100% level- this is much

better than the ROC performance from trial # 8.

2010_06_01_23_35_56_3

Classification Difference Cloud/No-Cloud

Ground Truth Land/Water Mask

Figure 101. HeBIS classification imagery and ROC for 2010_06_01_233_35_56_3 test. This is HeBIS selected trial

# 9.

Page 185:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 185 of 249

09/28/10

Finally, Figure 102 presents the summary imagery and ROC for selected HeBIS trial # 10 which has

the best overall classification accuracy of 99.15 %. This corresponds to C0 pixels being correctly

classified 98.51 % of the time and C1 pixels being correctly classified 99.99 % of the time. This trial

also has the lowest average (unweighted according to class balance) misclassification rate across the

two classes of 1.49 % for C0->C1 and 0.014 % for C1-> C0. These misclassification rates

correspond to 284 cloudy pixels being mistakenly classified as clear-weather and 202 clear-weather

pixels being misclassified as cloudy pixels. The corresponding improvement in recall to 98 % at the

0 % precision level which is the best of these trials examined in detail. The AUC is 0.997.

2010_06_10_07_35_57_34

Classification Difference Cloud/No-Cloud

Ground Truth Land/Water Mask

Figure 102. HeBIS classification imagery and ROC for 2010_06_10_07_35_57_34 test. This is HeBIS selected trial

# 10.

Page 186:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 186 of 249

09/28/10

Table 81 in the Appendix on page 233 lists the evolved genomes for the selected HeBIS trials in

graphical format. In this table, the image associated with each dataset displays the evolvable

parameters from each genome. The hotter colors, for example red, correspond to higher values for

that particular parameter and cooler colors (e.g. blue) correspond to lower relative values for the

parameter. Two color bars are shown because two different parameter ranges exist for different

parameter classes within the genome. For example, the real-valued range [0, 1.0] represents, for

example, expressed protein concentration and the lower and upper switch concentrations. The range

[1, 254] represents integral values that are used to define the 4 descriptive parameters for each

evolved protein in the genome.

Page 187:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 187 of 249

09/28/10

Table 48 lists the operational parameters and classification results that are used to create the graphical

correlation coefficient matrix in Figure 124 on page 240 in the Appendices and its corresponding p-

value matrix which is also in the Appendices in Figure 125 on page 240. These matrices are for the

279 HeBIS shotgun trials- 200 with protein chemistry deactivated and 79 with protein chemistry

activated.

Table 48. Feature and Result Indices for Correlation Coefficient and P-Value Matrices

Correlation Feature IndexCorrelation Feature IndexCorrelation Feature IndexCorrelation Feature Index FeatureFeatureFeatureFeature and Resultsand Resultsand Resultsand Results

1 Delta Sec

2 # Breeds

3 Max. # Iterations

4 # Particles

5 Reaction Probability

6 Output Protein Switch

7 Min. Correlation Value

8 Number Genes

9 SOFM SizeX

10 SOFM SizeY

11 Diffusion Rate Enviro

12 Diffusion Rate Cell

13 Statistics Flag

14 Fitness

15 # Correct All Valid

16 % Correct All Valid

17 # Incorrect All Valid

18 % Incorrect All Valid

19 # C0 Valid Correct

20 % C0 Valid Correct

21 # C0 Valid Incorrect

22 % C0 Valid Incorrect

23 # C1 Valid Correct

24 % C1 Valid Correct

25 # C1 Valid Incorrect

26 % C1 Valid Incorrect

Page 188:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 188 of 249

09/28/10

Referencing these figures and the corresponding correlation data, we see that the fitness parameter

(14) is positively correlated with the overall classification accuracy (16). The correlation coefficient

is 0.8183 with a corresponding significance greater than 99.0 % (Student’s t-test). The linear

correlation coefficient for fitness (14) and C0 classification accuracy (20) shows negligible

correlation with a coefficient equal to 0.0526, although the test of this significance is fairly low at

62.82% (Student’s t-test).

A moderate positive correlation of 0.4315 exists between the C0 classification accuracy (22) and the

C1 classification accuracy (24). The significance of this test is strong at greater than 99.0%

(Student’s t-test). Also, a strong correlation coefficient of 0.7140 exists between fitness (14) and

class C1 classification accuracy (24) and it has a corresponding p-value that is close to 0.0 (Student’s

t-test).

The environmental protein diffusion rate (11) is positively correlated with the overall classification

accuracy (16). The correlation coefficient is 0.6130 with a corresponding significance greater than

99.0% (Student’s t-test). However, this diffusion rate (11) is negatively correlated with the overall

misclassification rate (17). The correlation coefficient for this pair of parameters is -0.6130 and its

significance is greater than 99.0% (Student’s t-test). Finally, reaction probability (5) and the class C0

classification rate (20) are paired and examined. The correlation coefficient for this pairing shows a

moderate amount of negative correlation at -0.2525 with a significance that is greater than 99.0%

(Student’s t-test).

Figure 103 continues the shotgun analysis with a plot that compares HeBIS’ overall classification

accuracy with genome fitness for the 200 trials in which protein chemistry is deactivated. The

corresponding fit through the scatter data indicates a good linear fit because of the 0.8213 correlation

coefficient between the two parameters that has greater than 99.0% significance (Student’s t-test).

Page 189:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 189 of 249

09/28/10

Figure 103. Comparison plot of HeBIS classification accuracy vs. fitness of the genome for 200 trials. Protein

chemistry is deactivated.

Figure 104 plots the same parameters as in Figure 103, but this one is for the 79 trials in which

protein chemistry has been activated. The overall classification accuracy and fitness pair also has a

high positive correlation coefficient, 0.8214 with a significance that is greater than 99.0% (Student’s

t-test).

Figure 104. Comparison plot of HeBIS classification accuracy vs. fitness of the genome for 79 trials with protein

chemistry activated.

Page 190:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 190 of 249

09/28/10

In Figure 105, classification accuracy is plotted versus reaction probability with an activated protein

chemistry. Reaction probability is the probability that protein chemistry occurs between two proteins

which in turn creates a new third protein in the environmental lattice. Examining the correlation

between these parameters, we see that the correlation coefficient is small, 0.0617 with a significance

greater than 99.0% (Student’s t-test). Thus, little linear correlation is present and the fit has little

predictive value.

Figure 105. HeBIS classification accuracy vs. reaction probability for 79 trials. Protein chemistry is activated.

Conversely, Figure 106 and its correlation data present little significance (41.1 %) (Student’s t-test) to

the slight linear correlation (0.158) between reaction probability and fitness for genomes with

activated protein chemistries.

Page 191:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 191 of 249

09/28/10

Figure 106. HeBIS classification accuracy vs. reaction probability and fitness for 79 trials. Protein chemistry is

activated.

Another pairing of parameters for which there is no strong correlation is classification accuracy and

minimum protein correlation with protein chemistry activated. A plot of these parameters in addition

to classification accuracy is presented in Figure 107.

Page 192:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 192 of 249

09/28/10

Figure 107 HeBIS classification accuracy vs. reaction probability and minimum protein correlation for 79 trials.

Protein chemistry is activated.

There is little correlation (0.1070) between the two parameters and the significance of the test is only

65.21% (Student’s t-test)for the data in Figure 107.

Classification accuracy plotted versus environmental protein diffusion rate is shown in Figure 108 for

the 200 trials in which the protein chemistry was inactive. At 0.6313, there appears to be a moderate

amount of linear correlation between the members of the pair as indicated by a significance greater

than 99.0% (Student’s t-test).

Page 193:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 193 of 249

09/28/10

Figure 108. Classification accuracy vs. environmental diffusion rate for 200 shotgun trials. Protein chemistry is

deactivated. Now, in Figure 109, classification accuracy versus fitness and environmental diffusion rate is plotted.

We’ve already noticed a strong correlation between fitness and classification accuracy with prior data

and now the data shows a high correlation (0.6313) with a significance greater than 99.0% (Student’s

t-test) for the pairing of classification accuracy and environmental diffusion rate.

Figure 109. Classification accuracy vs. genome fitness and environmental diffusion rate for 200 shotgun trials.

Protein chemistry is deactivated.

Page 194:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 194 of 249

09/28/10

Finally, Figure 115 on page 234 in the Appendices displays a comparison plot of all 279 genomes that

were evolved for the HeBIS shotgun experiments. The color bar from Table 81 on page 233 is

applicable to the evolved elements of the genomes in this figure.

4.5.3.4.3. Experiment 14 conclusions

HeBIS’ aggregate classification accuracy ranges from 43.12 % to 99.15 given all 279 shotgun trials.

The worst HeBIS trial is 2010_05_28_07_58_58_61. The standalone SOFM’s aggregate

classification accuracy ranges from 63.41% to 97.23%. The highest overall HeBIS classification rate

occurs with the protein chemistry activated occurs with trial 2010_06_10_07_35_37_34. It would

appear that both the standalone SOFM and HeBIS have more difficulty with classifying the more

generic class, C1- the clear-weather class- compared to classifying the cloud class, C0.

The data show that classification accuracy of class C1 dips down to the 0% level whereas C0

classification accuracy is predominantly high and usually greater than 87.99 %. The increase in

overall classification rate for HeBIS occurs mostly because C0 classification improves from a low of

0% to a high of 99.61 % of all C0 pixels being classified correctly. Most C1 misclassification rates

are fairly low and in between 0 and 8.55 % for the selected examples.

The fitness function is working and providing the GRN with a basis for producing accurate

classifications of cloud pixels. This is indicated by the generally high level of correlation between the

fitness and classification accuracy parameters for the trials in which the protein chemistry was either

activated or deactivated. Successful classifications also point to the success of the fitness function

and show that the function can drive the evolution of a good genome and GRN for this simple

classification problem.

Aggregate classification accuracies compared to protein diffusion rates show the presence of a slight

positively correlated relationship. As the protein diffusion rate increases, the classification rate is

found to increase also. Thus, with the lower diffusion rates, the proteins remain for a longer period of

time in the environmental matrix and conversely with higher diffusion rates, the proteins disappear

more quickly.

Page 195:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 195 of 249

09/28/10

A summary of the best classification accuracies for the selected trials is listed in Table 49. The top

three highest classification rates were achieved with the HeBIS algorithm and the standalone SOFM

had the lowest classification accuracies.

Table 49. Best Classification Accuracies for the Selected HeBIS and SOFM Examples

Algorithm Trial Ranking Classification Accuracy [%]

HeBIS 1 99.15

HeBIS 2 98.38

HeBIS 3 98.10

SOFM 4 97.23

SOFM 5 97.20

SOFM 6 97.18

Page 196:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 196 of 249

09/28/10

4.6. Classification robustness Hybridization Background

Algorithm Discussion Remote Sensing Background

GRN Analyses GRN Training with PSO

GRN Action Analyses Remote Sensing Application – Analyses and Comparisons

Robustness Analyses

4.6.1. Introduction/Methodology

For this dissertation, classification robustness examines the ability of HeBIS to correctly classify a

corrupted test image. Corruption is divided into two types for these experiments and subsequent

analyses:

• Addition of noise to the multispectral pixel features,

• Features which are missing from the full multispectral set for each pixel.

Both HeBIS and a baseline SOFM are compared where possible.

4.6.2. Experiment 15 - Noise

The first experiment is the noise addition test.

4.6.2.1. Setup

Experiment 15

This tests a single “best” genome that is applied to the classification of a full test image over a

number of trials. Equivalently, a single “best” SOFM codebook is used to determine the SOFM’s

response to noise-corrupted input data.

An equivalent amount of noise is added to both the SOFM and HeBIS test datasets after both have

been trained on a noise-free baseline cloud/no-cloud dataset.

The noise-addition methodology is such that Gaussian noise is added to the spectral feature data with

0.1 probability across the 15 bands at varying levels of standard deviation for the noise. 10 datasets

are created for each of the 4 levels of standard deviation for the injected noise. Three classifications

are performed on each of these 10 datasets- one classification using the “best” HeBIS genome, a

classification with the “best” 2x1 SOFM codebook, and a classification with the “best” 3x1 SOFM

codebook. The noise is added to the datasets before log-normalization occurs. The 40 datasets are

constructed with the 4 noise parameter sets listed in Table 50.

Page 197:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 197 of 249

09/28/10

Table 50. Dataset definitions for Experiment 15.

Dataset # Mean Standard Deviation

1 0 0.0

2 0 0.01

3 0 0.1

4 0 0.2

The parameters for the HeBIS genome are listed in Table 51 and the parameters for the SOFM

codebooks are in Table 52 and Table 53.

Table 51. Simulation Parameters for the HeBIS "Best" Genome for Experiment 15.

Simulation Parameter Value

# Genes 3

Intracellular SOFM size 2x1

Protein diffusion rate 0.1

Minimum protein environmental concentration 0.01

Maximum number of network diffusion iterations 1000

PSO breedings 10

Topology and size of artificial protein environment 3x3x17

Protein correlation (affinity) 0.8

Reaction probability 0.0 %

# PSO particles 400

Table 52. Simulation Parameters for the 2x1 SOFM "Best" Codebook for Experiment 15.

Simulation Parameter Value

Kernel size 2x1

Number of trials for coarse initialization 1000

Number of trials for fine initialization 100000

Size of coarse neighborhood 2

Size of fine neighborhood 0.5

Table 53. Simulation Parameters for the 3x1 SOFM "Best" Codebook for Experiment 15.

Simulation Parameter Value

Kernel size 3x1

Number of trials for coarse initialization 200

Number of trials for fine initialization 2000

Size of coarse neighborhood 2

Size of fine neighborhood 0.5

Page 198:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 198 of 249

09/28/10

4.6.2.2. Experiment 15 results and discussion

Figure 110 summarizes the mean classification results for the three classifiers considered in this

experiment: the highest performing SOFM classifier with a 2x1 neuron topology, the highest

performing SOFM classifier with a 3x1 neuron topology and one of the highest performing HeBIS

classifiers from Experiment 14.

Baseline comparisons of the cases with a noise standard deviation of 0.0, essentially with noise

injection deactivated, rank the 3x1 SOFM classifier with the highest mean classification accuracy of

97.16 %. The next highest is the HeBIS classifier that classified 96.2 % of the C0 and C1 pixels

correctly. In last place came the 2x1 SOFM classifier with an aggregate classification rate of 63.78%.

Standard deviation of the averaged classification accuracies was equal to 0.0 for all three classifiers.

This occurred because the 10 datasets used for each of the classifiers are static and no protein

chemistry is active for the HeBIS classifier.

For a noise standard deviation of 0.01, the 3x1 SOFM classifier once again has the highest mean

classification accuracy of 97.18 %- slightly higher by 0.01 % when compared to its baseline with

deactivated noise injection. The HeBIS classifier attained 96.2 % overall classification accuracy

which was essentially unchanged from its baseline. The 2x1 SOFM classifier’s mean classification

accuracy also remained unchanged at 63.78 %. Standard deviations of the mean classification

accuracy were negligible with values of 8.1828x10-5 for the 3x1 SOFM, 7.4832x10-5 for HeBIS and

0.0 for the 2x1 SOFM.

With noise injected at a level with the standard deviation equal to 0.1, the HeBIS classifier attained

the best mean classification accuracy of 98.31 %. This was an improvement of approximately 2%

over its performance with the lower level of standard deviation for the noise. The 3x1 SOFM

achieved the next best mean classification accuracy with 95.07 % which was a reduction of 2.11 %

compared to the 3x1 SOFM classifier with a noise standard deviation of 0.01. The 2x1 SOFM

classifier maintained its performance with a 63.78 % mean classification accuracy. Standard

deviations for this level of noise injection were 0.0039 for the 3x1 SOFM, 0.0057 for HeBIS and 0.0

for the 2x1 SOFM classifier.

Page 199:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 199 of 249

09/28/10

At an injected noise level of 0.2, the HeBIS classifier maintained its lead position over the other two

classifiers by performing at an 85.87 % mean classification accuracy. This was a reduction of 9.2%

from its performance with the 0.1 inject noise level. The next best performer was the 3x1 SOFM in

which it attained 81.69 % classification accuracy- a reduction of 13.38 % from its performance with

the 0.1 noise level. The 2x1 SOFM classifier increased its mean classification accuracy to 67.29 %

from 63.78 % which it attained with the lower 0.1 noise level. Standard deviations of the

classification accuracies over the trials increased for all classifiers. The 3x1 SOFM data indicated a

standard deviation of accuracy of 0.031, 0.0731 for HeBIS and 0.0227 for the 2x1 SOFM classifier.

Figure 110. Noise comparison plot for classification accuracy vs. noise standard deviation for both HeBIS and

SOFM trials. Vertical bars correspond to the standard deviation of the sample mean of classification accuracy.

4.6.2.3. Experiment 15 conclusions

At the baseline noise level with a standard deviation of 0.0, the difference between the classification

averages is less than 1% between HeBIS and the 3x1 SOFM kernel. The 2x1 kernel is less resistant

to noise because its mapping is more coarse than the 3x1 kernel.

HeBIS maintains a significant 3.24 – 4.18 % improvement in classification accuracy over the next

best SOFM for noise levels that have a standard deviation that is greater than or equal to 0.1. Raw

data from these trials are included in the Appendices in Table 79 on page 232.

Page 200:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 200 of 249

09/28/10

4.6.3. Experiment 16 - Missing features

In the second robustness experiment, we examine performance with missing features in the

classification test dataset.

4.6.3.1. Setup

Experiment 16

This also tests a single “best” HeBIS genome and a “best” SOFM codebook on a full-sized test image

from the cloud/no-cloud dataset. In this case, the features associated with spectral wavelength 862-

877 nm (MODIS Band 16) and wavelength 2105-2155 nm (MODIS Band 7) are negated in separate

trials by zeroing the band value after log-normalization of the testing dataset has been applied. These

two bands are chosen because they are both used for cloud detection in NASA operational

processing, although we primarily use MODIS band 7 in our case study.

The same genome and SOFM codebooks (2x1 and 3x1) used in the noise test (Experiment 15) are

applied in this experiment. All the pixels in the test image have the chosen cloud-detection feature

“knocked out” and the resulting classification accuracies are collected for 10 trials of HeBIS and a

single SOFM trial for comparison. Also, HeBIS and SOFM classification results for this

genome/codebook applied to the non-knocked-out datasets with no injected noise from Experiment

15 are examined as baselines.

4.6.3.2. Experiment 16 results and discussion

The bar chart in Figure 111 presents a before-and-after comparison of mean classification accuracies

with the tests conducted with MODIS band 16 knocked out. The three classifiers examined are a

high performing HeBIS classifier with protein chemistry deactivated, the highest performing 3x1

SOFM classifier and the highest performing 2x1 SOFM classifier. The 3x1 SOFM has the best

before-knockout mean classification accuracy with a 97.78% success rate which falls by 7.73% to

89.45 % accuracy after the knockout. The HeBIS classifier has a baseline mean classification

accuracy of 96.2% and after the knockout it remains unchanged. Before the band 16 knockout, the

2x1 SOFM classifier achieved a 63.78% success rate in aggregate classification over the entire test

image, however after the knockout this decreased slightly to 63.44%.

Page 201:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 201 of 249

09/28/10

Figure 111. Comparison of before and after classification accuracy for the MODIS band 16 knockout.

These classification results for the MODIS band 16 knockout are summarized in Table 54.

Table 54. Missing Feature Comparison for Classification Accuracy Using HeBIS and SOFM

Algorithms with MODIS Band 16 Knockout

Algorithm Mean Standard Deviation

HeBIS Baseline (exp. 15, 0.0) 0.9620 0.0000

SOFM 2x1 Baseline (exp. 15, 0.0) 0.6378 0.0000

SOFM 3x1 Baseline (exp. 15, 0.0) 0.9718 0.0000

HeBIS Knockout (10 trials) 0.9620 0.0000

SOFM 2x1 Knockout 0.6344 0.0000

SOFM 3x1 Knockout 0.8945 0.0000

In Figure 112, a comparison of before-and-after classification results for the same three classifiers is

presented but for the MODIS band 7 knockout.

Page 202:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 202 of 249

09/28/10

Figure 112. Comparison of before and after classification accuracy for the MODIS band 7 knockout.

Each of the three classifiers essentially maintains its performance regardless of whether or not band 7

is modified. In decreasing order of mean classification performance, the 3x1 SOFM classifier

achieved 97.18 % classification accuracy before the knockout and it slightly increased by 0.20 %

after the modification. HeBIS came in as a close second with a mean classification accuracy before

the knockout of 96.20 % which it maintained after the knockout. The 2x1 SOFM came in last place

with a slight decrease from 63.78 % mean classification accuracy before the knockout to 63.68 %

accuracy after the knockout. Table 55 summarizes these results for the MODIS band 7 trials.

Standard deviations of the trials for both MODIS band 16 and 7 are 0 because the same dataset was

used and the knockouts were not random, but static and no protein chemistry was active.

Table 55. Missing Feature Comparison for Classification Accuracy Using HeBIS and SOFM

Algorithms with MODIS Band 7 Knockout

Algorithm Mean Standard Deviation

HeBIS Baseline (exp. 15, 0.0) 0.9620 0.0000

SOFM 2x1 Baseline (exp. 15, 0.0) 0.6378 0.0000

SOFM 3x1 Baseline (exp. 15, 0.0) 0.9718 0.0000

HeBIS Knockout (10 trials) 0.9620 0.0000

SOFM 2x1 Knockout 0.6368 0.0000

SOFM 3x1 Knockout 0.9738 0.0000

Page 203:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 203 of 249

09/28/10

4.6.3.3. Experiment 16 conclusions

It is apparent that the coarse 2x1 SOFM map has essentially no significant dependencies on MODIS

bands 7 or 16 although there was a very slight (0.10%) decrease in classification accuracy after

MODIS band 7 was modified. The coarse mapping of the 2x1 kernel helps in the cases of both band

7 or band 16 being modified. For the 3x1 SOFM kernel, its mapping does not depend on band 7,

however it is dependent on band 16 as indicated in Figure 111. The HeBIS GRN has derived an

algorithm that is not explicitly dependent on the band (or potential band) used to create the cloud/no-

cloud ground truth data.

Page 204:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 204 of 249

09/28/10

5. Summary and Conclusions

The focus of this thesis has been the novel hybridization of machine learning and biological

principles towards solving a relevant satellite image classification problem. This work proposed an

image classification system (HeBIS) built upon artificial protein diffusion, several principles of

evolutionary development, a control structure evolved from an artificial genetic regulatory network,

and the inclusion of a SOFM as a basic machine learning element. To our knowledge, this was the

first study of the evolution of genetic regulatory networks for control of cellular classification kernels

in the optical remote sensing domain.

The effectiveness of a genetic regulatory network applied to binary multi-dimensional classification

was examined in detail and the following contributions were made:

• Novel application of a biological construct to a practical computational classification problem

• Determination of the effectiveness of a simplified GRN applied to multi-dimensional classification

• Artificial proteins communicate classification information and results to and from the cellular machine learning kernels.

• Training of a GRN via particle swarm optimization (PSO).

• Application of a GRN-based classification system to a real-world multispectral remote sensing problem domain, i.e. cloud detection in optical satellite imagery.

• Performance comparison of HeBIS and a SOFM-only classification algorithm on a remotely-sensed multispectral dataset with unadulterated features, noisy features and deleted features.

These contributions follow from the experiments within the HeBIS environment that were conducted.

The literature review in Chapter 2 provides the theoretical basis for HeBIS. This literature review is

vital to the underpinnings of HeBIS as the system is based on the fusion of several, somewhat

disparate concepts: machine learning kernels, evolutionary development (evo-devo), genetic

regulatory networks and biological proteins, self-organization and evolutionary computation in the

form of particle swarm optimization.

Chapter 3 introduced the details of HeBIS.

The sixteen experiments in Chapter 4 were designed to characterize the behavior of HeBIS in a

detailed case study. Beginning with Experiment 1 and proceeding through Experiment 16, different

and increasingly more complex aspects were examined. This began with the investigation of the

Page 205:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 205 of 249

09/28/10

creation and diffusion of artificial proteins in an artificial lattice, proceeded to the mapping of an

artificial protein chemistry onto the created proteins, and examined the creation of simple genetic

regulatory networks. Then, the effects of PSO on the evolution of genomes and their corresponding

GRNs were investigated. The effects of different self-organizing and cellular behaviors were

examined and finally, these pieces were combined and studied under various classification scenarios.

In these classification tests, the fitness of the evolved GRNs was examined as well as classification

accuracy among the binary classes.

Experiments 1-4 examined protein creation, diffusion and the effects of a simple protein chemistry

overlaid on the basic genetic-based protein interactions. These experiments specifically targeted the

case in which only low-level spectral proteins were present in the environmental lattice because

protein activations/inhibitions were deactivated. Experiments 1 and 2 showed that the HeBIS

environmental proteins were able to diffuse through the artificial lattice whereas Experiments 3 and 4

showed that the protein chemistry varied the static protein interaction and diffusion performance that

was discovered in the first two experiments. Through these four experiments, it was determined that

the environmental proteins could interact in the lattice and create new proteins. This was an

important conclusion as the more complicated HeBIS behaviors all build upon these low-level protein

interactions.

Experiment 5 examined HeBIS in more detail with respect to gene activation and the creation of

simple genetic regulatory networks. This experiment determined if gene activation could occur and

whether it could be used as the basis for GRNs. The experiment showed that it was possible for

genes (and their resultant proteins) that were evolved with evolutionary computation (PSO in the case

of HeBIS) to become activated. A diversity of GRN behaviors were also shown to exist within the

evolved GRNs. This behavioral diversity consisted of differing genetic temporal activations for the

same gene, cellular cloning, and differing pixel responses at the GRN level. It was also determined

that the HeBIS fitness function is not necessarily geared towards simultaneous, multiple-gene

expression. In any case, the presence of these diverse behaviors provided the basis for the follow-on

experiments.

Genome/GRN optimization via evolutionary computation (PSO) was the subject of Experiment 6.

The utility of the PSO as an optimizer in HeBIS’ artificial protein environment was examined. It was

determined that the PSO was useful as an agent of optimization for HeBIS. In particular, search

diversity within the particle swarm was such that the overall fitness of the swarm increased both as

Page 206:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 206 of 249

09/28/10

the PSO iterated through successive breedings and as the number of particles in the swarm was

increased.

Self-organization and its effect on the GRN were examined in Experiments 7-10. The various self-

organization and cellular actions were examined individually and mutually exclusive of the other

actions in the set. The fitness of the GRN was evaluated for the following cases: cell location

(Experiment 7), cellular behaviors (Experiment 8), the injection of protein statistics into the

environment (Experiment 9), and the activation of output proteins (Experiment 10).

Experiment 7 varied the location of the initial starter cell and it was found that this choice did not

have a significant effect on the fitness of the evolved GRN- when this action was evaluated in

isolation of the other actions.

Experiment 8 tested the individual effects of various cellular actions. It was determined that the

independent use of these actions changed the GRN fitness when compared against the baseline in

which no cellular actions were initiated. Therefore, the activation and inhibition of these cellular

actions were found to have significant utility with respect to changes in the evolution of more

complicated classification behaviors.

In Experiment 9, the inclusion of statistical information into the artificial lattice was investigated.

This information consisted of protein analogs of local second-order statistics that were injected into

the environment during the simulation. Comparisons were made between the cases in which the

parameters of these statistics were either evolved by the PSO or were forced to remain static during

the simulation. The results indicate that bred statistical protein parameters were more likely to

significantly affect the fitness of the GRN than the static parameters were. This was potentially due

to the bred statistical proteins being subverted by the PSO to improve the fitness of the corresponding

GRN.

Experiment 10 compared the use of static versus bred output protein parameters. The results of this

experiment indicated that static output protein parameters helped to produce higher GRN fitnesses

than when bred output protein parameters were used.

Up to this point of the research, only individual actions and behaviors were tested in isolation so as to

determine the “impulse response”, per say, of HeBIS. With Experiments 11-16, HeBIS’ behaviors

Page 207:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 207 of 249

09/28/10

and modules were allowed to act in concert during the study of cloud/no-cloud classification of a full-

sized test image. In these experiments, genomes/GRNs were evolved for the purpose of

classification.

The effect of including more or less information from pixels that were geographically close to the test

pixel was examined in Experiment 11. For the limited number of cases studied, it was found that

classification accuracy actually decreased as the size of the included geographic region increased.

Experiment 12 surveyed the effect on classification accuracy due to changes in the size of the

intracellular SOFM kernel in HeBIS. Significant dependencies were found in the full-image test

classification that were based on the change in kernel size. In particular, misclassifications from class

to class were found to be the important component associated with classification success.

Experiment 13 added the before-studied protein chemistry to the gene-dictated protein interactions

from the GRN and examined the subsequent effects on classification accuracy. The conclusion was

that the mutations (new proteins) introduced into the HeBIS environment from the protein chemistry

were useful for improving classification accuracy on the full test image. However, it was found that

accuracy decreased if the probability of protein reaction increased beyond a threshold.

Fully-active HeBIS and standalone SOFMs were compared in Experiment 14. Analyses consisted of

confusion matrices, presentation of ROC curves, and visual examination of the classified imagery. In

this experiment, HeBIS was found to classify the test image as well or better than the standalone

SOFM trials. In particular, the simple HeBIS fitness function worked well enough to successfully

grade the evolved GRNs used for classification.

The final two experiments, Experiment 15 and 16, judged the robustness of HeBIS under varying

levels of injected noise and also with the information from certain features deleted entirely from the

simulation. In both cases, HeBIS was found to classify the full test image as good or better than the

standalone SOFM classifiers.

Throughout the development and research of HeBIS we have obtained results that are encouraging.

We have determined that biological inspiration hybridized with machine learning is definitely a

powerful paradigm with many interesting and useful qualities. The application of HeBIS to a real-

world classification problem in this case study demonstrates that these techniques can provide the

Page 208:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 208 of 249

09/28/10

basis for successfully solving relevant satellite image classification problems. There are many

theoretical and practical avenues for future research now that the HeBIS demonstration system has

been developed.

Page 209:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 209 of 249

09/28/10

6. Future research

This thesis opens up several possibilities for future research. In addition to applying the classifier to

other problem domains, these possibilities include the exploration of alternative fitness functions, the

inclusion of varied machine learning kernels on the cellular level, research into learned functional

modularity, the addition of new cellular actions, the use of fast processors to improve simulation turn-

around time, and finally the exploration of the potential for using protein communications in

biological systems.

6.1. Fitness functions

The current HeBIS fitness function produces simple GRNs that classify the cloud/no-cloud problem

well. However, classification accuracy vs. fitness scatter plots indicate that the current function is

somewhat noisy and that other fitness functions might improve classification performance. Also, the

HeBIS fitness function does not force a multiple-gene response and it could be interesting to explore

functions that evoke more complex GRNs such as this.

The application of Novelty Search may be useful in producing more complex GRNs [119]. Novelty

search encourages exploration of complexity without the explicit definition of fitness. In this case,

novelty search could be applied in place of the HeBIS fitness function and it could allow the system

to explore diverse operating points in an evaluated “action space”. With HeBIS, the PSO could still

remain the driver of diverse exploration under this potential avenue of research.

6.2. Additional machine learning kernels

Multiple, different kernels linked with the HeBIS GRN and its protein communications could be used

as the basis for protein-based aggregate classifiers. Machine learning kernels such as support vector

machines and artificial neural networks are obvious candidates for inclusion as both have been

extensively used in remote sensing applications [120, 121, 122].

6.3. Modularity and learned functionality

Modularity analyses could be used to uncover linkages between cluster of protein interactions along a

protein interaction pathway. These modularity analyses could be borrowed from computational

biology research [123, 124].

Additionally, learned modular functionality could be an interesting avenue of exploration through the

application of NeuroEvolution of Augmenting Topologies (NEAT) [125, 126]. Through NEAT,

complex control and classification problems can be split into separate sub-problems for which the

Page 210:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 210 of 249

09/28/10

first sub-problem is learned. This learned structure is then used as the basis for learning the next sub-

problem with learning of the remaining sub-problems following in an iterative fashion.

HeBIS already incorporates ideas of modularity with its independent cells that can have different

behaviors depending on the local protein concentrations and gradients. Modular learning in the

system could be invoked with separate sub-problem training such as that portrayed in Figure 113.

Figure 113. One method of presenting training/test data to HeBIS. Each behavior is trained separately, candidate

genomes are created, and the candidates then undergo evolutionary optimization in a final CV/GA loop.

6.4. Additional cellular actions

New cellular actions can be incorporated easily in HeBIS and would provide another venue for

potential future research. New cellular actions could include the application of scale-free network

theory to protein communications, the addition of the concept of mutual information for protein

interactions, research into new cellular instantiation algorithms, and adaptive sizing of the

intracellular SOFM kernel.

Page 211:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 211 of 249

09/28/10

6.4.1. Scale-free networks

The cell could be modified to promote scale-free protein communication links between other cells in

the HeBIS GRN (network). Scale-free networks are seen throughout biology, particularly in neural

connection networks [127, 128]. Also known as “1/f scaling”, where f is the distance between cells.

In future research, this distance could be correlated with inter-cell protein concentrations as follows:

The number of protein communication links is defined as the number of proteins from one cell that

are released into the environment that affect another cell in the environment. The number of these

communication links for cells that are close to each other in the environmental lattice is large and this

number decreases as the distance between communicating cells increases. However, this value never

falls completely to zero as the distance increases, instead it falls off as

(36)

where s is a scaling factor that accounts for the small network sizes used in the simulations.

The distance is defined as the distance between two cells in normalized units of length- 1 unit of

integral distance in the environmental lattice.

A cell in the HeBIS network could attempt to produce a scale-free environment by monitoring the

environmental proteins that are in its local neighborhood. The concentrations of these proteins could

be used to probabilistically determine which of these environmental proteins may be used as switch

templates for intracellular behaviors. Environmental proteins with high concentrations will have a

lower probability of being used as a switch template than those proteins with low concentrations.

A scale-free network could potentially allow information to efficiently move between upper and

lower levels of a network hierarchy as well as across the network.

6.4.2. Mutual information

Mutual information also provides a venue for future research [129]. Mutual information is a measure

of the amount of information that can be determined about one random variable given information

about another random variable.

Page 212:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 212 of 249

09/28/10

It is defined as

I X Y` a

=Xx,y

p y` a

p x|yb c

logp x|yb c

p x` a

ffffffffffffffffffffffh

l

j

i

m

k, (37)

and equivalently,

Xx,y

p x,y` a

logp x,y` a

p x` a

p y` a

fffffffffffffffffffffffffffffffffh

j

i

k. (38)

For this research, mutual information could be defined between different proteins and their

concentrations. The mathematical analog is that the protein is similar to a random variable where its

concentration is similar to the sampled value of the random variable. Through this protein mutual

information, the cell could attempt to use those proteins that appear to have more information content

than others. This process could occur purely from the perspective of the local cell and the

environmental proteins that it has seen in prior time steps. This is another method by which

irrelevant features in the protein feature vector may be discarded, thus potentially improving the

classification by decreasing the effect of the curse of dimensionality in large-dimensional spaces.

Another method through which this feature reduction could occur is through the GRN- both from the

unmapping of feature proteins that are not considered important by the network as well as the fact

that the added complexity of the GRN itself could allow for better management of the curse.

6.4.3. Cellular instantiation

Currently, new cells are instantiated purely when the corresponding switch for the ADDCELL action

is activated by specific evolved proteins within the artificial environment. One idea to extend this

paradigm is to always force cellular instantiation when protein concentration reaches a particular

level in the environment. Acting in conjunction (or separately) with the current ADDCELL action,

this type of instantiation could provide a more direct link to spectral texture information from the

input features as it is encoded by the prevalence of the artificial spectral proteins in different

geographic regions of the image. A second avenue of exploration is to base cellular instantiation at

the beginning of classification on the number of input domain features. This jumpstart could

potentially simplify the evolution of a suitable GRN for complex classification problems.

Page 213:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 213 of 249

09/28/10

6.5. Graphics Processing Unit

The amount of evolutionary computation required to evolve the HeBIS network requires a robust

system with high processing throughput. Because of this, future research is predicated on building a

processing system based on a high-performance graphics processing units (GPU) that can provide

higher sustained floating point processing throughput compared to a state-of-the-art CPU [12,113].

The computational hardware required for fast iteration through the HeBIS training and testing cycles

requires a processing capability that is several times greater than what is currently available in a state-

of-the-art i7-based desktop computer. This need can be addressed either through the acquisition of

time on a supercomputer, the creation of a relatively small processing cluster, or through the

innovative use of commodity computer graphics chips. By using a commodity graphics processing

unit (GPU) that is readily available for high-end desktop gaming and visualization computers, the

promise is that an order of magnitude increase in computation throughput can be achieved through

the use of a relatively inexpensive high-end graphics card [113]. As an example, harnessing the

specialized processor hardware on the $450 ATI X800 XT GPU will provide 63 GFLOPs of

sustained processing throughput compared to a peak of 14.8 GFLOPs for a 3.7 GHz Pentium 4

computer [113]. The “secret” behind this computational cornucopia is that the hardware structure in

the GPU is optimized for high-speed parallel mathematical operations on pixels whereas the typical

CPU is optimized for sequential code. This highlights the problem that is associated with the use of a

GPU as an inexpensive, desktop supercomputer: How does one efficiently map a general purpose

computation to hardware that has been designed for efficient graphics rendering and pixel

manipulation? This computational problem is being addressed by researchers and has resulted in

several high-level programming languages that are similar to C/C++. Examples include ‘sh’ from the

University of Waterloo and Brook from Stanford [113]. Although these (and other) languages are

still no panacea for the GPU computational mapping problem, they have recently been used for

evolutionary computation applications. [114] and [115] are examples of genetic algorithms that have

been ported to a GPU. Because of this, it is believed that future development and analysis of HeBIS-

inspired learning algorithms should be implemented on a GPU so as to enable quick iteration and

testing for research purposes.

Page 214:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 214 of 249

09/28/10

Figure 114. Processing model for research infrastructure.

6.6. Protein-based communications for artificial devices in a biological system

There is a potential that the HeBIS protein-based communications and classification could be used for

controlling human-made devices in biological systems, e.g. a human body. These devices could

conceivably be designed to respond to the local concentrations and gradients of specific proteins and

to perform internal computations based on a codebook of simple, artificial proteins.

Page 215:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 215 of 249

09/28/10

7. Appendices

Page 216:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 216 of 249

09/28/10

7.1. Data

Table 56. Experiment 6 Aggregate Breeding Data.

Breed Average

Best

Fitness

(1)

Best

Fitness

SDV

(1)

Average Best

Fitness

(100)

Best

Fitness

SDV

(100)

Average Best

Fitness

(250)

Best

Fitness

SDV

(250)

Average Best

Fitness

(500)

Best

Fitness

SDV

(500)

0 0 0 0 0 0 0 0 0

1 0.00027655 0.00177126 0.00464188 0.0117422 0.01555976 0.03278979 0.01398984 0.02187814

2 0.00085993 0.00487531 0.00901192 0.02481268 0.02385866 0.04189065 0.02643543 0.03746593

3 0.00098304 0.00506875 0.01277754 0.02968328 0.03147017 0.04936366 0.03700024 0.05196669

4 0.00118006 0.00548549 0.01675406 0.03625152 0.03616746 0.05304285 0.04489608 0.06000474

5 0.00137479 0.0059233 0.01812938 0.03766535 0.03998255 0.05612186 0.04783097 0.06140292

6 0.0015833 0.00645447 0.01954484 0.03929089 0.04166388 0.05620707 0.05216529 0.06440029

7 0.00168214 0.00655383 0.02140479 0.04321551 0.0434137 0.05645667 0.05899074 0.06897797

8 0.00176476 0.00667782 0.02267569 0.04443178 0.04584552 0.05630136 0.06503971 0.07238485

9 0.0018308 0.00677083 0.02385972 0.04586475 0.05028934 0.05904922 0.0723335 0.07971921

Table 57. Experiment 7 Data.

Feature Index Mean Fitness Fitness Standard Deviation

1 0.039564735 0.031173286

2 0.0510463 0.048527315

3 0.04498486 0.026785561

4 0.036350745 0.022174415

5 0.04720598 0.031625397

6 0.04069623 0.024782471

7 0.041230425 0.032343847

8 0.05012401 0.037711992

9 0.062870475 0.026917273

10 0.05576429 0.033785508

11 0.05222655 0.02367747

12 0.04195883 0.026432196

13 0.045210895 0.028171183

14 0.04615925 0.031231122

15 0.038886235 0.023972954

Page 217:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 217 of 249

09/28/10

Table 58. Experiment 8 Data.

Genetic Action Index Genetic Action Mean Fitness Fitness Standard Deviation

1 NO_ACTION 0.020973752 0.035764639

2 ADD_CELL 0.012882272 0.031065069

3 PRUNE_SELF 0.015769768 0.028084739

4 ACTIVATEENVIROPROTEIN 0.068884336 0.037255274

5 INHIBITENVIROPROTEIN 0.029064268 0.049464757

6 ACTIVATEREGPROTEIN 0.030676984 0.040561945

7 INHIBITREGPROTEIN 0.01569026 0.033994529

8 CHANGETOSOFMANDTRAIN 0.036585236 0.037505919

9 CLASSIFY 0.020820384 0.035469301

Table 59. Experiment 9 Data

Protein Statistics Index Protein Statistics Action Mean Fitness Fitness Standard Deviation

10 STATS_ACTIVE_STATIC 0.020746292 0.026351562

11 STATS_ACTIVE_BRED 0.303572754 0.048304406

Table 60. Experiment 10 Data.

Protein Output Index Protein Output Action Mean Fitness Fitness Standard Deviation

12 OUTPUT_ACTIVE_STATIC 0.406778162 0.067970135

13 OUTPUT_ACTIVE_BRED 0.18472541 0.054844318

Table 61. Experiment 11 Data for 3x3 and 5x5 Geographic Region Comparison

Number of Neurons Classification Accuracy Mean Classification Accuracy Standard Deviation

3x3 0.9147 0.0812

5x5 0.8396 0.1361

Page 218:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 218 of 249

09/28/10

Table 62. Experiment 11 Scatter Data for 3x3 Geographic Region

Trial Index Fitness Classification Accuracy

0 0.6210 0.9699

1 0.5719 0.9296

2 0.2208 0.9307

3 0.5971 0.8596

4 0.6307 0.9755

5 0.3914 0.9182

6 0.5943 0.9699

7 0.4180 0.9182

8 0.6386 0.9307

9 0.5956 0.8849

10 0.5461 0.9699

11 0.4488 0.8596

12 0.5418 0.9256

13 0.6250 0.9307

14 0.4414 0.5688

15 0.5816 0.9112

16 0.5040 0.9699

17 0.5790 0.9182

18 0.6234 0.9755

19 0.4105 0.8860

20 0.5035 0.9755

21 0.6005 0.8602

22 0.6473 0.9307

23 0.6057 0.9299

24 0.6042 0.9699

Page 219:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 219 of 249

09/28/10

Table 63. Experiment 11 Scatter Data for 5x5 Geographic Region

Trial Index Fitness Classification Accuracy

0 0.5444 0.8594

1 0.5146 0.9537

2 0.1974 0.9072

3 0.5073 0.9041

4 0.4320 0.7085

5 0.4898 0.5595

6 0.6136 0.9106

7 0.6091 0.9041

8 0.4648 0.5603

9 0.3714 0.5551

10 0.4566 0.9356

11 0.4599 0.8193

12 0.4796 0.8841

13 0.6337 0.9035

14 0.4145 0.5603

15 0.5397 0.9488

16 0.5960 0.9439

17 0.4394 0.8689

18 0.5477 0.9642

19 0.5349 0.9426

20 0.4859 0.8741

21 0.4761 0.8717

22 0.6164 0.9488

23 0.4752 0.8732

24 0.3505 0.8282

Table 64. Experiment 12 - Intracellular SOFM Data

# Neurons in Intracellular SOFM Classification Accuracy Mean Classification Accuracy Standard Deviation

0 0.8894 0.1123

1 0.8484 0.1678

4 0.8736 0.1371

9 0.9107 0.0921

81 0.8716 0.1362

Page 220:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 220 of 249

09/28/10

Table 65. Experiment 12 Scatter Data for 0x0 Intracellular SOFM

Trial Index Fitness Classification Accuracy

0 0.1300 0.9307

1 0.5400 0.8527

2 0.1600 0.5809

3 0.5700 0.9307

4 0.5200 0.9301

5 0.5400 0.8849

6 0.6000 0.9699

7 0.6400 0.8596

8 0.4500 0.8860

9 0.5300 0.9301

10 0.1700 0.9182

11 0.5000 0.8849

12 0.6200 0.9699

13 0.6500 0.9755

14 0.5700 0.9699

15 0.6200 0.9182

16 0.4800 0.9307

17 0.4500 0.9182

18 0.5900 0.9699

19 0.5100 0.5766

Table 66. Experiment 12 Scatter Data for 1x1 Intracellular SOFM

Trial Index Fitness Classification Accuracy

0 0.0800 0.5722

1 0.5800 0.9307

2 0.0700 0.5704

3 0.2700 0.9755

4 0.5600 0.9755

5 0.6100 0.9699

6 0.4900 0.5713

7 0.5600 0.9182

8 0.6200 0.9699

9 0.0800 0.5704

10 0.5200 0.9480

11 0.6000 0.9755

12 0.4700 0.9182

13 0.5900 0.9128

14 0.5100 0.8860

15 0.5600 0.9307

16 0.6200 0.9755

17 0.6000 0.9755

18 0.6000 0.8527

19 0.4200 0.5690

Page 221:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 221 of 249

09/28/10

Table 67. Experiment 12 Scatter Data for 2x2 Intracellular SOFM

Trial Index Fitness Classification Accuracy

0 0.1300 0.9182

1 0.4300 0.9182

2 0.4600 0.9307

3 0.6200 0.9182

4 0.6400 0.9755

5 0.3600 0.8527

6 0.5000 0.9301

7 0.5700 0.9755

8 0.5100 0.8860

9 0.2700 0.9307

10 0.1900 0.9307

11 0 0.5688

12 0.5300 0.8596

13 0.5700 0.9699

14 0.3300 0.5694

15 0.5400 0.9699

16 0.4000 0.5694

17 0.5000 0.9699

18 0.5800 0.9755

19 0.5300 0.8527

Table 68. Experiment 12 Scatter Data for 3x3 Intracellular SOFM

Trial Index Fitness Classification Accuracy

0 0.6000 0.9307

1 0.4700 0.9182

2 0.4800 0.9699

3 0.5700 0.9301

4 0.1700 0.9182

5 0.4800 0.9015

6 0.6200 0.8596

7 0.5200 0.9755

8 0.5100 0.9755

9 0.6300 0.9301

10 0.4400 0.8596

11 0.4900 0.9307

12 0.2400 0.9699

13 0.6100 0.9307

14 0.3300 0.5688

15 0.4200 0.9301

16 0.6300 0.9755

17 0.4800 0.9182

Page 222:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 222 of 249

09/28/10

Table 69. Experiment 12 Scatter Data for 9x9 Intracellular SOFM

Trial Index Fitness Classification Accuracy

0 0.1400 0.5809

1 0.5800 0.9699

2 0.5700 0.8934

3 0.6600 0.8849

4 0.5700 0.9755

5 0.6000 0.9755

6 0.5800 0.9755

7 0.5600 0.8527

8 0.4700 0.5689

9 0.5000 0.9301

10 0.5300 0.9699

11 0.5900 0.8602

12 0.5300 0.8602

13 0.4400 0.8527

14 0.5100 0.9699

15 0.5300 0.9182

16 0.5200 0.9182

17 0.6400 0.9307

18 0.3700 0.5691

19 0.5600 0.9755

Table 70. Experiment 13 Aggregate Classification Data

Classification Accuracy Mean Classification Accuracy Mean Classification Accuracy Standard Deviation

0.0 0.8710 0.1163

0.0010 0.9388 0.0397

0.0100 0.8547 0.1298

0.1000 0.8587 0.0590

Page 223:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 223 of 249

09/28/10

Table 71. Experiment 13 Data for 0.0 Reaction Probability

Trial Index Fitness Classification Accuracy

0 0.44 0.8527

1 0.65 0.9699

2 0.46 0.8527

3 0.51 0.8602

4 0.54 0.9699

5 0.58 0.8849

6 0.52 0.9755

7 0.45 0.9301

8 0.6 0.8849

9 0.59 0.9301

10 0.56 0.9307

11 0.52 0.8849

12 0.59 0.9699

13 0.3 0.7603

14 0.6 0.8596

15 0.61 0.9301

16 0.4 0.5692

17 0.59 0.8527

18 0.44 0.5768

19 0.57 0.9755

Table 72. Experiment 13 Data for 0.001 Reaction Probability

Trial Index Fitness Classification Accuracy

0 0.5200 0.9179

1 0.5900 0.9699

2 0.5700 0.9301

3 0.5400 0.9675

4 0.5700 0.9301

5 0.6600 0.9697

6 0.5300 0.9755

7 0.6300 0.9307

8 0.6000 0.9979

9 0.4600 0.9698

10 0.5400 0.9591

11 0.6500 0.9685

12 0.6600 0.8618

13 0.5300 0.9755

14 0.6600 0.9263

15 0.4600 0.8587

16 0.4500 0.9185

17 0.4600 0.8988

18 0.5000 0.8848

19 0.6000 0.9649

Page 224:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 224 of 249

09/28/10

Table 73. Experiment 13 Data for 0.01 Reaction Probability

Trial Index Fitness Classification Accuracy

0 0.6100 0.8905

1 0.5400 0.9733

2 0.6400 0.9024

3 0.2600 0.6263

4 0.2500 0.9685

5 0.3600 0.8812

6 0.5300 0.7710

7 0.2900 0.6460

8 0.5200 0.9670

9 0.5800 0.9141

10 0.3900 0.8814

11 0.6100 0.9746

12 0.6500 0.8537

13 0.4900 0.5754

14 0.5100 0.8971

15 0.6500 0.9257

16 0.6500 0.9695

17 0.6400 0.8626

18 0.3000 0.6457

19 0.6400 0.9682

Table 74. Experiment 13 Data for 0.1 Reaction Probability

Trial Index Fitness Classification Accuracy

0 0.4200 0.7717

1 0.5600 0.8577

2 0.4500 0.8443

3 0.5600 0.9217

4 0.4400 0.9075

5 0.4000 0.8372

6 0.6000 0.8839

7 0.5100 0.7744

8 0.5000 0.7864

9 0.4700 0.9323

10 0.4600 0.9343

11 0.6000 0.8880

12 0.5600 0.8590

13 0.6100 0.8876

14 0.4600 0.7847

15 0.4000 0.9026

16 0.5800 0.8928

17 0.5800 0.8441

18 0.5200 0.9194

19 0.5200 0.7437

Page 225:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 225 of 249

09/28/10

Table 75. Experiment 14 - Statistical Summary Data for HeBIS Classification Accuracy and

Fitness Scatter Data

Statistic Fitness Classification Accuracy

min 0 0.4312

max 0.82 0.9631

mean 0.3034 0.7123

median 0.265 0.6361

mode 0 0.5688

std 0.2425 0.1515

range 0.82 0.5319

Table 76. Experiment 14 HeBIS Scatter Data for Fitness and Classification Accuracy

Trial Index Fitness Classification Accuracy

0 0.35 0.8684

1 0.36 0.8838

2 0 0.5688

3 0.18 0.5688

4 0.37 0.5688

5 0 0.57

6 0.39 0.674

7 0.67 0.7889

8 0.2 0.6198

9 0 0.5699

10 0 0.5688

11 0.12 0.621

12 0.18 0.9203

13 0.07 0.6121

14 0 0.5716

15 0.17 0.6427

16 0.36 0.569

17 0.56 0.919

18 0.12 0.569

19 0.24 0.5884

20 0.07 0.5711

21 0.47 0.9288

22 0.3 0.7117

23 0.37 0.7632

24 0.26 0.6105

25 0 0.5688

26 0 0.5688

27 0.73 0.8951

28 0 0.5688

29 0.15 0.6726

30 0.08 0.5831

31 0.32 0.6716

32 0.1 0.5953

33 0.59 0.6755

Page 226:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 226 of 249

09/28/10

34 0.22 0.5688

35 0.64 0.9343

36 0.82 0.9444

37 0.2 0.6895

38 0 0.569

39 0.65 0.8639

40 0.06 0.6134

41 0.27 0.7418

42 0.32 0.569

43 0.65 0.9479

44 0.38 0.6085

45 0.47 0.9029

46 0.02 0.5852

47 0.64 0.8562

48 0.07 0.6295

49 0.21 0.7469

50 0.79 0.9366

51 0 0.5688

52 0.31 0.8761

53 0.55 0.9232

54 0.42 0.8246

55 0.06 0.5688

56 0.76 0.9446

57 0.49 0.8262

58 0.51 0.8336

59 0.54 0.9073

60 0.38 0.6833

61 0 0.4312

62 0.63 0.8887

63 0.7 0.8272

64 0.15 0.5688

65 0.26 0.9243

66 0.42 0.8811

67 0 0.5793

68 0.05 0.5691

69 0.06 0.5902

70 0.08 0.5869

71 0.71 0.9339

72 0.53 0.9042

73 0 0.569

74 0.8 0.8719

75 0.5 0.9631

76 0.4 0.8898

77 0.17 0.5688

78 0.17 0.9009

79 0.39 0.5688

80 0.69 0.9294

81 0.53 0.9136

82 0.08 0.6078

83 0.69 0.9284

84 0 0.5688

85 0 0.5688

86 0.13 0.607

87 0.26 0.7946

88 0.24 0.5727

89 0.07 0.5734

90 0.69 0.9229

91 0.45 0.5688

Page 227:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 227 of 249

09/28/10

92 0.28 0.5688

93 0.55 0.8152

94 0.22 0.7591

95 0.16 0.5932

96 0.29 0.5688

97 0.43 0.903

98 0.3 0.569

99 0.05 0.569

100 0.3512 0.569

101 0.1675 0.5688

102 0.8094 0.7428

103 0.5382 0.9838

104 0.384 0.937

105 0.0635 0.5697

106 0.3124 0.6927

107 0.0075 0.5688

108 0.3182 0.5904

109 0 0.569

110 0.2841 0.7454

111 0.0955 0.639

112 0 0.5688

113 0.597 0.8123

114 0.0541 0.5945

115 0 0.5688

116 0.644 0.9519

117 0.6829 0.8592

118 0.2669 0.7003

119 0.5179 0.8729

120 0.4633 0.8294

121 0.2206 0.7428

122 0.2768 0.767

123 0.0236 0.5881

124 0 0.5688

125 0.5824 0.8975

126 0.753 0.7552

127 0.5752 0.9372

128 0.3422 0.5713

129 0.6215 0.9183

130 0 0.5688

131 0.8747 0.9308

132 0.5956 0.893

133 0.1371 0.721

134 0.4704 0.7995

135 0 0.5688

136 0.7757 0.8628

137 0.2016 0.5688

138 0 0.5691

139 0.1274 0.5688

140 0.5313 0.7463

141 0.0852 0.5688

142 0.072 0.5689

143 0.0406 0.5761

144 0.6003 0.9181

145 0 0.569

146 0.2368 0.7162

147 0.7508 0.9179

148 0.5126 0.7956

149 0.1095 0.5689

Page 228:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 228 of 249

09/28/10

150 0.1983 0.5814

151 0.0687 0.5814

152 0.2488 0.7254

153 0.0701 0.5892

154 0.5141 0.8566

155 0.076 0.6006

156 0.1471 0.576

157 0.5882 0.8395

158 0 0.5688

159 0.6008 0.6286

160 0.1597 0.6639

161 0 0.5688

162 0.5688 0.9198

163 0.6854 0.9321

164 0.2625 0.655

165 0.4991 0.981

166 0.0682 0.5688

167 0.2776 0.5748

168 0.6011 0.9228

169 0.3835 0.5767

170 0.6597 0.8832

171 0.1813 0.5688

172 0.4722 0.8957

173 0.1865 0.6232

174 0.4205 0.758

175 0.6355 0.7591

176 0 0.5688

177 0.0717 0.5811

178 0.611 0.9339

179 0.4897 0.768

180 0.2623 0.6364

181 0.1793 0.5688

182 0.4865 0.782

183 0.2089 0.6704

184 0.1413 0.6657

185 0.6919 0.9279

186 0 0.5714

187 0.852 0.9811

188 0.3809 0.9444

189 0.0795 0.5688

190 0 0.5696

191 0.6799 0.8897

192 0.4964 0.6963

193 0.2776 0.5764

194 0.3512 0.8427

195 0.7043 0.9405

196 0.1496 0.6303

197 0.3157 0.6895

198 0.5535 0.9443

199 0 0.5752

200 0.5244 0.9232

201 0.1725 0.5688

202 0.0944 0.5193

203 0.5127 0.87

204 0.4279 0.8524

205 0.16 0.5688

206 0.1586 0.5706

207 0.6497 0.9426

Page 229:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 229 of 249

09/28/10

208 0.5101 0.8879

209 0.0601 0.5688

210 0.5363 0.9291

211 0.1667 0.6643

212 0.5847 0.8678

213 0.1706 0.6508

214 0.4105 0.7503

215 0.393 0.8961

216 0.1094 0.5708

217 0.2609 0.6211

218 0.2866 0.7646

219 0.2873 0.7791

220 0.4912 0.6764

221 0.0912 0.5365

222 0.7354 0.9081

223 0.7637 0.9428

224 0.5285 0.9397

225 0.0925 0.6266

226 0.4085 0.6736

227 0.3312 0.7676

228 0.5134 0.9169

229 0.3269 0.6624

230 0.3437 0.5688

231 0.2975 0.5728

232 0.2059 0.6038

233 0.3235 0.5688

234 0.7135 0.9915

235 0.1953 0.5688

236 0.5159 0.9009

237 0.1986 0.5714

238 0.066 0.5688

239 0.301 0.923

240 0.2326 0.6269

241 0.0978 0.5688

242 0.6236 0.9259

243 0.1691 0.5688

244 0.3569 0.5792

245 0.6249 0.7031

246 0.7232 0.9076

247 0.5947 0.8508

248 0.3011 0.7383

249 0.5918 0.7605

250 0.8685 0.9284

251 0.1299 0.7746

252 0.3794 0.9368

253 0.1588 0.6353

254 0.2952 0.6571

255 0.4624 0.8175

256 0.4037 0.8389

257 0.0248 0.5703

258 0.2766 0.6795

259 0.3142 0.5751

260 0.1762 0.5279

261 0.1342 0.5688

262 0.219 0.6046

263 0.4781 0.9275

264 0.7801 0.8865

265 0.6403 0.8943

Page 230:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 230 of 249

09/28/10

266 0.3154 0.6276

267 0.1297 0.5738

268 0.552 0.91

269 0.2652 0.7138

270 0.5755 0.8408

271 0.0558 0.4681

272 0.2192 0.5688

273 0.6579 0.931

274 0.473 0.8005

275 0.4082 0.9295

276 0.8855 0.8565

277 0.1781 0.5141

278 0.1648 0.5763

Table 77. Experiment 14 - Aggregate Classification Accuracy Results for SOFM.

Trial

#

Classification

Accuracy

SOFM

Topology

# of

Neurons

1st stage

#iterations

1st

stage

alpha

1st

stage

radius

2nd stage

#iterations

2nd

stage

alpha

2nd

stage

radius

600 0.637726 2x1 2 100 0.05 2 1000 0.02 0.5

601 0.634084 2x1 2 1000 0.05 2 100000 0.02 0.5

602 0.634084 2x1 2 1000 0.05 2 100000 0.02 0.5

603 0.971786 3x1 3 200 0.05 2 2000 0.02 0.5

604 0.971159 2x2 4 200 0.05 2 2000 0.02 0.5

605 0.953395 3x2 6 200 0.05 2 2000 0.02 0.5

606 0.948468 5x4 20 200 0.05 2 2000 0.02 0.5

607 0.971995 5x5 25 200 0.05 2 2000 0.02 0.5

608 0.972294 8x7 56 200 0.05 2 2000 0.02 0.5

Page 231:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 231 of 249

09/28/10

Table 78. Selected Results from Experiment 14

Test Dataset c_50 c_61 c_75 c_76 d_3 d_85 e_23 e_34 e_38 e_77

Delta_sec 3310 1497 447 906 36881 206 73 460 377 1417

superBreedingCounterMax

16 19 15 17 17 4 5 11 10 19

counterMax 348 2770 779 1184 591 2943 1264 3859 3431 2999

numberParticles 426 197 104 110 436 270 24 209 120 442

reactionProbability 0 0 0 0 0 0 0 0.03 0.02 0.02

outputProteinValueSwitch

0 0 1 1 0 0 0 0 1 1

minCorrelationValue 0.95 0.97 0.95 0.96 0.91 0.99 0.98 0.98 0.98 0.91

numberGenes 19 6 9 19 11 10 15 4 14 13

sofmSizeX 3 3 1 1 2 0 1 1 1 0

sofmSizeY 50 61 75 76 3 85 23 34 38 77

diffusionRate 0.94 0.04 0.56 0.12 0.02 0.95 0.91 0.86 0.43 0.37

diffusionRateCell 0.85 0.75 0.08 0.43 0.29 0.94 0.47 0.88 0.51 0.51

statFlagCell 0 0 1 1 0 1 1 1 1 0

Fitness 0.79 0 0.5 0.4 0.54 0.69 0.76 0.71 0.07 0.18

#Correct_all_valid 31372 14441 32259 29803 32952 31078 31577 33209 19053 17220

%Correct_all_valid 93.66 43.12 96.31 88.98 98.38 92.79 94.28 99.15 56.88 51.41

#Incorrect_all_valid 2122 19053 1235 3691 542 2416 1917 285 14441 16274

%Incorrect_all_valid 6.34 56.88 3.69 11.02 1.62 7.21 5.72 0.85 43.12 48.59

#C0_valid_correct 16931 0 19053 15554 18979 16765 17581 18770 19053 2940

%C0_valid_correct 88.86 0 100 81.64 99.61 87.99 92.27 98.51 100 15.43

#C0_valid_incorrect 2122 19053 0 3499 74 2288 1472 283 0 16113

%C0_valid_incorrect 11.14 100 0 18.36 0.39 12.01 7.73 1.49 0 84.57

#C1_valid_correct 14441 14441 13206 14249 13973 14313 13996 14439 0 14280

%C1_valid_correct 100 100 91.45 98.67 96.76 99.11 96.92 99.99 0 98.89

#C1_valid_incorrect 0 0 1235 192 468 128 445 2 14441 161

%C1_valid_incorrect 0 0 8.55 1.33 3.24 0.89 3.08 0.01 100 1.11

Page 232:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 232 of 249

09/28/10

Table 79. Data for Experiment 15- Classification Accuracy for 0.1 Probability Noise Injection

with Varying Noise Standard Deviations

Trial 0.0_2x1 0.01_2x1 0.1_2x1 0.2_2x1 0.0_3x1 0.01_3x1 0.1_3x1 0.2_3x1 0.0_hebis 0.01_hebis 0.1_hebis 0.2_hebis

0 0 0.6378 0.6378 0.6378 0.7258 0.9718 0.9716 0.9502 0.7841 0.962 0.9607 0.9783 1 1 0.6378 0.6378 0.6378 0.6875 0.9718 0.9716 0.9476 0.8135 0.962 0.9606 0.9849 2 2 0.6378 0.6378 0.6378 0.6488 0.9718 0.9717 0.9505 0.8455 0.962 0.9605 0.9862 3 3 0.6378 0.6378 0.6378 0.662 0.9718 0.9717 0.9525 0.8193 0.962 0.9607 0.9813 4 4 0.6378 0.6378 0.6378 0.6703 0.9718 0.9716 0.9528 0.8329 0.962 0.9606 0.9818 5 5 0.6378 0.6378 0.6378 0.6554 0.9718 0.9716 0.9554 0.8377 0.962 0.9607 0.9836 6 6 0.6378 0.6378 0.6378 0.6609 0.9718 0.9718 0.957 0.8342 0.962 0.9607 0.9942 7 7 0.6378 0.6378 0.6378 0.6691 0.9718 0.9717 0.9459 0.8094 0.962 0.9607 0.9837 8 8 0.6378 0.6378 0.6378 0.6896 0.9718 0.9718 0.945 0.7469 0.962 0.9606 0.9854 9 9 0.6378 0.6378 0.6378 0.66 0.9718 0.9718 0.9506 0.8451 0.962 0.9606 0.9719

Table 80. Data for Experiment 16 – Classification Accuracy with MODIS Band 15 Knocked

Out

Trial 2x1 SOFM 3x1 SOFM HeBIS

0 0.634387 0.894503 0.962017

1 0.634387 0.894503 0.962017

2 0.634387 0.894503 0.962017

3 0.634387 0.894503 0.962017

4 0.634387 0.894503 0.962017

5 0.634387 0.894503 0.962017

6 0.634387 0.894503 0.962017

7 0.634387 0.894503 0.962017

8 0.634387 0.894503 0.962017

9 0.634387 0.894503 0.962017

Page 233:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor

Table 81. Comparison of Selected Genomes for Shotgu

HeBIS

Sel.

Trial

#

Dataset

0 2010_05_28_07_58_58_61

1 2010_06_11_01_18_20_77

2 2010_06_10_08_03_09_38

2a 2010_06_09_16_37_07_17

3 2010_05_28_12_49_29_76

4 2010_06_03_10_39_23_85

5 2010_05_28_00_32_27_50

6 2010_06_09_17_43_13_23

7 2010_05_28_12_42_02_75

8 2010_06_01_233_35_56_3

9 2010_06_10_07_35_57_34

. Comparison of Selected Genomes for Shotgun Experiments

Genome

233 of 249

n Experiments

Colorbars

Page 234:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor

7.2. Genome mappings for

Figure 115. Parameter activation maps for genomes discovere

Table 81 is applicable to the evolved elements of the genomes in this figure.

Genome mappings for shotgun data

. Parameter activation maps for genomes discovered during shotgun experiments. The color bar from

is applicable to the evolved elements of the genomes in this figure.

234 of 249

d during shotgun experiments. The color bar from

Page 235:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor

7.3. Protein diffusion example

Figure 116. Single protein diffusion from four sites within an

initial sites of protein activation at the beginning of the simulation

Figure 117. Frame #2 in the simulation

iteration 2 of the diffusion simulation

two green sites are decaying. The light blue color represents sites within the lattice that hav

protein concentrations at this point in the simulation.

Protein diffusion example

ffusion from four sites within an 11x11x11 cubic environment. This frame

of protein activation at the beginning of the simulation

#2 in the simulation This frame is a snapshot of activity in the environmental lattice after

of the diffusion simulation. The two red protein sites are still actively producing proteins whereas the

The light blue color represents sites within the lattice that hav

protein concentrations at this point in the simulation.

235 of 249

11x11x11 cubic environment. This frame shows four

environmental lattice after

. The two red protein sites are still actively producing proteins whereas the

The light blue color represents sites within the lattice that have the lowest non-zero

Page 236:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 236 of 249

09/28/10

Figure 118. Frame #6. The sites colored red are still actively producing whereas the light-blue-colored and dark-

blue-colored sites possess lower concentrations of the simulated protein. The darker blue sites contain lower

concentrations of protein than the light-blue sites. This is the snapshot from iteration 6 of the simulation.

Figure 119. Frame #26. After 26 iterations, the artificial protein has diffused throughout a large portion of the

11x11x11 environmental matrix. Hotter colors (e.g. red, yellow, green) correspond to higher concentrations of the

proteins whereas cooler colors (e.g. light blue, blue) correspond to areas of relatively low concentrations.

Page 237:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 237 of 249

09/28/10

Figure 120. Frame #40. At iteration 40, the protein has diffused throughout the environmental lattice. The red

sites are the locations of the original and continuing protein sources. Hotter colors correspond to higher protein

concentrations whereas cooler colors correspond to lower concentrations.

Page 238:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 238 of 249

09/28/10

7.4. HeBIS fitness function details

Figure 121. Regions of equivalent CorrC0 max

for the fitness function.

Figure 122. Θcorr

2 portion of fitness function.

Page 239:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 239 of 249

09/28/10

Figure 123. Magcorr

2 portion of fitness function.

Page 240:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 240 of 249

09/28/10

7.5. HeBIS Shotgun Experiment Correlation Maps and PDF

Figure 124. Correlation coefficient grid for processing parameters and classification results. Parameters and

results are numbered from 1 to 26.

Figure 125. Significance p-value grid for processing parameters and classification results obtained with a

Student’s t-test. Parameters and results are numbered from 1 to 26.

Page 241:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 241 of 249

09/28/10

7.6. HeBIS training cycle details

Figure 126. HeBIS classification training cycle.

Page 242:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 242 of 249

09/28/10

Bibliography [1] D.A. Landgrebe, “Multispectral Land Sensing: Where From, Where to?”, IEEE Transactions on Geoscience and Remote Sensing, vol. 43, no. 3, pp. 414-421, Mar. 2005. [2] A. Ranganathan and Z. Kira, “Self-Organization in Artificial Intelligence and the Brain”, College of Computing, Georgia Institute of Technology, 2003. [3] P. Bentley, “Evolving Fractal Gene Regulatory Networks for Robot Control”, Department of Computer Science, University College London, 2003. [4] J. Bongard, “Evolving Modular Genetic Regulatory Networks”, Artificial Intelligence Laboratory, University of Zurich. [5] S. B. Carroll, “Endless Forms Most Beautiful”, W.M. Norton & Company, New York, 2005. [6] T.G. Dietterich, “Machine Learning”, Department of Computer Science, Oregon State University, 2001. [7] R.P. Lippmann, Pattern Recognition, class, MIT Lincoln Laboratory, 2001. [8] Alek Kolcz, System Architect, AOL, private communication, 2004. [9] Y. Yingjie, C. Hinde, and D. Gillingwater, “Improved neural network training using redundant structure,” Proc. of the Int. Joint Conference on Neural Networks 2003, vol. 3, pp. 2023-2027, 2003. [10] R. Reed, “Pruning algorithms- a survey”, IEEE Transactions on Neural Networks, vol. 4, no. 5, pp. 740-747, Sept. 1993. [11] S. Osowski and K. Siwek, “Regularisation of neural networks for improved load forecasting in the power system”, IEEE Proceedings on Generation, Transmission and Distribution, vol. 149, no. 3, pp. 340-344, May 2002. [12] C.-A. Bohn, Kohonen, "Feature Mapping through Graphics Hardware”, Computational Intelligence and Neuroscience (ICCIN ’98), 1998. [13] H. Weiming, D. Xie, T. Tieniu, and S. Maybank, “Learning activity patterns using fuzzy self-organizing neural network”, IEEE Trans. On Systems, Man and Cybernetics, Part B, vo. 34, no. 3, pp. 1618-1626, June 2004. [14] W.S. McCulloch and W.S. Pitts, “A Logical Calculus of the Ideas Immanent in Nervous Activity”, Bulletin of Mathematical Biophysics, vol. 5, pp. 115-133, 1943. [15] M.L. Minsky, “Steps Toward Artificial Intelligence”, Proceedings of the Institute of Radio Engineers, vol. 49, no. 1, pp. 8-30, January 1961. [16] J.J. Hopfield, “Neural Networks and Physical Systems with Emergent Collective Computational Abilities”, Proceedings of the National Academy of Sciences, USA, vol. 79, no. 8, pp. 2554-2558, 1982. [17] D.O. Hebb, “The Organization of Behavior: A Neuropsychological Theory”, Wiley, New York, 1949. [18] D.E. Rumelhart and J.L. McClelland, “Parallel Distributed Processing: Explorations in the Microstructure of Cognition Vol. 1”, MIT Press, Cambridge, MA, 1986. [19] G.A. Carpenter and S. Grossberg, “ART 3: Hierarchical Search Using Chemical Transmitters in Self-Organizing Pattern Recognition Architectures”, Neural Networks, vol. 3, pp.129-152, 1990. [20] S. Haykin, “Neural Networks: A Comprehensive Foundation”, Prentice Hall, Upper Saddle River, NJ, 1999.

Page 243:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 243 of 249

09/28/10

[21] “Pulsed Neural Networks”, ed. W. Maass and C.M. Bishop, MIT Press, Cambridge, MA, 1999. [22] Gomez, J., “Co-Evolving Recurrent Neurons Learn Deep Memory POMDPs”, GECCO ’05, June 25-29, 2005, Washington, DC, 2005. [23] T. Kohonen, “Self-Organizing Maps”, Springer-Verlag, Berlin, 2

nd. Edition, 1997.

[24] D.E. Rumelhart, G.E. Hinton, and R.J. Williams, “Learning Internal Representations by Error Propagation,” Parallel Distributed Processing: Explorations in The Microstructure of Cognition, D.E. Rumelhart and J.L. McClelland, eds., vol. 1, pp. 318-362, Cambridge, Mass.: MIT Press, 1986. [25] M. Kukar and I. Kononenko, “Cost-Sensitive Learning with Neural Network,” Proc. 13

th European Conf.

Artificial Intelligence, pp. 445-449, 1998. [26] S. Lawrence, I. Burns, A. Back, A.C. Tsoi, and C.L. Giles,” Neural Network Classification and Prior Class Probabilities,” Lecture Notes in Computer Science 1524, G.B. Orr and K.-R. Muller, eds., pp. 299-313, Berlin: Spring, 1998. [27] Z.-H. Zhou and X.-Y. Liu, “Training Cost-Sensitive Neural Networks with Methods Addressing the Class Imbalance Problem,” IEEE Transactions on Knowledge and Data Engineering, vol. 18, no. 1, pp. 63-77, January 2006. [28] J. Sum, C.-S. Leung, G.H. Young, and K. Wing-Kay, “On the Kalman filtering method in neural network training”, IEEE Transactions on Neural Networks, vol. 10, no. 1, pp. 161-166, Jan. 1999. [29] G. Lera and M. Pinzolas, “Neighborhood based Levenberg-Marquardt algorithm for neural network training”, IEEE Transactions on Neural Networks, vol. 13, no. 5, pp. 1200-1203, Sep 2002. [30] J. Go, G. Han, H. Kim, and C. Lee, “Multigradient: A New Neural Network Learning Algorithm for Pattern Classification”, IEEE Transactions on Geoscience and Remote Sensing, vol. 39, no. 5, pp. 986-993, May 2001. [31] J.F.G. de Freitas et al, “Sequential Monte Carlo Methods to Train Neural Network Models”, Neural Computation, vol. 12, 2000, pp. 955-993. [32] A. Vieira, “An Iterative Artificial Neural Network for High Dimensional Data Analysis”, Springer Verlag, Berlin, 2005. [33] J.C. Bongard, “Incremental Approaches to the Combined Evolution of a Robot’s Body and Brain”, University of Zurich, dissertation, 2003. [34] Peter J. Bentley, “Fractal Proteins”, Genetic Programming and Evolvable Machines Journal, Kluwer Academic Publishers, London, v5, pp. 71-101. [35] J.M. Benyus, “Biomimicry: Innovation Inspired by Nature”, Harper Perennial, London, 2002. [36] S. A. Brueckner et al., editors, “Engineering Self-Organising Systems – Methodologies and Applications”, Lecture Notes in Artificial Intelligence, LNAI 3464, Springer, Heidelberg, 2005. [37] S. Kumar and P.J. Bentley, “On Growth, Form and Computers”, Elsevier Academic Press, Amsterdam, 2003. [38] A. Lindenmayer, “Mathematical models for cellular interaction in development, parts I and II”, Journal of Theoretical Biology, vol. 18, pp. 280-315. [39] Hunt, J.E., & Cooke, D.E., “Learning Using an Artificial Immune System”, Journal of Network and Computer Applications, vol. 19, no. 2, pp. 189-212, 1996. [40] R. Calabretta, et al, “Duplication of Modules Facilitates the Evolution of Functional Specialization”, Artificial Life, vol. 6, pp. 69-84, 2000.

Page 244:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 244 of 249

09/28/10

[41] Sung Young Jung, “A Topographical Method for the Development of Neural Networks for Artificial Brain Evolution”, Artificial Life, vol. 11, pp. 293-316, 2005. [42] A. Horzyk and R. Tadeusiewicz, “Self-Optimizing Neural Networks”, Lecture Notes in Computer Science, vol. 3173, pp. 150-155, Springer Verlag, Berlin, 2004. [43] A. Horzyk and R. Tadeusiewicz, “A new Extension of Self-Optimizing Neural Networks for Topology Estimation”, ICANN 2005, pp. 415-420, Springer Verlag, Berlin, 2005. [44] H. Kitano, “Neurogenetic learning: an integrated method of designing and training neural networks using genetic algorithms”, Physica D, pp. 225-238, Aug. 1, 1994. [45] H. Kitano, “Designing Neural Networks Using Genetic Algorithms with Graph Generation System”, Complex Systems, vol. 4, pp. 461-476, 1990. [46] F. Del Frate et al, “Wheat cycle monitoring using radar data and a neural network trained by a model”, IEEE Transactions on Geoscience and Remote Sensing, vol. 42, no. 1, pp.35-44, Jan. 2004. [47] P. Cipollini, G. Corsini, M. Diani, and R. Grasso, “Retrieval of Sea Water Optically Active Parameters from Hyperspectral Data by Means of Generalized Radial Basis Function Neural Networks”, IEEE Transactions on Geoscience and Remote Sensing, vol. 39, no. 7, pp. 1508-1524, July 2001. [48] R. Lepage, R.G. Rouhana, B. St.-Onge, R. Noumeir, and R. Desjardins, “Cellular Neural Network for Automated Detection of Geological Lineaments on Radarsat Images”, IEEE Transactions on Geoscience and Remote Sensing, vol. 38, no. 3, pp. 1224-1233, May 2000. [49] V.M. Krasnopolsky and H. Schiller, “Some Neural Network Applications in Environmental Sciences. Part I: Forward and Inverse Problems in Geophysical Remote Measurements”, Neural Networks, vol. 16, no. 3-4, pp. 321-334, April-May 2003. [50] A. Niang, L. Gross, S. Thiria, F. Badran, and C. Moulin, “Automatic Neural Classification of Ocean Colour Reflectance Spectra at the Top of the Atmosphere”, Remote Sensing of Environment, vol. 86, no. 2, pp. 257-271, 30 July 2003. [51] Z. Liu, A. Liu, C. Wang, and Z. Niu, “Evolving Neural Network Using Real Coded Genetic Algorithm (GA) for Multispectral Image Classification”, Future Generation of Computer Systems, vol. 20, no. 7, pp. 1119-1129, 1 October 2004. [52] H.-M. Chi and O.K. Ersoy, “A Statistical Self-Organizing Learning System for Remote Sensing Classification”, IEEE Transactions on Geoscience and Remote Sensing, vol. 43, no. 8, pp. 1890-1900, Aug. 2005. [53] Y.-H. Pao and Y. Takefuji, “Functional-Link Net Computing: Theory, System Architecture, and Functionalities”, Computer, pp. 76-79, May 1992. [54] E. Erwin, K. Obermayer, and K. Schulten, “Self-organizing maps: ordering, convergence properties and energy functions”, Biological Cybernetics, vol. 67, pp. 47-55, 1992. [55] T. Kohenen, “Speech recognition based on topology preserving neural maps”, Neural Computing Architectures, ed. I. Aleksander, pp. 26-40, MIT Press, Cambridge, 1989. [56] F. Favata and R. Walker, “A Study of the Application of Kohonen-type Neural Networks to the Traveling Salesman Problem”, Biological. Cybernetics, vol. 64, pp. 463-468, 1991. [57] D. Alahakoon, S.K. Halgamuge, and B. Srinivasan, “Dynamic Self-Organizing Maps with Controlled Growth for Knowledge Discovery”, IEEE Trans. On Neural Networks, vol. 11, no. 3, pp. 601-614, May 2000. [58] K.H. Low, W.K. Leow, and M.H. Ang, “An Ensemble of Cooperative Extended Kohonen Maps for Complex Robot Motion Tasks”,Neural Computation, vol. 17, pp. 1411-1445, June 2005.

Page 245:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 245 of 249

09/28/10

[59] T.M. Martinetz and K.J. Schulten, “A ‘neural-gas’ network learns topologies”, Artificial Neural Networks, ed. Kohonen, T. Makisara, K., Simula, O., and Kangas, J., pp. 397-402, North-Holland, Amsterdam, 1991. [60] T. Villman, et al, “Topology Preservation in Self-Organizing Feature Maps: Exact Definition and Measurement”, IEEE Trans. On Neural Networks, vol. 8, no. 2, pp. 256-266, March 1997. [61] A. Neme and P. Miramontes, “Statistical Properties of Lattices Affect Topographic Error in Self-organizing Maps”, Springer-Verlag, Berlin, 2005. [62] S. Strogatz, “Exploring Complex Networks”, Nature, vol. 410, pp. 268-276, 2001. [63] D. Watts and S. Strogatz, “Collective dynamics of ‘small-world’ networks”, Nature, vol. 393, pp. 440-442, 1998. [64] P. Somervuo, Kohonen, T., “Self-Organizing Maps and Learning Vector Quantization for Feature Sequences”, Neural Processing Letters, vol. 10, pp. 151-159, 1999. [65] A. Rauber, D. Merkl, and M. Dittenbach, “ The Growing Hierarchical Self-Organizing Map: Exploratory Analysis of High-Dimensional Data”, IEEE Trans. On Neural Network, vol. 13, no. 6, pp. 1331-1341, November 2002. [66] B. Fritzke, “Growing Grid – A Self-Organizing Network with Constant Range and Adaptation Strengths”, Neural Processing Letters, vol. 2, no. 5, pp. 1-5, 1995. [67] B. Fritzke, “Growing Neural Gas Network Learns Topologies”, Advances in Neural Information Processing Systems 7, ed. G. Tesauro, MIT Press, Cambridge MA, 1995. [68] S. Marsland, J. Shapiro, U. Nehmzow, “A self-organising network that grows when required”, Neural Networks, Vol. 15, 2002, pp. 1041-1058. [69] A. Kurz, “Constructing Maps for Mobile Robot Navigation Based on Ultrasonic Range Data”, IEEE Trans. On Systems, Man and Cybernetics – Part B: Cybernetics, vol. 26, no. 2, pp. 233-242, 1996. [70] C.M. Bishop, M. Svensen, and C.K.I. Williams, “GTM: A Principled Alternative to the Self-Organizing Map”, Advances in Neural Information Processing Systems 9, MIT Press, 1997. [71] T. Furukawa, “SOM of SOMs: Self-organizing Map Which Maps a Group of Self-organizing Maps”, Springer-Verlag, Berlin, 2005. [72] T. Kohonen, “The Hypermap Architecture”, Artificial Neural Networks, ed. K.Mkisara, Ol. Simula, and J. Kangas, pp. 1357-1360, 1991. [73] G. Vauche, N. Mozayyani, V. Alanou, and J.F. Dreyfus, “A Spatio-Temporal Data-Coding Applied to Kohonen Maps”, Proc. of the International Conference on Artificial Neural Networks, 2:75-79, 1995. [74] P. Campoy and C.J. Vicente, “Residual Activity in the Neurons Allows SOMs to Learn Temporal Order”, Springer-Verlag, Berlin, 2005. [75] C.H. Chen and B. Shrestha, “Classification of Multi-Sensor Remote Sensing Images Using Self-Organizing Feature Maps and Radial Basis Function Networks”, IGARSS 2000, vol. 2, 24-28, pp. 711-713, July 2000. [76] B. Solaiman and M.C. Mouchot, “A Comparative Study of Conventional and Neural Network Classification of Multispectral Data”, IGARSS ’94, vo. 3, 8-12, pp. 1413-1415, August 1994. [77] C.I. Christodoulou, S.C. Michaelides, and C.S. Pattichis, “Multifeature Texture Analysis for the Classification of Clouds in Satellite Imagery”, IEEE Trans. on Geoscience and Remote Sensing, vol. 41, no. 11, pp. 2662-2668, November 2003.

Page 246:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 246 of 249

09/28/10

[78] E.J. Kwiatkowska, “Cloud Cover Classification using SOM”, dissertation, University of Bradford, UK, 1997. [79] A.P. Engelbrecht, “Computational Intelligence: An Introduction”, Wiley, New York, 2002. [80] D.B. Fogel, “An Introduction to Simulated Evolutionary Optimization”, IEEE Trans. On Neural Networks, vol. 5, no. 1, pp. 3-14, January 1994. [81] J. Kennedy and R. Eberhart, “Particle Swarm Optimization”, Proc. IEEE INt. Conference on Neural Networks, vol. IV, Perth, Australia, 1995, pp. 1942-1948. [82] P. Bentley, “Evolving Fractal Proteins”, Department of Computer Science, University College London, 2003. [83] Amor, H.B., Rettinger, A., “Intelligent Exploration for Genetic Algorithms – Using Self-Organizing Maps in Evolutionary Computation”, Proc. GECCO ’05, 2005. [84] D.B. Fogel (ed.), “Evolutionary Computation: The Fossil Record”, IEEE Press, 1998. [85] J.H. Holland, “Adaptation in Natural and Artificial Systems”, 2

nd Edition, MIT Press, Cambridge, MA, 1992.

[86] W. Atmar, “Notes on the Simulation of Evolution”, IEEE Trans. On Neural Networks, vol. 5, no. 1, pp. 130-147, January 1994. [87] C. A. Coello, “A Short Tutorial on Evolutionary Multiobjective Optimization”, eds. E. Zitzler et al, Lecture Notes in Computer Science 1993, pp. 21-40, Springer-Verlag, Berlin, 2001. [88] N.R. Harvey et al, “Comparison of GENIE and Conventional Supervised Classifiers for Multispectral Image Feature Extraction”, IEEE Trans. on Geoscience and Remote Sensing, vol. 40, no. 2, pp. 393-404, February 2002. [89] B.C.K. Tso and P.M. Mather, “Classification of Multisource Remote Sensing Imagery Using a Genetic Algorithm and Markov Random Fields”, IEEE Trans. on Geoscience and Remote Sensing, vol. 37, no. 3, pp. 1255-1266, May 1990. [90] R.C. Eberhart and Y. Shi, ed., Special Issue on Particle Swarm Optimization, IEEE Trans. on Evolutionary Computation, vol. 8, no.3, June 2004. [91] A.S. Fraser, “Simulation of Genetic Systems”, J. of Theor. Biology, vol. 2, pp. 329-346, 1962. [92] H.J. Bremermann, “Optimization Through Evolution and Recombination,” in Self-Organizing Systems, eds. M.C. Yovits, G.T. Jacobi, and G.D. Goldstone, Spartan Books, Washington, D.C., pp. 93-106, 1962. [93] Y. Shi and R.C. Eberhart, “Empirical Study of Particle Swarm Optimization”, Proc. of the 1999 Congress on Evolutionary Computation, Washington, DC, vol. 3, pp. 1945-1950, July 1999. [94] C.A. Coello, G.T. Pulido, and M.S. Lechuga, “Handling Multiple Objectives with Particle Swarm Optimization”, IEEE Trans. on Evolutionary Computation, vol. 8, no. 3, pp. 256-279, June 2004. [95] W.H. Slade, H.W. Ressom, M.T. Mujavi, and R.L. Miller, “Inversion of Ocean Color Observationis Using Particle Swarm Optimization”, IEEE Trans. on Geoscience and Remote Sensing, vol. 42, no. 9, pp. 1915-1923, September 2004. [96] C.-F. Juang, “A Hybrid of Genetic Algorithm and Particle Swarm Optimization for Recurrent Network Design”, IEEE. Trans. on Systems, Man, and Cybernetics – Part B: Cybernetics, vol. 34, no. 2, pp. 997-1006, April 2004. [97] K. P. White, “Functional Genomics and the Study of Development, Variation, and Evolution”, Nature Reviews Genetics, July 2001, vol. 2, pp. 528-537. [98] T. D. Pollard and W.C. Earnshaw, “Cell Biology”, Saunders, Philadelphia, 2002.

Page 247:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 247 of 249

09/28/10

[99] S.B. Carroll, “Endless forms: the evolution of gene regulation and morphological diversity”, Cell, vol. 101, pp. 577-580, 2000. [100] H. Kitano, “A Simple Model of Neurogenesis and Cell Differentiation based on Evolutionary Large-Scale Chaos”, Artificial Life, vol. 2, no. 1, 1995. [101] D. Madina et al, “Cellular Evolution in a 3D Lattice Artificial Chemistry, 2004. [102] S. Kauffman, The Origins of Order, Oxford University Press, 1993. [103] T. Reil, “Dynamics of Gene Expression in an Artificial Genome – Implications for Biological and Artificial Ontogeny”, Proceedings of the 5

th European Conference on Advances in Artificial Life, Lecture Notes in

Computer Science, vol. 1674, 1999, pp. 457-466, Springer-Verlag, London. [104] N. Geard, “Modeling Gene Regulatory Networks: Systems Biology to Complex Systems”, ACCS Draft Technical Report, May 6, 2004. [105] Peter J. Bentley, “Evolving beyond perfection: an investigation of the effects of long-term evolution on fractal gene regulatory networks”, BioSystems, vol. 76, 2004, pp. 291-301. [106] L. N. de Castro and F.J. Von Zuben, “aiNet: An Artificial Immune Network for Data Analysis”, chapter in Data Mining: A Heuristic Approach, H. A. Abbass ed., Idea Group Publishing, USA, March 2001. [107] Y. Zhong, L. Zhang, B. Huang, and P. Li, “An Unsupervised Artificial Immune Classifier for Multi/Hyperspectral Remote Sensing Imagery”, IEEE Trans. On Geoscience and Remote Sensing, vol. 44, no. 2, February 2006. [108] L.A.N. Amaral, et al, “Emergence of complex dynamics in a simple model of signaling networks”, Proceedings of the National Academy of Sciences, vol. 101, no. 44, 2004. [109] J.C. Astor and C. Adami, “Development and Evolution of Neural Networks in an Artificial Chemistry”, arXiv:adap-org/9807003, Jul. 17, 1998. [110] R. Kohavi, “A Study of Cross-Validation and Boostrap for Accuracy Estimation and Model Selection”, International Joint Conference on Artificial Intelligence (IJCAI), 1985, Los Angeles, California, USA. [111] NASA MODIS website, http://www.modis.gsfc.nasa.gov, acquired July 30, 2009. [112] H.R. Gordon and M. Wang, “Retrieval of water-leaving radiance and aerosol optical thickness over the oceans with SeaWiFS: a preliminary algorithm”, Applied Optics, vol. 33, no. 3, January 1994. [113] J. D. Owens, et al, “A Survey of General-Purpose Computation on Graphics Hardware”, EUROGRAPHICS 2005. [114] Qizhi Yu, Chongcheng Chen, Zhigeng Pan, Parallel Genetic Algorithms on Programmable Graphics Hardware, Lecture Notes in Computer Science, vol. 3612, pp. 1051 – 1059, Jul 2005. [115] K.-L. Fok, T.-T Wong, and M.-L. Wong, “Evolutionary Computing on Consumer-Level Graphics Hardware”, to appear in IEEE Intelligent Systems, 2005. [116] Conversation with Ewa Kwiatkowska, August 19, 2009. [117] A.P. Bradley, “The Use Of The Area Under The ROC Curve In The Evaluation of Machine Learning Algorithms”, Pattern Recognition, vol. 30, no. 7, pp. 1145-1159, 1997. [118] P.A. Flasch, “Tutorial on The Many Faces of ROC Analysis in Machine Learning”, ICML 2004, July 4-8, 2004, Banff, Alberta, Canada.

Page 248:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 248 of 249

09/28/10

[119] J. Lehman and K.O. Stanley, “Revising the Evolutionary Computation Abstraction: Minimal Criteria Novelty Search”, Proceedings of GECCO 2010, July 7-11, Portland, Oregon USA. [120] B. Scholkpf, C.J.C. Burges, and A.J. Smola, “Advances in Kernel Methods: Support Vector Learning,” MIT Press, Cambridge MA, 1998. [121] A.N. Srivastava, N.C. Oza, and J. Stroeve, “Virtual Sensors: Using Data Mining Techniques to Efficiently Estimate Remote Sensing Spectra,” IEEE Trans. On Geoscience and Remote Sensing, vol. 43, no. 3, pp. 590-600, March 2005. [122] E.J. Kwiatkowska, and J.S. Fargion, “Application of Machine-Learning Techniques Toward the Creation of a Consistent and Calibrated Global Chlorophyll Concentration Baseline Dataset”, IEEE Trans. On Geoscience and Remote Sensing, vol. 41, no. 12, pp. 2844-2860. [123] K.Y. Yip, et al, “Improved Reconstruction of In Silico Gene Regulatory Networks by Integrating Knockout and Perturbation Data”, PLoS ONE , vol. 5, no. 1, January 2010. [124] S. Pinkert, et al, “Protein Interaction Networks- More than Mere Modules”, PloS Computation Biology, vol. 6, no. 1, January 2010. [125] K.O. Stanley and R. Miikkulainen, “Competitive Coevolution Trhough Evolutionary Complexification”, Journal of Artificial Intelligence Research, vol. 21, pp. 63-100, 2004. [126] K.O. Stanley, et al “Real-time Neuroevolution in the NERO Video Game”, IEEE Transactions on Evolutionary Computation, vol. 9, no. 6, pp. 653-668, December 2005. [127] A.-L. Barabasi and E. Bonabeau, “Scale-Free Networks”, Scientific American, pp. 60-69, May 2003. [128] Watts, D. and Strogatz, S., “Collective dynamics of ‘small-world’ networks”, Nature, vol. 393, pp. 440-442, 1998. [129] J.D. Gibson, Principles of Digital and Analog Communications, Second Edition, Prentice Hall, 1993.

Page 249:  · Taylor 4 of 249 09/28/10 ABSTRACT This research develops and investigates an algorithm for the self-organized development of a classification network. This idea for this classific

Taylor 249 of 249

09/28/10