23
©2021 the authors. This is a preprint version. The final version of this paper is available at https://dl.acm.org/doi/10.1145/3449726.3463131 (DOI:10.1145/3449726.3463131). Trustworthy AI for Process Automation on a Chylla-Haase Polymerization Reactor Daniel Hein ? and Daniel Labisch ?? ? Siemens AG, Technology, Munich, Germany (e-mail: [email protected]). ?? Siemens AG, Digital Industries, Karlsruhe, Germany (e-mail: [email protected]). August 31, 2021 Abstract In this paper, genetic programming reinforcement learning (GPRL) is utilized to generate human-interpretable control policies for a Chylla- Haase polymerization reactor. Such continuously stirred tank reactors (CSTRs) with jacket cooling are widely used in the chemical industry, in the production of fine chemicals, pigments, polymers, and medi- cal products. Despite appearing rather simple, controlling CSTRs in real-world applications is quite a challenging problem to tackle. GPRL utilizes already existing data from the reactor and generates fully au- tomatically a set of optimized simplistic control strategies, so-called policies, the domain expert can choose from. Note that these policies are white-box models of low complexity, which makes them easy to val- idate and implement in the target control system, e.g., SIMATIC PCS 7. However, despite its low complexity the automatically-generated policy yields a high performance in terms of reactor temperature con- trol deviation, which we empirically evaluate on the original reactor template. 1 Introduction Utilizing AI-based methods in real-world industrial applications requires a certain amount of trust in the automatically-generated solution. One way to 1 arXiv:2108.13381v1 [cs.AI] 30 Aug 2021

Trustworthy AI for Process Automation on a Chylla-Haase

  • Upload
    others

  • View
    4

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Trustworthy AI for Process Automation on a Chylla-Haase

©2021 the authors. This is a preprint version. The final version of thispaper is available at https://dl.acm.org/doi/10.1145/3449726.3463131(DOI:10.1145/3449726.3463131).

Trustworthy AI for Process Automation on aChylla-Haase Polymerization Reactor

Daniel Hein? and Daniel Labisch??

?Siemens AG, Technology, Munich, Germany (e-mail: [email protected]).??Siemens AG, Digital Industries, Karlsruhe, Germany (e-mail:

[email protected]).

August 31, 2021

Abstract

In this paper, genetic programming reinforcement learning (GPRL) isutilized to generate human-interpretable control policies for a Chylla-Haase polymerization reactor. Such continuously stirred tank reactors(CSTRs) with jacket cooling are widely used in the chemical industry,in the production of fine chemicals, pigments, polymers, and medi-cal products. Despite appearing rather simple, controlling CSTRs inreal-world applications is quite a challenging problem to tackle. GPRLutilizes already existing data from the reactor and generates fully au-tomatically a set of optimized simplistic control strategies, so-calledpolicies, the domain expert can choose from. Note that these policiesare white-box models of low complexity, which makes them easy to val-idate and implement in the target control system, e.g., SIMATIC PCS7. However, despite its low complexity the automatically-generatedpolicy yields a high performance in terms of reactor temperature con-trol deviation, which we empirically evaluate on the original reactortemplate.

1 Introduction

Utilizing AI-based methods in real-world industrial applications requires acertain amount of trust in the automatically-generated solution. One way to

1

arX

iv:2

108.

1338

1v1

[cs

.AI]

30

Aug

202

1

Page 2: Trustworthy AI for Process Automation on a Chylla-Haase

Figure 1: SIMATIC PCS 7 Chylla-Haase polymerization reactor template [1]

build trust, is to have machine learning (ML) methods present the reasons fortheir decisions in a human-understandable way [11]. On the one hand, onecan attempt to interpret any useful system, even if it inherently is a black-box system, e.g., by explanation techniques for classifiers [24] or visualizationmethods for neural networks (NNs) [4]. On the other hand, interpretabilitycan be evaluated via a quantifiable proxy, where at first, a model class, e.g.,algebraic equations, is claimed to be interpretable and then algorithms aredeveloped in order to optimize within this class [6, 23, 27].

For the latter, recently a new method called genetic programming rein-forcement learning (GPRL) has been proposed [17] and evaluated on chal-lenging benchmark problems, like a hardware cart-pole demonstrator [18] or areal-world inspired industrial benchmark [16]. In the present paper, we applyGPRL for the first time on a SIMATIC PCS 7 Chylla-Haase polymerizationreactor template (see Fig. 1) [1], to improve the process automation con-troller while keeping the policy solution interpretable and thus trustworthy.The template exemplifies the conventional control strategy using a realisticbut simple simulation model directly built in SIMATIC PCS 7. It aims as atemplate for real applications.

2

Page 3: Trustworthy AI for Process Automation on a Chylla-Haase

If a process plant runs safely and reliably during operation, the operatoralways strives to maximize the economic yield. Existing degrees of freedomare used to design the operation as optimally as possible without violatingexisting limitations and boundary conditions. This can be done throughmanual intervention by the operator based on experience, or through the useof various optimization methods.

Optimization methods can basically be divided into two classes, simulation-based and data-based tools, although mixed forms are also possible. Strictlyspeaking, the optimization works with a simulation model in the first caseand with a data-based model in the second case. An optimization basedon a general PROcess Modelling System (gPROMS) is an example for thefirst case. Most data-based approaches create black-box results like neuralnetworks. These models can show impressive results, but most of them aretoo complex to be interpreted or validated.

Hence, available optimization methods for process industries are eitherinterpretable expert-generated process controllers or non-interpretable AI-generated process controllers. Herein, we show how the ML method geneticprogramming can be utilized together with supervised AI-based regressionmodels to automatically generate interpretable process controllers. By doingso, the manual effort to generate interpretable process controllers can bereduced significantly. Furthermore, novel unknown control strategies can befound, which improve the quality key performance indicators (KPIs) of thecontroller.

In contrast to simulation-based approaches, this method significantly re-duces the modeling efforts. Manually creating a simulation model is verycostly in general. Even if a model already exists, e.g., from the planningphase, the adaption and parametrization of the model to represent the realbehavior is still very time-consuming. Furthermore, many simulation modelsfrom the planning phase are just static ones which couldn’t be easily extendedto a dynamic version. The data-based approach GPRL does not require allthese efforts.

Furthermore, generating black-box controllers by ML can be prone tooverfit to some process model inaccuracies. Black-box controllers normallyhave hundreds of parameters which are automatically tuned to yield a highperformance with respect to the predicted control quality. However, in manycases model inaccuracies and loopholes can be exploited by the ML processwhich result in controllers that only performing well on the imperfect data-based models but not on the real process. Interpretable controllers by geneticprogramming, formulated using the established controller building blocksare more robust to process model inaccuracies. Furthermore, experts of therespective process are able to identify contradictions and implausibilities in

3

Page 4: Trustworthy AI for Process Automation on a Chylla-Haase

interpretable controllers.The resulting process controllers can either directly replace existing con-

trol structures or can also be used as a superimposed optimization strategy.This flexibility combined with the interpretable controller function enablesthe process owner to balance the conflicting goals of optimizing performanceand guaranteeing safety and reliability of the process.

In Sections 2 and 3, the interface of the reactor template and the GPRLmethod are introduced, respectively. In Section 4, we show how GPRLuses available simulation data from a Chylla-Haase reactor to generate inter-pretable policies. In the evaluation phase of the experiment, we show thatthe newly found policies reduce the control error by up to 37 %.

2 Chylla-Haase polymerization reactor

Continuously stirred tank reactors (CSTRs) with jacket cooling are widelyused in the chemical industry, in the production of fine chemicals, pigments,polymers, and medical products. CSTRs can be run in batch, semi-batch orcontinuous operation. [1]

To perform an accurate control, online optimization and monitoring ofprocesses, precise knowledge of the status of the respective process is requiredat all times. However, only in rare cases it is possible to measure all the statesof the reactor using real sensor equipment. Often, some process variables cannot be determined at all or only with great technical effort.

Simulating the dynamic process model in a soft sensor in parallel to thereal process, allows the automation system to observe all modeled inner statesof the reactor even if they cannot be registered by measurement equipment.The Chylla-Haase reactor, considered in our experiments, contains a uni-versal stirred tank reactor that is employed for the production of diversepolymers with different recipes. Here, an extended Kalman filter serves as asoft sensor for monitoring the chemical reaction.

The values calculated online by the soft sensor can be used in variousways [1]:

• Calculation of the current speed of reaction and the heat released bythe exothermic reaction,

• Calculation of the monomer mass remaining in the reactor, and

• Calculation of heat transfer from reactor to cooling jacket in order todetect the build-up of deposition.

4

Page 5: Trustworthy AI for Process Automation on a Chylla-Haase

A frequently considered control problem statement for Chylla-Haase re-actors, is to maintain reaction temperature of a semi-batch polymerizationreactor at a setpoint in spite of a variety of conditions ranging from rawmaterial impurity, product variety, repeated batching and varying ambienttemperature. Although the reactor configuration and control requirementsappear rather simple, in real-world applications it is quite a challenging prob-lem to tackle. [5]

The rector template (SIMATIC PCS 7 project and documentation) weused in our experiments is available online1. The production of polymer ismaintained by an sequential function chart (SFC) recipe, where the simula-tion of one batch takes between 10 to 40 minutes, depending on the processparameters. The operator screen as user interface of the template is depictedin Fig. 1.

The reactor template can be influenced by two variables either given bythe operator or using a superimposed optimization strategy during the batch:

• Reactor temperature setpoint T , and

• Monomer feed setpoint M .

These variables are called decision variables or actions.The task is to process a certain amount of monomer into polymer, given

a certain reactor setpoint. According to the recipe provided, the reactor isheated up to a certain start temperature. If this temperature is reached, themonomer is added to the stirred tank reactor at a certain monomer feed rateM . As soon as a pre-defined amount of monomer is processed , the batchis finished and the tank gets drained. Note that the reactor temperaturehas to be controlled constantly by heating and cooling, since the reactiontemperature is highly dependent on the amount of polymer inside the tank.

The default method to control the temperature is to use a proportional–integral–derivative (PID) controller. Today PID control is still one of themost common control strategies, since it is very often applied at the lowestlevel of the control hierarchy in industrial systems [3]. In 2002, Desboroughand Miller conducted a survey of more than 11,000 controllers in the refining,chemicals, and pulp and paper industries which revealed that 97 % of regula-tory controllers had a PID structure [10]. However, PID controllers can havesevere drawbacks. They use only limited process information and the designcriterion sometimes yields closed loop systems with poor robustness [2].

Hence, it is expected that an RL-based controller, which is potentiallynon-linear and can utilize all available process variables, can outperform

1https://support.industry.siemens.com/cs/de/de/view/109756215

5

Page 6: Trustworthy AI for Process Automation on a Chylla-Haase

Variable Description Unit Range

a

T Reactor temperature setpoint K 352-365

M Monomer feed setpoint kg/s 0.005-0.015

s

S Intended setpoint K 352-365T Reactor temperature K -M Monomer mass kg -P Polymer mass kg -UA Heat loss coefficient kW/K 0-1Q Reaction heat flow kW -

Table 1: Reactor variables of action a and state s

standard PID regulation on many real-world use cases. The action and statevariables of the reactor template, which are available for an RL controller,are given in Table 1. The action variables in a are the decision variables ofthe reactor template, whereas the state variables in s contain the intendedsetpoint S as well as sensor measurements from the reactor. Note that S isnot a variable modeled by the reactor simulation, but it represents the trueintended temperature setpoint of the operator. The default controller setsT = S, i.e., the operator’s intended setpoint is directly used as input for theunderlying PID regulator.

3 Genetic programming reinforcement learn-

ing

The GPRL approach learns policy representations which are basic algebraicequations of low complexity [17]. Given that GPRL can find rather short(non-complex) equations, it is expected to reveal substantial knowledge aboutunderlying coherencies between available state variables and well-performingcontrol policies for a certain RL problem. By the term complexity of a policy,we refer to its human-interpretable form, e.g., a simple algebraic equationcan be easily understood by a domain expert, hence it is of low complexity,whereas a non-linear system of equations, like a neural network, cannot bevalidated by a human expert, yielding a high complexity.

Note that GPRL is not only optimizing a certain trade-off between pol-icy complexity and performance, but rather it evolves a whole optimizedPareto front from which the domain expert can choose a trustworthy and

6

Page 7: Trustworthy AI for Process Automation on a Chylla-Haase

well-performing solution.

3.1 Population-based reinforcement learning

Generally, reinforcement learning (RL) is distinguished from other computa-tional approaches by its emphasis on an agent learning from direct interactionwith its environment. This approach, which does not rely on exemplary su-pervision or complete models of the environment, is referred to as onlinelearning. However, for many real-world problems online learning is prohib-ited for safety reasons. For example, it is not advisable to deploy an onlineRL agent, who starts by applying an arbitrary initial policy, which is subse-quently improved by exploitation and exploration, on safety-critical domainslike process industries or power plants. For this reason, offline learning is amore suitable approach for applying RL methods on already collected train-ing data and system models, to yield RL policies.

Offline RL is often referred to as batch learning because it is based solelyon a previously generated batch of transition samples from the environment.The batch data set D contains transition tuples of the form (st,at, st+1, rt+1),where the application of action at in state st resulted in a transition to statest+1 and yielded a reward rt+1, where t denotes a discrete time step.

Herein, we use model-based value estimation to estimate the performanceof a policy. In the first step, supervised ML is applied to learn approximatemodels of the underlying environment from transition tuples (st,at, st+1, rt+1)as follows:

g(st,at)← st+1, r(st,at, st+1)← rt+1. (1)

Using models g and r, the value for policy π of each state s in the databatch can be estimated by computing the value function vπ(s). Hence, usingmodel-based value estimation means performing trajectory rollouts on systemmodels for different starting states:

vπ(st) =∞∑k=0

γkr(st+k,at+k, st+k+1), (2)

with st+k+1 = g(st+k,at+k), at+k = π(st+k), (3)

and discount factor 0 ≤ γ ≤ 1.Since we are searching for interpretable solutions, the resulting policies

have to be represented in an explicit form. Policy search yields explicitpolicies which can be enforced to be of an interpretable form. Furthermore,policy search is inherently well-suited for being used with population-basedoptimization techniques, like genetic programming (GP).

7

Page 8: Trustworthy AI for Process Automation on a Chylla-Haase

<latexit sha1_base64="c2+IruHlk400z420Qh0I89XlCrU=">AAAC0nichVFLS8NAEJ7GV1tfVY9egkXwICWRgh4LPvAiVOhL2lI26TYuTZOwSQtt6UG8evOqP0x/iwe/rFHQIt2wmdlvvvl2ZscKXBFGhvGW0paWV1bX0pns+sbm1nZuZ7cW+kNp86rtu75sWCzkrvB4NRKRyxuB5Gxgubxu9c/jeH3EZSh8rxKNA94eMMcTPWGzCNBdKxCdqXFszjq5vFEw1NLnHTNx8pSssp97pxZ1ySebhjQgTh5F8F1iFOJrkkkGBcDaNAUm4QkV5zSjLHKHYHEwGNA+/g5OzQT1cI41Q5Vt4xYXWyJTp0PsK6VogR3fyuGHsB/YE4U5/94wVcpxhWNYC4oZpXgDPKJ7MBZlDhLmdy2LM+OuIurRmepGoL5AIXGf9o/OBSISWF9FdLpUTAcaljqP8AIebBUVxK/8raCrjruwTFmuVLxEkUFPwsavj3owZvPvUOed2knBNArmbTFfKiYDT9M+HdARpnpKJbqmMuqwUckzvdCrVtEm2oP2+EXVUknOHv1a2tMnkwqR7A==</latexit>

<latexit sha1_base64="CoNJRUr+gisPcTdWr2HtD+vGU3I=">AAAC7XichVFNa9tAEH1RkjZx01ZJj70Ym0IOwUjFkB4NbUIvhTTEHxAHs1I29mJ9Ia0DqfC9f6C30GtvvbZ/pf0tOeTtRgmkpnjFat6+mXk7sxNkkSq05/1ZcVbX1p883disPdt6/uKlu73TK9JZHspumEZpPghEISOVyK5WOpKDLJciDiLZD6bvjb9/KfNCpcmJvsrkWSzGibpQodCkRm5jmNJtssthLPQkFFF5PJ+PymGmRqW35xO7Ta/l2VVfBH4FmqjWUer+xRDnSBFihhgSCTRxBIGC3yl8eMjInaEklxMp65eYo8bcGaMkIwTZKf9jnk4rNuHZaBY2O+QtEXfOzDrecB9axYDR5lZJXNDecH+x3Pi/N5RW2VR4RRtQcdMqfiKvMWHEssy4iryvZXmm6UrjAu9sN4r1ZZYxfYYPOh/oyclNraeOAxs5pkZgz5d8gYS2ywrMK98r1G3H57TCWmlVkkpRUC+nNa/Pejhm/9+hLoLe25bvtfzP7WanXQ18A6/RwC6nuo8OPuKIdYT4ip/4hd9O6nxzrp3vd6HOSpXzCo+W8+MWDBWdpg==</latexit>

<latexit sha1_base64="h6y2ZmFOnY2ox2tebxYybKrqYx0=">AAAC0nichVFLS8NAEJ7GV1tfVY9egkXwVBIp6LHgAy9Chb6kLbJJtzV08yDZlj7wIF69edUfpr/Fg1/WVNAi3bCZ2W+++XZmxwqEE0nDeE9pS8srq2vpTHZ9Y3NrO7ezW4v8QWjzqu0LP2xYLOLC8XhVOlLwRhBy5lqC163+WRyvD3kYOb5XkeOAt13W85yuYzMJ6HbUEh1fRvroLpc3CoZa+rxjJk6eklX2cx/Uog75ZNOAXOLkkYQviFGEr0kmGRQAa9MUWAjPUXFOD5RF7gAsDgYD2se/h1MzQT2cY81IZdu4RWCHyNTpEPtSKVpgx7dy+BHsJ/ZEYb1/b5gq5bjCMawFxYxSvAYu6R6MRZluwpzVsjgz7kpSl05VNw7qCxQS92n/6JwjEgLrq4hOF4rZg4alzkO8gAdbRQXxK88UdNVxB5Ypy5WKlygy6IWw8eujHozZ/DvUead2XDCNgnlTzJeKycDTtE8HdISpnlCJrqiMOmxU8kKv9KZVtIn2qD19U7VUkrNHv5b2/AW255Jl</latexit>

<latexit sha1_base64="bU2aoiD1hG+bww6/5l2U5Ni1Ny4=">AAAC0XichVFLS8NAEJ7GV1tfVY9egkXwFBIR9FjQihehon1AWyRJtzU0LzbbYi0F8erNq/4x/S0e/HZNBS3SDZuZ/eabb2d2nNj3EmGa7xltYXFpeSWby6+urW9sFra2a0k04C6rupEf8YZjJ8z3QlYVnvBZI+bMDhyf1Z3+qYzXh4wnXhTeiFHM2oHdC72u59oCUKMl2L0Ylye3haJpmGrps46VOkVKVyUqfFCLOhSRSwMKiFFIAr5PNiX4mmSRSTGwNo2BcXieijOaUB65A7AYGDbQPv49nJopGuIsNROV7eIWH5sjU6d97HOl6IAtb2XwE9hP7AeF9f69YayUZYUjWAeKOaV4CVzQHRjzMoOUOa1lfqbsSlCXTlQ3HuqLFSL7dH90zhDhwPoqolNZMXvQcNR5iBcIYauoQL7yVEFXHXdgbWWZUglTRRt6HFa+PurBmK2/Q511aoeGZRrW1VGxZKQDz9Iu7dEBpnpMJbqgCuqQ03yhV3rTrrWR9qg9fVO1TJqzQ7+W9vwFD1uSIQ==</latexit> <latexit sha1_base64="e42w+peHwDYlJrr1/t4WluWSwoQ=">AAAC0XichVFLS8NAEJ7GV1tfVY9egkXwFBIR9FjwgR6EivYBbZEk3dbQvNhsi7UUxKs3r/rH9Ld48Ns1FbRIN2xm9ptvvp3ZcWLfS4Rpvme0ufmFxaVsLr+8srq2XtjYrCZRn7us4kZ+xOuOnTDfC1lFeMJn9ZgzO3B8VnN6xzJeGzCeeFF4I4YxawV2N/Q6nmsLQPWmYPdidDG+LRRNw1RLn3as1ClSuspR4YOa1KaIXOpTQIxCEvB9sinB1yCLTIqBtWgEjMPzVJzRmPLI7YPFwLCB9vDv4tRI0RBnqZmobBe3+NgcmTrtYp8pRQdseSuDn8B+Yj8orPvvDSOlLCscwjpQzCnFS+CC7sCYlRmkzEktszNlV4I6dKS68VBfrBDZp/ujc4IIB9ZTEZ1OFbMLDUedB3iBELaCCuQrTxR01XEb1laWKZUwVbShx2Hl66MejNn6O9Rpp7pvWKZhXR0US0Y68Cxt0w7tYaqHVKJzKqMOOc0XeqU37Vobao/a0zdVy6Q5W/Rrac9fGPeSJQ==</latexit>

<latexit sha1_base64="Gi37WHsFfe/OxrEzyk0dZSxAjxM=">AAAC0nichVFLS8NAEJ7GV1tfVY9egkXwICUpBT0WfOBFqNCXtKUk6TYuzYtNWmhLD+LVm1f9YfpbPPhlTQUt0g2bmf3mm29ndszA4WGkae8pZWV1bX0jnclubm3v7Ob29uuhPxQWq1m+44umaYTM4R6rRTxyWDMQzHBNhzXMwUUcb4yYCLnvVaNxwDquYXu8zy0jAnTfDnh3qp0WZ91cXitocqmLjp44eUpWxc99UJt65JNFQ3KJkUcRfIcMCvG1SCeNAmAdmgIT8LiMM5pRFrlDsBgYBtAB/jZOrQT1cI41Q5lt4RYHWyBTpWPsa6logh3fyuCHsJ/YE4nZ/94wlcpxhWNYE4oZqXgLPKIHMJZluglzXsvyzLiriPp0LrvhqC+QSNyn9aNziYgANpARla4k04aGKc8jvIAHW0MF8SvPFVTZcQ/WkJZJFS9RNKAnYOPXRz0Ys/53qItOvVjQtYJ+V8qXS8nA03RIR3SCqZ5RmW6ogjosVPJCr/SmVJWJ8qg8fVOVVJJzQL+W8vwFlXGR7Q==</latexit>

<latexit sha1_base64="ArHKTagTro+F+vcryTkSzxddgJw=">AAAC0nichVFLS8NAEJ7GV1tfVY9egkXwICXRgh4LPvAiVOhL2lI26TYuzYskLbSlB/Hqzav+MP0tHvyypoIW6YbNzH7zzbczO4ZvizDStPeUsrS8srqWzmTXNza3tnM7u7XQGwQmr5qe7QUNg4XcFi6vRiKyecMPOHMMm9eN/kUcrw95EArPrUQjn7cdZrmiJ0wWAbpv+aIz0Y5Pp51cXitocqnzjp44eUpW2ct9UIu65JFJA3KIk0sRfJsYhfiapJNGPrA2TYAF8ISMc5pSFrkDsDgYDGgffwunZoK6OMeaocw2cYuNHSBTpUPsa6logB3fyuGHsJ/YY4lZ/94wkcpxhSNYA4oZqXgLPKIHMBZlOglzVsvizLiriHp0LrsRqM+XSNyn+aNziUgArC8jKl1JpgUNQ56HeAEXtooK4leeKaiy4y4sk5ZLFTdRZNALYOPXRz0Ys/53qPNO7aSgawX9rpgvFZOBp2mfDugIUz2jEt1QGXWYqOSFXulNqShj5VF5+qYqqSRnj34t5fkLl9iR7g==</latexit>

<latexit sha1_base64="sj5HG0SxNAxALhBOubmqE6P2PU0=">AAAC0nichVFLS8NAEJ7GV1tfVY9egkXwICWRgh4LPvCiVOhL2lKSdBuX5kWyLbSlB/Hqzav+MP0tHvyypoIW6YbNzH7zzbczO2bg8Eho2ntKWVpeWV1LZ7LrG5tb27md3VrkD0KLVS3f8cOGaUTM4R6rCi4c1ghCZrimw+pm/zyO14csjLjvVcQoYG3XsD3e45YhAN23At6ZaMe3004urxU0udR5R0+cPCWr7Oc+qEVd8smiAbnEyCMB3yGDInxN0kmjAFibJsBCeFzGGU0pi9wBWAwMA2gffxunZoJ6OMeakcy2cIuDHSJTpUPsK6logh3fyuBHsJ/YY4nZ/94wkcpxhSNYE4oZqXgDXNADGIsy3YQ5q2VxZtyVoB6dyW446gskEvdp/ehcIBIC68uISpeSaUPDlOchXsCDraKC+JVnCqrsuAtrSMukipcoGtALYePXRz0Ys/53qPNO7aSgawX9rpgvFZOBp2mfDugIUz2lEl1TGXVYqOSFXulNqShj5VF5+qYqqSRnj34t5fkL2LWSCQ==</latexit>

<latexit sha1_base64="Fbb/nlneYPY4rZpxPd6eWhxhc54=">AAAC7XichVFNSyNBEH0ZdY3uupt1j15CgrCHJcyIsB4F18WLEMWYgJHQM7axyXwx0xF0yN0/sDfx6s3r+lf0t3jwdTsKuyL20FOvX1W9rury01Dl2nXvKs7U9MyH2erc/MdPC5+/1L4u7ufJOAtkJ0jCJOv5IpehimVHKx3KXppJEfmh7PqjDePvnsosV0m8p89SeRiJYayOVSA0qUGt0U/oNtlFPxL6JBBhsTuZDIp+qgaF+2OFuNZ0W65d9dfAK0ET5WontXv0cYQEAcaIIBFDE4cQyPkdwIOLlNwhCnIZkbJ+iQnmmTtmlGSEIDvif8jTQcnGPBvN3GYHvCXkzphZxzL3b6voM9rcKolz2gfuc8sN37yhsMqmwjNan4pzVnGbvMYJI97LjMrI51rezzRdaRxjzXajWF9qGdNn8KLzi56M3Mh66ti0kUNq+PZ8yheIaTuswLzys0LddnxEK6yVViUuFQX1Mlrz+qyHY/b+H+prsL/S8tyWt7PaXF8tB17FEhr4zqn+xDq20GYdAS5wg7+4dRLnj3PpXD2FOpUy5xv+Wc71Iw59nac=</latexit>

<latexit sha1_base64="BQJxzXifRem5IW/A8iQeJgOXw6w=">AAAC7XichVFNS+RAEH1m/f4c9ehlcBA8yJDowO5RcF28CCqOCo4MndiOzeSLpEfQMPf9A3tbvHrzqn9Ff4sHX7dRUBE7dOr1q6rXVV1+Gqpcu+7DgPNjcGh4ZHRsfGJyanqmMjt3kCe9LJDNIAmT7MgXuQxVLJta6VAepZkUkR/KQ7+7YfyHFzLLVRLv68tUnkSiE6szFQhNql1ZbCV0m+yiFQl9Hoiw2Ov320UrVe3CXVkjrtTcumtX9TPwSlBDuXaSyiNaOEWCAD1EkIihiUMI5PyO4cFFSu4EBbmMSFm/RB/jzO0xSjJCkO3y3+HpuGRjno1mbrMD3hJyZ8ysYon7j1X0GW1ulcQ57RP3leU6X95QWGVT4SWtT8Uxq7hNXuOcEd9lRmXkay3fZ5quNM7wy3ajWF9qGdNn8Kbzm56MXNd6qti0kR1q+PZ8wReIaZuswLzyq0LVdnxKK6yVViUuFQX1Mlrz+qyHY/Y+DvUzOFite27d223U1hvlwEexgEUsc6o/sY4t7LCOAH9xizvcO4nzz/nvXL+EOgNlzjzeLefmGRDlnag=</latexit>

<latexit sha1_base64="vbsof7IKzTRwc4ppEXp+c1skZtw=">AAAC7XichVHPaxNBFP6y1tqm2kY99hISCh4k7EqgHgOt4kWppWkCTQiz20kyZH+xOwnUJXf/AW/i1ZtX/Vfq3+LBb6abQg3SWWbfN99775v35vlpqHLtutcV58HGw81HW9vVncdPdvdqT5+d58k8C2Q3SMIk6/sil6GKZVcrHcp+mkkR+aHs+bMj4+8tZJarJD7TV6kcRmISq7EKhCY1qjUGCd0muxhEQk8DERany+WoGKRqVLgvPxDXmm7Ltau+DrwSNFGuk6T2GwNcIkGAOSJIxNDEIQRyfhfw4CIlN0RBLiNS1i+xRJW5c0ZJRgiyM/4nPF2UbMyz0cxtdsBbQu6MmXUccL+1ij6jza2SOKf9w/3JcpP/3lBYZVPhFa1PxW2r+J68xpQR92VGZeSqlvszTVcaY7y23SjWl1rG9Bnc6hzTk5GbWU8db2zkhBq+PS/4AjFtlxWYV14p1G3Hl7TCWmlV4lJRUC+jNa/Pejhm79+hroPzVy3PbXkf281Ouxz4FvbRwAtO9RAdvMMJ6wjwGT/wE7+cxPnifHW+3YQ6lTLnOe4s5/tfUd2dww==</latexit>

<latexit sha1_base64="h6y2ZmFOnY2ox2tebxYybKrqYx0=">AAAC0nichVFLS8NAEJ7GV1tfVY9egkXwVBIp6LHgAy9Chb6kLbJJtzV08yDZlj7wIF69edUfpr/Fg1/WVNAi3bCZ2W+++XZmxwqEE0nDeE9pS8srq2vpTHZ9Y3NrO7ezW4v8QWjzqu0LP2xYLOLC8XhVOlLwRhBy5lqC163+WRyvD3kYOb5XkeOAt13W85yuYzMJ6HbUEh1fRvroLpc3CoZa+rxjJk6eklX2cx/Uog75ZNOAXOLkkYQviFGEr0kmGRQAa9MUWAjPUXFOD5RF7gAsDgYD2se/h1MzQT2cY81IZdu4RWCHyNTpEPtSKVpgx7dy+BHsJ/ZEYb1/b5gq5bjCMawFxYxSvAYu6R6MRZluwpzVsjgz7kpSl05VNw7qCxQS92n/6JwjEgLrq4hOF4rZg4alzkO8gAdbRQXxK88UdNVxB5Ypy5WKlygy6IWw8eujHozZ/DvUead2XDCNgnlTzJeKycDTtE8HdISpnlCJrqiMOmxU8kKv9KZVtIn2qD19U7VUkrNHv5b2/AW255Jl</latexit>

<latexit sha1_base64="51NKWXtK0QWXqt2dG54BYI5I8Jk=">AAAC0nichVFLS8NAEJ7GV1tfVY9egkXwICWRgh4LPvAiVOhL2lKSdBuXpklItoW29CBevXnVH6a/xYPfrlHQIt2wmdlvvvl2ZscOPR4Lw3hLaUvLK6tr6Ux2fWNzazu3s1uLg2HksKoTeEHUsK2YedxnVcGFxxphxKyB7bG63T+X8fqIRTEP/IoYh6w9sFyf97hjCUB3rZB3puaxOevk8kbBUEufd8zEyVOyykHunVrUpYAcGtKAGPkk4HtkUYyvSSYZFAJr0xRYBI+rOKMZZZE7BIuBYQHt4+/i1ExQH2epGatsB7d42BEydTrEvlKKNtjyVgY/hv3AnijM/feGqVKWFY5hbShmlOINcEH3YCzKHCTM71oWZ8quBPXoTHXDUV+oENmn86NzgUgErK8iOl0qpgsNW51HeAEftooK5Ct/K+iq4y6spSxTKn6iaEEvgpWvj3owZvPvUOed2knBNArmbTFfKiYDT9M+HdARpnpKJbqmMupwUMkzvdCrVtEm2oP2+EXVUknOHv1a2tMnlXOR7Q==</latexit>

<latexit sha1_base64="9aJ/QYQiaMnKaN/mimpRjNEPDlc=">AAAC0nichVFLS8NAEJ7GV1tfVY9egkXwICUpBT0WfOBFqNCXtKUk6TYuzYtNWmhLD+LVm1f9YfpbPPhlTQUt0g2bmf3mm29ndszA4WGkae8pZWV1bX0jnclubm3v7Ob29uuhPxQWq1m+44umaYTM4R6rRTxyWDMQzHBNhzXMwUUcb4yYCLnvVaNxwDquYXu8zy0jAnTfDnh3qp8WZ91cXitocqmLjp44eUpWxc99UJt65JNFQ3KJkUcRfIcMCvG1SCeNAmAdmgIT8LiMM5pRFrlDsBgYBtAB/jZOrQT1cI41Q5lt4RYHWyBTpWPsa6logh3fyuCHsJ/YE4nZ/94wlcpxhWNYE4oZqXgLPKIHMJZluglzXsvyzLiriPp0LrvhqC+QSNyn9aNziYgANpARla4k04aGKc8jvIAHW0MF8SvPFVTZcQ/WkJZJFS9RNKAnYOPXRz0Ys/53qItOvVjQtYJ+V8qXS8nA03RIR3SCqZ5RmW6ogjosVPJCr/SmVJWJ8qg8fVOVVJJzQL+W8vwFl9qR7g==</latexit>

<latexit sha1_base64="dvRKB5jdkaiTkXTXuNNRC/rfZMY=">AAAC0nichVFLS8NAEJ7GV1tfVY9egkXwICXRgh4LPvAiVOhL2lI26TYuzYskLbSlB/Hqzav+MP0tHvyypoIW6YbNzH7zzbczO4ZvizDStPeUsrS8srqWzmTXNza3tnM7u7XQGwQmr5qe7QUNg4XcFi6vRiKyecMPOHMMm9eN/kUcrw95EArPrUQjn7cdZrmiJ0wWAbpv+aIz0Y9Pp51cXitocqnzjp44eUpW2ct9UIu65JFJA3KIk0sRfJsYhfiapJNGPrA2TYAF8ISMc5pSFrkDsDgYDGgffwunZoK6OMeaocw2cYuNHSBTpUPsa6logB3fyuGHsJ/YY4lZ/94wkcpxhSNYA4oZqXgLPKIHMBZlOglzVsvizLiriHp0LrsRqM+XSNyn+aNziUgArC8jKl1JpgUNQ56HeAEXtooK4leeKaiy4y4sk5ZLFTdRZNALYOPXRz0Ys/53qPNO7aSgawX9rpgvFZOBp2mfDugIUz2jEt1QGXWYqOSFXulNqShj5VF5+qYqqSRnj34t5fkLmkGR7w==</latexit>

<latexit sha1_base64="Zn9EyRxiJbR7oqgW9/UHvQMKokc=">AAAC0nichVFLS8NAEJ7GV1tfVY9egkXwICWRgh4LPvCiVOhL2lKSdBuX5kWyLbSlB/Hqzav+MP0tHvyypoIW6YbNzH7zzbczO2bg8Eho2ntKWVpeWV1LZ7LrG5tb27md3VrkD0KLVS3f8cOGaUTM4R6rCi4c1ghCZrimw+pm/zyO14csjLjvVcQoYG3XsD3e45YhAN23At6Z6Me3004urxU0udR5R0+cPCWr7Oc+qEVd8smiAbnEyCMB3yGDInxN0kmjAFibJsBCeFzGGU0pi9wBWAwMA2gffxunZoJ6OMeakcy2cIuDHSJTpUPsK6logh3fyuBHsJ/YY4nZ/94wkcpxhSNYE4oZqXgDXNADGIsy3YQ5q2VxZtyVoB6dyW446gskEvdp/ehcIBIC68uISpeSaUPDlOchXsCDraKC+JVnCqrsuAtrSMukipcoGtALYePXRz0Ys/53qPNO7aSgawX9rpgvFZOBp2mfDugIUz2lEl1TGXVYqOSFXulNqShj5VF5+qYqqSRnj34t5fkL2x6SCg==</latexit>

<latexit sha1_base64="h6y2ZmFOnY2ox2tebxYybKrqYx0=">AAAC0nichVFLS8NAEJ7GV1tfVY9egkXwVBIp6LHgAy9Chb6kLbJJtzV08yDZlj7wIF69edUfpr/Fg1/WVNAi3bCZ2W+++XZmxwqEE0nDeE9pS8srq2vpTHZ9Y3NrO7ezW4v8QWjzqu0LP2xYLOLC8XhVOlLwRhBy5lqC163+WRyvD3kYOb5XkeOAt13W85yuYzMJ6HbUEh1fRvroLpc3CoZa+rxjJk6eklX2cx/Uog75ZNOAXOLkkYQviFGEr0kmGRQAa9MUWAjPUXFOD5RF7gAsDgYD2se/h1MzQT2cY81IZdu4RWCHyNTpEPtSKVpgx7dy+BHsJ/ZEYb1/b5gq5bjCMawFxYxSvAYu6R6MRZluwpzVsjgz7kpSl05VNw7qCxQS92n/6JwjEgLrq4hOF4rZg4alzkO8gAdbRQXxK88UdNVxB5Ypy5WKlygy6IWw8eujHozZ/DvUead2XDCNgnlTzJeKycDTtE8HdISpnlCJrqiMOmxU8kKv9KZVtIn2qD19U7VUkrNHv5b2/AW255Jl</latexit>

<latexit sha1_base64="bU2aoiD1hG+bww6/5l2U5Ni1Ny4=">AAAC0XichVFLS8NAEJ7GV1tfVY9egkXwFBIR9FjQihehon1AWyRJtzU0LzbbYi0F8erNq/4x/S0e/HZNBS3SDZuZ/eabb2d2nNj3EmGa7xltYXFpeSWby6+urW9sFra2a0k04C6rupEf8YZjJ8z3QlYVnvBZI+bMDhyf1Z3+qYzXh4wnXhTeiFHM2oHdC72u59oCUKMl2L0Ylye3haJpmGrps46VOkVKVyUqfFCLOhSRSwMKiFFIAr5PNiX4mmSRSTGwNo2BcXieijOaUB65A7AYGDbQPv49nJopGuIsNROV7eIWH5sjU6d97HOl6IAtb2XwE9hP7AeF9f69YayUZYUjWAeKOaV4CVzQHRjzMoOUOa1lfqbsSlCXTlQ3HuqLFSL7dH90zhDhwPoqolNZMXvQcNR5iBcIYauoQL7yVEFXHXdgbWWZUglTRRt6HFa+PurBmK2/Q511aoeGZRrW1VGxZKQDz9Iu7dEBpnpMJbqgCuqQ03yhV3rTrrWR9qg9fVO1TJqzQ7+W9vwFD1uSIQ==</latexit> <latexit sha1_base64="e42w+peHwDYlJrr1/t4WluWSwoQ=">AAAC0XichVFLS8NAEJ7GV1tfVY9egkXwFBIR9FjwgR6EivYBbZEk3dbQvNhsi7UUxKs3r/rH9Ld48Ns1FbRIN2xm9ptvvp3ZcWLfS4Rpvme0ufmFxaVsLr+8srq2XtjYrCZRn7us4kZ+xOuOnTDfC1lFeMJn9ZgzO3B8VnN6xzJeGzCeeFF4I4YxawV2N/Q6nmsLQPWmYPdidDG+LRRNw1RLn3as1ClSuspR4YOa1KaIXOpTQIxCEvB9sinB1yCLTIqBtWgEjMPzVJzRmPLI7YPFwLCB9vDv4tRI0RBnqZmobBe3+NgcmTrtYp8pRQdseSuDn8B+Yj8orPvvDSOlLCscwjpQzCnFS+CC7sCYlRmkzEktszNlV4I6dKS68VBfrBDZp/ujc4IIB9ZTEZ1OFbMLDUedB3iBELaCCuQrTxR01XEb1laWKZUwVbShx2Hl66MejNn6O9Rpp7pvWKZhXR0US0Y68Cxt0w7tYaqHVKJzKqMOOc0XeqU37Vobao/a0zdVy6Q5W/Rrac9fGPeSJQ==</latexit>

<latexit sha1_base64="h6y2ZmFOnY2ox2tebxYybKrqYx0=">AAAC0nichVFLS8NAEJ7GV1tfVY9egkXwVBIp6LHgAy9Chb6kLbJJtzV08yDZlj7wIF69edUfpr/Fg1/WVNAi3bCZ2W+++XZmxwqEE0nDeE9pS8srq2vpTHZ9Y3NrO7ezW4v8QWjzqu0LP2xYLOLC8XhVOlLwRhBy5lqC163+WRyvD3kYOb5XkeOAt13W85yuYzMJ6HbUEh1fRvroLpc3CoZa+rxjJk6eklX2cx/Uog75ZNOAXOLkkYQviFGEr0kmGRQAa9MUWAjPUXFOD5RF7gAsDgYD2se/h1MzQT2cY81IZdu4RWCHyNTpEPtSKVpgx7dy+BHsJ/ZEYb1/b5gq5bjCMawFxYxSvAYu6R6MRZluwpzVsjgz7kpSl05VNw7qCxQS92n/6JwjEgLrq4hOF4rZg4alzkO8gAdbRQXxK88UdNVxB5Ypy5WKlygy6IWw8eujHozZ/DvUead2XDCNgnlTzJeKycDTtE8HdISpnlCJrqiMOmxU8kKv9KZVtIn2qD19U7VUkrNHv5b2/AW255Jl</latexit> <latexit sha1_base64="h6y2ZmFOnY2ox2tebxYybKrqYx0=">AAAC0nichVFLS8NAEJ7GV1tfVY9egkXwVBIp6LHgAy9Chb6kLbJJtzV08yDZlj7wIF69edUfpr/Fg1/WVNAi3bCZ2W+++XZmxwqEE0nDeE9pS8srq2vpTHZ9Y3NrO7ezW4v8QWjzqu0LP2xYLOLC8XhVOlLwRhBy5lqC163+WRyvD3kYOb5XkeOAt13W85yuYzMJ6HbUEh1fRvroLpc3CoZa+rxjJk6eklX2cx/Uog75ZNOAXOLkkYQviFGEr0kmGRQAa9MUWAjPUXFOD5RF7gAsDgYD2se/h1MzQT2cY81IZdu4RWCHyNTpEPtSKVpgx7dy+BHsJ/ZEYb1/b5gq5bjCMawFxYxSvAYu6R6MRZluwpzVsjgz7kpSl05VNw7qCxQS92n/6JwjEgLrq4hOF4rZg4alzkO8gAdbRQXxK88UdNVxB5Ypy5WKlygy6IWw8eujHozZ/DvUead2XDCNgnlTzJeKycDTtE8HdISpnlCJrqiMOmxU8kKv9KZVtIn2qD19U7VUkrNHv5b2/AW255Jl</latexit>

<latexit sha1_base64="KUJbB/akks5kM6qJ6V0KBP0YKbc=">AAAC7XichVFNa9tAEH1WkjZ209ZJj70Ym0IOwUglkB4DbUIvhTTEdsA2ZiWvncX6QlobXOF7/0Bvodfeem3/SvtbcsjbjRJIQ8mK1bx9M/N2ZsdPQ5Vr1/1TcdbWN5483azWnm09f/Gyvr3TzZN5FshOkIRJdu6LXIYqlh2tdCjP00yKyA9lz5+9N/7eQma5SuIzvUzlMBLTWE1UIDSpUb05SOg22cUgEvoiEGFxulqNikGqRoW35xHXW27btavxEHglaKFcJ0n9LwYYI0GAOSJIxNDEIQRyfn14cJGSG6IglxEp65dYocbcOaMkIwTZGf9TnvolG/NsNHObHfCWkDtjZgNvuI+tos9oc6skzmmvuL9YbvrfGwqrbCpc0vpUrFrFT+Q1LhjxWGZURt7W8nim6Upjgne2G8X6UsuYPoM7nQ/0ZORm1tPAkY2cUsO35wVfIKbtsALzyrcKDdvxmFZYK61KXCoK6mW05vVZD8fs/TvUh6D7tu25be/zfutwvxz4Jl6jiV1O9QCH+IgT1hHgK37iF347ifPNuXS+34Q6lTLnFe4t58c1Dn+dpw==</latexit>

<latexit sha1_base64="8pGKKGwbVJAsKytKWGl9423OQuc=">AAAC7XichVFNa9tAEH1W0zZOv9z2mIuxKfRQjBQC7TGQtuQSSEIcG2JjVsraXqwvpLXBEb73D/RWes0t1+avJL8lh7zdyIHWlKxYzds3M29ndvw0VLl23euK82Tt6bPn69WNFy9fvX5Te/vuJE+mWSDbQRImWdcXuQxVLNta6VB200yKyA9lx5/sGn9nJrNcJfGxnqeyH4lRrIYqEJrUoNboJXSb7KIXCT0ORFgcLRaDopeqQeF92iKuNd2Wa1d9FXglaKJcB0ntBj2cIUGAKSJIxNDEIQRyfqfw4CIl10dBLiNS1i+xwAZzp4ySjBBkJ/yPeDot2Zhno5nb7IC3hNwZM+v4wP3dKvqMNrdK4pz2lvvccqP/3lBYZVPhnNanYtUq7pPXGDPiscyojFzW8nim6UpjiC+2G8X6UsuYPoMHna/0ZOQm1lPHNxs5ooZvzzO+QEzbZgXmlZcKddvxGa2wVlqVuFQU1MtozeuzHo7Z+3eoq+Bkq+W5Le9wu7mzXQ58HZto4COn+hk72MMB6wjwA5f4gysncX46v5zf96FOpcx5j7+Wc3EHEOedqA==</latexit>

<latexit sha1_base64="g804vR+Zy1N80G3c0Ib1HjIhTTQ=">AAAC7XichVHPaxNBFP6ytjZpq6b16CU0FDxI2NWAHgu14kWIYtJCUsLsZpIM2V/sTgrpkrv/gDfptbde67+if4sHv5luBC2SWWbfN99775v35vlpqHLtuj8qzoONzYdb1dr2zu6jx0/qe/u9PJlngewGSZhkZ77IZahi2dVKh/IszaSI/FCe+rNj4z+9kFmukvizXqTyPBKTWI1VIDSpYf1gkNBtsotBJPQ0EGHxabkcFoNUDQvvxSvietNtuXY17gOvBE2Uq5PUf2KAERIEmCOCRAxNHEIg59eHBxcpuXMU5DIiZf0SS2wzd84oyQhBdsb/hKd+ycY8G83cZge8JeTOmNnAIfc7q+gz2twqiXPaX9yXlpv894bCKpsKF7Q+FWtW8QN5jSkj1mVGZeSqlvWZpiuNMd7YbhTrSy1j+gz+6LylJyM3s54GTmzkhBq+PV/wBWLaLiswr7xSaNiOR7TCWmlV4lJRUC+jNa/Pejhm79+h3ge9ly3PbXkf282jdjnwKp7hAM851dc4wnt0WEeAL7jBLb47ifPV+eZc3YU6lTLnKf5azvVvE0+dqQ==</latexit>

<latexit sha1_base64="XNPUgPlX2k1acpdnflcbLw1egBI=">AAAC7XichVHPaxNBFP6ytprEVqMeewkJggcJu6Wgx0Bb8aLE0rSFpoTZ7SQZsr/YnQTiknv/AW/i1ZtX/Vfq3+LBb6abQltKZ5l933zvvW/em+enocq1615WnEdr64+fVGv1pxubz543Xrw8ypNZFsh+kIRJduKLXIYqln2tdChP0kyKyA/lsT/dNf7jucxylcSHepHKs0iMYzVSgdCkho3WIKHbZBeDSOhJIMLiYLkcFoNUDQvv7WfiRtvtuHY17wKvBG2Uq5c0/mKAcyQIMEMEiRiaOIRAzu8UHlyk5M5QkMuIlPVLLFFn7oxRkhGC7JT/MU+nJRvzbDRzmx3wlpA7Y2YTr7k/WEWf0eZWSZzT/uP+arnxvTcUVtlUuKD1qVizip/Ia0wY8VBmVEauank403SlMcJ7241ifallTJ/Btc4ePRm5qfU0sW8jx9Tw7XnOF4hp+6zAvPJKoWk7PqcV1kqrEpeKgnoZrXl91sMxe7eHehccbXc8t+N92Wl3d8qBV7GFFt5wqu/QxUf0WEeAC/zCb/xxEueb8935cRXqVMqcV7ixnJ//AVRHncQ=</latexit>

<latexit sha1_base64="PdKhvUaNfE6D002IPRe2pS+5hcw=">AAAC0nichVFLS8NAEJ7GV1tfVY9egkXwICUpBT0WfOBFqNCXtKUk6TYuzYtNWmhLD+LVm1f9YfpbPPhlTQUt0g2bmf3mm29ndszA4WGkae8pZWV1bX0jnclubm3v7Ob29uuhPxQWq1m+44umaYTM4R6rRTxyWDMQzHBNhzXMwUUcb4yYCLnvVaNxwDquYXu8zy0jAnTfDnh3WjzVZ91cXitocqmLjp44eUpWxc99UJt65JNFQ3KJkUcRfIcMCvG1SCeNAmAdmgIT8LiMM5pRFrlDsBgYBtAB/jZOrQT1cI41Q5lt4RYHWyBTpWPsa6logh3fyuCHsJ/YE4nZ/94wlcpxhWNYE4oZqXgLPKIHMJZluglzXsvyzLiriPp0LrvhqC+QSNyn9aNziYgANpARla4k04aGKc8jvIAHW0MF8SvPFVTZcQ/WkJZJFS9RNKAnYOPXRz0Ys/53qItOvVjQtYJ+V8qXS8nA03RIR3SCqZ5RmW6ogjosVPJCr/SmVJWJ8qg8fVOVVJJzQL+W8vwFl9yR7g==</latexit>

<latexit sha1_base64="E66nj/wNHLRSp4ZTJ51qZvpYras=">AAAC0nichVFLS8NAEJ7GV1tfVY9egkXwICUpBT0WfOBFqNCXtKUk6TYuzYskLbSlB/Hqzav+MP0tHvx2TQUt0g2bmf3mm29ndszA4VGsae8pZWV1bX0jnclubm3v7Ob29uuRPwwtVrN8xw+bphExh3usFvPYYc0gZIZrOqxhDi5EvDFiYcR9rxqPA9ZxDdvjfW4ZMaD7dsC70+JpcdbN5bWCJpe66OiJk6dkVfzcB7WpRz5ZNCSXGHkUw3fIoAhfi3TSKADWoSmwEB6XcUYzyiJ3CBYDwwA6wN/GqZWgHs5CM5LZFm5xsENkqnSMfS0VTbDFrQx+BPuJPZGY/e8NU6ksKhzDmlDMSMVb4DE9gLEs002Y81qWZ4quYurTueyGo75AIqJP60fnEpEQ2EBGVLqSTBsapjyP8AIebA0ViFeeK6iy4x6sIS2TKl6iaEAvhBWvj3owZv3vUBederGgawX9rpQvl5KBp+mQjugEUz2jMt1QBXVYqOSFXulNqSoT5VF5+qYqqSTngH4t5fkLmkOR7w==</latexit>

<latexit sha1_base64="JJR075LuOWu33xLg7aVxFaFVPw8=">AAAC0nichVFLS8NAEJ7GV1tfVY9egkXwICWpBT0WfOBFqNCXtKVs0m1cmhdJWmhLD+LVm1f9YfpbPPhljYIW6YbNzH7zzbczO4ZvizDStLeUsrS8srqWzmTXNza3tnM7u/XQGwYmr5me7QVNg4XcFi6vRSKyedMPOHMMmzeMwXkcb4x4EArPrUZjn3ccZrmiL0wWAbpr+6I7LR6fzLq5vFbQ5FLnHT1x8pSsipd7pzb1yCOThuQQJ5ci+DYxCvG1SCeNfGAdmgIL4AkZ5zSjLHKHYHEwGNAB/hZOrQR1cY41Q5lt4hYbO0CmSofYV1LRADu+lcMPYT+wJxKz/r1hKpXjCsewBhQzUvEGeET3YCzKdBLmdy2LM+OuIurTmexGoD5fInGf5o/OBSIBsIGMqHQpmRY0DHke4QVc2BoqiF/5W0GVHfdgmbRcqriJIoNeABu/PurBmPW/Q5136sWCrhX021K+XEoGnqZ9OqAjTPWUynRNFdRhopJneqFXpapMlAfl8YuqpJKcPfq1lKdPnKqR8A==</latexit>

<latexit sha1_base64="iFnvrIGRqpETUkVF8T4XFP3Eaac=">AAAC0nichVFLS8NAEJ7GV1tfVY9egkXwICUpBT0WfOBFqdCXtKUk6TYuzYtkW2hLD+LVm1f9YfpbPPhljYIW6YbNzH7zzbczO2bg8Eho2ltKWVpeWV1LZ7LrG5tb27md3XrkD0OL1Szf8cOmaUTM4R6rCS4c1gxCZrimwxrm4CyON0YsjLjvVcU4YB3XsD3e55YhAN21A96dFo9vZt1cXitocqnzjp44eUpWxc+9U5t65JNFQ3KJkUcCvkMGRfhapJNGAbAOTYGF8LiMM5pRFrlDsBgYBtAB/jZOrQT1cI41I5lt4RYHO0SmSofYl1LRBDu+lcGPYD+wJxKz/71hKpXjCsewJhQzUvEauKB7MBZlugnzu5bFmXFXgvp0KrvhqC+QSNyn9aNzjkgIbCAjKl1Ipg0NU55HeAEPtoYK4lf+VlBlxz1YQ1omVbxE0YBeCBu/PurBmPW/Q5136sWCrhX021K+XEoGnqZ9OqAjTPWEynRFFdRhoZJneqFXpapMlAfl8YuqpJKcPfq1lKdP3YeSCw==</latexit>

<latexit sha1_base64="bU2aoiD1hG+bww6/5l2U5Ni1Ny4=">AAAC0XichVFLS8NAEJ7GV1tfVY9egkXwFBIR9FjQihehon1AWyRJtzU0LzbbYi0F8erNq/4x/S0e/HZNBS3SDZuZ/eabb2d2nNj3EmGa7xltYXFpeSWby6+urW9sFra2a0k04C6rupEf8YZjJ8z3QlYVnvBZI+bMDhyf1Z3+qYzXh4wnXhTeiFHM2oHdC72u59oCUKMl2L0Ylye3haJpmGrps46VOkVKVyUqfFCLOhSRSwMKiFFIAr5PNiX4mmSRSTGwNo2BcXieijOaUB65A7AYGDbQPv49nJopGuIsNROV7eIWH5sjU6d97HOl6IAtb2XwE9hP7AeF9f69YayUZYUjWAeKOaV4CVzQHRjzMoOUOa1lfqbsSlCXTlQ3HuqLFSL7dH90zhDhwPoqolNZMXvQcNR5iBcIYauoQL7yVEFXHXdgbWWZUglTRRt6HFa+PurBmK2/Q511aoeGZRrW1VGxZKQDz9Iu7dEBpnpMJbqgCuqQ03yhV3rTrrWR9qg9fVO1TJqzQ7+W9vwFD1uSIQ==</latexit> <latexit sha1_base64="e42w+peHwDYlJrr1/t4WluWSwoQ=">AAAC0XichVFLS8NAEJ7GV1tfVY9egkXwFBIR9FjwgR6EivYBbZEk3dbQvNhsi7UUxKs3r/rH9Ld48Ns1FbRIN2xm9ptvvp3ZcWLfS4Rpvme0ufmFxaVsLr+8srq2XtjYrCZRn7us4kZ+xOuOnTDfC1lFeMJn9ZgzO3B8VnN6xzJeGzCeeFF4I4YxawV2N/Q6nmsLQPWmYPdidDG+LRRNw1RLn3as1ClSuspR4YOa1KaIXOpTQIxCEvB9sinB1yCLTIqBtWgEjMPzVJzRmPLI7YPFwLCB9vDv4tRI0RBnqZmobBe3+NgcmTrtYp8pRQdseSuDn8B+Yj8orPvvDSOlLCscwjpQzCnFS+CC7sCYlRmkzEktszNlV4I6dKS68VBfrBDZp/ujc4IIB9ZTEZ1OFbMLDUedB3iBELaCCuQrTxR01XEb1laWKZUwVbShx2Hl66MejNn6O9Rpp7pvWKZhXR0US0Y68Cxt0w7tYaqHVKJzKqMOOc0XeqU37Vobao/a0zdVy6Q5W/Rrac9fGPeSJQ==</latexit>

<latexit sha1_base64="h6y2ZmFOnY2ox2tebxYybKrqYx0=">AAAC0nichVFLS8NAEJ7GV1tfVY9egkXwVBIp6LHgAy9Chb6kLbJJtzV08yDZlj7wIF69edUfpr/Fg1/WVNAi3bCZ2W+++XZmxwqEE0nDeE9pS8srq2vpTHZ9Y3NrO7ezW4v8QWjzqu0LP2xYLOLC8XhVOlLwRhBy5lqC163+WRyvD3kYOb5XkeOAt13W85yuYzMJ6HbUEh1fRvroLpc3CoZa+rxjJk6eklX2cx/Uog75ZNOAXOLkkYQviFGEr0kmGRQAa9MUWAjPUXFOD5RF7gAsDgYD2se/h1MzQT2cY81IZdu4RWCHyNTpEPtSKVpgx7dy+BHsJ/ZEYb1/b5gq5bjCMawFxYxSvAYu6R6MRZluwpzVsjgz7kpSl05VNw7qCxQS92n/6JwjEgLrq4hOF4rZg4alzkO8gAdbRQXxK88UdNVxB5Ypy5WKlygy6IWw8eujHozZ/DvUead2XDCNgnlTzJeKycDTtE8HdISpnlCJrqiMOmxU8kKv9KZVtIn2qD19U7VUkrNHv5b2/AW255Jl</latexit>

<latexit sha1_base64="rLoE5Bujxtl/2z+Sk6r1j8ziL3s=">AAAC7XichVFNa9tAEH1W0zZOv9z2mIuxKfRQjBQC7TGQtuQSSEIcG2JjVsraXqwvpLXBEb73D/RWes0t1+avJL8lh7zdyIHWlKxYzds3M29ndvw0VLl23euK82Tt6bPn69WNFy9fvX5Te/vuJE+mWSDbQRImWdcXuQxVLNta6VB200yKyA9lx5/sGn9nJrNcJfGxnqeyH4lRrIYqEJrUoNboJXSb7KIXCT0ORFgcLRaDopeqQbH1ySOuNd2Wa1d9FXglaKJcB0ntBj2cIUGAKSJIxNDEIQRyfqfw4CIl10dBLiNS1i+xwAZzp4ySjBBkJ/yPeDot2Zhno5nb7IC3hNwZM+v4wP3dKvqMNrdK4pz2lvvccqP/3lBYZVPhnNanYtUq7pPXGDPiscyojFzW8nim6UpjiC+2G8X6UsuYPoMHna/0ZOQm1lPHNxs5ooZvzzO+QEzbZgXmlZcKddvxGa2wVlqVuFQU1MtozeuzHo7Z+3eoq+Bkq+W5Le9wu7mzXQ58HZto4COn+hk72MMB6wjwA5f4gysncX46v5zf96FOpcx5j7+Wc3EHEOmdqA==</latexit>

<latexit sha1_base64="K/Zx1oNTs5xVBYC0aG4FTnLFgQM=">AAAC7XichVHPSxtBFP5crSa21VSPvQRDoYcSdkNAj0Jb8VKIxRjBSJjdTOKQ/cXuREiX3PsPeCu9euvV/iv1b+mh34yroKE4y+z75nvvffPePD8NVa5d98+Ss7zyYnWtUl1/+er1xmbtzdZJnkyzQHaDJEyyU1/kMlSx7GqlQ3maZlJEfih7/uSj8fcuZZarJD7Ws1SeR2Icq5EKhCY1qO30E7pNdtGPhL4IRFh8nc8HRT9Vg6L1oUVca7hN1676IvBK0EC5OkntFn0MkSDAFBEkYmjiEAI5vzN4cJGSO0dBLiNS1i8xxzpzp4ySjBBkJ/yPeTor2Zhno5nb7IC3hNwZM+t4x31gFX1Gm1slcU77l/ub5cb/vaGwyqbCGa1PxapV/EJe44IRz2VGZeR9Lc9nmq40Rtiz3SjWl1rG9Bk86HyiJyM3sZ46PtvIMTV8e77kC8S0XVZgXvleoW47HtIKa6VViUtFQb2M1rw+6+GYvadDXQQnrabnNr2jdmO/XQ68grfYwXtOdRf7OESHdQT4jl+4wW8nca6cH87Pu1BnqczZxqPlXP8DE1GdqQ==</latexit>

<latexit sha1_base64="z37xcO4SwF7vPdAE3vrO7JWqJHs=">AAAC7XichVHPSxtBFP7captorWl79BIMBQ8l7EbBHoW24kVIxaiQSJhdx2TI/mJ3IqRL7v0HvIlXb73qv6J/Sw9+M65CG4qzzL5vvvfeN+/N89NQ5dp17+acV/MLr99UqotLb5ffrdTefzjMk3EWyE6QhEl27ItchiqWHa10KI/TTIrID+WRP/pq/EfnMstVEh/oSSpPIjGI1ZkKhCbVr631ErpNdtGLhB4GIiz2p9N+0UtVv2h93iCuNdyma1d9FnglaKBc7aR2jx5OkSDAGBEkYmjiEAI5vy48uEjJnaAglxEp65eYYpG5Y0ZJRgiyI/4HPHVLNubZaOY2O+AtIXfGzDo+ce9YRZ/R5lZJnNP+4f5pucF/byissqlwQutTsWoV98hrDBnxUmZURj7V8nKm6UrjDF9sN4r1pZYxfQbPOt/oyciNrKeO7zZyQA3fns/5AjFthxWYV35SqNuOT2mFtdKqxKWioF5Ga16f9XDM3r9DnQWHrabnNr0fm43tzXLgFaxiDeuc6ha2sYs26wjwC79xg1sncS6cS+fqMdSZK3M+4q/lXD8AFbmdqg==</latexit>

<latexit sha1_base64="11PboPdyJDsAnRLWeU8CFBpwwzI=">AAAC7XichVHPSxtBFP5ca5torakevQRDwUMJu0HQo9BWvLREMSokEmbXMRmyv9idCOmSu/+AN/Ham9f2X6l/iwe/GddCG4qzzL5vvvfeN+/N89NQ5dp1f885868WXr+pVBeX3i6/W6m9Xz3Ok3EWyE6QhEl26otchiqWHa10KE/TTIrID+WJP/pk/CeXMstVEh/pSSrPIjGI1YUKhCbVr230ErpNdtGLhB4GIiwOp9N+0UtVv2h9/EZca7hN1676LPBK0EC52kntHj2cI0GAMSJIxNDEIQRyfl14cJGSO0NBLiNS1i8xxSJzx4ySjBBkR/wPeOqWbMyz0cxtdsBbQu6MmXV84N6zij6jza2SOKd94P5uucF/byissqlwQutTsWoVv5LXGDLipcyojHyu5eVM05XGBXZsN4r1pZYxfQZ/dD7Tk5EbWU8dX2zkgBq+PV/yBWLaDiswr/ysULcdn9MKa6VViUtFQb2M1rw+6+GYvX+HOguOW03PbXoHW43drXLgFaxjA5uc6jZ2sY826whwhTv8xC8nca6dG+f2KdSZK3PW8NdyfjwCVrGdxQ==</latexit>

<latexit sha1_base64="h6y2ZmFOnY2ox2tebxYybKrqYx0=">AAAC0nichVFLS8NAEJ7GV1tfVY9egkXwVBIp6LHgAy9Chb6kLbJJtzV08yDZlj7wIF69edUfpr/Fg1/WVNAi3bCZ2W+++XZmxwqEE0nDeE9pS8srq2vpTHZ9Y3NrO7ezW4v8QWjzqu0LP2xYLOLC8XhVOlLwRhBy5lqC163+WRyvD3kYOb5XkeOAt13W85yuYzMJ6HbUEh1fRvroLpc3CoZa+rxjJk6eklX2cx/Uog75ZNOAXOLkkYQviFGEr0kmGRQAa9MUWAjPUXFOD5RF7gAsDgYD2se/h1MzQT2cY81IZdu4RWCHyNTpEPtSKVpgx7dy+BHsJ/ZEYb1/b5gq5bjCMawFxYxSvAYu6R6MRZluwpzVsjgz7kpSl05VNw7qCxQS92n/6JwjEgLrq4hOF4rZg4alzkO8gAdbRQXxK88UdNVxB5Ypy5WKlygy6IWw8eujHozZ/DvUead2XDCNgnlTzJeKycDTtE8HdISpnlCJrqiMOmxU8kKv9KZVtIn2qD19U7VUkrNHv5b2/AW255Jl</latexit>

<latexit sha1_base64="h6y2ZmFOnY2ox2tebxYybKrqYx0=">AAAC0nichVFLS8NAEJ7GV1tfVY9egkXwVBIp6LHgAy9Chb6kLbJJtzV08yDZlj7wIF69edUfpr/Fg1/WVNAi3bCZ2W+++XZmxwqEE0nDeE9pS8srq2vpTHZ9Y3NrO7ezW4v8QWjzqu0LP2xYLOLC8XhVOlLwRhBy5lqC163+WRyvD3kYOb5XkeOAt13W85yuYzMJ6HbUEh1fRvroLpc3CoZa+rxjJk6eklX2cx/Uog75ZNOAXOLkkYQviFGEr0kmGRQAa9MUWAjPUXFOD5RF7gAsDgYD2se/h1MzQT2cY81IZdu4RWCHyNTpEPtSKVpgx7dy+BHsJ/ZEYb1/b5gq5bjCMawFxYxSvAYu6R6MRZluwpzVsjgz7kpSl05VNw7qCxQS92n/6JwjEgLrq4hOF4rZg4alzkO8gAdbRQXxK88UdNVxB5Ypy5WKlygy6IWw8eujHozZ/DvUead2XDCNgnlTzJeKycDTtE8HdISpnlCJrqiMOmxU8kKv9KZVtIn2qD19U7VUkrNHv5b2/AW255Jl</latexit>

<latexit sha1_base64="h6y2ZmFOnY2ox2tebxYybKrqYx0=">AAAC0nichVFLS8NAEJ7GV1tfVY9egkXwVBIp6LHgAy9Chb6kLbJJtzV08yDZlj7wIF69edUfpr/Fg1/WVNAi3bCZ2W+++XZmxwqEE0nDeE9pS8srq2vpTHZ9Y3NrO7ezW4v8QWjzqu0LP2xYLOLC8XhVOlLwRhBy5lqC163+WRyvD3kYOb5XkeOAt13W85yuYzMJ6HbUEh1fRvroLpc3CoZa+rxjJk6eklX2cx/Uog75ZNOAXOLkkYQviFGEr0kmGRQAa9MUWAjPUXFOD5RF7gAsDgYD2se/h1MzQT2cY81IZdu4RWCHyNTpEPtSKVpgx7dy+BHsJ/ZEYb1/b5gq5bjCMawFxYxSvAYu6R6MRZluwpzVsjgz7kpSl05VNw7qCxQS92n/6JwjEgLrq4hOF4rZg4alzkO8gAdbRQXxK88UdNVxB5Ypy5WKlygy6IWw8eujHozZ/DvUead2XDCNgnlTzJeKycDTtE8HdISpnlCJrqiMOmxU8kKv9KZVtIn2qD19U7VUkrNHv5b2/AW255Jl</latexit>

<latexit sha1_base64="h6y2ZmFOnY2ox2tebxYybKrqYx0=">AAAC0nichVFLS8NAEJ7GV1tfVY9egkXwVBIp6LHgAy9Chb6kLbJJtzV08yDZlj7wIF69edUfpr/Fg1/WVNAi3bCZ2W+++XZmxwqEE0nDeE9pS8srq2vpTHZ9Y3NrO7ezW4v8QWjzqu0LP2xYLOLC8XhVOlLwRhBy5lqC163+WRyvD3kYOb5XkeOAt13W85yuYzMJ6HbUEh1fRvroLpc3CoZa+rxjJk6eklX2cx/Uog75ZNOAXOLkkYQviFGEr0kmGRQAa9MUWAjPUXFOD5RF7gAsDgYD2se/h1MzQT2cY81IZdu4RWCHyNTpEPtSKVpgx7dy+BHsJ/ZEYb1/b5gq5bjCMawFxYxSvAYu6R6MRZluwpzVsjgz7kpSl05VNw7qCxQS92n/6JwjEgLrq4hOF4rZg4alzkO8gAdbRQXxK88UdNVxB5Ypy5WKlygy6IWw8eujHozZ/DvUead2XDCNgnlTzJeKycDTtE8HdISpnlCJrqiMOmxU8kKv9KZVtIn2qD19U7VUkrNHv5b2/AW255Jl</latexit>

<latexit sha1_base64="h6y2ZmFOnY2ox2tebxYybKrqYx0=">AAAC0nichVFLS8NAEJ7GV1tfVY9egkXwVBIp6LHgAy9Chb6kLbJJtzV08yDZlj7wIF69edUfpr/Fg1/WVNAi3bCZ2W+++XZmxwqEE0nDeE9pS8srq2vpTHZ9Y3NrO7ezW4v8QWjzqu0LP2xYLOLC8XhVOlLwRhBy5lqC163+WRyvD3kYOb5XkeOAt13W85yuYzMJ6HbUEh1fRvroLpc3CoZa+rxjJk6eklX2cx/Uog75ZNOAXOLkkYQviFGEr0kmGRQAa9MUWAjPUXFOD5RF7gAsDgYD2se/h1MzQT2cY81IZdu4RWCHyNTpEPtSKVpgx7dy+BHsJ/ZEYb1/b5gq5bjCMawFxYxSvAYu6R6MRZluwpzVsjgz7kpSl05VNw7qCxQS92n/6JwjEgLrq4hOF4rZg4alzkO8gAdbRQXxK88UdNVxB5Ypy5WKlygy6IWw8eujHozZ/DvUead2XDCNgnlTzJeKycDTtE8HdISpnlCJrqiMOmxU8kKv9KZVtIn2qD19U7VUkrNHv5b2/AW255Jl</latexit>

Figure 2: Population-based policy search iterating policy evaluation (E) andpolicy improvement (I)

The goal of using policy search for learning interpretable RL policies is tofind the best policy among a set of policies that is spanned by a parametervector x ∈ X , where X is the space of all interpretable policy parameters forthis RL task. Herein, a policy corresponding to a particular parameter valuex is denoted by π[x]. The policy’s performance, when starting from st ismeasured by its value function given in Eq. (2). Furthermore, including onlya finite number of T > 1 future rewards for the model-based value estimationfrom Eq. (2) yields

vπ[x](st) =T−1∑k=0

γkr(st+k,at+k, st+k+1),

with st+k+1 = g(st+k,at+k) and at+k = π[x](st+k).

(4)

To rate the performance of policy π[x], the value function is used asfollows:

Rπ[x] =1

|S|∑s∈S

vπ[x](s), (5)

with Rπ[x] being the average discounted return of π[x] on a representativeset of test states S.

In population-based RL, the interest lies in populations of policies, whichmeans that multiple different policies exist at the same time in one policyiteration step. Fig. 2 depicts how the return values of all population membersdrive the improvement step (I) of the whole population. The evaluation step(E) is performed individually for every member. Applying population-basedevolutionary methods to policy search has already been successfully achievedin the past [8, 9, 15, 28].

3.2 Genetic programming

GP has been utilized for creating system controllers since its introduction [21].Since then, the field of GP has grown significantly and has produced numer-

8

Page 9: Trustworthy AI for Process Automation on a Chylla-Haase

ous results that can compete with human-produced solutions [22], includingsystem controllers [20, 26], game playing [14, 28], and robotics [12, 19].

GP encodes computer programs as sets of genes and then modifies (evolves)them using a so-called genetic algorithm (GA) to drive the optimization ofthe population, by applying selection and reproduction to the population.The basis for both concepts is a fitness value, which represents the qualityof performing the predefined task for each individual. Selection means, thatonly the best portion of the current generation will survive each iterationand continue existing in the next generation. Analogous to biological sex-ual breeding, two individuals are selected for reproduction based on theirfitness, and two offspring individuals are created by crossing their chromo-somes. Technically, this is realized by selecting compatible cutting points inthe function trees and subsequently interchanging the subtrees beneath thesecuts. The two resulting individuals are introduced to the population of thenext generation. Herein, we applied tournament selection [7] for selectingthe individuals to be crossed.

To rate the quality of each policy candidate, a fitness value has to beprovided in order for the GP algorithm to advance. For GPRL, the fitness ofeach individual is calculated by generating trajectories using the model-basedreturn estimation from Eq. (5).

The overall GA used in the experiments is given as follows:

1. Randomly initialize the population of size N

2. Determine fitness value of each individual using Eq. (5) (in parallel)

3. Evolve next generation

(a) Crossover (depending on crossover ratio rc)

i. Select individuals by tournament selection

ii. Cross two tournament winners

iii. Add resulting individuals to new population

(b) Reproduction (depending on reproduction ratio rr)

i. Select individuals by tournament selection

ii. Add tournament winner to new population

(c) Automatic cancelation and terminal mutation (depending onauto cancel ratio ra and terminal mutation ration rm)

i. Apply automatic cancelation on all individuals

ii. Add canceled individuals according to ra

9

Page 10: Trustworthy AI for Process Automation on a Chylla-Haase

iii. Select best individuals for each complexity level of old popu-lation

iv. Randomly mutate float terminals using normal distributionN : z′ ∼ z+ 0.1z ·N (0, 1), where z and z′ are the original andthe mutated float terminals, respectively; and create N · raadjusted individuals from each best

v. Determine fitness value of each individual (in parallel)

vi. Add best adjusted individuals to new population accordingto rm

(d) Fill new population with new randomly generated individuals(new individuals ratio rn)

(e) Determine fitness value of each individual using Eq. (5) (inparallel)

(f) If none of the stopping criteria is met

i. Go back to 3.

4. Return best individual found so far for each complexity level

For the experiments described in Section 4, the following new populationratios have been used: rc = 0.45, rr = 0.05, rm = 0.1, ra = 0.1, rn =0.3. Stopping criteria can be a maximum number of iterations, a certainalgorithm runtime, a pre-defined performance value, etc. In our experiments,we stopped the GA after 100 iterations.

GPRL has been implemented using the open source evolutionary com-putation framework DEAP2 which provides an excellent and widely-usedPython implementation of genetic programing using prefix trees [13]. A de-tailed explaination of the GA utilized in GPRL, including a description ofautomatic cancelation and parameter motivation, can be found in [16].

4 Interpretable process controllers for a Chylla-

Haase polymerization reactor

The framework of learning interpretable process controllers contains two MLsteps. Firstly, available data from the process is used to generate a processmodel. Here, we used supervised ML to train weights of a recurrent neuralnetwork as system identification. Secondly, GPRL is performed on this re-current model to produce several Pareto-optimized process controllers. The

2https://github.com/DEAP/deap

10

Page 11: Trustworthy AI for Process Automation on a Chylla-Haase

Process

Process data

Supervised ML regression

Process surrogate

model

GPRLProcess KPIs Controller building blocks

Control strategies

Controller

Domain expert

Validation and selection

Selected strategy

Figure 3: EPC of generating interpretable process controllers using GPRL.Note that the Domain expert is integrated into the optimization loop. How-ever, generating interpretable control strategies is fully automatized.

controllers are optimized towards the process KPIs utilizing the platform-specific process control building blocks. An event-driven process chain (EPC)flowchart of applying GPRL for process optimization is depicted in Fig. 3.

4.1 Recurrent system identification

To generate a batch data set D with sufficient experience, 100 different explo-ration process recipes have been used. In the classic control strategy, usinga PID temperature controller, the controller receives a constant setpoint. Asthe controller needs to handle both, heating and cooling during the produc-tion, the performance is not optimal and significant control deviations couldbe observed. This will be shown in Section 4.4 together with the results. Forsafety and acceptance reasons, GPRL is not intended to replace the PID con-troller but to be superimposed on it. Therefore, the setpoint temperature T

11

Page 12: Trustworthy AI for Process Automation on a Chylla-Haase

of the PID controller becomes the action of GPRL and a new overall setpointS is introduced (see Table 1). Modifying T reveals new degrees of freedomnot used before in the operation of the reactor. Therefore, new training dataneeds to be created wherein T is changed. For our experiments, the set-point has been changed once per recipe between 352 and 365 K at a randombatch runtime between 100 and 600 s. In process optimization, such stepattempts are a common method of safely exploring in a bounded subspaceof the control problem.

The response of the process on these changes is then recorded and ex-ported to the data batch. In the next step, the data is subsampled on a 10 sgrid and then normalized, i.e., average 0 and standard deviation 1.

The system model g is a recurrent neural network (RNN) with the taskto predict the values of st+1 = (Tt+1,Mt+1, Pt+1, UAt+1, Qt+1), st+2, . . . , st+Ffrom the inputs st = (Tt,Mt, Pt, UAt, Qt), st−1, . . ., st−H+1 and at = (Tt, Mt),at−1, . . . ,at−H+1. The RNN is unfolded H = 10 time steps into the past andF = 10 time steps into the future. In each time step, the observable variablesof the past and present are inputs. Whereas, in the future branch of the RNNonly the control variables T and M are used as input. The topology of theRNN is described in [25] as Markov decision process extraction network. Themodel was implemented as a tensorflow graph using RNN cells with twohidden layers and 20 tanh activation neurons on each layer.

Note that an F = 10 overshooting is used, which means that the regres-sion error used to optimize the network weights, is not only computed on thenext time step, but as an average error over the next 10 time steps. In ourexperience, this generally helps to yield more robust predictions for modelswhich are used in closed loop simulations, where predictions have to be madeon the models own, possibly inaccurate, previous predictions.

After 10,000 training episodes, the quadratic loss of the prediction hasfallen to 0.0017 on the trainings set, 0.0022 on the validations set, and 0.0021on the generalization set (see Fig. 4). The sets contained 70, 20, and 10 stepattempt time series for training, validation, and generalization, respectively.Fig. 5 depicts the average absolute loss for the predicted reactor temperatureT over the 10 overshooting time steps. It is shown, how the prediction qualityis decreasing for more distant time steps. However, a validation error increasefrom 0.019 to 0.026 on a prediction horizon of 100 seconds (10 time steps)still yields an adequate system model.

4.2 Learning an interpretable process controller

The RNN, trained in the previous step, can now be used as model g togetherwith the reward function r to rate the performance of policy candidates. In

12

Page 13: Trustworthy AI for Process Automation on a Chylla-Haase

Figure 4: Learning curve of the RNN surrogate model training

Figure 5: Prediction error of the trained model RNN surrogate model overten future time steps

13

Page 14: Trustworthy AI for Process Automation on a Chylla-Haase

case of the temperature control task we consider here, the reward is computedby r(st) = rt = −(S − Tt)2, where St is the desired setpoint and Tt is thepredicted reactor temperature at time step t. According to Eq. (4), thecontroller parameters x are tested on 100 starting states S drawn from thedata batch D. The trajectories produced by closed loop evaluation of policyπ with parameters x and system model g, have a length of T = 10. Hence,the performance of x can be estimated by Eq. (5).

The building blocks of the genetic programming algorithm are basic alge-braic functions {+,−, ∗, /}, state and action variables {s,a}, and dead timeblocks storing past values of the variables up to 50 seconds {�t,�t−10, . . . ,�t−50}, with � an arbitrary variable from s or a. Here, we are searching for apolicy that serves as a setpoint process controller which dynamically changessetpoint T to minimize the distance between the desired setpoint S and theactual reactor temperature T . The monomer setpoint M has been fixed tothe value 0.015. Note that the policy can incorporate state and action valuesfrom the past and is evaluated in a closed loop manor for ten time stepsduring training. Consequently, it will not greedily try to reach the desiredsetpoint in the next step, but in the long run try to avoid overshooting andmaximize the reward instead.

After 100 iterations of GPRL with 500 individuals in each generation, theperformance of the best policies found are visualized in Fig. 6. It is easy tosee that individuals of lower complexity have a higher penalty (lower reward)compared to individuals of higher complexity. However, the estimated per-formance seems to not improve significantly for individuals of complexity 10or higher. After discussion with domain experts, an individual of complexity9 and an estimated penalty of around 392 has been selected (highlighted inred). The selected individual can be represented as the following equation:

Tt = Tt−30 − 2Tt + 2S − 1. (6)

This equation can easily be interpreted thanks to its simple structure.The reactor temperature setpoint is dynamically changed with respect tothe reactor temperature 30 s ago, the current temperature, and the intendedsetpoint. Note that all variables (T , T , and S) in the equation have beennormalized by average 359.12 and standard deviation 6.47.

4.3 SIMATIC PCS 7 integration

The policy found by GPRL and represented by Eq. (6), can easily be imple-mented in SIMATIC PCS 7, since GPRL utilized only components from thepre-defined building blocks which are available in the target system. Fig. 7

14

Page 15: Trustworthy AI for Process Automation on a Chylla-Haase

0 5 10 15 20 25 30 35complexity

380

390

400

410

420

430

440

450se

tpoi

nt p

enal

ty

Figure 6: Pareto front generated by GPRL. According to the domain expert,the selected policy (red dot) is a good compromise between setpoint penalty(392) and complexity (9).

shows a screenshot of the respective structured control language (SCL) code.Note, that since this is the native language of the system at hand, the newpolicy is easy to integrate and fast in execution. Domain experts are ableto fully understand the code and can easily make adjustments. All of thiswould be impossible if we were to use a black-box policy.

After generating the SCL code, the new control strategy can be loadedinto a standard continuous function chart (CFC) building block. Fig. 8 showsthe CFC diagram with our new policy block (FB10) and the additional deadtime block (DeadTime) with 30 s below. The policy is now fully integratedinto the automation system and can be evaluated.

Note that for applying the function to a real production plant, somesafety logic and switching between different modes like manual and automaticwould be required. Integration of this is straight forward and neglected forthe following evaluation using the simulation model.

4.4 Evaluation

To evaluate and visualize the performance of the policy, we tested it onfour exemplary intended setpoints, i.e., 352, 358, 362, 365 K. The intendedsetpoint S is plotted in red. The action of the policy T is plotted in blue.The resulting reactor temperature T , using the default PID regulator T = S,is plotted in orange. The resulting reactor temperature T , using the new RL

15

Page 16: Trustworthy AI for Process Automation on a Chylla-Haase

Figure 7: SCL script with the implemented GPRL control strategy

16

Page 17: Trustworthy AI for Process Automation on a Chylla-Haase

Figure 8: CFC diagram with the building block (FB10) containing theGPRL policy

policy given in Eq. (6), is plotted in green.Fig. 9 depicts the results for S = 352 K. In the beginning the RL policy

changes the temperature setpoint to the maximum value of 365 K, to speedup the heating process of the reactor. After approximately 80 s, it reducesthe setpoint to the minimum value of 352 K, to keep the reactor temperatureas close as possible to the intended setpoint. However, since the setpoint cannot be lowered below the minimum of 352 K, the reactor is constantly abovethe intended setpoint. In this example, the new RL policy could slightlyreduce control deviation by 0.3 %.

A real improvement can be seen for intended setpoint S = 358 K. Fig. 10shows that the default controller produces huge overshooting, whereas thenew RL policy stays close to the intended setpoint over the whole batchtime. Again, the RL policy is increasing the setpoint to the maximum at thebeginning just to reduce it to the minimum after 100 s. After 400 s we cansee that the RL policy is starting to slightly increase the setpoint again tocounteract the decrease in reactor temperature at the end of the batch. Thecontrol deviation can be reduced by 33.5 %.

In Fig. 11, we see a similar behavior for intended setpoint S = 362 K.The new RL policy changes the setpoint dynamically over the whole batchtime to reduce control deviation by 37.7 %.

17

Page 18: Trustworthy AI for Process Automation on a Chylla-Haase

Figure 9: Setpoint controls and resulting trajectories for default and GPRLpolicies. Results are shown for an intended reactor setpoint of 352 K.

Figure 10: Setpoint controls and resulting trajectories for default and GPRLpolicies. Results are shown for an intended reactor setpoint of 358 K.

18

Page 19: Trustworthy AI for Process Automation on a Chylla-Haase

Figure 11: Setpoint controls and resulting trajectories for default and GPRLpolicies. Results are shown for an intended reactor setpoint of 362 K.

The final example is intended setpoint S = 365 K in Fig. 12. Here,the new RL policy reduces control deviation by 30.1 %. Note that in thisexample, it is very important not to overshoot too much, since depending onthe product in the reactor, too high temperatures can spoil the process andruin the quality of the final product.

To conclude, we can say that the control deviation has been reduced for allintended setpoints tested. The policy was automatically learned from safelygenerated batch data, easy to implement, and reduced control deviation byup to 37 %.

5 Conclusion

The experiments using GPRL on a SIMATIC PCS 7 polymerization reactortemplate show that it is possible to generate human-interpretable policiesfrom existing process data fully automatically. In contrast to widely-knownRL methods, which produce black-box value functions and/or policies, withGPRL, the domain experts can stay in the optimization loop, since the re-sults are convenient to validate and implement. Utilizing such trustworthyAI methods will help bring state-of-the-art ML algorithms into real-worldindustry applications, to leverage the optimization potentials which are tobe expected in many domains.

19

Page 20: Trustworthy AI for Process Automation on a Chylla-Haase

Figure 12: Setpoint controls and resulting trajectories for default and GPRLpolicies. Results are shown for an intended reactor setpoint of 365 K.

Acknowledgments

The project this report is based on was supported with funds from theGerman Federal Ministry of Education and Research under project num-ber 01IS18049A. The sole responsibility for the report’s contents lies withthe authors.

References

[1] Siemens AG. Pcs 7 unit template ”stirred tank reactor with kalmanfilter” using the example of the chemical industry, 2018. URL https:

//support.industry.siemens.com/cs/de/de/view/109756215.

[2] K.J. Astrom and T. Hagglund. The future of pid control. Controlengineering practice, 9(11):1163–1175, 2001.

[3] K.J. Astrom and T. Hagglund. Revisiting the ziegler–nichols step re-sponse method for pid control. Journal of process control, 14(6):635–650,2004.

[4] S. Bach, A. Binder, G. Montavon, F. Klauschen, K.-R. Muller, andW. Samek. On pixel-wise explanations for non-linear classifier decisionsby layer-wise relevance propagation. PloS one, 10(7):e0130140, 2015.

20

Page 21: Trustworthy AI for Process Automation on a Chylla-Haase

[5] A. Bhat and R.N. Banavar. The chylla-haase problem: a neural networkcontroller. In Proceedings of the 1998 IEEE International Conference onControl Applications (Cat. No. 98CH36104), volume 1, pages 192–196.IEEE, 1998.

[6] B. Bischoff, D. Nguyen-Tuong, T. Koller, H. Markert, and A. Knoll.Learning throttle valve control using policy search. In Joint Euro-pean Conference on Machine Learning and Knowledge Discovery inDatabases, pages 49–64. Springer, 2013.

[7] T. Blickle and L. Thiele. A mathematical analysis of tournament selec-tion. In ICGA, pages 9–16, 1995.

[8] H.S. Chang, J. Hu, M.C. Fu, and S.I. Marcus. Population-based evolu-tionary approaches. In Simulation-Based Algorithms for Markov Deci-sion Processes, chapter 3, pages 61–87. Springer, 2007.

[9] H.H. Chin and A.A. Jafari. Genetic algorithm methods for solving thebest stationary policy of finite Markov decision processes. In SystemTheory, 1998. Proceedings of the Thirtieth Southeastern Symposium on,pages 538–543. IEEE, 1998.

[10] L. Desborough and R. Miller. Increasing customer value of industrialcontrol performance monitoring - Honeywell’s experience. In AIChEsymposium series, number 326, pages 169–189. New York; AmericanInstitute of Chemical Engineers; 1998, 2002.

[11] F. Doshi-Velez and B. Kim. Towards a rigorous science of interpretablemachine learning. arXiv preprint arXiv:1702.08608, 2017.

[12] K.L. Downing. Adaptive genetic programs via reinforcement learning. InProceedings of the 3rd Annual Conference on Genetic and EvolutionaryComputation, GECCO’01, pages 19–26, San Francisco, CA, USA, 2001.Morgan Kaufmann Publishers Inc.

[13] F.-A. Fortin, F.-M. De Rainville, M.-A. Gardner, M. Parizeau, andC. Gagne. Deap: Evolutionary algorithms made easy. The Journalof Machine Learning Research, 13(1):2171–2175, 2012.

[14] C. Gearhart. Genetic programming as policy search in Markov deci-sion processes. In J. R. Koza, editor, Genetic Algorithms and GeneticProgramming at Stanford, pages 61–67. Stanford Bookstore, Stanford,USA, 2003.

21

Page 22: Trustworthy AI for Process Automation on a Chylla-Haase

[15] F. Gomez, J. Schmidhuber, and R. Miikkulainen. Efficient non-linearcontrol through neuroevolution. In European Conference on MachineLearning, pages 654–662. Springer, 2006.

[16] D. Hein. Interpretable Reinforcement Learning Policies by EvolutionaryComputation. PhD thesis, Technische Universitat Munchen, 2019.

[17] D. Hein, S. Udluft, and T.A. Runkler. Interpretable policies for rein-forcement learning by genetic programming. Engineering Applicationsof Artificial Intelligence, 76:158–169, 2018.

[18] D. Hein, S. Limmer, and T.A. Runkler. Interpretable control by rein-forcement learning. IFAC-PapersOnLine, 53(2):8082–8089, 2020. 21thIFAC World Congress.

[19] S. Kamio and H. Iba. Adaptation technique for integrating geneticprogramming and reinforcement learning for real robots. Trans. Evol.Comp, 9(3):318–333, June 2005.

[20] M.A. Keane, J.R. Koza, and M.J. Streeter. Automatic synthesis usinggenetic programming of an improved general-purpose controller for in-dustrially representative plants. In Proceedings of the 2002 NASA/DoDConference on Evolvable Hardware (EH’02), EH ’02, pages 113–123,Washington, DC, USA, 2002. IEEE.

[21] J.R. Koza. Genetic Programming: On the Programming of Computersby Means of Natural Selection. MIT Press, Cambridge, MA, USA, 1992.

[22] J.R. Koza. Human-competitive results produced by genetic program-ming. Genetic Programming and Evolvable Machines, 11(3):251–284,2010.

[23] F. Maes, R. Fonteneau, L. Wehenkel, and D. Ernst. Policy search in aspace of simple closed-form formulas: Towards interpretability of rein-forcement learning. Discovery Science, pages 37–50, 2012.

[24] M.T. Ribeiro, S. Singh, and C. Guestrin. Why should I trust you?:Explaining the predictions of any classifier. In Proceedings of the 22ndACM SIGKDD International Conference on Knowledge Discovery andData Mining, pages 1135–1144. ACM, 2016.

[25] A.M. Schafer and S. Udluft. Solving partially observable reinforcementlearning problems with recurrent neural networks. In Workshop Proc.of the European Conf. on Machine Learning, pages 71–81, 2005.

22

Page 23: Trustworthy AI for Process Automation on a Chylla-Haase

[26] H. Shimooka and Y. Fujimoto. Generating equations with genetic pro-gramming for control of a movable inverted pendulum. In Selected Pa-pers from the Second Asia-Pacific Conference on Simulated Evolutionand Learning on Simulated Evolution and Learning, SEAL’98, pages179–186, London, UK, 1999. Springer-Verlag.

[27] T. Wang, C. Rudin, F. Velez-Doshi, Y. Liu, E. Klampfl, and P. Mac-Neille. Bayesian rule sets for interpretable classification. In Data Min-ing (ICDM), 2016 IEEE 16th International Conference on, pages 1269–1274. IEEE, 2016.

[28] D.G. Wilson, S. Cussat-Blanc, H. Luga, and J.F. Miller. Evolving sim-ple programs for playing Atari games. In Proceedings of the Geneticand Evolutionary Computation Conference, GECCO ’18, pages 229–236,New York, NY, USA, 2018. ACM.

23