14
Applied Mathematical Modelling 56 (2018) 275–288 Contents lists available at ScienceDirect Applied Mathematical Modelling journal homepage: www.elsevier.com/locate/apm Uncertain dynamical system-based decision making with application to production-inventory problems Linxue Sheng a,b , Yuanguo Zhu b,, Kai Wang c a State Key Laboratory of Air Traffic Management System and Technology, The 28th Research Institute of China Electronics Technology Group Corporation, Nanjing, Jiangsu 210014, China b School of Science, Nanjing University of Science and Technology, Nanjing, Jiangsu 210094, China c Science and Technology on Information Systems Engineering Laboratory, The 28th Research Institute of China Electronics Technology Group Corporation, Nanjing, Jiangsu 210007, China a r t i c l e i n f o Article history: Received 8 April 2016 Revised 29 November 2017 Accepted 5 December 2017 Available online 8 December 2017 Keywords: Uncertainty theory Production-inventory model Optimal control Control constraint critical value a b s t r a c t Uncertainty theory is a new methodology to describe subjective indeterminacy. This pa- per discusses an infinite-horizon production-inventory problem with constraints that the production rates shall be restricted to an appropriate interval, and the indeterminacy of the decision system is extracted and analyzed through uncertainty theory. To solve such a problem, an uncertain linear quadratic control model is investigated with a critical value criterion. Applying the equation of optimality, it is shown that an optimal feedback control exists. In addition, the behavior of the value function that is characterized benefits from numerical illustrations. Furthermore, we compare these optimal solutions with the solu- tions of an unconstrained problem. Then some economic interpretations are provided to reveal the significance of the model to practical applications. © 2017 Elsevier Inc. All rights reserved. 1. Introduction Analysis of dynamical systems is a classical work in system sciences, mathematics, and cybernetics. In many real-life decision-making situations, optimizing a utility for a dynamical system is classified as an optimal control problem. However, the implementations of many control systems are always intractable because, in general, real system dynamics will involve some degree of data indeterminacy, and possibly some parameters or external disturbances as well. In fact, indeterminacy is absolute whereas determinacy is relative, thus how to deal with indeterminacy is an essential research subject not only in mathematics, but also in data science and engineering. In probability theory, normally indeterminate or imprecise data could be described as random variables and, hence, when the dynamic control system is considered, the state of the system over time would be a stochastic process, and we will surely be faced with the stochastic optimal control problem. In 1965, Zadeh [1] initiated the concept of the fuzzy set via a membership function. The proposed fuzzy set theory is, in contrast, used to depict vagueness especially when there is data ambiguity as a result of subjective or linguistic judgments. As for fuzzy optimization, fuzziness in train timetable problems [2] has been studied through a fuzzy programming model, and an effective branch-and-bound algorithm has been proposed to obtain robust solutions. The idea of applying fuzzy sets to control problems was explicitly presented by Chang and Zadeh [3] for the first time. Consequently, the most impor- tant research on fuzzy controllers by Mamdani and Assilian [4] and Mamdani [5] marked the start of fuzzy control theory. Corresponding author. E-mail addresses: [email protected], [email protected] (Y. Zhu). https://doi.org/10.1016/j.apm.2017.12.006 0307-904X/© 2017 Elsevier Inc. All rights reserved.

Applied Mathematical Modelling - orsc.edu.cn · Applied Mathematical Modelling 56 (2018) ... over time would be a stochastic process, and we will surely be faced with the stochastic

  • Upload
    ledieu

  • View
    220

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Applied Mathematical Modelling - orsc.edu.cn · Applied Mathematical Modelling 56 (2018) ... over time would be a stochastic process, and we will surely be faced with the stochastic

Applied Mathematical Modelling 56 (2018) 275–288

Contents lists available at ScienceDirect

Applied Mathematical Modelling

journal homepage: www.elsevier.com/locate/apm

Uncertain dynamical system-based decision making with

application to production-inventory problems

Linxue Sheng

a , b , Yuanguo Zhu

b , ∗, Kai Wang

c

a State Key Laboratory of Air Traffic Management System and Technology, The 28th Research Institute of China Electronics Technology

Group Corporation, Nanjing, Jiangsu 210014, China b School of Science, Nanjing University of Science and Technology, Nanjing, Jiangsu 210094, China c Science and Technology on Information Systems Engineering Laboratory, The 28th Research Institute of China Electronics Technology

Group Corporation, Nanjing, Jiangsu 210 0 07, China

a r t i c l e i n f o

Article history:

Received 8 April 2016

Revised 29 November 2017

Accepted 5 December 2017

Available online 8 December 2017

Keywords:

Uncertainty theory

Production-inventory model

Optimal control

Control constraint

critical value

a b s t r a c t

Uncertainty theory is a new methodology to describe subjective indeterminacy. This pa-

per discusses an infinite-horizon production-inventory problem with constraints that the

production rates shall be restricted to an appropriate interval, and the indeterminacy of

the decision system is extracted and analyzed through uncertainty theory. To solve such a

problem, an uncertain linear quadratic control model is investigated with a critical value

criterion. Applying the equation of optimality, it is shown that an optimal feedback control

exists. In addition, the behavior of the value function that is characterized benefits from

numerical illustrations. Furthermore, we compare these optimal solutions with the solu-

tions of an unconstrained problem. Then some economic interpretations are provided to

reveal the significance of the model to practical applications.

© 2017 Elsevier Inc. All rights reserved.

1. Introduction

Analysis of dynamical systems is a classical work in system sciences, mathematics, and cybernetics. In many real-life

decision-making situations, optimizing a utility for a dynamical system is classified as an optimal control problem. However,

the implementations of many control systems are always intractable because, in general, real system dynamics will involve

some degree of data indeterminacy, and possibly some parameters or external disturbances as well. In fact, indeterminacy

is absolute whereas determinacy is relative, thus how to deal with indeterminacy is an essential research subject not only

in mathematics, but also in data science and engineering. In probability theory, normally indeterminate or imprecise data

could be described as random variables and, hence, when the dynamic control system is considered, the state of the system

over time would be a stochastic process, and we will surely be faced with the stochastic optimal control problem.

In 1965, Zadeh [1] initiated the concept of the fuzzy set via a membership function. The proposed fuzzy set theory is, in

contrast, used to depict vagueness especially when there is data ambiguity as a result of subjective or linguistic judgments.

As for fuzzy optimization, fuzziness in train timetable problems [2] has been studied through a fuzzy programming model,

and an effective branch-and-bound algorithm has been proposed to obtain robust solutions. The idea of applying fuzzy sets

to control problems was explicitly presented by Chang and Zadeh [3] for the first time. Consequently, the most impor-

tant research on fuzzy controllers by Mamdani and Assilian [4] and Mamdani [5] marked the start of fuzzy control theory.

∗ Corresponding author.

E-mail addresses: [email protected] , [email protected] (Y. Zhu).

https://doi.org/10.1016/j.apm.2017.12.006

0307-904X/© 2017 Elsevier Inc. All rights reserved.

Page 2: Applied Mathematical Modelling - orsc.edu.cn · Applied Mathematical Modelling 56 (2018) ... over time would be a stochastic process, and we will surely be faced with the stochastic

276 L. Sheng et al. / Applied Mathematical Modelling 56 (2018) 275–288

Recently, some research [6–8] has begun to handle dynamic optimization problems via the expected-value-based fuzzy op-

timal control model.

In fact, the sources of data indeterminacy may be inexhaustible and more complex. While solving practical problems, we

often lack samples or cannot afford numerous experiments to obtain statistical data due to economic reasons or technical

difficulties, leading to failure in building a probability distribution. Instead, some domain experts are invited to evaluate a

belief degree measuring such an indeterminate quantity. A belief degree represents the strength with which we believe the

event will happen. However, people tend to put too much weight on unlikely events. Thus, subjective probability sometimes

fails to model the belief degree unless some observed data are obtained to revise the belief degree. Some theories have

been proposed to deal with the belief degree such as possibility theory [9] and Dempster–Shafer theory [10,11] . To deal

with belief degrees rationally, an uncertainty theory was constructed by Liu [12,13] based on the uncertain measure in

2007, and then became a branch of axiomatic mathematics systems for modeling the indeterminacy. As counterparts of

stochastic processes and Brownian motion, Liu [13] proposed uncertain processes and canonical processes, respectively. The

notion of an uncertain differential equation driven by a canonical process was presented in 2008 [13] . After that, Yao and

Li [14] proposed an uncertain alternating renewal process Gao and Yao [15] studied continuous dependence theorems in

uncertain differential equations. Nowadays, uncertainty theory has become much more mature in a variety of aspects such as

uncertain programming, uncertain statistics, and uncertain propositional logic. In practical applications, Gao et al. [16] were

the first to introduce the uncertainty theory into rail transportation problems.

When a system is affected by uncertain factors, it can be characterized by an uncertain differential equation. To inves-

tigate an uncertain control system that is guided by the uncertain differential equation, Zhu [17] was the first to study

an uncertain optimal control problem in which the expected value operator was used to optimize the uncertain objec-

tive function. The solution methodology was dynamic programming and an equation of optimality as a counterpart of the

Hamilton–Jacobi–Bellman (HJB) equation was provided and subsequently studied by some researchers [18–22] .

From another aspect, with the greater use of methods and results in mathematics and computer science, optimal con-

trol theory has achieved great progress and a variety of practical areas have increased tremendously such as engineering,

medicine, and management science. In particular, production planning problems have emerged as central to controlling the

production management systems in practice. Among them, the quadratic production control models are very important, as

made famous by Holt et al. in 1960 [23] . They considered both production costs and inventory holding costs over time,

and the analytical study was based on a calculus of variations method. Since then, a great number of production-inventory

problems have been investigated by many researchers. A model was proposed and analyzed by applying the maximum prin-

ciple in the work of Thompson and Sethi [24] . In their model, they determined production rates over time to minimize an

integral of a discounted quadratic loss function. The turnpike points were further obtained when the horizon is infinite. In

[25,26] , different types of production planning problems with some constraints have been developed. In addition, Kleindor-

fer et al. [27] solved the discrete version of production planning problems using optimal control theory. Data indeterminacy

is also pervasive in production management issues. For instance, some research into the stochastic version of production-

inventory problems was presented by Sethi and Thompson [28] , Sethi and Zhang [29] , Bensoussan et al. [30] , and van der

Laan and Salomon [31] . An infinite-horizon stochastic production planning problem [30] was considered in which the ad-

missible control must be non-negative. From [29] , it is specified that production planning problems where production is

done by machines that are unreliable or failure-prone could be formulated as stochastic optimal control problems involving

jump Markov processes.

In this article, we consider a production-inventory problem involving uncertain data, with constraints that the production

rates are restricted to an appropriate interval. The infinite-horizon case is primarily treated. As mentioned above, usually the

best suggestion is to maximize or minimize the expected value of the indeterminate objective function. However, owing to

the complexity of the control systems, optimality problems under the customary criterion are not always favorable owing

to the different requirements from people. Parameters of the objective function are not easy to measure so that the object

is difficult to achieve in reality. On the other hand, decision-makers may have different preferences in that some may be

cautious while others may be adventurous. Therefore, in the model formulation, we make use of the critical-value-based

criterion, which was investigated in our previous work, Sheng et al. [32,33] and used successfully in saddle point prob-

lems [34] , switched linear systems [35] , and stability analysis [36] .

It is worth noting that the preceding works on uncertain optimal control are reasonably well behaved with unrestricted

admissible control, except for the bang–bang control problem whose objective function is linear in the state variable. The

problem introduced in this paper could be mathematically abstracted by an uncertain linear quadratic optimal control model

with bounded control. Borrowing ideas from dynamic programming techniques, this kind of production-inventory problem

under uncertainty is solved along with some economic interpretations, including optimal feedback control and the charac-

terization of the value function. Furthermore, some numerical illustrations are offered to demonstrate the efficiency of the

proposed model and the solution procedure.

2. Theoretical background

For reference, we summarize some basic knowledge of uncertainty theory and uncertain optimal control models that will

be frequently used in the following.

Page 3: Applied Mathematical Modelling - orsc.edu.cn · Applied Mathematical Modelling 56 (2018) ... over time would be a stochastic process, and we will surely be faced with the stochastic

L. Sheng et al. / Applied Mathematical Modelling 56 (2018) 275–288 277

2.1. Uncertainty theory

Mathematically, let � be a nonempty set, let L be a σ -algebra over �, and let M be an uncertain measure. Then the

triplet ( �, L , M ) is called an uncertainty space [12] . An uncertain variable is a measurable function from an uncertainty

space ( �, L , M ) to the set of real numbers. The distribution �: R → [0, 1] of an uncertain variable ξ is defined by �(x ) =M{ γ ∈ � | ξ (γ ) ≤ x } for any real number x ; in many cases, it is sufficient to know the uncertainty distribution rather than

the uncertain variable itself.

For an uncertain variable ξ , the expected value of ξ is defined by

E [ ξ ] =

∫ + ∞

0

M { ξ ≥ r } d r −∫ 0

−∞

M { ξ ≤ r } d r

provided that at least one of the two integrals is finite. This expected value represents the size of an uncertain variable.

When dealing with control problems with nondeterministic interference, the expected value is usually used as an effective

measurement to optimize the objective function. In reality, people tend to take into account the possibility of reliability

or risk, favorable or adverse events. For example, for the same investment, there are two return options, 1 million yuan

with reliability 1 or 10 million yuan with reliability 0.1. At this point, investors have a choice: risk-averse investors will

choose low-risk investments, i.e. the first option; risk takers will largely choose the latter option; but for people who do

not care about risk, the options will appear the same. That is, the expectation model is not always applicable in practical

problems. Correspondingly, to handle the uncertain optimal control problems in an alternative manner, critical values have

been designed by Liu based on the uncertain measure. Let β ∈ (0, 1]. Then

ξsup (β) = sup { r|M { ξ ≥ r } ≥ β} , ξinf (β) = inf { r|M { ξ ≤ r } ≥ β} are called the β-optimistic value to ξ and the β-pessimistic value to ξ , respectively. From the definition, we can immediately

infer that ξsup (β) = �−1 (1 − β) and ξinf (β) = �−1 (β) . For instance, assume that ξ is an uncertain income giving a high

confidence level 0.8. Then for each scheme, ξsup (0 . 8) is the same as saying that the uncertain return can achieve up to

ξsup (0 . 8) , i.e. its reliability can reach at least 80%; ξinf (0 . 8) = ξsup (0 . 2) is the same as saying that the uncertain return

can achieve up to ξsup (0 . 2) , but its reliability is only 20%. Thus, when the reliability is high, the optimistic value gains

denote that the decision maker is cautious in the assessment of income to obtain a conservative result; on the other hand,

pessimistic values indicate that the decision maker is willing to take risks in predicting earnings, and may be able to take

greater risks with a high return at the same time.

The uncertain variables ξ1 , . . . , ξn are said to be independent if

M

{

n ⋂

i =1

{ ξ ∈ B i } }

=

n ∧

i =1

M{ ξ ∈ B i }

for any Borel sets B 1 , B 2 , . . . , B n . This means that some uncertain variables are independent if they can be separately defined

on different uncertainty spaces. If ξ and η are independent uncertain variables with finite expected values, then for any real

numbers a and b , we have E [ aξ + bη] = a E [ ξ ] + bE [ η] . Meanwhile, the following properties of critical values hold:

( a ) if a ≥ 0, then ( aξ ) sup (β) = aξsup (β) , and ( aξ ) inf (β) = aξinf (β) ;

( b ) if a < 0, then ( aξ ) sup (β) = aξinf (β) , and ( aξ ) inf (β) = aξsup (β) ;

( c ) ( ξ + η) sup (β) = ξsup (β) + ηsup (β) , ( ξ + η) inf (β) = ξinf (β) + ηinf (β) .

Further, the variation of uncertain phenomena over time is studied by uncertain calculus using the concept of an un-

certain differential equation as a type of differential equation driven by a canonical process. An uncertain process C t is said

to be a canonical Liu process if: (i) C 0 = 0 and almost all sample paths are Lipschitz continuous; (ii) C t has stationary and

independent increments; (iii) every increment C s + t − C s is a normal uncertain variable with expected value 0 and variance

t 2 , whose uncertainty distribution is

�(x ) =

(1 + exp

(−πx √

3 t

))−1

, x ∈ R.

For each γ ∈ �, the function C t ( γ ) is called a sample path of canonical process C t . If f and g are some given functions,

d X t = f (t, X t )d t + g(t, X t )d C t

is called an uncertain differential equation. A solution is a Liu process X t that satisfies the equation identically in t . We

let α be a number with 0 < α < 1. The former uncertain differential equation is said to have an α-path X αt if it solves the

corresponding ordinary differential equation

d X

αt = f (t, X

αt )d t + | g(t, X

αt ) | �−1 (α)d t,

Page 4: Applied Mathematical Modelling - orsc.edu.cn · Applied Mathematical Modelling 56 (2018) ... over time would be a stochastic process, and we will surely be faced with the stochastic

278 L. Sheng et al. / Applied Mathematical Modelling 56 (2018) 275–288

where �−1 (α) is the inverse standard normal uncertainty distribution, i.e. �−1 (α) =

3 π ln

α1 −α . Furthermore, we have

M { X t ≤ X

αt , ∀ t } = α

M { X t > X

αt , ∀ t } = 1 − α,

which could be called the Yao–Chen Formula. Then the solution X t has an inverse uncertainty distribution: �−1 t (α) = X αt .

The Yao–Chen formula relates uncertain differential equations and ordinary differential equations, in the same way that the

Feynman–Kac formula relates stochastic differential equations and partial differential equations.

For more detailed properties and interpretations of uncertainty theory, such as critical values or uncertain differential

equations, the interested reader is referred to Liu [12,13,37] .

2.2. Critical value model of uncertain optimal control

Inspired by the Hurwicz decision criterion, which was designed by Hurwicz [38] in 1951, we composited the optimistic

value criterion and the pessimistic value criterion to generate a new model of uncertain optimal control problems in [33] :⎧ ⎨

J(t, x ) ≡ sup

u ∈ U { ρF sup (α) + (1 − ρ) F inf (β) }

subject to

dX s = b(s, X s , u ) ds + σ(s, X s , u ) dC s and X t = x

(1)

where F =

∫ T t f (s, X s , u ) ds + h (T , X T ) , and J ( t , x ) is called the value function. A selected coefficient ρ ∈ (0, 1) denotes the

optimism degree and we have a predetermined confidence level β ∈ (0, 1). The vector X s and the vector u ∈ U are the state

variable and the control variable, respectively. Assume that C t = (C t1 , C t2 , . . . , C tk ) τ, where C t1 , C t2 , . . . , C tk are independent

canonical processes, and the symbol v τ denotes the transpose of a vector or a matrix v .

To find the best control such that the given objective functional is optimized, we presented the following equation of

optimality.

Theorem 2.1. Suppose J(t, x ) ∈ C 2 ([0 , T ] × R n ) . Then we have

−J t (t, x ) = sup

u ∈ U

{f (t, x , u ) + ∇ x J(t , x ) τ b(t , x , u ) + (2 ρ − 1)

(√

3

πln

1 − β

β

)‖

∇ x J(t , x ) τσ(t , x , u ) ‖ 1

}, (2)

where J t ( t , x ) is the partial derivative of the function J ( t , x ) in t , ∇ x J ( t , x ) is the gradient of J ( t , x ) in x , and ‖ · ‖ 1 is the 1-norm

for vectors, that is, ‖ p‖ 1 =

∑ n i =1 | p i | for p = (p 1 , p 2 , . . . , p n ) .

3. Problem specification

This section is devoted to the model assumptions. We introduce the mathematical description of the production-

inventory system with nondeterminacy items, and we consider the lack of historical data when optimizing production level

and inventories of new products. This case study offers the dual purpose of conducting some verifications of our proposed

uncertain optimal control theory and presenting an important avenue to advance its own extension.

Consider a factory producing homogeneous products that it stores in an inventory warehouse: guided by the previous

work and the practical circumstances, we plug uncertain variables into this dynamic system. Let us define the following

quantities:

• X ( t ), the inventory level at time t (state variable); • P ( t ), the production rate at time t (control variable); • S , the constant demand rate at time t, S > 0; • ˆ X , the factory-optimal inventory level, which can be interpreted as a safety stock; • ˆ P , the factory-optimal production rate, that denotes the most favorable level to run the factory; • h , the inventory holding cost coefficient, h > 0; • c , the production cost coefficient, c > 0; • δ, the constant positive discount rate; • C t , the canonical processes (uncertain variable); • σ , the constant disturbance coefficient.

Note that the control constraint is required, which keeps the production rate P ( t ) between a specified lower bound 0 and

a specified upper bound M . Note further that the inventory level is not always nonnegative, i.e. X ( t ) < 0 cannot denote the

situation such as backlogging of demand. We now present the dynamics of the inventory level described by an uncertain

difference equation

d X (t) = (P (t) − S)d t + σd C t , X (0) = x, (3)

where x denotes the initial inventory level. The expression σd C t is considered as a sale sudden fluctuation of the new

product, such as the sale return, product defect, or inventory spoilage, that can be modeled by some uncertain variables.

Page 5: Applied Mathematical Modelling - orsc.edu.cn · Applied Mathematical Modelling 56 (2018) ... over time would be a stochastic process, and we will surely be faced with the stochastic

L. Sheng et al. / Applied Mathematical Modelling 56 (2018) 275–288 279

Fig. 1. Frequency, probability, and uncertainty [45] .

Next, if the controller has to balance one’s desire to minimize the cost due to the difference between the factory-optimal

levels and the real state values and hold the system under stable control, we adopt a conceptual quadratic objective:

min

0 ≤P(t) ≤M

F x,P(t) = min

0 ≤P(t) ≤M

∫ ∞

0

e −δt [

h

(X (t) − ˆ X

)2 + c (P (t) − ˆ P

)2 ]

d t. (4)

Because the dynamic system operation involves uncertain variables, the crux is how to measure the object converting the

uncertain objective into its crisp equivalent. As many methods are discussed in [39] involving this scheme, we plan to

take advantage of the critical value criterion. The underlying philosophy is that for two uncertain variables ξ and η, we

say that ξ > η if and only if ξsup ( inf ) (β) > ηsup ( inf ) (β) for some predetermined confidence level β . Since there are so many

indeterminate factors in the real world, no decisions can guarantee 100% success rate. Thus, we can make decisions under

some given success rate. Let β ∈ (0, 1), we formulate the objective function

min

0 ≤P(t) ≤M

H

λβ

{F x,P(t)

}= min

0 ≤P(t) ≤M

{λF x,P(t) sup

(β) + (1 − λ) F x,P(t) inf (β)

}, (5)

where λ∈ [0, 1] specifies a selected coefficient that means the optimism degree 1 − λ surely represents a measure of the

controller’s pessimism. With the confidence level β , the uncertain variable F x, P ( t ) will reach a lower bound F x,P(t) inf (β) ;

meanwhile, F x, P ( t ) could reach an upper bound F x,P(t) sup (β) . Then their weighted average need to be minimized in the

presence of an appropriate control policy. Of course, according to the properties of critical values, we have min H

λβ

{F x,P(t)

}=

− max H

1 −λβ

{−F x,P(t)

}.

Here, we emphasize why the uncertainty theory arises in this situation. Several researchers have argued that nonde-

terminacy in production-inventory problems could be modeled by random variables based on probability theory. Recently,

Widyadana and Wee [40] analyzed the effect of machine breakdown in a production-inventory system with stochastic repair

time. Pan and Li [41] employed stochastic optimal control theory to model the deteriorating items in a production-inventory

system with environmental constraints. Other stochastic models could be found, for example, in [42,43] . However, a funda-

mental premise of applying probability theory is that the estimated probability is close enough to the long-run frequency,

no matter whether the probability is interpreted as subjective or objective. Otherwise, the law of large numbers is no longer

valid and probability theory is no longer applicable.

The reasons for a lack of observed data about the unknown state of nature may not only be economic, but also technical.

How can we deal with this case? We have to invite some domain experts to evaluate their belief degree that each event will

occur. Since humans usually give undue weightage to unlikely events [44] , the belief degree may have much larger variance

than the real frequency, therefore we should deal with it by uncertainty theory.

Fig. 1 shows that when the sample size is large enough, the estimated probability (left curve) is close enough to the

cumulative frequency (left histogram) and probability theory is applicable. When the sample size is too small or in the event

there are no samples, the belief degree (right curve) usually has much larger variance than the cumulative frequency (right

histogram: it is actually unknown); in this case, uncertainty theory is the more appropriate approach. For more detailed

interpretations, the interested reader is referred to Liu [45] .

4. Uncertain optimal control approach

The above model is solved via the equation of optimality presented by Sheng and Zhu [33] , which is satisfied by a certain

value function. To simplify the mathematics, we suppose that ˆ X =

ˆ P = 0 and h = c = 1 . The assumption is without loss of

generality as the following analysis can be extended in a parallel manner for the case without this assumption. We can

rewrite F x, P ( t ) in (4) as ∫ ∞

e −δt [X

2 (t) + P 2 (t) ]d t. (6)

0

Page 6: Applied Mathematical Modelling - orsc.edu.cn · Applied Mathematical Modelling 56 (2018) ... over time would be a stochastic process, and we will surely be faced with the stochastic

280 L. Sheng et al. / Applied Mathematical Modelling 56 (2018) 275–288

Define V = V (x ) to represent the current value of the linear quadratic optimal control problem (5), (6) , subject to the

state Eq. (3) . First, inspired by the preceding work and the linearity of the system, we characterize the behavior of V ( x ) as

follows.

Theorem 4.1. The value function V ( x ) is strictly convex.

The proof can be found in the Appendix.

Subsequently, we choose V ( x ) in strictly convex form with quadratic growth. According to the equation of optimality, V ( x )

satisfies

SV x (x ) + δV (x ) + σ (1 − 2 λ)

(√

3

πln

1 − β

β

)| V x (x ) | = x 2 + min

0 ≤P≤M

(P 2 + P V x (x )) . (7)

In [33] , we called this flexible criterion the Hurwicz criterion. In particular, by setting the coefficient λ to 1 and 0, the

criterion reduces to the optimistic value criterion and the pessimistic value criterion, respectively; if λ = 0 . 5 , the above

equation becomes the same as the result in the case of the expected model [17] .

First, we give the unconstrained solutions P ∗, V

∗. By an argument similar to the proof of Theorem 4.1 , V

∗ is strictly convex,

and we have to solve

SV

∗x (x ) + δV

∗(x ) + ρ| V

∗x (x ) | = x 2 + min (P 2 + P V

∗x (x )) ,

where ρ = σ (1 − 2 λ)( √

3 π ln

1 −ββ

) , and let ρ < S . Setting d(P 2 + P V ∗x ) / d P = 0 , we obtain P ∗ = −V ∗x / 2 . It follows from the con-

vexity that there exists a unique x ′ such that V ∗x (x ′ ) = 0 , and V

∗( x ) satisfies

(S + ρ) V

∗x + δV

∗ = x 2 − V

∗2 x

4

, x ≥ x ′ and V

∗x ≥ 0 ; (8a)

(S − ρ) V

∗x + δV

∗ = x 2 − V

∗2 x

4

, x ≤ x ′ and V

∗x ≤ 0 . (8b)

For these two equations, taking derivatives in x on both sides, we obtain (S + ρ +

V

∗x

2

)V

∗xx = 2 x − δV

∗x , x ≥ x ′ ;(

S − ρ +

V

∗x

2

)V

∗xx = 2 x − δV

∗x , x ≤ x ′ .

Since V xx is positive, we infer that x ′ ≥ 0 and V

∗ is quadratic growth for x ≥ x ′ . However, towards the case x ≤ x ′ , the

general solutions of the differential equation cannot be guaranteed to represent quadratic growth. Here, we could suppose

that V ∗(x ) = ax 2 + bx + c, where a, b , and c are real numbers. Substituting the expression into (8b) and collecting the terms,

we obtain

[ a 2 + aδ − 1] x 2 + [2 a (S − ρ) + ab + bδ] x + [ b(S − ρ) + b 2 / 4 + cδ] = 0 ,

which must hold for any value of x ≤ x ′ , so we must have

a 2 + aδ − 1 = 0 ,

2 a (S − ρ) + ab + bδ = 0 ,

b(S − ρ) + b 2 / 4 + cδ = 0 .

We obtain

a =

δ2 + 4 − δ

2

> 0 , b = −2 a 2 (S − ρ) , c =

−b 2 − 4 b(S − ρ)

4 δ,

and 2 ax ′ + b = 0 , which yields x ′ = (S − ρ)( √

δ2 + 4 − δ) / 2 . Further, V

∗ is quadratic growth. For x ≥ x ′ , V

∗ satisfies an Abelian

equation that only has an implicit solution. Fortunately, it can be solved numerically to express its manner. For numerical

illustration, we set δ = 0 . 2 , S = 2 , β = 0 . 98 , σ = 0 . 7 , and λ = 0 . 6 to generate Fig. 2 .

Note that, in this section, we do not restrict the production rate to be nonnegative or below some upper bound. In other

words, we permit disposal and unlimited production capacity, for mathematical expedience.

In the next section, we discuss the solutions of the optimal control model (3) –(6) . As defined previously, V ( x ) is the value

function for the uncertain optimal control problem. We restate it as

SV x + δV + ρ| V x | = x 2 + min

0 ≤P≤M

(P 2 + P V x ) .

Because of the strict convexity of V , the optimal feedback control (denoted by P x ) is given by

P x =

{

0 , if V x ≥ 0 ;−V x / 2 , if − 2 M ≤ V x ≤ 0 ;

M, if V x ≤ −2 M,

or P x =

{

0 , if x ≥ x 0 ;−V x / 2 , if x m

≤ x ≤ x 0 ;M, if x ≤ x m

.

(10)

Page 7: Applied Mathematical Modelling - orsc.edu.cn · Applied Mathematical Modelling 56 (2018) ... over time would be a stochastic process, and we will surely be faced with the stochastic

L. Sheng et al. / Applied Mathematical Modelling 56 (2018) 275–288 281

Fig. 2. Behavior of V ∗ .

In the above formulation, x 0 and x m

are separately the unique roots of V x (x 0 ) = 0 and V x (x m

) = −2 M. We emphasize that

the switching points x 0 and x m

and the optimal objective value V are all M -dependent.

(I) We first verify the conclusion under the case that M > (S − ρ) .

For x ≥ x 0 , V satisfies the differential equation (S + ρ) V x + δV = x 2 , and its general quadratic growth solution is

V (x ) =

1

δx 2 − 2(S + ρ)

δ2 x +

2(S + ρ) 2

δ3 + c̄ · e −

δx (S+ ρ) , (11)

since V x (x 0 ) = 0 , c̄ is determined as

c̄ =

[2(S + ρ)

δ2 x 0 − 2(S + ρ) 2

δ3

]e

δx 0 (S+ ρ) .

Similarly, for x ≤ x m

we have (S − ρ − M) V x + δV = x 2 + M

2 . Then

V (x ) =

1

δx 2 − 2(S − ρ − M)

δ2 x +

2(S − ρ − M) 2

δ3 +

M

2

δ+

¯̄c · e −δx

(S−ρ−M) , (12)

and

¯̄c =

[2(S − ρ − M)

δ2 x m

− 2(S − ρ − M) 2

δ3 +

2 M(S − ρ − M)

δ

]e

δx m (S−ρ−M) .

Following from the condition V xx > 0, we can infer that x 0 > 0 and x m

< −δM. When x takes the value between the switching

points x m

and x 0 , the solution of (7) is also considered as follows,

V (x ) =

˜ a x 2 + ̃

b x +

˜ c , (13)

where ˜ a , ˜ b , and ˜ c are real numbers. Then V x = 2 ̃ a x + ̃

b , x 0 = −˜ b / 2 ̃ a , and x m

= (−2 M − ˜ b ) / ̃ a . We have calculated that ˜ a = a,˜ b = b, and ˜ c = c. In addition, it is easily verified that

x 0 = (S − ρ)

δ2 + 4 − δ

2

> 0 , x m

=

(S − ρ − M)( √

δ2 + 4 − δ)

2

− δM < −δM.

To summarize the points that we have just made, a strictly convex solution V ( x ) with quadratic growth has been obtained

in a piecewise manner.

To visualize the results, we adopt the parameters that have been specified in the unconstrained case. In addition, we give

M = 3 . Fig. 3 shows different initial state levels and their corresponding optimal objective values.

As displayed in Fig. 3 , the solid line denotes the value function that meets Eq. (7) ; the dashed line V

∗ represents the

unconstrained value function; the dashed lines V and V m

, respectively, indicate the value functions for production-inventory

0
Page 8: Applied Mathematical Modelling - orsc.edu.cn · Applied Mathematical Modelling 56 (2018) ... over time would be a stochastic process, and we will surely be faced with the stochastic

282 L. Sheng et al. / Applied Mathematical Modelling 56 (2018) 275–288

Fig. 3. Behavior of V , with M = 3 .

problem with P ≡ 0 and P ≡ M = 3 . When x takes values between x 0 and x m

, we can note that V ∗(x ) = V (x ) . Further, from

the quadratic growth and the observation of the behavior of V , we may infer that

V → V 0 as x → + ∞; V → V m

as x → −∞ .

Intuitively, the justification of this result is that for large inventory levels, it is naturally optimal to produce nothing, i.e. set

P = 0 . Similarly, if there are large-scale shortages, it is optimal to produce at the highest possible rate, i.e. set P = M.

(II) When M ≤ (S − ρ) is required, we perform the following analysis.

For x ≥ x 0 and x ≤ x m

, the value functions have the same structures as (11) and (12) , respectively. However, since (S − ρ −M) ≥ 0 , if the quadratic growth of V ( x ) is to be guaranteed, we have to set ¯̄c = 0 . Therefore, for x ≤ x m

,

V (x ) =

1

δx 2 − 2(S − ρ − M)

δ2 x +

2(S − ρ − M) 2

δ3 +

M

2

δ, (14)

and x m

= (S − ρ − M) /δ − δM. For x m

≤ x ≤ x 0 , the function V ( x ) satisfies

(S − ρ) V x + δV = x 2 − V

2

4

, V x (x m

) = −2 M,

and x 0 results from V x (x 0 ) = 0 . Yet this differential equation only has an implicit solution. This situation looks similar to

the unconstrained case in that we must resort to numerical execution for the problem. Here, we still take the previous

parameters and set M = 1 , and we obtain the value function presented in Fig. 4 .

Referring to the computation illustrated in Fig. 4 , x m

= 3 . 3 can be calculated directly, and via the numerical execution

we find that x 0 takes a value of about 3.639. Further, the value-function whose dependent variable acts as x m

≤ x ≤ x 0 (the

dash–dotted line) can be bound by the quadratic growth functions.

At the end of this section, we present the results of some numerical experiments to illustrate the advantage of adopting

the critical value criterion, rather than the expected value criterion. In this case, the parameters are initialized as: M = 3 ,

δ = 0 . 2 , S = 2 , β = 0 . 9 , σ = 0 . 02 , λ = 0 (that means the controller chooses the pessimistic criterion), and X(0) = 1 . Since

the uncertain variable is equipped with an uncertainty space (�, L , M ) , in which some sample paths can be realized due

to several γ ∈ �.

As displayed in Fig. 5 , we compare these realization points of uncertain cost (based on the sample path of uncertain

variables) with the pessimistic value of 0.9 and the realization points of uncertain cost and its expected value. It can be

observed that the minimal total abatement cost is larger with the pessimistic criterion than with the expected criterion.

Although the minimal expected cost is optimal, the effective cost realized can be far from this value,whereas the pessimistic

cost may be easily reached. Minimizing the pessimistic cost is cautious to some extent: it actually provides the least bad

cost with a belief degree of 0.90. However, in this problem, the target is not only to reduce the cost, but also to stabilize

the inventory level to satisfy the demand and other suddenly incomplete predicted situations. From this perspective, using

the critical value criterion may be more flexible and realistic.

Page 9: Applied Mathematical Modelling - orsc.edu.cn · Applied Mathematical Modelling 56 (2018) ... over time would be a stochastic process, and we will surely be faced with the stochastic

L. Sheng et al. / Applied Mathematical Modelling 56 (2018) 275–288 283

Fig. 4. Behavior of V , with M = 1 .

Fig. 5. Discounted costs realizations.

Remark 4.1. Since the optimal control gives a piecewise continuous form, it is difficult to substitute the optimal control in

Eq. (3) to obtain the optimal state and the closed-form expressions for the optimal value functions. In this case, we decided

to discretize the dynamical system in time. We consider a given time interval �t , and define a recursive equation

X (i · �t) = X ((i − 1) · �t) + [ P ((i − 1) · �t) − S]�t + σηi , (15)

where i = 1 , 2 , . . . denotes an index, ηi is the realization of the uncertain variable �C t , following from the definition of a

canonical process, �C t is a normal uncertain variable with expected value 0 and variance �t 2 , and the uncertainty distribu-

tion is denoted by �. Thus, ηi can be produced by ηi = �−1 (r i ) for an arbitrarily generated number r i from the interval [0,

1]. In addition, the selected planning period N ·�t can ensure the difference (between the optimal value in N ·�t and the

optimal value in (N − 1) · �t approaching zero.

Page 10: Applied Mathematical Modelling - orsc.edu.cn · Applied Mathematical Modelling 56 (2018) ... over time would be a stochastic process, and we will surely be faced with the stochastic

284 L. Sheng et al. / Applied Mathematical Modelling 56 (2018) 275–288

5. Discussion and extension

5.1. Discussion

Recall the figures that were presented in the previous section. For the value function V , we now rewrite it as V

M . From

the previous figures, for the same given x we may infer the following result.

Theorem 5.1. Let M > N ≥ 0 . Then 0 ≤ V

∗( x ) ≤ V

M ( x ) ≤ V

N ( x ) ≤ V 0 ( x ) .

Proof. In this section, only the case M > (S − ρ) ≥ N ≥ 0 is considered. Other cases can be easily verified in an analogous

way.

(I) 0 ≤ V

∗ ≤ V

M . From the previous section, we have x ′ = x 0 and V ∗(x ) = V M (x ) if x 0 ≤ x ≤ x m

. Let ˜ V M = V M − V ∗, clearly ˜ V M

is quadratic growth. For x ≥ x 0 , there exists

(S + ρ) ̃ V

M

x + δ ˜ V

M = h (x ) ≥ 0 , ˜ V

M

x (x 0 ) = 0 ,

where h (x ) = − min −∞≤P≤+ ∞

(P 2 + P V ∗x ) . The solution of this initial-value problem can be expressed as

˜ V

M = e −δx

(S+ ρ)

[∫ x

x 0

h (z)

(S + ρ) e

δz (S+ ρ) d z

].

Because of h (z) ≥ 0 ⇒

h (z) (S+ ρ)

e δz

(S+ ρ) ≥ 0 , we have that ∫ h (z)

(S+ ρ) e

δz (S+ ρ) d z is an increasing function with respect to z ( x ≥ x 0 ).

Hence, ˜ V M ≥ 0 ⇒ V M ≥ V ∗. Likewise, when x ≤ x m

, V

M ≥ V

∗ could also be derived. Eventually, we should point out that in x ′ (i.e. x 0 ), V

∗ has the minimum value x ′ 2 / δ ≥ 0.

(II) V

M ≤ V

N . In this section, the minimal points for V

M and V

N are denoted by x M

0 and x N

0 , respectively. Optimal feedback

control is separated as P M

and P N . Here x M

0 ≤ x N

0 ( x M

0 > x N

0 can be analyzed in the same manner). Define ˜ V N = V N − V M , there

are three cases: for x ≥ x N 0 , we have

(S + ρ) ̃ V

N x + δ ˜ V

N = 0 , ˜ V

N x (x N 0 ) ≤ 0 ,

the general solution of which is given by ˜ V N = ̃

k e − δx

(S+ ρ) . Following the initial-value condition, ˜ k ≥ 0 . Then

˜ V N ≥ 0 ⇒ V N ≥ V M .

For x M

0 ≤ x ≤ x N

0 , we have

(S − ρ) ̃ V

N x + δ ˜ V

N ≥ −ρ(V

N x + V

M

x ) + S(V

N x − V

M

x ) + δ(V

N − V

M )

= min

0 ≤P≤N (P 2 + P V

N x ) − min

0 ≤P≤M

(P 2 + P V

M

x )

≥ P 2 N + P N V

N x − P 2 N − P N V

M

x

= P N ̃ V

N x .

Therefore, (S − ρ − P N ) ̃ V N x + δ ˜ V N ≥ 0 with (S − ρ − P N ) > 0 since P N ≤ N ≤ (S − ρ) . Returning to the analysis in (I), we obtain

V

N ≥ V

M . Where x ≤ x M

0 , the situation is similar to the case of x M

0 ≤ x ≤ x N

0 .

(III) V

N ≤ V 0 . For this comparison, we summarize that with the methods used in (I) and (II), the result is easy to prove.

We do not repeat the procedure. �

5.2. Extension

Here we have studied an infinite-horizon production planning problem in an uncertain environment. If this problem

needs to be optimized over a fixed period of time T , such a complex issue will not be easy to analyze with the same

method. Since the optimal production rate depends on not only the level of inventory x , but also the distance from the

current time s to the end time T , the value function would be indicated by V ( s, x ). However, an extension to the infinite case

might be natural. Consider

V (s, x ) = min

0 ≤P(t) ≤M

{λF T sup (β) + (1 − λ) F T inf (β)

},

subject to dynamic system (1), where F T =

∫ T s

[X(t) 2 + P (t) 2

]d t and M > (S − ρ) . The value function V ( s, x ) obeys the equa-

tion of optimality

−V s (s, x ) + SV x (s, x ) + ρ| V x (s, x ) | = x 2 + min

0 ≤P≤M

(P 2 + P V x (s, x )) .

Suppose −2 M ≤ V x (s, x ) ≤ 0 , we have optimal control P s,x = −V x (s, x ) / 2 . For a finite T , in the plane ( s, x ) there are two loci

(separately denoted by x m

( s ) and x 0 ( s )), along which the optimal control will switch from the boundary value M to values in

the interior of the interval [0, M ], and then change from the interior, i.e. −V x / 2 , to the other boundary 0. When the optimal

control takes values between 0 and M , the equation of optimality becomes

−V s (s, x ) + (S − ρ) V x (s, x ) = x 2 − V

2 x (s, x )

4

Page 11: Applied Mathematical Modelling - orsc.edu.cn · Applied Mathematical Modelling 56 (2018) ... over time would be a stochastic process, and we will surely be faced with the stochastic

L. Sheng et al. / Applied Mathematical Modelling 56 (2018) 275–288 285

and the expressions of those loci can be determined as

V x (s, x m

) = −2 M and V x (s, x 0 ) = 0 .

Let V (s, x ) = A (s ) x 2 + B (s ) x + D (s ) , where A ( s ), B ( s ) and D ( s ) are deterministic functions of s . Following from this form, we

have V s =

d A (s ) d s

x 2 +

d B (s ) d s

x +

d D (s ) d s

and V x = 2 A (s ) x + B (s ) . Substituting these into the equation of optimality above, we calcu-

late functions A ( s ), B ( s ), and D ( s ). Then

x m

=

2e T + s (S − ρ) + (e 2 T + e 2 s )(M − S + ρ)

(e 2 T − e 2 s ) , x 0 =

(e T − e s )(S − ρ)

(e T + e s ) .

6. Conclusion

In this paper, we have explored a production-inventory problem in an uncertain environment with bounded production

rates. An uncertain optimal control model with Hurwicz criterion has been formulated. Drawing support from the equa-

tion of optimality, we have provided the optimal feedback policy P x and characterized the value function V ( x ), which we

then simulated successfully. The complexity of the real world makes the control systems that we face take numerous dif-

ferent forms. The essential difference between uncertain optimal control and stochastic optimal control is that the former

is concerned with the study of dynamic uncertain phenomena while the latter is about the study of dynamic stochastic

phenomena. Uncertainty is regarded as subjective indeterminacy, and it is estimated by experts when there is no or very

limited historical data on hand. This means that our production-inventory model will be useful when there is a lack of

historical data, such as the problems of production-inventory management for new products. On the contrary, stochastic

planning model may be applied when the managers have enough data. In other words, the two types are, respectively, suit-

able for different cases, and it may be unreasonable to evaluate stochastic model and uncertain model by using the same

criteria because they are proposed for different purposes.

We now discuss the limitations of the proposed approach: (1) this paper discusses an infinite-horizon production-

inventory problem. If this problem needs to be optimized over a fixed period of time, such a complex issue will not be

easy to be analyzed with the same method. Even though we give an extended result of such case, it is necessary to further

develop the complete process of calculation and proof; (2) With respect to the uncertainty distribution of an uncertain pro-

cess, we note that, for a fixed time moment, it must be estimated by domain experts. In our work, uncertainty distributions

are assumed to be known just as probability distributions are assumed to be known in a stochastic production-inventory

model. This presents the practitioner with the challenge to estimate the uncertain difference equation, which describes the

dynamics of the inventory level.

In future research, the authors are intending to continue investigating optimal control for uncertain systems and more

practical application aspects, such as: (1) an uncertain production planning problems with the constraints on both the inven-

tory level and the production rate; (2) Uncertain optimal control for partially observable nonlinear systems; (3) Maximum

principle approach of uncertain optimal control theory.

Acknowledgments

This work is supported by the National Natural Science Foundation of China (No. 61673011 ).

Appendix

Let us give a proof of Theorem 4.1 .

Proof. Firstly, we prove that H

λβ

{F x,P(t)

}is strictly convex in x and P ( · ). Assume x 1 and x 2 , P 1 ( · ) and P 2 ( · ) be arbitrarily and

differently chosen. Let θ ∈ (0, 1). Define

x θ = θx 1 + (1 − θ ) x 2 , and P θ = θP 1 + (1 − θ ) P 2 ,

thus the solutions of (1) with different initial conditions

X 1 (t) = x 1 +

∫ t

0

(P 1 − S)d s +

∫ t

0

σd C 1 s ,

X 2 (t) = x 2 +

∫ t

0

(P 2 − S)d s +

∫ t

0

σd C 2 s ,

X θ (t) = x θ +

∫ t

0

(P θ − S)d s +

∫ t

0

σd C θs ,

Page 12: Applied Mathematical Modelling - orsc.edu.cn · Applied Mathematical Modelling 56 (2018) ... over time would be a stochastic process, and we will surely be faced with the stochastic

286 L. Sheng et al. / Applied Mathematical Modelling 56 (2018) 275–288

where C 1 s , C 2 s and C θs are treated as identically distributed independence canonical processes. It follows from the properties

of α-path, we obtain

X

α1 (t) = x 1 +

∫ t

0

(P 1 − S)d s + σ

∫ t

0

�−1 (α)d s,

X

α2 (t) = x 2 +

∫ t

0

(P 2 − S)d s + σ

∫ t

0

�−1 (α)d s,

X

αθ (t) = x θ +

∫ t

0

(P θ − S)d s + σ

∫ t

0

�−1 (α)d s.

Then, the α-path X αθ(t) can be rewritten as

X

αθ (t) = θX

α1 (t) + (1 − θ ) X

α2 (t) = [ θX 1 (t) + (1 − θ ) X 2 (t) ]

α.

It means that X θ ( t ) and [ θX 1 (t) + (1 − θ ) X 2 (t) ] have the same uncertainty distribution. Further, if we respectively denote

uncertainty distributions of X 2 θ(t) and [ θX 1 (t) + (1 − θ ) X 2 (t) ]

2 by � t and � ′

t , we have �t = � ′ t . The detail calculations can

be found in Zhu [46] . Consider the critical values of uncertain variables introduced in the second section, we obtain [X

2 θ (t)

]inf

(α) =

[[ θX 1 (t) + (1 − θ ) X 2 (t) ]

2 ]

inf (α) .

Since

[ θX 1 (t) + (1 − θ ) X 2 (t) ] 2 = θ2 X

2 1 (t) + (1 − θ )

2 X

2 2 (t) + 2 θ (1 − θ ) X 1 (t) X 2 (t)

<

[θ2 + θ (1 − θ )

]X

2 1 (t) +

[(1 − θ )

2 + θ (1 − θ ) ]X

2 2 (t)

= θX

2 1 (t) + (1 − θ ) X

2 2 (t) ,

similarly, we also obtain P 2 θ

< θP 2 1

+ (1 − θ ) P 2 2

.

Drawing support from the critical value criterion, we have [θX

2 1 (t)

]inf

(α) +

[(1 − θ ) X

2 2 (t)

]inf

(α) >

[[ θX 1 (t) + (1 − θ ) X 2 (t) ]

2 ]

inf (α) =

[X

2 θ (t)

]inf

(α) .

If we use ϕ1, t and ϕ2, t to represent the uncertainty distribution of X 2 1 (t) and X 2

2 (t) , respectively. Then

θϕ

−1 1 ,t (α) + (1 − θ ) ϕ

−1 2 ,t (α) > ϕ

−1 θ,t

(α) .

For {F x θ ,P θ (t)

}inf

(β) =

{

∫ ∞

0

e −δt [X

2 θ (t) + P 2 θ (t)

]d t

}

inf

(β) ,

due to {

∫ ∞

0

e −δt [X

2 θ (t) + P 2 θ (t)

]d t ≤

∫ ∞

0

e −δt [ϕ

−1 θ,t

(β) + P 2 θ (t) ]d t

}

⊃{

X

2 θ (t) ≤ ϕ

−1 θ,t

(β) },

we can obtain

M

{

∫ ∞

0

e −δt [X

2 θ (t) + P 2 θ (t)

]d t ≤

∫ ∞

0

e −δt [ϕ

−1 θ,t

(β) + P 2 θ (t) ]d t

}

≥ M

{X

2 θ (t) ≤ ϕ

−1 θ,t

(β) }

= β.

Similarly,

M

{

∫ ∞

0

e −δt [X

2 θ (t) + P 2 θ (t)

]d t >

∫ ∞

0

e −δt [ϕ

−1 θ,t

(β) + P 2 θ (t) ]d t

}

≥ M

{X

2 θ (t) > ϕ

−1 θ,t

(β) }

= 1 − β.

Following the axiom of self-duality, we have

M

{

∫ ∞

0

e −δt [X

2 θ (t) + P 2 θ (t)

]d t ≤

∫ ∞

0

e −δt [ϕ

−1 θ,t

(β) + P 2 θ (t) ]d t

}

= β.

Therefore, the time integral

∫ ∞

0 e −δt

[X 2 θ (t) + P 2 θ (t)

]d t has an inverse uncertainty distribution

�−1 (β) =

∫ ∞

e −δt [ϕ

−1 θ,t

(β) + P 2 θ (t) ]d t,

0

Page 13: Applied Mathematical Modelling - orsc.edu.cn · Applied Mathematical Modelling 56 (2018) ... over time would be a stochastic process, and we will surely be faced with the stochastic

L. Sheng et al. / Applied Mathematical Modelling 56 (2018) 275–288 287

and then according to the property of the critical value, which has been mentioned in Section 2 , we obtain {F x θ ,P θ (t)

}inf

(β) = �−1 (β) =

{

∫ ∞

0

e −δt [ϕ

−1 θ,t

(β) + P 2 θ (t) ]d t

}

.

Similarly, {F x 1 ,P 1 (t)

}inf

(β) =

{

∫ ∞

0

e −δt [ϕ

−1 1 ,t (β) + P 2 1 (t)

]d t

}

,

{F x 2 ,P 2 (t)

}inf

(β) =

{

∫ ∞

0

e −δt [ϕ

−1 2 ,t (β) + P 2 2 (t)

]d t

}

.

From the above, {F x θ ,P θ (t)

}inf

(β)

=

{

∫ ∞

0

e −δt [ϕ

−1 θ,t

(β) + P 2 θ (t) ]d t

}

<

{

∫ ∞

0

e −δt [θϕ

−1 1 ,t (β) + (1 − θ ) ϕ

−1 2 ,t (β) + P 2 θ (t)

]d t

}

< θ{

∫ ∞

0

e −δt [ϕ

−1 1 ,t (β) + P 2 1 (t)

]d t

}

+ (1 − θ ) {

∫ ∞

0

e −δt [ϕ

−1 2 ,t (β) + P 2 2 (t)

]d t

}

= θ{

F x 1 ,P 1 (t)

}inf

(β) + (1 − θ ) {

F x 2 ,P 2 (t)

}inf

(β) . (A.1)

With the same derivation method, we can also get {F x θ ,P θ (t)

}sup

(β) < θ{

F x 1 ,P 1 (t)

}sup

(β) + (1 − θ ) {

F x 2 ,P 2 (t)

}sup

(β) . (A.2)

Now, synthesize Eq. (A.1) and Eq. (A.2) , we obtain

H

λβ

{F x θ ,P θ (t)

}= λ

{F x θ ,P θ (t)

}sup

(β) + (1 − λ) {

F x θ ,P θ (t)

}inf

(β)

<

{

λθ{

F x 1 ,P 1 (t)

}sup

(β) + (1 − λ) θ{

F x 1 ,P 1 (t)

}inf

(β)) }

+

{

λ(1 − θ ) {

F x 2 ,P 2 (t)

}sup

(β) + (1 − λ)(1 − θ ) {

F x 2 ,P 2 (t)

}inf

(β)) }

= θH

λβ

{F x 1 ,P 1 (t)

}+ (1 − θ ) H

λβ

{F x 2 ,P 2 (t)

}.

Next, let P ∗1 (t) and P ∗

2 (t) denote the optimal controls for initial states x 1 and x 2 , respectively. Therefore

V (x θ ) = V (θx 1 + (1 − θ ) x 2 ) = min

0 ≤P(t) ≤M

H

λβ

{F x θ ,P(t)

}≤ H

λβ

{F x θ ,θP ∗

1 (t)+(1 −θ ) P ∗

2 (t)

}< θH

λβ

{F x 1 ,P ∗1 (t)

}+ (1 − θ ) H

λβ

{F x 2 ,P ∗2 (t)

}= θV (x 1 ) + (1 − θ ) V (x 2 ) .

This proves the strict convexity of V ( x ). �

References

[1] L.A. Zadeh , Fuzzy sets, Inf. Control 8 (1965) 338–353 . [2] L. Yang , P. Li , Y. Gao , Train timetable problem on a single-line railway with fuzzy passenger demand, IEEE Trans. Fuzzy Syst. 17 (2009) 617–629 .

[3] S.S. Chang , L.A. Zadeh , On fuzzy mapping and control, IEEE Trans. Syst. Man Cybern. 2 (1972) 30–34 . [4] E.H. Mamdani , S. Assilian , An experiment in linguistic synthesis with a fuzzy logic controller, Int. J. Man-Mach. Stud. 7 (1975) 1–13 .

[5] E.H. Mamdani , Advances in the linguistic synthesis of fuzzy controllers, Int. J. Man-Mach. Stud. 8 (1976) 669–678 .

[6] Y. Zhu , Fuzzy optimal control for multistage fuzzy systems, IEEE Trans. Syst. Man Cybern. B Cybern. 41 (2011) 964–975 . [7] B. Liu , Y. Liu , Expected value of fuzzy variable and fuzzy expected value models, IEEE Trans. Fuzzy Syst. 10 (2002) 445–450 .

[8] J. Liu , Y. Zhu , Multi-dimensional multistage fuzzy optimal control, in: Proceedings of the Eighth International Conference on Fuzzy System and Knowl-edge Discovery, FSKD, Shanghai, China, 2011, pp. 188–191 .

[9] L.A. Zadeh , Fuzzy sets as the basis for a theory of possibility, Fuzzy Set. Syst. 1 (1978) 3–28 . [10] A.P. Dempster , Upper and lower probabilities induced by a multivalued mapping, Ann. Math. Stat. 38 (1967) 325–339 .

[11] G. Shafer , A Mathematical Theory of Evidence, Princeton University Press, New Jersey, 1976 .

[12] B. Liu , Uncertainty Theory, second ed., Springer-Verlag, Berlin, 2007 . [13] B. Liu , Fuzzy process, hybrid process and uncertain process, J. Uncertain Syst. 2 (2008) 3–16 .

[14] K. Yao , X. Li , Uncertain alternating renewal process and its application, IEEE Trans. Fuzzy Syst. 20 (2012) 1154–1160 . [15] Y. Gao , K. Yao , Continuous dependence theorems on solutions of uncertain differential equations, Appl. Math. Model. 38 (2014) 3031–3037 .

[16] Y. Gao , L. Yang , S. Li , Uncertain models on railway transportation planning problem, Appl. Math. Model. 40 (2016) 4 921–4 934 . [17] Y. Zhu , Uncertain optimal control with application to a portfolio selection model, Cybern. Syst. 41 (2010) 535–547 .

Page 14: Applied Mathematical Modelling - orsc.edu.cn · Applied Mathematical Modelling 56 (2018) ... over time would be a stochastic process, and we will surely be faced with the stochastic

288 L. Sheng et al. / Applied Mathematical Modelling 56 (2018) 275–288

[18] X. Xu , Y. Zhu , Uncertain bang–bang control for continuous time model, Cybern. Syst. 43 (2012) 515–527 . [19] Y. Yan , Y. Zhu , Bang–bang control model for uncertain switched systems, Appl. Math. Model. 39 (2015) 2994–3002 .

[20] L. Deng , Y. Zhu , Uncertain optimal control with jump, ICIC Express Lett. B 3 (2012) 419–424 . [21] R. Chen , Y. Zhu , An optimal control model for uncertain systems with time-delay, Oper. Res. Soc. Jpn. 56 (2013) 243–256 .

[22] X. Yang , J. Gao , Uncertain differential games with application to capitalism, J. Uncertain. Anal. Appl. 1 (2013) 1–11 . [23] C.C. Holt , F. Modigliani , J.F. Muth , H.A. Simon , Planning Production, Inventories and Work Force, Prentice Hall, Englewood Cliffs, New Jersey, 1960 .

[24] G.L. Thompson , S.P. Sethi , Turnpike horizons for production planning, Manag. Sci. 26 (1980) 229–241 .

[25] G. Yuan , Two-stage fuzzy production planning expected value model and its approximation method, Appl. Math. Model. 36 (2012) 2429–2445 . [26] M. Rong , M. Maiti , On an EOQ model with service level constraint under fuzzy-stochastic demand and variable lead-time, Appl. Math. Model. 39

(2015) 5230–5240 . [27] P.R. Kleidorfer , C.H. Kriebels , G.L. Thompson , G.G. Kleindorfer , Discrete optimal control of production plans, Manag. Sci. 22 (1975) 261–273 .

[28] S.P. Sethi , G.L. Thompson , Optimal Control Theory: Applications to Management Science and Economics, second ed., Springer-Verlag, 20 0 0 . [29] S.P. Sethi , Q. Zhang , Asymptotic optimal controls in stochastic manufacturing systems with machine failures dependent on production rates, Stochast.

Stochast. Rep. 48 (1994) 97–121 . [30] A. Bensoussan , S.P. Sethi , R. Vickson , N. Dewko , Stochastic production planning with production constraints, SIAM J. Control Optim. 22 (1984) 920–935 .

[31] E. van der Laan , M. Salomon , Production planning and inventory control with remanufacturing and disposal, Eur. J. Oper. Res. 102 (1997) 264–278 .

[32] L. Sheng , Y. Zhu , Optimistic value model of uncertain optimal control, Int. J. Uncertain. Fuzziness 21 (2013) 75–87 . [33] L. Sheng , Y. Zhu , T. Hamalainen , An uncertain optimal control model with Hurwicz criterion, Appl. Math. Comput. 224 (2013) 412–421 .

[34] Y. Sun , Y. Zhu , Bang–bang property for an uncertain saddle point problem, J. Intell. Manuf. 28 (2017) 605–613 . [35] H. Yan , Y. Zhu , Bang–bang control model with optimistic value criterion for uncertain switched systems, J. Intell. Manuf. 28 (2017) 527–534 .

[36] N. Tao , Y. Zhu , Stability and attractivity in optimistic value for dynamical systems with uncertainty, Int. J. Gen. Syst. 45 (2016) 418–433 . [37] B. Liu , Uncertainty Theory: A Branch of Mathematics for Modeling Human Uncertainty, Springer-Verlag, Berlin, 2010 .

[38] L. Hurwicz , Some specification problems and application to econometric models, Econometrica 19 (1951) 343–344 .

[39] B. Liu , Theory and Practice of Uncertain Programming, Springer-Verlag, Berlin, 2009 . [40] G.A. Widyadana , H.M. Wee , Optimal deteriorating items production inventory models with random machine breakdown and stochastic repair time,

Appl. Math. Model. 35 (2011) 3495–3508 . [41] X. Pan , S. Li , Optimal control of a stochastic production-inventory system under deteriorating items and environmental constraints, Int. J. Prod. Res. 53

(2015) 607–628 . [42] H.J. Lin , Two-critical-number control policy for a stochastic production inventory system with partial backlogging, Int. J. Prod. Res. 55 (2017)

4123–4135 .

[43] J.P. Gayon , S. Vercraene , S.D.P. Flapper , Optimal control of a production-inventory system with product returns and two disposal options, Eur. J. Oper.Res. 262 (2017) 499–508 .

[44] A. Tversky , D. Kahneman , Rational choice and the framing of decisions, J. Bus. 59 (1986) 251–278 . [45] B. Liu , Uncertainty Theory, fourth ed., Springer-Verlag, Berlin, 2015 .

[46] Y. Zhu , Functions of uncertain variables and uncertain programming, J. Uncertain Syst. 6 (2012) 278–288 .