204
Neurodynamic Optimization: Models and Applications Jun Wang [email protected] Department of Mechanical & Automation Engineering The Chinese University of Hong Kong Shatin, New Territories, Hong Kong http://www.mae.cuhk.edu.hk/ ˜ jwang Computational Intelligence Laboratory, CUHK – p. 1/14

Neurodynamic Optimization: Models and Applications

  • Upload
    others

  • View
    35

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Neurodynamic Optimization: Models and Applications

Neurodynamic Optimization:Models and Applications

Jun [email protected]

Department of Mechanical & Automation Engineering

The Chinese University of Hong Kong

Shatin, New Territories, Hong Kong

http://www.mae.cuhk.edu.hk/jwang

Computational Intelligence Laboratory, CUHK – p. 1/145

Page 2: Neurodynamic Optimization: Models and Applications

Outline• Introduction

Computational Intelligence Laboratory, CUHK – p. 2/145

Page 3: Neurodynamic Optimization: Models and Applications

Outline• Introduction• Problem Formulation

Computational Intelligence Laboratory, CUHK – p. 2/145

Page 4: Neurodynamic Optimization: Models and Applications

Outline• Introduction• Problem Formulation• Dynamic Optimization

Computational Intelligence Laboratory, CUHK – p. 2/145

Page 5: Neurodynamic Optimization: Models and Applications

Outline• Introduction• Problem Formulation• Dynamic Optimization• Exiting Approaches

Computational Intelligence Laboratory, CUHK – p. 2/145

Page 6: Neurodynamic Optimization: Models and Applications

Outline• Introduction• Problem Formulation• Dynamic Optimization• Exiting Approaches• Neurodynamic Models

Computational Intelligence Laboratory, CUHK – p. 2/145

Page 7: Neurodynamic Optimization: Models and Applications

Outline• Introduction• Problem Formulation• Dynamic Optimization• Exiting Approaches• Neurodynamic Models• Design Procedure

Computational Intelligence Laboratory, CUHK – p. 2/145

Page 8: Neurodynamic Optimization: Models and Applications

Outline• Introduction• Problem Formulation• Dynamic Optimization• Exiting Approaches• Neurodynamic Models• Design Procedure• Winners Take All

Computational Intelligence Laboratory, CUHK – p. 2/145

Page 9: Neurodynamic Optimization: Models and Applications

Outline• Introduction• Problem Formulation• Dynamic Optimization• Exiting Approaches• Neurodynamic Models• Design Procedure• Winners Take All• Linear Assignment

Computational Intelligence Laboratory, CUHK – p. 2/145

Page 10: Neurodynamic Optimization: Models and Applications

Outline• Introduction• Problem Formulation• Dynamic Optimization• Exiting Approaches• Neurodynamic Models• Design Procedure• Winners Take All• Linear Assignment• Machine Learning

Computational Intelligence Laboratory, CUHK – p. 2/145

Page 11: Neurodynamic Optimization: Models and Applications

Outline• Introduction• Problem Formulation• Dynamic Optimization• Exiting Approaches• Neurodynamic Models• Design Procedure• Winners Take All• Linear Assignment• Machine Learning• Robot Control

Computational Intelligence Laboratory, CUHK – p. 2/145

Page 12: Neurodynamic Optimization: Models and Applications

Outline• Introduction• Problem Formulation• Dynamic Optimization• Exiting Approaches• Neurodynamic Models• Design Procedure• Winners Take All• Linear Assignment• Machine Learning• Robot Control• Concluding Remarks

Computational Intelligence Laboratory, CUHK – p. 2/145

Page 13: Neurodynamic Optimization: Models and Applications

Outline• Introduction• Problem Formulation• Dynamic Optimization• Exiting Approaches• Neurodynamic Models• Design Procedure• Winners Take All• Linear Assignment• Machine Learning• Robot Control• Concluding Remarks• Future Works Computational Intelligence Laboratory, CUHK – p. 2/145

Page 14: Neurodynamic Optimization: Models and Applications

IntroductionOptimization is ubiquitous in nature and society.

Computational Intelligence Laboratory, CUHK – p. 3/145

Page 15: Neurodynamic Optimization: Models and Applications

IntroductionOptimization is ubiquitous in nature and society.

Optimization arises in a wide variety of scientificproblems.

Computational Intelligence Laboratory, CUHK – p. 3/145

Page 16: Neurodynamic Optimization: Models and Applications

IntroductionOptimization is ubiquitous in nature and society.

Optimization arises in a wide variety of scientificproblems.

Optimization is an important tool for design,planning, control, operation, and management ofengineering systems.

Computational Intelligence Laboratory, CUHK – p. 3/145

Page 17: Neurodynamic Optimization: Models and Applications

Problem FormulationConsider a general optimization problem:

OP1 : Minimize f(x)

subject to c(x) ≤ 0,

d(x) = 0,

wherex ∈ ℜn is the vector of decision variables,f(x)

is an objective function,c(x) = [c1(x), . . . , cm(x)]T isa vector-valued function, andd(x) = [d1(x), . . . , dp(x)]T a vector-valued function.

Computational Intelligence Laboratory, CUHK – p. 4/145

Page 18: Neurodynamic Optimization: Models and Applications

Problem FormulationConsider a general optimization problem:

OP1 : Minimize f(x)

subject to c(x) ≤ 0,

d(x) = 0,

wherex ∈ ℜn is the vector of decision variables,f(x)

is an objective function,c(x) = [c1(x), . . . , cm(x)]T isa vector-valued function, andd(x) = [d1(x), . . . , dp(x)]T a vector-valued function.If f(x) andc(x) are convex andd(x) is affine, thenOP is a convex programming problem CP. Otherwise,it is a nonconvex program.

Computational Intelligence Laboratory, CUHK – p. 4/145

Page 19: Neurodynamic Optimization: Models and Applications

Quadratic Programs

QP1 : minimize1

2xTQx + qTx

subject to Ax = b,

l ≤ Cx ≤ h,

whereQ ∈ ℜn×n , q ∈ ℜn, A ∈ ℜm×n,b ∈ ℜm, C ∈ ℜn×n, l ∈ ℜn, h ∈ ℜn.

Computational Intelligence Laboratory, CUHK – p. 5/145

Page 20: Neurodynamic Optimization: Models and Applications

Quadratic Programs

QP1 : minimize1

2xTQx + qTx

subject to Ax = b,

l ≤ Cx ≤ h,

whereQ ∈ ℜn×n , q ∈ ℜn, A ∈ ℜm×n,b ∈ ℜm, C ∈ ℜn×n, l ∈ ℜn, h ∈ ℜn.Whenl = 0, h = ∞, C = I, QP1 becomes a standardQP:

QP2 : minimize1

2xTQx + qTx

subject to Ax = b, x ≥ 0Computational Intelligence Laboratory, CUHK – p. 5/145

Page 21: Neurodynamic Optimization: Models and Applications

Linear ProgramsWhenQ = 0, andC = I, QP1 becomes a linearprogram with bound constraints:

LP1 : minimize qTx

subject to Ax = b,

l ≤ x ≤ h

Computational Intelligence Laboratory, CUHK – p. 6/145

Page 22: Neurodynamic Optimization: Models and Applications

Linear ProgramsWhenQ = 0, andC = I, QP1 becomes a linearprogram with bound constraints:

LP1 : minimize qTx

subject to Ax = b,

l ≤ x ≤ h

In addition, whenl = 0, andh = +∞, LP1 becomes astandard linear program:

LP2 : minimize qTx

subject to Ax = b,

x ≥ 0

Computational Intelligence Laboratory, CUHK – p. 6/145

Page 23: Neurodynamic Optimization: Models and Applications

Dynamic OptimizationIn many applications (e.g., online pattern recognition,robot motion control, and onboard signal processing),real-time solutions to optimization problems arenecessary or desirable.

Computational Intelligence Laboratory, CUHK – p. 7/145

Page 24: Neurodynamic Optimization: Models and Applications

Dynamic OptimizationIn many applications (e.g., online pattern recognition,robot motion control, and onboard signal processing),real-time solutions to optimization problems arenecessary or desirable.For such applications, classical optimizationtechniques may not be competent due to the problemdimensionality and stringent requirement oncomputational time.

Computational Intelligence Laboratory, CUHK – p. 7/145

Page 25: Neurodynamic Optimization: Models and Applications

Dynamic OptimizationIn many applications (e.g., online pattern recognition,robot motion control, and onboard signal processing),real-time solutions to optimization problems arenecessary or desirable.For such applications, classical optimizationtechniques may not be competent due to the problemdimensionality and stringent requirement oncomputational time.It is computationally challenging when optimizationprocedures have to be performed in real time tooptimize the performance of dynamical systems.

Computational Intelligence Laboratory, CUHK – p. 7/145

Page 26: Neurodynamic Optimization: Models and Applications

Dynamic OptimizationIn many applications (e.g., online pattern recognition,robot motion control, and onboard signal processing),real-time solutions to optimization problems arenecessary or desirable.For such applications, classical optimizationtechniques may not be competent due to the problemdimensionality and stringent requirement oncomputational time.It is computationally challenging when optimizationprocedures have to be performed in real time tooptimize the performance of dynamical systems.One very promising approach to dynamicoptimization is to apply artificial neural networks.

Computational Intelligence Laboratory, CUHK – p. 7/145

Page 27: Neurodynamic Optimization: Models and Applications

Neurodynamic OptimizationBecause of the inherent nature of parallel anddistributed information processing in neural networks,the convergence rate of the solution process is notdecreasing as the size of the problem increases.

Computational Intelligence Laboratory, CUHK – p. 8/145

Page 28: Neurodynamic Optimization: Models and Applications

Neurodynamic OptimizationBecause of the inherent nature of parallel anddistributed information processing in neural networks,the convergence rate of the solution process is notdecreasing as the size of the problem increases.

Neural networks can be implemented physically indesignated hardware such as ASICs whereoptimization is carried out in a truly parallel anddistributed manner.

Computational Intelligence Laboratory, CUHK – p. 8/145

Page 29: Neurodynamic Optimization: Models and Applications

Neurodynamic OptimizationBecause of the inherent nature of parallel anddistributed information processing in neural networks,the convergence rate of the solution process is notdecreasing as the size of the problem increases.

Neural networks can be implemented physically indesignated hardware such as ASICs whereoptimization is carried out in a truly parallel anddistributed manner.

This feature is particularly desirable for dynamicoptimization in decentralized decision-makingsituations.

Computational Intelligence Laboratory, CUHK – p. 8/145

Page 30: Neurodynamic Optimization: Models and Applications

Existing ApproachesIn their seminal work, Tank and Hopfield (1985,1986) applied the Hopfield networks for solving alinear program and the traveling salesman problem.

Computational Intelligence Laboratory, CUHK – p. 9/145

Page 31: Neurodynamic Optimization: Models and Applications

Existing ApproachesIn their seminal work, Tank and Hopfield (1985,1986) applied the Hopfield networks for solving alinear program and the traveling salesman problem.

Kennedy and Chua (1988) developed a neural networkfor nonlinear programming, which contains finitepenalty parameters and thus its equilibrium pointscorrespond to approximate optimal solutions only.

Computational Intelligence Laboratory, CUHK – p. 9/145

Page 32: Neurodynamic Optimization: Models and Applications

Existing ApproachesIn their seminal work, Tank and Hopfield (1985,1986) applied the Hopfield networks for solving alinear program and the traveling salesman problem.

Kennedy and Chua (1988) developed a neural networkfor nonlinear programming, which contains finitepenalty parameters and thus its equilibrium pointscorrespond to approximate optimal solutions only.

The two-phase optimization networks by Maa andShanblatt (1992).

Computational Intelligence Laboratory, CUHK – p. 9/145

Page 33: Neurodynamic Optimization: Models and Applications

Existing ApproachesIn their seminal work, Tank and Hopfield (1985,1986) applied the Hopfield networks for solving alinear program and the traveling salesman problem.

Kennedy and Chua (1988) developed a neural networkfor nonlinear programming, which contains finitepenalty parameters and thus its equilibrium pointscorrespond to approximate optimal solutions only.

The two-phase optimization networks by Maa andShanblatt (1992).

The Lagrangian networks for quadratic programmingby Zhang and Constantinides (1992) and Zhang, et al.(1992).

Computational Intelligence Laboratory, CUHK – p. 9/145

Page 34: Neurodynamic Optimization: Models and Applications

Existing Approaches (cont’d)A recurrent neural network for quadratic optimizationwith bounded variables only by Bouzerdoum andPattison (1993).

Computational Intelligence Laboratory, CUHK – p. 10/145

Page 35: Neurodynamic Optimization: Models and Applications

Existing Approaches (cont’d)A recurrent neural network for quadratic optimizationwith bounded variables only by Bouzerdoum andPattison (1993).

The deterministic annealing network for linear andconvex programming by Wang (1993, 1994).

Computational Intelligence Laboratory, CUHK – p. 10/145

Page 36: Neurodynamic Optimization: Models and Applications

Existing Approaches (cont’d)A recurrent neural network for quadratic optimizationwith bounded variables only by Bouzerdoum andPattison (1993).

The deterministic annealing network for linear andconvex programming by Wang (1993, 1994).

The primal-dual networks for linear and quadraticprogramming by Xia (1996, 1997).

Computational Intelligence Laboratory, CUHK – p. 10/145

Page 37: Neurodynamic Optimization: Models and Applications

Existing Approaches (cont’d)A recurrent neural network for quadratic optimizationwith bounded variables only by Bouzerdoum andPattison (1993).

The deterministic annealing network for linear andconvex programming by Wang (1993, 1994).

The primal-dual networks for linear and quadraticprogramming by Xia (1996, 1997).

The projection networks for solving projectionequations, constrained optimization, etc by Xia andWang (1998, 2002, 2004) and Liang and Wang(2000).

Computational Intelligence Laboratory, CUHK – p. 10/145

Page 38: Neurodynamic Optimization: Models and Applications

Existing Approaches (cont’d)The dual networks for quadratic programming by Xiaand Wang (2001), Zhang and Wang (2002).

Computational Intelligence Laboratory, CUHK – p. 11/145

Page 39: Neurodynamic Optimization: Models and Applications

Existing Approaches (cont’d)The dual networks for quadratic programming by Xiaand Wang (2001), Zhang and Wang (2002).

A two-layer network for convex programming subjectto nonlinear inequality constraints by Xia and Wang(2004).

Computational Intelligence Laboratory, CUHK – p. 11/145

Page 40: Neurodynamic Optimization: Models and Applications

Existing Approaches (cont’d)The dual networks for quadratic programming by Xiaand Wang (2001), Zhang and Wang (2002).

A two-layer network for convex programming subjectto nonlinear inequality constraints by Xia and Wang(2004).

A simplified dual network for quadratic programmingby Liu and Wang (2006)

Computational Intelligence Laboratory, CUHK – p. 11/145

Page 41: Neurodynamic Optimization: Models and Applications

Existing Approaches (cont’d)The dual networks for quadratic programming by Xiaand Wang (2001), Zhang and Wang (2002).

A two-layer network for convex programming subjectto nonlinear inequality constraints by Xia and Wang(2004).

A simplified dual network for quadratic programmingby Liu and Wang (2006)

Two one-layer networks with discontinuous activationfunctions for linear and quadratic programming byLiu and Wang (2007).

Computational Intelligence Laboratory, CUHK – p. 11/145

Page 42: Neurodynamic Optimization: Models and Applications

General Design ProcedureA design procedure begins with a given objectivefunction and constraint(s).

The next step involves the derivation of aneurodynamic equation which prescribes the motionof the activation states of the neural network.

Computational Intelligence Laboratory, CUHK – p. 12/145

Page 43: Neurodynamic Optimization: Models and Applications

General Design ProcedureA design procedure begins with a given objectivefunction and constraint(s).

The next step involves the derivation of aneurodynamic equation which prescribes the motionof the activation states of the neural network.

The derivation of a neurodynamic equation is crucialfor success of the neural network approach tooptimization.

Computational Intelligence Laboratory, CUHK – p. 12/145

Page 44: Neurodynamic Optimization: Models and Applications

General Design ProcedureA design procedure begins with a given objectivefunction and constraint(s).

The next step involves the derivation of aneurodynamic equation which prescribes the motionof the activation states of the neural network.

The derivation of a neurodynamic equation is crucialfor success of the neural network approach tooptimization.

A properly derived neurodynamic equation can ensurethat the state of neural network reaches an equilibriumand the equilibrium satisfies the constraints andoptimizes the objective function.

Computational Intelligence Laboratory, CUHK – p. 12/145

Page 45: Neurodynamic Optimization: Models and Applications

General Design Procedure(cont’d)In general, there are two approaches to designneurodynamic equations.

Computational Intelligence Laboratory, CUHK – p. 13/145

Page 46: Neurodynamic Optimization: Models and Applications

General Design Procedure(cont’d)In general, there are two approaches to designneurodynamic equations.

The first approach is based on an defined energyfunction.

Computational Intelligence Laboratory, CUHK – p. 13/145

Page 47: Neurodynamic Optimization: Models and Applications

General Design Procedure(cont’d)In general, there are two approaches to designneurodynamic equations.

The first approach is based on an defined energyfunction.

The second approach is based on the existingoptimality conditions.

Computational Intelligence Laboratory, CUHK – p. 13/145

Page 48: Neurodynamic Optimization: Models and Applications

General Design ProcedureThe first approach starts with the formulation of anenergy function based on a given objective functionand constraints

Computational Intelligence Laboratory, CUHK – p. 14/145

Page 49: Neurodynamic Optimization: Models and Applications

General Design ProcedureThe first approach starts with the formulation of anenergy function based on a given objective functionand constraints

It plays an important role in neurodynamicoptimization.

Computational Intelligence Laboratory, CUHK – p. 14/145

Page 50: Neurodynamic Optimization: Models and Applications

General Design ProcedureThe first approach starts with the formulation of anenergy function based on a given objective functionand constraints

It plays an important role in neurodynamicoptimization.

Ideally, the minimum of a formulated energy functioncorresponds to the optimal solution of the originaloptimization problem.

Computational Intelligence Laboratory, CUHK – p. 14/145

Page 51: Neurodynamic Optimization: Models and Applications

General Design ProcedureThe first approach starts with the formulation of anenergy function based on a given objective functionand constraints

It plays an important role in neurodynamicoptimization.

Ideally, the minimum of a formulated energy functioncorresponds to the optimal solution of the originaloptimization problem.

For constrained optimization, the minimum of theenergy function has to satisfy a set of constraints.

Computational Intelligence Laboratory, CUHK – p. 14/145

Page 52: Neurodynamic Optimization: Models and Applications

General Design Procedure(cont’d)The majority of the existing approaches formulates anenergy function by incorporating objective functionand constraints through functional transformation andnumerical weighting.

Computational Intelligence Laboratory, CUHK – p. 15/145

Page 53: Neurodynamic Optimization: Models and Applications

General Design Procedure(cont’d)The majority of the existing approaches formulates anenergy function by incorporating objective functionand constraints through functional transformation andnumerical weighting.

Functional transformation is usually used to convertconstraints to a penalty function to penalize theviolation of constraints; e.g.,p(x) =

∑mi=1[−ci(x)]+2 +

∑pj=1[dj(x)]2, where

[y]+ = max0, y.

Computational Intelligence Laboratory, CUHK – p. 15/145

Page 54: Neurodynamic Optimization: Models and Applications

General Design Procedure(cont’d)The majority of the existing approaches formulates anenergy function by incorporating objective functionand constraints through functional transformation andnumerical weighting.

Functional transformation is usually used to convertconstraints to a penalty function to penalize theviolation of constraints; e.g.,p(x) =

∑mi=1[−ci(x)]+2 +

∑pj=1[dj(x)]2, where

[y]+ = max0, y.

Numerical weighting is often used to balanceconstraint satisfaction and objective optimization;e.g.,E(x) = f(x) + wp(x) wherew is a positiveweight. Computational Intelligence Laboratory, CUHK – p. 15/145

Page 55: Neurodynamic Optimization: Models and Applications

General Design Procedure(cont’d)The majority of the existing approaches formulates anenergy function by incorporating objective functionand constraints through functional transformation andnumerical weighting.

Functional transformation is usually used to convertconstraints to a penalty function to penalize theviolation of constraints; e.g.,p(x) =

∑mi=1[−ci(x)]+2 +

∑pj=1[dj(x)]2, where

[y]+ = max0, y.

Numerical weighting is often used to balanceconstraint satisfaction and objective optimization;e.g.,E(x) = f(x) + wp(x) wherew is a positiveweight. Computational Intelligence Laboratory, CUHK – p. 15/145

Page 56: Neurodynamic Optimization: Models and Applications

General Design Procedure(cont’d)Neurodynamic equations are usually derived as thenegative gradient of the energy function:

dx(t)

dt∝ −∇E(x(t)).

Computational Intelligence Laboratory, CUHK – p. 16/145

Page 57: Neurodynamic Optimization: Models and Applications

General Design Procedure(cont’d)Neurodynamic equations are usually derived as thenegative gradient of the energy function:

dx(t)

dt∝ −∇E(x(t)).

If the enegery function is bounded blow, the stabilityof the neurodynamics can be ensured.

Computational Intelligence Laboratory, CUHK – p. 16/145

Page 58: Neurodynamic Optimization: Models and Applications

General Design Procedure(cont’d)Second approach: Neurodynamic equations of somerecent neural networks for optimization are derivedbased on optimality conditions (e.g.,Karush-Kuhn-Tucker condition) and projectionequations.

Computational Intelligence Laboratory, CUHK – p. 17/145

Page 59: Neurodynamic Optimization: Models and Applications

General Design Procedure(cont’d)Second approach: Neurodynamic equations of somerecent neural networks for optimization are derivedbased on optimality conditions (e.g.,Karush-Kuhn-Tucker condition) and projectionequations.

Stability analysis is needed explicitly to ensure thethat resulting neural network is stable.

Computational Intelligence Laboratory, CUHK – p. 17/145

Page 60: Neurodynamic Optimization: Models and Applications

General Design Procedure(cont’d)Second approach: Neurodynamic equations of somerecent neural networks for optimization are derivedbased on optimality conditions (e.g.,Karush-Kuhn-Tucker condition) and projectionequations.

Stability analysis is needed explicitly to ensure thethat resulting neural network is stable.

All equilibria of a stable neural network satisfy theoptimality condition.

Computational Intelligence Laboratory, CUHK – p. 17/145

Page 61: Neurodynamic Optimization: Models and Applications

General Design Procedure(cont’d)Second approach: Neurodynamic equations of somerecent neural networks for optimization are derivedbased on optimality conditions (e.g.,Karush-Kuhn-Tucker condition) and projectionequations.

Stability analysis is needed explicitly to ensure thethat resulting neural network is stable.

All equilibria of a stable neural network satisfy theoptimality condition.

If the problem is a convex program, an equilibriumpoint represents an optimal solution.

Computational Intelligence Laboratory, CUHK – p. 17/145

Page 62: Neurodynamic Optimization: Models and Applications

General Design Procedure(cont’d)The next step is to determine the architecture of theneural network in terms of the neurons andconnections based on the derived dynamical equation.

Computational Intelligence Laboratory, CUHK – p. 18/145

Page 63: Neurodynamic Optimization: Models and Applications

General Design Procedure(cont’d)The next step is to determine the architecture of theneural network in terms of the neurons andconnections based on the derived dynamical equation.

An activation function models importantcharacteristics of a neuron.

Computational Intelligence Laboratory, CUHK – p. 18/145

Page 64: Neurodynamic Optimization: Models and Applications

General Design Procedure(cont’d)The next step is to determine the architecture of theneural network in terms of the neurons andconnections based on the derived dynamical equation.

An activation function models importantcharacteristics of a neuron.

The range of an activation function usually prescribesthe the state space of the neural network.

Computational Intelligence Laboratory, CUHK – p. 18/145

Page 65: Neurodynamic Optimization: Models and Applications

General Design Procedure(cont’d)The next step is to determine the architecture of theneural network in terms of the neurons andconnections based on the derived dynamical equation.

An activation function models importantcharacteristics of a neuron.

The range of an activation function usually prescribesthe the state space of the neural network.

The activation function depends on the feasible regiondelimited by the constraints.

Computational Intelligence Laboratory, CUHK – p. 18/145

Page 66: Neurodynamic Optimization: Models and Applications

General Design Procedure(cont’d)The next step is to determine the architecture of theneural network in terms of the neurons andconnections based on the derived dynamical equation.

An activation function models importantcharacteristics of a neuron.

The range of an activation function usually prescribesthe the state space of the neural network.

The activation function depends on the feasible regiondelimited by the constraints.

Specifically, it is necessary for the state space toinclude the feasible region.

Computational Intelligence Laboratory, CUHK – p. 18/145

Page 67: Neurodynamic Optimization: Models and Applications

General Design Procedure(cont’d)Any explicit bounds on decision variables can berealized by properly selecting the range of activationfunctions.

Computational Intelligence Laboratory, CUHK – p. 19/145

Page 68: Neurodynamic Optimization: Models and Applications

General Design Procedure(cont’d)Any explicit bounds on decision variables can berealized by properly selecting the range of activationfunctions.

If the gradient-based method is adopted in derivingthe dynamical equation, then the convex energyfunction results in an increasing activation function.

Computational Intelligence Laboratory, CUHK – p. 19/145

Page 69: Neurodynamic Optimization: Models and Applications

General Design Procedure(cont’d)Any explicit bounds on decision variables can berealized by properly selecting the range of activationfunctions.

If the gradient-based method is adopted in derivingthe dynamical equation, then the convex energyfunction results in an increasing activation function.

Precisely, if the steepest descent method is used, theactivation function is equal to the derivative of theenergy function.

Computational Intelligence Laboratory, CUHK – p. 19/145

Page 70: Neurodynamic Optimization: Models and Applications

General Design Procedure(cont’d)Any explicit bounds on decision variables can berealized by properly selecting the range of activationfunctions.

If the gradient-based method is adopted in derivingthe dynamical equation, then the convex energyfunction results in an increasing activation function.

Precisely, if the steepest descent method is used, theactivation function is equal to the derivative of theenergy function.

The last step is usually devoted to simulation to testthe performance of the neural network numerically orphysically.

Computational Intelligence Laboratory, CUHK – p. 19/145

Page 71: Neurodynamic Optimization: Models and Applications

Kennedy-Chua NetworkThe Kennedy-Chua network for solving OPa:

ǫdx

dt= −∇f(x)−s ·h(c(x))T∇c(x)−s ·d(x)T∇d(x),

whereǫ > 0 is a scaling parameter,x ∈ ℜn is the statevector,s > 0 is a penalty parameter,h(r) = (h(r1), ..., h(rn))T , andh(ri) = max0, ri.

aM. P. Kennedy and L. O. Chua, “Neural networks for nonlinear programming,”IEEE Trans-

actions on Circuits and Systems, vol. 35, no. 5, pp. 554–562, May 1988.

Computational Intelligence Laboratory, CUHK – p. 20/145

Page 72: Neurodynamic Optimization: Models and Applications

Kennedy-Chua NetworkThe Kennedy-Chua network for solving OPa:

ǫdx

dt= −∇f(x)−s ·h(c(x))T∇c(x)−s ·d(x)T∇d(x),

whereǫ > 0 is a scaling parameter,x ∈ ℜn is the statevector,s > 0 is a penalty parameter,h(r) = (h(r1), ..., h(rn))T , andh(ri) = max0, ri.

With a finite penalty parameters, the network isglobally convergent to a near-optimal solution to anOP even though CP.

aM. P. Kennedy and L. O. Chua, “Neural networks for nonlinear programming,”IEEE Trans-

actions on Circuits and Systems, vol. 35, no. 5, pp. 554–562, May 1988.

Computational Intelligence Laboratory, CUHK – p. 20/145

Page 73: Neurodynamic Optimization: Models and Applications

Deterministic Annealing Net-workThe deterministic annealing network for solving OPa:

ǫdx

dt= −T (t)∇f(x)−h(c(x))T∇c(x)−d(x)T∇d(x),

whereǫ > 0 is a scaling parameter,x ∈ ℜn is the statevector,T (t) ≥ 0 is a temperature parameter,h(r) = (h(r1), ..., h(rn))T , andh(ri) = max0, ri.

aJ. Wang, “A deterministic annealing neural network for convex programming,”Neural Net-

works, vol. 7, no. 4, pp. 629-641, 1994.

Computational Intelligence Laboratory, CUHK – p. 21/145

Page 74: Neurodynamic Optimization: Models and Applications

Deterministic Annealing Net-workThe deterministic annealing network for solving OPa:

ǫdx

dt= −T (t)∇f(x)−h(c(x))T∇c(x)−d(x)T∇d(x),

whereǫ > 0 is a scaling parameter,x ∈ ℜn is the statevector,T (t) ≥ 0 is a temperature parameter,h(r) = (h(r1), ..., h(rn))T , andh(ri) = max0, ri.If limt→∞ T (t) = 0, then the network is globallyconvergent to a feasible near-optimal solution to CP.

aJ. Wang, “A deterministic annealing neural network for convex programming,”Neural Net-

works, vol. 7, no. 4, pp. 629-641, 1994.

Computational Intelligence Laboratory, CUHK – p. 21/145

Page 75: Neurodynamic Optimization: Models and Applications

Deterministic Annealing Net-workThe deterministic annealing network for solving OPa:

ǫdx

dt= −T (t)∇f(x)−h(c(x))T∇c(x)−d(x)T∇d(x),

whereǫ > 0 is a scaling parameter,x ∈ ℜn is the statevector,T (t) ≥ 0 is a temperature parameter,h(r) = (h(r1), ..., h(rn))T , andh(ri) = max0, ri.If limt→∞ T (t) = 0, then the network is globallyconvergent to a feasible near-optimal solution to CP.If T (t) decreases gradually to 0, then the network isglobally convergent to an optimal solution to CP.

aJ. Wang, “A deterministic annealing neural network for convex programming,”Neural Net-

works, vol. 7, no. 4, pp. 629-641, 1994.

Computational Intelligence Laboratory, CUHK – p. 21/145

Page 76: Neurodynamic Optimization: Models and Applications

Primal-Dual NetworkThe primal-dual network for solving LP2a:

ǫdx

dt= −(qTx − bTy)q − AT (Ax − b) + x+,

ǫdy

dt= −(qTx − bTy)b,

whereǫ > 0 is a scaling parameter,x ∈ ℜn is theprimal state vector,y ∈ ℜm is the dual (hidden) statevector,x+ = (x+

1 ), ..., x+n )T , andx+

i = max0, xi.

aY. Xia, “A new neural network for solving linear and quadratic programming problems,”

IEEE Transactions on Neural Networks, vol. 7, no. 6, 1544-1548, 1996.

Computational Intelligence Laboratory, CUHK – p. 22/145

Page 77: Neurodynamic Optimization: Models and Applications

Primal-Dual NetworkThe primal-dual network for solving LP2a:

ǫdx

dt= −(qTx − bTy)q − AT (Ax − b) + x+,

ǫdy

dt= −(qTx − bTy)b,

whereǫ > 0 is a scaling parameter,x ∈ ℜn is theprimal state vector,y ∈ ℜm is the dual (hidden) statevector,x+ = (x+

1 ), ..., x+n )T , andx+

i = max0, xi.The network is globally convergent to an optimalsolution to LP1.

aY. Xia, “A new neural network for solving linear and quadratic programming problems,”

IEEE Transactions on Neural Networks, vol. 7, no. 6, 1544-1548, 1996.

Computational Intelligence Laboratory, CUHK – p. 22/145

Page 78: Neurodynamic Optimization: Models and Applications

Lagrangian Network for QPIf C = 0 in QP1:

ǫd

dt

(

x

y

)

=

(

−Qx(t) − ATy(t) − q,

Ax − b

)

.

whereǫ > 0, x ∈ ℜn, y ∈ ℜm.

It is globally exponentially convergent to the optimalsolutiona.

aJ. Wang, Q. Hu, and D. Jiang, “A Lagrangian network for kinematic control of redundant

robot manipulators,”IEEE Transactions on Neural Networks, vol. 10, no. 5, pp. 1123-1132, 1999.

Computational Intelligence Laboratory, CUHK – p. 23/145

Page 79: Neurodynamic Optimization: Models and Applications

Projection NetworkA recurrent neural network called the projectionnetwork was developed for optimization with boundconstraints onlya b

ǫdx

dt= −x + g(x −∇f(x)).

aY.S. Xia and J. Wang, “On the stability of globally projecteddynamic systems,”J. of Opti-

mization Theory and Applications, vol. 106, no. 1, pp. 129-150, 2000.bY.S. Xia, H. Leung, and J. Wang, “A projection neural networkand its application to con-

strained optimization problems,”IEEE Trans. Circuits and Systems I, vol. 49, no. 4, pp. 447-458,

2002.

Computational Intelligence Laboratory, CUHK – p. 24/145

Page 80: Neurodynamic Optimization: Models and Applications

Convex ProgramConsider a convex programming problem withoutequality constraints:

CP2 : minimize f(x)

subject to c(x) ≤ 0, x ≥ 0

wheref(x) andc(x) = (c1(x), ..., cm(x))T areconvex,m ≤ n.

Computational Intelligence Laboratory, CUHK – p. 25/145

Page 81: Neurodynamic Optimization: Models and Applications

Equivalent ReformulationThe Karush-Kuhn-Tucker (KKT) conditions for CP:

y ≥ 0, c(x) ≤ 0, x ≥ 0

∇f(x) + ∇c(x)y ≥ 0, yT c(x) = 0

According to the projection method, the KKTcondition is equivalent to:

h(x − α(∇f(x) + ∇c(x)y)) − x = 0

h(y + αc(x)) − y = 0,(1)

whereh(r) = (h(r1), ..., h(rn))T ,

h(ri) = max0, ri, andα is any positive constant.Computational Intelligence Laboratory, CUHK – p. 26/145

Page 82: Neurodynamic Optimization: Models and Applications

Two-layer networkBased on an equivalent formulation, a two-layerneural network was developed for OPa is then givenby

ǫd

dt

(

x

y

)

=

(

−x + g(x − (∇f(x) + ∇c(x)y))

−y + h(y + c(x))

)

,

wherex ∈ ℜn andy ∈ ℜm.aY.S. Xia and J. Wang, “A recurrent neural network for nonlinear convex optimization subject

to nonlinear inequality constraints,”IEEE Trans. Circuits and Systems I, vol. 51, no. 7, pp. 1385-

1394, 2004.

Computational Intelligence Laboratory, CUHK – p. 27/145

Page 83: Neurodynamic Optimization: Models and Applications

Model Architecture

Computational Intelligence Laboratory, CUHK – p. 28/145

Page 84: Neurodynamic Optimization: Models and Applications

Convergence ResultsFor anyx(t0) andy(t0), x(t) andy(t) are continuousand unique.u(t) ≥ 0 if u(t0) ≥ 0. The equilibriumpoint solves CP2.

If ∇2f(x) +∑n

i=1 yi∇2ci(x) is positive definite on

ℜn+m+ , then the two-layer neural network is globally

convergent to the KKT point(x∗, y∗), wherex∗ is theoptimal solution to CP2.

Computational Intelligence Laboratory, CUHK – p. 29/145

Page 85: Neurodynamic Optimization: Models and Applications

Two-layer Neural Network forQPIf C = I in QP1, let α = 1 in the two-layer neuralnetwork for CP:

ǫd

dt

(

x

y

)

=

(

−x + g((I − Q)x + ATy − q)

−Ax + b

)

.

whereǫ > 0, x ∈ ℜn, y ∈ ℜm,g(x) = [g(x1), ..., g(xn)]T

g(xi) =

li xi < liui li ≤ xi ≤ hi

hi xi > hi.

It is globally asymptotically convergent to the optimalsolution. Computational Intelligence Laboratory, CUHK – p. 30/145

Page 86: Neurodynamic Optimization: Models and Applications

Illustrative Example

minimize1

4x4

1 + 0.5x21 +

1

4x4

2 + 0.5x22 − 0.9x1x2

subject to Ax ≤ b, x ≥ 0

where

A =

1 1

−1 1

1 −3

and b =

2

2

−2

.

This problem has an optimal solutionx∗ = [0.427, 0.809]T .

Computational Intelligence Laboratory, CUHK – p. 31/145

Page 87: Neurodynamic Optimization: Models and Applications

Simulation Results

−3 −2 −1 0 1 2 3−3

−2

−1

0

1

2

3

P1(−3,−2)

P2(−1,−3) P

3(−2,−3) P

4(3,−3)

P5(3,3)P

6(−1,3)

P7(−3,2)

x*

Computational Intelligence Laboratory, CUHK – p. 32/145

Page 88: Neurodynamic Optimization: Models and Applications

Illustrative Example

minimize x21 + 2x1x2 + x2

2 + (x1 − 1)4 + (x2 − 3)4

subject to x ≥ 0, ci(x) ≤ 0 (i = 1, 2, 3),

where

c1(x) = x21 + x2

2 − 64,

c2(x) = (x1 + 3)2 + (x2 + 4)2 − 36,

c3(x) = (x1 − 3)2 + (x2 + 4)2 − 36.

This problem has an optimal solutionx∗ = (0, 1.96)T .

Computational Intelligence Laboratory, CUHK – p. 33/145

Page 89: Neurodynamic Optimization: Models and Applications

Simulation Results

−6 −4 −2 0 2 4 6

−8

−6

−4

−2

0

2

4

6

. x*

Ω

P1(−4,−9) P

2(4,−9)

P3(6,−6)

P4(6,6)

P5(−6,5)

P6(−6,1)

P7(−6,−4)

Computational Intelligence Laboratory, CUHK – p. 34/145

Page 90: Neurodynamic Optimization: Models and Applications

Illustrative Example

minimize (x1 − x2)2 + (x2 − x3)

2 + (x3 − x4)4

subject to x ≥ 0, ci(x) ≤ 0 (i = 1, 2),

where

c1(x) = x21 + x2

2 + x23 + x2

4 − 9

c1(x) = (x1 − 4)2 + (x2 + 4)2 + (x3 − 1)2 + (x4 + 1)4

This problem has an optimal solutionx∗ = (3.013, 0, 0.766, 0)T .

Computational Intelligence Laboratory, CUHK – p. 35/145

Page 91: Neurodynamic Optimization: Models and Applications

Simulation Results

0 0.5 1 1.5−3

−2

−1

0

1

2

3

4

time

t

x1(t)

x3(t)

x2(t), x

4(t)

Computational Intelligence Laboratory, CUHK – p. 36/145

Page 92: Neurodynamic Optimization: Models and Applications

Dual Network for QP 2

For strictly convex QP2, Q is invertible. The dynamicequation of the dual network:

ǫdy(t)

dt= −CQ−1CTy + g

(

CQ−1CTy − y − Cq)

+Cq + b,

x(t) = Q−1CTy − q,

whereǫ > 0.It is also globally exponentially convergent to theoptimal solutiona b.

aY. Xia and J. Wang, “A dual neural network for kinematic control of redundant robot manip-

ulators,”IEEE Trans. on Systems, Man, and Cybernetics, vol. 31, no. 1, pp. 147-154, 2001.bY. Zhang and J. Wang, “A dual neural network for convex quadratic programming subject to

linear equality and inequality constraints,”Physics Letters A, pp. 271-278, 2002.Computational Intelligence Laboratory, CUHK – p. 37/145

Page 93: Neurodynamic Optimization: Models and Applications

Simplified Dual Network forQP1

For strictly convex QP1, Q is invertible. The dynamicequation of the simplified dual networka:

ǫdu

dt= −Cx + g(Cx − u),

x = Q−1(ATy + CTu − q),

y = (AQ−1AT )−1[

−AQ−1CTu + AQ−1q + b]

,

whereu ∈ Rn is the state vector,ǫ > 0.

It is proven to be globally asymptotically convergentto the optimal solution.

aS. Liu and J. Wang, “A simplified dual neural network for quadratic programming with its

KWTA application,” IEEE Trans. Neural Networks, vol. 17, no. 6, pp. 1500-1510, 2006.

Computational Intelligence Laboratory, CUHK – p. 38/145

Page 94: Neurodynamic Optimization: Models and Applications

Illustrative Example

minimize 3x21 + 3x2

2 + 4x23 + 5x2

4 + 3x1x2 + 5x1x3+

x2x4 − 11x1 − 5x4

subject to 3x1 − 3x2 − 2x3 + x4 = 0,

4x1 + x2 − x3 − 2x4 = 0,

−x1 + x2 ≤ −1,

−2 ≤ 3x1 + x3 ≤ 4.

Computational Intelligence Laboratory, CUHK – p. 39/145

Page 95: Neurodynamic Optimization: Models and Applications

Illustrative Example (cont’d)

Q =

6 3 5 0

3 6 0 1

5 0 8 0

0 1 0 10

, q =

−11

0

0

−5

,

A =

3 −3 −2 1

4 1 −1 −2

, b =

0

0

,

C =

−1 1 0 0

3 0 1 0

, l =

−∞

−2

, h =

−1

4

.

The simplified dual neural network for solving thisquadratic programming problem needs only twoneurons, whereas the Lagrange neural network needstwelve neurons, the primal-dual neural network needsnine neurons, the dual neural network needs fourneurons.

Computational Intelligence Laboratory, CUHK – p. 40/145

Page 96: Neurodynamic Optimization: Models and Applications

Illustrative Example (cont’d)

0 0.2 0.4 0.6 0.8 1

x 10−5

−80

−70

−60

−50

−40

−30

−20

−10

0

10

t (sec)

u

u1

u2

Transientbehaviors of the state vectoru.

Computational Intelligence Laboratory, CUHK – p. 41/145

Page 97: Neurodynamic Optimization: Models and Applications

Illustrative Example (cont’d)

0 0.2 0.4 0.6 0.8 1

x 10−5

−2

−1.5

−1

−0.5

0

0.5

1

1.5

2

t (sec)

x

x1

x2

x3

x4

Transientbehaviors of the output vectorx.

Computational Intelligence Laboratory, CUHK – p. 42/145

Page 98: Neurodynamic Optimization: Models and Applications

Illustrative Example (cont’d)

−5 −4 −3 −2 −1 0 1 2 3−6

−5

−4

−3

−2

−1

0

1

2

x∗

x1

x2

Trajectories ofx1 andx2 from different initial states.

Computational Intelligence Laboratory, CUHK – p. 43/145

Page 99: Neurodynamic Optimization: Models and Applications

Illustrative Example (cont’d)

−3 −2 −1 0 1 2 3 4 5−6

−5

−4

−3

−2

−1

0

1

2

x∗

x3

x4

Trajectories ofx3 andx4 from different initial states.

Computational Intelligence Laboratory, CUHK – p. 44/145

Page 100: Neurodynamic Optimization: Models and Applications

A New Model for LPA new recurrent neural network model with adiscontinuous activation function was recentlydeveloped for linear programming LP1.

Computational Intelligence Laboratory, CUHK – p. 45/145

Page 101: Neurodynamic Optimization: Models and Applications

A New Model for LPA new recurrent neural network model with adiscontinuous activation function was recentlydeveloped for linear programming LP1.

The dynamic equation of the new model is describedas follows:

ǫdx

dt= −Px − σ(I − P )g(x) + s, (3)

whereg(x) = (g1(x1), g2(x2), . . . , gn(xn))T is the

vector-valued activation function,ǫ is a positivescaling constant,σ is a nonnegative gain parameter,P = AT (AAT )−1A, ands = −(I − P )q + AT (AAT )−1b.

Computational Intelligence Laboratory, CUHK – p. 45/145

Page 102: Neurodynamic Optimization: Models and Applications

Activation FunctionThe following activation function is defineda: Fori = 1, 2, . . . , n;

gi(xi) =

1, if xi > hi,

[0, 1], if xi = hi,

0, if xi ∈ (li, hi),

[−1, 0], if xi = li,

−1, if xi < li.

aQ. Liu, and J. Wang, ”A one-layer recurrent neural network with a discontinuous activation

function for linear programming,” Neural Computation, in press, 2007.

Computational Intelligence Laboratory, CUHK – p. 46/145

Page 103: Neurodynamic Optimization: Models and Applications

Activation Function (cont’d)

-

6

0 xi

gi(xi)

1

−1

li hi

Computational Intelligence Laboratory, CUHK – p. 47/145

Page 104: Neurodynamic Optimization: Models and Applications

Convergence ResultsThe neural network is globally convergent to anoptimal solution of LP1 with C = I, if Ω ⊂ Ω, whereΩ is the equilibrium point set andΩ = x|l ≤ x ≤ h.The neural network is globally convergent to anoptimal solution of LP1 with C = I, if it has a uniqueequilibrium point andσ ≥ 0 when(I − P )c = 0 orone of the following conditions holds when(I − P )c 6= 0:

(i) σ ≥ ‖(I − P )c‖p/min+γ∈X ‖(I − P )γ‖p for

p = 1, 2,∞, or

(ii) σ ≥ cT (I − P )c/min+γ∈X|c

T (I − P )γ|,

whereX = −1, 0, 1n

Computational Intelligence Laboratory, CUHK – p. 48/145

Page 105: Neurodynamic Optimization: Models and Applications

Simulation ResultsConsider the following LP problem:

minimize 4x1 + x2 + 2x3,

subject to x1 − 2x2 + x3 = 2,

−x1 + 2x2 + x3 = 1,

−5 ≤ x1, x2, x3 ≤ 5.

According to the above condition, the lower bound ofσ is 9

Computational Intelligence Laboratory, CUHK – p. 49/145

Page 106: Neurodynamic Optimization: Models and Applications

Simulation Results (cont’d)

0 0.2 0.4 0.6 0.8 1

x 10−5

−10

−5

0

5

time (sec)

state trajectories

x1

x2

x3

0 0.2 0.4 0.6 0.8 1

x 10−5

−6

−5

−4

−3

−2

−1

0

1

2

3

4

5

time (sec)

state trajectories

x1

x2

x3

0 0.2 0.4 0.6 0.8 1

x 10−5

−6

−5

−4

−3

−2

−1

0

1

2

3

4

5

time (sec)

state trajectories

x1

x2

x3

0 0.2 0.4 0.6 0.8 1

x 10−5

−6

−5

−4

−3

−2

−1

0

1

2

3

4

5

time (sec)

state trajectories

x1

x2

x3

Transient behaviors of the states with four different values ofσ ∈ 3, 5, 9, 15. Computational Intelligence Laboratory, CUHK – p. 50/145

Page 107: Neurodynamic Optimization: Models and Applications

k Winners Take All OperationThek-winners-take-all (kWTA) operation is to selectthek largest inputs out ofn inputs (1 ≤ k ≤ n).

Computational Intelligence Laboratory, CUHK – p. 51/145

Page 108: Neurodynamic Optimization: Models and Applications

k Winners Take All OperationThek-winners-take-all (kWTA) operation is to selectthek largest inputs out ofn inputs (1 ≤ k ≤ n).

ThekWTA operation has important applications inmachine learning, such ask-neighborhoodclassification,k-means clustering, etc.

Computational Intelligence Laboratory, CUHK – p. 51/145

Page 109: Neurodynamic Optimization: Models and Applications

k Winners Take All OperationThek-winners-take-all (kWTA) operation is to selectthek largest inputs out ofn inputs (1 ≤ k ≤ n).

ThekWTA operation has important applications inmachine learning, such ask-neighborhoodclassification,k-means clustering, etc.

As the number of inputs increases and/or the selectionprocess should be operated in real time, parallelalgorithms and hardware implementation aredesirable.

Computational Intelligence Laboratory, CUHK – p. 51/145

Page 110: Neurodynamic Optimization: Models and Applications

kWTA Problem FormulationsThekWTA function can be defined as:

xi = f(ui) =

1, if ui ∈ k largest elements ofu,0, otherwise,

whereu ∈ Rn andx ∈ R

n is the input vector andoutput vector, respectively.

Computational Intelligence Laboratory, CUHK – p. 52/145

Page 111: Neurodynamic Optimization: Models and Applications

kWTA Problem FormulationsThekWTA function can be defined as:

xi = f(ui) =

1, if ui ∈ k largest elements ofu,0, otherwise,

whereu ∈ Rn andx ∈ R

n is the input vector andoutput vector, respectively.ThekWTA solution can be determined by solving thefollowing linear integer program:

minimize −n∑

i=1

uixi,

subject ton∑

i=1

xi = k,

xi ∈ 0, 1, i = 1, 2, . . . , n.Computational Intelligence Laboratory, CUHK – p. 52/145

Page 112: Neurodynamic Optimization: Models and Applications

kWTA Problem Formulations(cont’d)If the kth and(k + 1)th largest elements ofu aredifferent (denoted asuk anduk+1 respectively), thekWTA problem is equivalent to the following LP orQP problems:

minimize −uTx or a2x

Tx − uTx,

subject ton∑

i=1

xi = k,

0 ≤ xi ≤ 1, i = 1, 2, . . . , n,

wherea ≤ uk − uk+1 is a positive constant.

Computational Intelligence Laboratory, CUHK – p. 53/145

Page 113: Neurodynamic Optimization: Models and Applications

The Primal-Dual Network forkWTAThe primal-dual network based on the QP formulationneeds3n + 1 neurons and6n + 2 connections, and itsdynamic equations can be written as:

dxdt

= −(1 + a)(x − (x + ve + w − ax + u)+)

−(eTx − k)e − x − y + edydt

= −y + (y + w)+ − x − y + edvdt

= −eT (x − (x + ve + w − ax + u)+)

+eTx − kdwdt

= −x + (x + ve + w − ax + u)+

−y + (y + w)+ + x + y − e

wherex, y, w ∈ Rn, v ∈ R, e = (1, 1, . . . , 1)T ∈ R

n,x+ = (x+

1 , . . . , x+n )T , andx+

i = max0, xi(i = 1, . . . , n).

Computational Intelligence Laboratory, CUHK – p. 54/145

Page 114: Neurodynamic Optimization: Models and Applications

The Projection Network forkWTAThe projection neural network forkWTA operationbased on the QP formulation needsn + 1 neurons and2n + 2 connections, which dynamic equations can bewritten as:

dxdt

= λ [−x + f(x − η(ax − u − ve))]dvdt

= λ(−eTx + k).

wherex ∈ Rn, v ∈ R, λ andη are positive constants,

f(x) = (f(x1), . . . , f(xn))T and

f(xi) =

0, if xi < 0,

xi, if 0 ≤ xi ≤ 1,

1, if xi > 1.Computational Intelligence Laboratory, CUHK – p. 55/145

Page 115: Neurodynamic Optimization: Models and Applications

The Simplified Dual Networkfor kWTAThe simplified dual neural network forkWTAoperation based on the QP formulationa needsnneurons and3n connections, and its dynamic equationcan be written as:

dydt

= λ [−My + f((M − I)y − s) − s]

x = My + s,

wherex, y ∈ Rn, M = 2(I − eeT/n)/a,

s = Mu + ke/n, I is an identity matrix,λ andf aredefined as before.

aS. Liu and J. Wang, “A simplified dual neural network for quadratic programming with its

KWTA application,” IEEE Trans. Neural Networks, vol. 17, no. 6, pp. 1500-1510, 2006.

Computational Intelligence Laboratory, CUHK – p. 56/145

Page 116: Neurodynamic Optimization: Models and Applications

The Simplified Dual Networkfor kWTA

1/a

1/a

1/a

v1

v2

· · · · · ·

vn

1/a

1/a

1/a

u1

u2

· · · · · ·

un

v

u

xn

1/ǫ∫ uu

+

+

+

−+

s

Mu

Computational Intelligence Laboratory, CUHK – p. 57/145

Page 117: Neurodynamic Optimization: Models and Applications

A Static ExampleLet the inputs arevi = i (i = 1, 2, · · · , n),n = 10, k = 2, ǫ = 10−8, anda = 0.25.

0 0.5 1 1.5 2

x 10−7

−5

−4

−3

−2

−1

0

1

2

3

4

5

t (sec)

u

Computational Intelligence Laboratory, CUHK – p. 58/145

Page 118: Neurodynamic Optimization: Models and Applications

A Static Example (cont’d)Let the inputs arevi = i (i = 1, 2, · · · , n),n = 10, k = 2, ǫ = 10−8, anda = 0.25.

0 0.5 1 1.5 2

x 10−7

−2

−1.5

−1

−0.5

0

0.5

1

1.5

2

t (sec)

x

Computational Intelligence Laboratory, CUHK – p. 59/145

Page 119: Neurodynamic Optimization: Models and Applications

A Static Example (cont’d)

10−11

10−10

10−9

10−8

0

0.5

1

1.5

2

2.5

t (sec)

u1

a = 0.05a = 0.1a = 0.2a = 0.4

Computational Intelligence Laboratory, CUHK – p. 60/145

Page 120: Neurodynamic Optimization: Models and Applications

A Static Example (cont’d)

0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1

x 10−8

0

2

4

6

8

10

12

14

16

18

20

t (sec)

u1

n = 5, a = 0.05n = 10, a = 0.05n = 20, a = 0.05n = 40, a = 0.05

Computational Intelligence Laboratory, CUHK – p. 61/145

Page 121: Neurodynamic Optimization: Models and Applications

A Dynamic ExampleLet inputs be 4 sinusoidal input signals (i.e.,n = 4)vi(t) = 10 sin[2π(1000t + 0.2(i − 1)], andk = 2.

0 0.2 0.4 0.6 0.8 1−10

−5

0

5

10

v

v1

v2

v3

v4

0 0.2 0.4 0.6 0.8 1

0

0.5

1

x1

0 0.2 0.4 0.6 0.8 1

0

0.5

1

x2

0 0.2 0.4 0.6 0.8 1

0

0.5

1

x3

0 0.2 0.4 0.6 0.8 1

0

0.5

1

x4

Computational Intelligence Laboratory, CUHK – p. 62/145

Page 122: Neurodynamic Optimization: Models and Applications

A One-layer kWTA NetworkThe dynamic equation of a new LP-basedkWTAnetwork model is described as follows:

ǫdx

dt= −Px − σ(I − P )g(x) + s, (4)

whereP = eeT/n, s = u − Pu + ke/n, ǫ is a positivescaling constant,σ is a nonnegative gain parameter,andg(x) = (g(x1), g(x2), . . . , g(xn))

T is adiscontinuous vector-valued activation function.

Computational Intelligence Laboratory, CUHK – p. 63/145

Page 123: Neurodynamic Optimization: Models and Applications

Activation FunctionA discontinuous activation function is defined asfollows:

g(xi) =

1, if xi > 1,

[0, 1], if xi = 1,

0, if 0 < xi < 1,

[−1, 0], if xi = 0,

−1, if xi < 0.

Computational Intelligence Laboratory, CUHK – p. 64/145

Page 124: Neurodynamic Optimization: Models and Applications

Activation Function (cont’d)

-

6

0 xi

g(xi)

1

−1

1

Computational Intelligence Laboratory, CUHK – p. 65/145

Page 125: Neurodynamic Optimization: Models and Applications

Convergence ResultsThe network (4) can perform thekWTA operation ifΩ ⊂ x ∈ R

n : 0 ≤ x ≤ 1, whereΩ is the set ofequilibrium point(s).The network (4) can perform thekWTA operation if ithas a unique equilibrium point andσ ≥ 0 when(I − eeT/n)u = 0 or one of the following conditionsholds when(I − eeT/n)u 6= 0:

(i) σ ≥∑ n

i=1 |ui−∑ n

j=1 uj/n|

2n−2, or

(ii) σ ≥ n

ni=1

(ui−∑

nj=1

uj/n)2

n(n−1), or

(iii) σ ≥ 2maxi |ui −∑n

j=1 uj/n|, or,

(iv) σ ≥

ni=1

(ui−∑

nj=1

uj/n)2

min+

γi∈−1,0,1

|∑

ni=1

(ui−∑

nj=1

uj/n)γi| .

Computational Intelligence Laboratory, CUHK – p. 66/145

Page 126: Neurodynamic Optimization: Models and Applications

Model Comparisons

model type Eqn(s). neurons connectionsPrimal-dual neural network(??) 3n + 1 6n +

Projection neural network (??) n + 1 2n +

Simplified dual network (??) n 3n

Neural network herein (4) n 2n

Neural network herein (??)(??) n 3n

Table 1: Comparison of related neural networks in

terms of model complexity.

Computational Intelligence Laboratory, CUHK – p. 67/145

Page 127: Neurodynamic Optimization: Models and Applications

Simulation ResultsConsider akWTA problem with input vectorui = i (i = 1, 2, . . . , n), n = 5, k = 3.

0 0.2 0.4 0.6 0.8 1

x 10−5

−1.5

−1

−0.5

0

0.5

1

1.5

2

2.5

time (sec)

x1,x

2

x3,x

4,x

5

state trajectories

Transient behaviors of thekWTA networkσ = 6.Computational Intelligence Laboratory, CUHK – p. 68/145

Page 128: Neurodynamic Optimization: Models and Applications

Convergence Results (cont’d)

0 0.2 0.4 0.6 0.8 1

x 10−5

−1.5

−1

−0.5

0

0.5

1

1.5

2

2.5

time (sec)

state trajectories

Transient behaviors of thekWTA network withσ = 2.Computational Intelligence Laboratory, CUHK – p. 69/145

Page 129: Neurodynamic Optimization: Models and Applications

Convergence Results (cont’d)

0 0.5 1 1.5 2 2.5 3 3.5 4

x 10−7

−0.1

0

0.1

0.2

0.3

0.4

0.5

0.6

time (sec)

x 1

n=5,σ=4n=10,σ=9n=15,σ=14n=20,σ=19

Convergence behavior of thekWTA network withrespect to different values ofn. Computational Intelligence Laboratory, CUHK – p. 70/145

Page 130: Neurodynamic Optimization: Models and Applications

Linear Assignment ProblemThe linear assignment problem is to find an optimalsolution to the following linear integer programmingproblem:

minimizen∑

i=1

n∑

j=1

cijxij,

subject ton∑

j=1

xij = 1, i = 1, 2, . . . , n,

n∑

i=1

xij = 1, j = 1, 2, . . . , n,

xij ∈ 0, 1, i, j = 1, 2, . . . , n.

Computational Intelligence Laboratory, CUHK – p. 71/145

Page 131: Neurodynamic Optimization: Models and Applications

Linear Assignment Problem(cont’d)If the optimal solution to problem (71) is unique, thenit is equivalent to the following linear programmingproblem:

minimizen∑

i=1

n∑

j=1

cijxij,

subject ton∑

j=1

xij = 1, i = 1, 2, . . . , n,

n∑

i=1

xij = 1, j = 1, 2, . . . , n,

0 ≤ xij ≤ 1, i, j = 1, 2, . . . , n.

Computational Intelligence Laboratory, CUHK – p. 72/145

Page 132: Neurodynamic Optimization: Models and Applications

Simulation ResultsConsider a linear assignment problem with

C =

4 2 5

3 1.5 2

4 2.5 1

.

A lower bound ofσ is 13.

Computational Intelligence Laboratory, CUHK – p. 73/145

Page 133: Neurodynamic Optimization: Models and Applications

Simulation Results (cont’d)Let ǫ = 10−6 andσ = 15.

0 0.2 0.4 0.6 0.8 1

x 10−5

−0.4

−0.2

0

0.2

0.4

0.6

0.8

1

1.2

1.4

time (sec)

state trajectories

x11

,x13

,x22

,x23

,x31

,x32

x12

,x21

,x33

Computational Intelligence Laboratory, CUHK – p. 74/145

Page 134: Neurodynamic Optimization: Models and Applications

Support Vector MachineConsider a set of training examples

(x1, y1), (x2, y2), ..., (xN , yN )

where thei-th examplexi ∈ Rn belongs to one of twoseparate classes labeled byyi ∈ −1, 1.

A support vector machine provides an optimalpartition with maximum possible margin for patternclassification.

Computational Intelligence Laboratory, CUHK – p. 75/145

Page 135: Neurodynamic Optimization: Models and Applications

SVM Primal Problem

min1

2wTw + c

N∑

i=1

ξi

s.t.

yi[wTφ(xi) + b] ≥ 1 − ξi, i = 1, · · · , N

ξi ≥ 0, i = 1, · · · , N.

wherec > 0 is a regularization parameter for thetradeoff between model complexity and training error,andξi measures the (absolute) difference betweenwTz + b andyi.

Computational Intelligence Laboratory, CUHK – p. 76/145

Page 136: Neurodynamic Optimization: Models and Applications

SVM Dual Problem

max −1

2

N∑

i=1

N∑

j=1

yiyjφ(xi)Tφ(xj)αiαj +

N∑

i=1

αi

s.t.

∑N

i=1 αiyi = 0

0 ≤ αi ≤ c, i = 1, · · · , N.

Computational Intelligence Laboratory, CUHK – p. 77/145

Page 137: Neurodynamic Optimization: Models and Applications

SVM Dual ProblemFor convenient computation here, letai = αiyi. Thenthe SVM dual problem can be equivalently written as

min1

2

N∑

i=1

N∑

j=1

aiajK(xi, xj) −N

i=1

aiyi

s.t.

∑N

i=1 ai = 0

c−i ≤ ai ≤ c+i , i = 1, · · · , N.

wherec−i = c · sgn(1 − yi) andc+i = c · sgn(1 + yi)

for i = 1, ..., N .

Computational Intelligence Laboratory, CUHK – p. 78/145

Page 138: Neurodynamic Optimization: Models and Applications

SVM Learning Network

ǫd

dt

(

a

µ

)

=

(

−a + h(a − (Qa + eµ − y))

−eTa

)

whereǫ > 0, a ∈ ℜN , andµ ∈ ℜ, e = (1, . . . , 1)T .

ǫdai

dt= −ai + h(

N∑

k=1

wikak − µ + yi), i = 1, ..., N ;

ǫdµ

dt= −

N∑

k=1

ak,

whereQ = [qij] = [K(x(i), y(j))], wik = δik − qik.a

aY. Xia and J. Wang, “A one-layer recurrent neural network forsupport vector machine learn-

ing,” IEEE Transactions on Systems, Man and Cybernetics, vol. 34, no. 2, pp. 1261-1269, 2004.

Computational Intelligence Laboratory, CUHK – p. 79/145

Page 139: Neurodynamic Optimization: Models and Applications

Network Architecture

Computational Intelligence Laboratory, CUHK – p. 80/145

Page 140: Neurodynamic Optimization: Models and Applications

Iris Benchmark ProblemThe data of the iris problem are characterized withfour attributes (i.e., the petal length and width, setallength and width).

The dataset consists of 150 samples belonging tothree classes (i.e., viginica, versilcolor, setosa), eachclass has 50 samples.

120 samples for training and the remaining 30 fortesting.

We usec = 0.25 and the polynomial kernel functionK(x, y) = (xTy + 1)p, with p = 2 andp = 4.

Computational Intelligence Laboratory, CUHK – p. 81/145

Page 141: Neurodynamic Optimization: Models and Applications

Simulation Results

0 0.01 0.02 0.03 0.04 0.05 0.06 0.07 0.08 0.09 0.1−0.4

−0.2

0

0.2

0.4

0.6

0.8

1

Time (sec)

Figure 1: Convergence of the SVM Learning neural

network withǫ = 1/150 andp = 2Computational Intelligence Laboratory, CUHK – p. 82/145

Page 142: Neurodynamic Optimization: Models and Applications

Simulation Results

0 0.01 0.02 0.03 0.04 0.05 0.06 0.07 0.08 0.09 0.1−0.25

−0.2

−0.15

−0.1

−0.05

0

0.05

0.1

0.15

0.2

0.25

Time (sec)

Figure 2: Convergence of the proposed neural network

with ǫ = 1/150 andp = 4Computational Intelligence Laboratory, CUHK – p. 83/145

Page 143: Neurodynamic Optimization: Models and Applications

Simulation Results

−1.5 −1 −0.5 0 0.5 1 1.5−1.5

−1

−0.5

0

0.5

1

1.5

peta

l wid

th

petal length

Class IIIClass IISupport point of class IIISupport point of class II

Figure 3: Support vectors of SVC using the proposed

neural network with a polynomial kernelp = 2Computational Intelligence Laboratory, CUHK – p. 84/145

Page 144: Neurodynamic Optimization: Models and Applications

Simulation Results

−1.5 −1 −0.5 0 0.5 1 1.5−1.5

−1

−0.5

0

0.5

1

1.5

petal length

peta

l wid

th

Class IIIClass IISupport point of class IIISupport point of class II

Figure 4: Support vectors of SVC using the proposed

neural network with a polynomial kernel andp = 4Computational Intelligence Laboratory, CUHK – p. 85/145

Page 145: Neurodynamic Optimization: Models and Applications

Adult Benchmark ProblemThe UCI adult benchmark task is to predict whether ahousehold has an income greater than $50,000 basedon 14 other fields in a census form.

Eight of those fields are categorical, while six arecontinuous. The six fields are quantized into quintile,which yields a total of 123 sparse binary features.

1605 training samples and 2000 testing samples.

Gaussian RBF kernel with width of10 andc = 0.5.

Let ǫ = 0.1, and the initial pointz0 ∈1606 with theelement being1.

Computational Intelligence Laboratory, CUHK – p. 86/145

Page 146: Neurodynamic Optimization: Models and Applications

Adult Benchmark Problem

Method iterations SVs Testing accuracySOR 924 635 84.06

SMO 3230 633 84.06

SVM-light 294 634 84.25

NN 567 633 84.15

Table 2: Comparisons of results of the SOR, SMO,

SVM-light, and proposed neural network algorithm

Computational Intelligence Laboratory, CUHK – p. 87/145

Page 147: Neurodynamic Optimization: Models and Applications

Support Vector Regression(SVR)Consider the regression problem of approximating aset of data

(x1, y1), (x2, y2), . . . , (xN , yN )

with a regression function as

φ(x) =N

i=1

αiΦi(x) + ς,

whereΦi(x)(i = 1, 2, . . . , n) are the feature functionsdefined in a high-dimensional space,αi(i = 1, 2, . . . , n) andς are parameters of the modelto be estimated.

Computational Intelligence Laboratory, CUHK – p. 88/145

Page 148: Neurodynamic Optimization: Models and Applications

SVR (cont’d)By utilizing Huber loss function, the above regressionfunction can be represented as

φ(x) =N

i=1

θiK(x, xi) + ς, (5)

whereK(x, y) is a kernel function satisfyingK(x, y) = Φ(x)TΦ(x).

Computational Intelligence Laboratory, CUHK – p. 89/145

Page 149: Neurodynamic Optimization: Models and Applications

SVR (cont’d)θi (i = 1, 2, . . . , N) can be obtained from thefollowing quadratic program:

min1

2

N∑

i=1

N∑

j=1

θiθjK(xi, xj) −N

i=1

θiyi +ε

N∑

i=1

θ2i ,

s.t.N

i=1

θi = 0,

−µ ≤ θi ≤ µ, i = 1, 2, . . . , N ;

whereε > 0 is an accuracy parameter required for theapproximation,µ > 0 is a pre-specified parameter.

Computational Intelligence Laboratory, CUHK – p. 90/145

Page 150: Neurodynamic Optimization: Models and Applications

SVR (cont’d)The neural network with a discontinuous activationfunction for solving the above quadratic program:

ǫdz

dt= −Pz + [PQ +

α

NeeT ]g(z) + q,

θ = (PQ +α

NeeT )−1(Pz − q),

wheree = [1, 1, . . . , 1]T ,P = I − eeT/N,Q = K(xi, xj)N×N + εI/µ,q = (I − eeT/N)y with y = −(y1, y2, . . . , yn), andh = −l = µe in the activation function.

Computational Intelligence Laboratory, CUHK – p. 91/145

Page 151: Neurodynamic Optimization: Models and Applications

SVR (cont’d)Moreover,ς can be obtained from

ς = −1

N(eT (Q − I)θ∗ + eT c − eTz∗),

wherez∗ is an equilibrium point andθ∗ is an outputvector corresponding toz∗.

Compared with existing neural networks for SVMlearning, the existing neural networks need eithertwo-layer structure andn + 1 neurons.

In contrast, the neural network herein has one-layerstructure andn neurons only.

Computational Intelligence Laboratory, CUHK – p. 92/145

Page 152: Neurodynamic Optimization: Models and Applications

SVR (cont’d)For the SVR learning by using the proposed neuralnetwork based on titanium regression dataa. Let thekernel be a Gaussian function:

K(x, y) = exp

(

−‖x − y‖2

2σ2

)

ε = 0.01, µ = 100 andσ = 6.aP. Dierckx,Curve and Surface Fitting with Splines, Clarendon

Press, Oxford, 1993.

Computational Intelligence Laboratory, CUHK – p. 93/145

Page 153: Neurodynamic Optimization: Models and Applications

Regression Result

500 600 700 800 900 1000 11000.4

0.6

0.8

1

1.2

1.4

1.6

1.8

2

2.2

2.4

x

y

Computational Intelligence Laboratory, CUHK – p. 94/145

Page 154: Neurodynamic Optimization: Models and Applications

Inverse Kinematics ProblemBecauseθ is underdetermined in a kinematicallyredundant manipulator, one way to determineθ(t)without the need for computing the pseudoinverse isto solve:

minimize1

2θ(t)TWθ(t) + cT θ(t),

subject to J(θ(t))θ(t) = xd(t),

η−6 θ 6 η+

whereW is a positive-definite weighting matrix,c isan column vector, andη± are upper and lower boundsof the joint velocity vector.

Computational Intelligence Laboratory, CUHK – p. 95/145

Page 155: Neurodynamic Optimization: Models and Applications

Lagrangian Network DynamicsLet the state vectors of output neurons and hiddenneurons be denoted byv(t) andu(t), representingestimatedθ(t) and estimatedλ(t), respectively.

The dynamic equation of the two-layer Lagrangiannetwork can be expressed as:

ǫ1dv(t)

dt= −Wv(t) − J(θ(t))Tu(t) − c,

ǫ2du(t)

dt= J(θ(t))v(t) − xd(t),

whereǫ1 > 0 andǫ2 > 0.

Computational Intelligence Laboratory, CUHK – p. 96/145

Page 156: Neurodynamic Optimization: Models and Applications

Lagrangian Network Architec-ture

Computational Intelligence Laboratory, CUHK – p. 97/145

Page 157: Neurodynamic Optimization: Models and Applications

7-DOF PA10 Manipulator

Computational Intelligence Laboratory, CUHK – p. 98/145

Page 158: Neurodynamic Optimization: Models and Applications

Coordinate system of PA10 ma-nipulator

Computational Intelligence Laboratory, CUHK – p. 99/145

Page 159: Neurodynamic Optimization: Models and Applications

Circular Motion of the PA10Manipulator

−0.4 −0.2 0 0.2 0.4−0.4

−0.2

0

0.2

0.4

x

y

−0.4 −0.2 0 0.2 0.4

0.6

0.8

1

1.2

x

z

−0.20

0.2−0.20

0.2

0.5

1

−0.4 −0.2 0 0.2 0.4

0.6

0.8

1

1.2

yz

xy

z

Computational Intelligence Laboratory, CUHK – p. 100/145

Page 160: Neurodynamic Optimization: Models and Applications

Simulation Results

0 1 2 3 4−0.5

0

0.5

1

0 1 2 3 4

−0.5

0

0.5

0 1 2 3 4

−1

−0.5

0

0.5

1

0 1 2 3 40

0.5

1

1.5

2

0 1 2 3 4−2

−1

0

1

2

0 1 2 3 40.5

1

1.5

2

2.5

0 1 2 3 4−0.01

−0.005

0

0.005

0.01

Time(second)0 1 2 3 4

−4

−2

0

2

4

6x 10

−3

Time(second)

x y

z

xω yω zωx.

y.

z.

x.ω y

.ω z.ω

θ1.

θ2.

θ3.

θ4.

θ5.

θ6.

θ7.

IIθ.(t)II2

θ1

θ2

θ3θ4

θ5

θ6

θ7

IIθ(t)II2

xerr

yerr

zerr

xerrω

yerrω

zerrω

xerr.

yerr.

zerr.

xerr.ω

yerr.ω

zerr.ω

(a) (b)

(c) (d)

(e) (f)

(g) (h)

Computational Intelligence Laboratory, CUHK – p. 101/145

Page 161: Neurodynamic Optimization: Models and Applications

Dual Network DynamicsTo reduce the number of neurons to minimum, nextwe propose a dual neural network with its dynamicequation and output equation defined as

ǫdu(t)

dt= −J(θ(t))W−1JT (θ))u + xd(t),

v(t) = JT (θ(t))u(t);

whereu is the dual state variable,v is the outputvariable.The Lagrangian network containsn + m neurons. Butthe dual network contains onlym neurons, wheren isthe number of joints andm is the dimension of thecartesian space (i.e., 6).

Computational Intelligence Laboratory, CUHK – p. 102/145

Page 162: Neurodynamic Optimization: Models and Applications

Dual Network

Computational Intelligence Laboratory, CUHK – p. 103/145

Page 163: Neurodynamic Optimization: Models and Applications

Simulation Results

−0.4 −0.3 −0.2 −0.1 0 0.1

−0.3

−0.25

−0.2

−0.15

−0.1

−0.05

0

0.05

0.1

0.15

0.2

−0.4−0.2

0

−0.2

0

0.2

0

0.2

0.4

0.6

0.8

−0.4 −0.3 −0.2 −0.1 0 0.1

−0.1

0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

−0.3 −0.2 −0.1 0 0.1 0.2

−0.1

0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

x

xx

y

y

y

zz

z

Computational Intelligence Laboratory, CUHK – p. 104/145

Page 164: Neurodynamic Optimization: Models and Applications

Simulation Results

0 2 4 6 8 10−0.5

0

0.5

1

0 2 4 6 8 10−0.2

−0.1

0

0.1

0.2

0 2 4 6 8 10

−0.2

0

0.2

0.4

0 2 4 6 8 100

0.2

0.4

0.6

0 2 4 6 8 10−2

−1

0

1

2

0 2 4 6 8 10

8

10

12

14

0 2 4 6 8 10−4

−2

0

2

4

6x 10

−8

0 2 4 6 8 10

−1

−0.5

0

0.5

1

1.5x 10

−4

x

y

z

θ1 θ2θ3

θ4θ5

θ6

θ7

θ1

θ2

θ3

θ4 θ5

θ6θ7

‖θ‖ 2‖e‖2

ex

ey

‖e‖2ex

ey

ez

x

yz

ez

cond

(J)

t (sec)t (sec)Computational Intelligence Laboratory, CUHK – p. 105/145

Page 165: Neurodynamic Optimization: Models and Applications

Bounded Inverse KinematicsThe dual neural network with the following dynamicequation and output equation

ǫ1dx

dt= −JW−1JTx + xd

ǫ2dy

dt= −W−1y + g((W−1 − I)y)

v = JTx + y

where the piecewise linear activation function

gi(ui) =

η−i , if ui < η−

i

ui, if η−i 6 ui 6 η+

i

η+i , if ui > η+

iComputational Intelligence Laboratory, CUHK – p. 106/145

Page 166: Neurodynamic Optimization: Models and Applications

Piecewise Linear ActivationFunction

Computational Intelligence Laboratory, CUHK – p. 107/145

Page 167: Neurodynamic Optimization: Models and Applications

PA10 Drift-free Circular Motion

00.1

0.20.3

0.40.5

0.60.7

−0.2

−0.1

0

0.1

0.2−0.4

−0.3

−0.2

−0.1

0

0.1

0.2

0.3

0.4

0.5

xy

z

Computational Intelligence Laboratory, CUHK – p. 108/145

Page 168: Neurodynamic Optimization: Models and Applications

PA10 Joint Variables

0 1 2 3.84 5 6 7 8 9 10−1

−0.5

0

0.5

1

1.5

2

2.68

θ1

θ2

θ3

θ4

θ5

θ6

θ7

t (second)

Computational Intelligence Laboratory, CUHK – p. 109/145

Page 169: Neurodynamic Optimization: Models and Applications

Joint Velocities and Dual StateVariables

0 1 2 3 4 5 6 7 8 9 10

−2.5

−2

−1.5

−1

−0.5

0

0.5

1

1.5

θ1θ2θ3

θ4θ5

θ6

θ7

t (second)(a) Joint rate variables in rad/sec

0 1 2 3 4 5 6 7 8 9 10−80

−60

−40

−20

0

20

40

60

80

u1

u2

u3

u7

t (second)(b) Nonzero dual decision vari-

ables

Computational Intelligence Laboratory, CUHK – p. 110/145

Page 170: Neurodynamic Optimization: Models and Applications

Euclidean Norm vs. InfinityNormThe Euclidean norm (or 2-norm) is widely used oftenbecause of its analytical tractability.

Minimizing the 2-norm of the joint velocities does notnecessarily minimize the magnitudes of the individualjoint velocities.

This is undesirable in situations where the individualjoint velocities are of primary interest.

Minimizing the infinity norm of velocity variables canminimize the maximum velocity.

Computational Intelligence Laboratory, CUHK – p. 111/145

Page 171: Neurodynamic Optimization: Models and Applications

Inverse Kinematics ProblemMinimizing the infinity norm ofθ subject to thekinematic constraint:

minθ

∥θ∥

∞= min

θ

max1≤j≤n

|eTj θ|,

s.t. J(θ(t))θ(t) = xd(t),

whereej is thej-th column of the identity matrix.

Computational Intelligence Laboratory, CUHK – p. 112/145

Page 172: Neurodynamic Optimization: Models and Applications

Inverse Kinematics ProblemLet

s = max1≤j≤n

|eTj θ|.

The inverse kinamatic problem can be written as

minθn

s

s.t. |eTj θ| ≤ s, j = 1, 2, . . . , n

J(θ(t))θ(t) = xd(t).

Computational Intelligence Laboratory, CUHK – p. 113/145

Page 173: Neurodynamic Optimization: Models and Applications

Inverse Kinematics Problem Re-formulationThe inverse kinematics problem can be summarized ina matrix form:

min s

s.t.

[

−I In

I In

] [

θ

s

]

[

0

0

]

J(θ)θ = xd(t),

whereIn = (1, 1, . . . , 1)T ∈ Rn andI is the identitymatrix.

Computational Intelligence Laboratory, CUHK – p. 114/145

Page 174: Neurodynamic Optimization: Models and Applications

Primal Inverse KinematicsProblem FormulationLet y = (yT

1 , y2)T , y1 = θ, y2 = s, then a final form of

the problem can be derived as

min cTy

s.t. A1y ≥ 0,

A2y = b(t),

where

A1 =

[

−I In

I In

]

, A2(t) = [J(θ(t), 0],

b(t) = xd(t), cT = [0, 0, . . . , 1].

Computational Intelligence Laboratory, CUHK – p. 115/145

Page 175: Neurodynamic Optimization: Models and Applications

Dual Inverse Kinematics Prob-lem FormulationThe dual problem of the preceding linear program isdefined as follows:

max bTz2

s.t. AT1 z1 + AT

2 z2 = c,

z1 ≥ 0,

wherez = (zT1 , zT

2 )T is the dual decision variable.

Computational Intelligence Laboratory, CUHK – p. 116/145

Page 176: Neurodynamic Optimization: Models and Applications

Energy FunctionAn energy function to be minimized can be definedbased on the primal and dual formulation:

E(y, z) =1

2(cTy − bz2)

2 +1

2‖A2y − b‖2

2 +

1

2‖AT

1 z1 + AT2 z2 − c‖2

2 +

1

4(A1y)T (A1y − |A1y|) +

1

4zT1 (z1 − |z1|).

Computational Intelligence Laboratory, CUHK – p. 117/145

Page 177: Neurodynamic Optimization: Models and Applications

Primal-Dual NetworkThe dynamic equation of the primal-dual network:

ǫ1y = −c(cTy − bTz2) + AT1 h(−A1y) +

AT2 (A2y − b),

ǫ2z1 = −h(−z1) + A1(AT1 z1 + AT

2 z2 − c),

ǫ3z2 = −b(cTy − bTz2) + A2(AT1 z1 + AT

2 z2 − c),

wherey, z1, z2, are state vectors;h(x) = max0, x;andǫi are positive scaling constants.

Computational Intelligence Laboratory, CUHK – p. 118/145

Page 178: Neurodynamic Optimization: Models and Applications

Primal-Dual Network Architec-ture

Computational Intelligence Laboratory, CUHK – p. 119/145

Page 179: Neurodynamic Optimization: Models and Applications

Desired Position of PA10 End-Effector

0 1 2 3 4 5 6 7 8 9 10−0.2

−0.15

−0.1

−0.05

0

0.05

0.1

0.15

0.2

Time (second)

Des

ired

end−

effe

ctor

vel

ocity

(ra

d/s)

PSfrag repla ements

_x_y_z

Computational Intelligence Laboratory, CUHK – p. 120/145

Page 180: Neurodynamic Optimization: Models and Applications

PA10 Circular Motion

−0.5−0.4

−0.3−0.2

−0.10

0.10.2

−0.2

−0.1

0

0.1

0.20

0.2

0.4

0.6

0.8

1

1.2

1.4

x (m)y (m)

z (m

)

Computational Intelligence Laboratory, CUHK – p. 121/145

Page 181: Neurodynamic Optimization: Models and Applications

Joint Velocities from the La-grangian Network

0 1 2 3 4 5 6 7 8 9 10−0.4

−0.3

−0.2

−0.1

0

0.1

0.2

0.3

0.4

0.5

Time (second)

Join

t Vel

ocity

(ra

d/s)

PSfrag repla ements_1_2_3_4_5_6_7

Computational Intelligence Laboratory, CUHK – p. 122/145

Page 182: Neurodynamic Optimization: Models and Applications

Joint Velocities from the Primal-Dual Network

0 1 2 3 4 5 6 7 8 9 10−0.3

−0.2

−0.1

0

0.1

0.2

0.3

0.4

Time (second)

Join

t Vel

ocity

(ra

d/s)

PSfrag repla ements_1_2_3_4_5_6_7

Computational Intelligence Laboratory, CUHK – p. 123/145

Page 183: Neurodynamic Optimization: Models and Applications

Infinity Norm of Joint Velocities

0 1 2 3 4 5 6 7 8 9 100

0.05

0.1

0.15

0.2

0.25

0.3

0.35

0.4

0.45

0.5

Time (second)

Max

imum

join

t vel

ocity

(ra

d/s)

2−norm minimization∞−norm minimization

Computational Intelligence Laboratory, CUHK – p. 124/145

Page 184: Neurodynamic Optimization: Models and Applications

Transients of Energy Function

0 0.2 0.4 0.6 0.8 1 1.2 1.4

x 10−6

0

0.05

0.1

0.15

0.2

0.25

0.3

0.35

0.4

0.45

0.5

Time (second)

Ene

rgy

Computational Intelligence Laboratory, CUHK – p. 125/145

Page 185: Neurodynamic Optimization: Models and Applications

Bi-criteria Kinematic ControlThe bi-criteria redundancy resolution scheme subjectto joint limits:

minimize1

2

α‖θ‖22 + (1 − α)‖θ‖2

subject to J(θ)θ = xd

η−6 θ 6 η+

whereη± denote upper and lower limits of jointvelocities respectively.

Computational Intelligence Laboratory, CUHK – p. 126/145

Page 186: Neurodynamic Optimization: Models and Applications

Problem ReformulationWith ej denoting thejth column of identity matrixI,

‖θ‖∞ = max|θ1|, |θ2|, · · · , |θn| = max16j6n

|eTj θ|.

With s(t) := ‖θ(t)‖∞, the term(1 − α)‖θ(t)‖2∞/2

equals

min. 1−α2 s2(t)

s.t. |eTj θ| 6 s(t)

=⇒

min. 1−α2 s2(t)

s.t.

[

I −1

−I −1

] [

θ(t)

s(t)

]

6

[

0

0

]

Computational Intelligence Laboratory, CUHK – p. 127/145

Page 187: Neurodynamic Optimization: Models and Applications

Problem FormulationWith y := [θT , s]T , the bi-criteria problem becomes:

minimize1

2yTQy

subject to Ay 6 b

Cy = d

y− 6 y 6 y+

where Q :=

αI

(1 − α)

, A :=

I −1

−I −1

, b := 0 ∈ R2n,

C :=[

J(θ) 0

]

, d := xd(t), y− :=

η−

0

, y+ :=

η+

maxη±

Computational Intelligence Laboratory, CUHK – p. 128/145

Page 188: Neurodynamic Optimization: Models and Applications

Problem FormulationTreat equality and inequality constraints as boundconstraints:

ξ− =

b−

d

y−

, ξ+ =

b

d

y+

, E =

A

C

I

with b− sufficiently negative to represent−∞. Thenthe bicriteria kinematic control problem can berewritten as

minimize1

2yTQy

subject to ξ− 6 Ey 6 ξ+.

Computational Intelligence Laboratory, CUHK – p. 129/145

Page 189: Neurodynamic Optimization: Models and Applications

Dual Network Dynamics

ǫdu(t)

dt= −EQ−1ETu(t) + g((EQ−1ET − I)u(t)),

y(t) = Q−1ETu(t).

Computational Intelligence Laboratory, CUHK – p. 130/145

Page 190: Neurodynamic Optimization: Models and Applications

Simulation Results

−0.4−0.3

−0.2−0.1

00.1

−0.3

−0.2

−0.1

0

0.1

0.2−0.4

−0.2

0

0.2

0.4

0.6

0.8

1

xy

z

Computational Intelligence Laboratory, CUHK – p. 131/145

Page 191: Neurodynamic Optimization: Models and Applications

Joint Velocity

0

0.2

0.4

−0.5

0

0.5

−0.2

0

0.2

−0.5

0

0.5

−0.1

0

0.1

0 1.48 3.45 4.75 6.87 8 9 10−0.2

0

0.2

θ1

θ2

θ3

θ4

θ5

θ6

time (sec)Computational Intelligence Laboratory, CUHK – p. 132/145

Page 192: Neurodynamic Optimization: Models and Applications

Norm Comparison

0

0.05

0.1

0.15

0.2

0.25

0.3

0 1 2 3 4 5 6 7 8 9 100

0.1

0.2

0.3

0.4

0.5

0.6

‖θ‖∞

‖θ‖2

time (sec)Computational Intelligence Laboratory, CUHK – p. 133/145

Page 193: Neurodynamic Optimization: Models and Applications

Grasping Force OptimizationConsider a multifingered robot hand grasping a singleobject in a3-dimensional workspace withm pointcontacts between the grasped object and the fingers.

The problem of the grasp force optimization is to finda set of contact forces such that the object is held atthe desired position and external forces arecompensated.

A grasping forcexi is applied by each finger to holdthe object without slippage and to balance anyexternal forces.

Computational Intelligence Laboratory, CUHK – p. 134/145

Page 194: Neurodynamic Optimization: Models and Applications

Grasping Force OptimizationTo ensure non-slipping at a contact point, the graspingforcexi should satisfyx2

i1 + x2i2 ≤ µix

2i3, where

µi > 0 is the friction coefficient at fingeri, andxi1, xi2, andxi3 are components of contact forcexi inthe contact coordinate frame.

Besides the form-closure constraints, to balance anyexternal wrenchfext to maintain a stable grasp, eachfinger must apply a grasping forcexi = [xi1, xi2, xi3]to the object such thatGx = −fext, whereG ∈ R6×3m

is the grasp transformation matrix andx = [x1, ..., xm]T ∈ R3m is the grasping force vector.

Computational Intelligence Laboratory, CUHK – p. 135/145

Page 195: Neurodynamic Optimization: Models and Applications

Grasping Force OptimizationThe optimal grasping force optimization can beformulated as the following quadtatic minimizationproblem with linear and quadratic constraints:

minimize f(x) =1

2xTQx

subject to ci(x) ≤ 0, i = 1, ...,m;

Gx = −fext

whereq ∈ R3m, Q is a3m × 3m positive definitematrix, andci(x) =

x2i1 + x2

i2 − µixi3.

Computational Intelligence Laboratory, CUHK – p. 136/145

Page 196: Neurodynamic Optimization: Models and Applications

Neurodynamic Optimization ofGasping ForceBased on the problem formulation, we develop thethree-layer recurrent neural network for gasping forceoptimization

ǫd

dt

x

y

z

=

−Qx −∇c(x)y + GTz

−y + h(c(x) + y)

−Gx − fext

,

wherex ∈ R3m, y ∈ Rm, z ∈ R6, andǫ > 0 is ascaling parameter.The neural network is globally convergent to the KKTpoint (x∗, y∗, z∗), wherex∗ is the optimal gaspingforce.

Computational Intelligence Laboratory, CUHK – p. 137/145

Page 197: Neurodynamic Optimization: Models and Applications

Neurodynamic Optimization ofGasping ForceConsider a minimum norm forcef(x) = 1

2‖x‖2.

A polyhedral object withM = 0.1kg is grasped by athree-fingered robotic hand.

Let the robotic hand move along a circular trajectoryof radiusr = 0.5m with a constant velocityv =0.2m/s.

The time-varying external wrench applied to thecenter of mass of the object isfext = [0, fc sin(θ(t)),−Mg + fc cos(θ(t)), 0, 0, 0]T ,whereg = 9.8(m/s2), θ ∈ [0, 2π], andfc = Mv2/r.

Computational Intelligence Laboratory, CUHK – p. 138/145

Page 198: Neurodynamic Optimization: Models and Applications

Simulation Results

mg W

r mv

f c

2

Figure 5: Three-finger grasp example

Computational Intelligence Laboratory, CUHK – p. 139/145

Page 199: Neurodynamic Optimization: Models and Applications

Simulation Results

Figure 6: Motion of the three-finger grasp

Computational Intelligence Laboratory, CUHK – p. 140/145

Page 200: Neurodynamic Optimization: Models and Applications

Simulation Results

0 0.5 1 1.5 2 2.5 3 3.5 4 4.5 5−0.2

0

0.2

0.4

0.6

0.8

1

1.2

1.4

1.6

Time (sec)

For

ce (

New

ton)

x11

x12

x13

x21

x22

x23

x31

x32

x33

Computational Intelligence Laboratory, CUHK – p. 141/145

Page 201: Neurodynamic Optimization: Models and Applications

Simulation Results

0 0.5 1 1.5 2 2.5 3 3.5 4 4.5 5

x 10−3

0

1

2

3

4

5

6

7

8

9

10

Time (sec)

Ene

rgy

Figure 7: Convergence of the energy function withǫ =

0.0001 Computational Intelligence Laboratory, CUHK – p. 142/145

Page 202: Neurodynamic Optimization: Models and Applications

Simulation Results

0 0.5 1 1.5 2 2.5 3 3.5 4 4.5 50

0.5

1

1.5

2

2.5

3

Time (sec)

Euc

lidea

n N

orm

CurrentPrimal DualMatlab

Figure 8: Comparison of Euclidean norm of optimal

forces using three different methodsComputational Intelligence Laboratory, CUHK – p. 143/145

Page 203: Neurodynamic Optimization: Models and Applications

Concluding RemarksNeurodynamic optimization has been demonstrated tobe a powerful alternative approach to manyoptimization problems.

For convex optimization, recurrent neural networksare available with global convergence to the optimalsolution.

Neurodynamic optimization approaches provideparallel distributed computational models moresuitable for real-time applications.

Computational Intelligence Laboratory, CUHK – p. 144/145

Page 204: Neurodynamic Optimization: Models and Applications

Future WorksThe existing neurodynamic optimization model canstill be improved to reduce their model complexity orincrease their convergence rate.

The available neurodynamic optimization model canbe applied to more areas such as control, robotics, andsignal processing.

Neurodynamic approaches to global optimization anddiscrete optimization are much more interesting andchallenging.

It is more needed to develop neurodynamic modelsfor nonconvex optimization and combinatorialoptimization.

Computational Intelligence Laboratory, CUHK – p. 145/145