41
Chapter4_Part3 Chapter4_Part3 Differential Heb learning Differential Heb learning & & Differential Competitive Differential Competitive learning learning Tutor : Prof. Gao Tutor : Prof. Gao Reporter : WangYing Reporter : WangYing

Chapter4_Part3 Differential Heb learning & Differential Competitive learning

  • Upload
    zorina

  • View
    43

  • Download
    1

Embed Size (px)

DESCRIPTION

Chapter4_Part3 Differential Heb learning & Differential Competitive learning. Tutor : Prof. Gao Reporter : WangYing. Review. Signal Heb Learning Law Competitive Learning Law. Part I: Differential Heb Learning. Learning law Its simpler version - PowerPoint PPT Presentation

Citation preview

Page 1: Chapter4_Part3 Differential Heb learning &  Differential Competitive learning

Chapter4_Part3Chapter4_Part3Differential Heb learning & Differential Heb learning &

Differential Competitive Differential Competitive learninglearning

Tutor : Prof. Gao Reporter : WaTutor : Prof. Gao Reporter : WangYingngYing

Page 2: Chapter4_Part3 Differential Heb learning &  Differential Competitive learning

2006.10.30

ReviewSignal Heb Learning

Law

Competitive Learning Law

ij ij i jm m S S

ij j ij i jm S m S S ij j i ijm S S m

Page 3: Chapter4_Part3 Differential Heb learning &  Differential Competitive learning

2006.10.30

Part I: Differential Heb Learning

Learning law Its simpler version Hebbian correlations promote spurious causal associations among concurrently active units. Differential correlations estimate the concurrent and presumably causal variation among active units.

ij ij i j i jm m S S S S

ij ij i jm m S S

Page 4: Chapter4_Part3 Differential Heb learning &  Differential Competitive learning

2006.10.30

Differential Heb Learning

Fuzzy Cognitive Maps (FCMs)Adaptive Causal InferenceKlopf’s Drive Reinforcement ModelConcomitant Variation as Statistical CovariancePulse-Coded Differential Hebbian Learning

Page 5: Chapter4_Part3 Differential Heb learning &  Differential Competitive learning

2006.10.30

Fuzzy Cognitive Maps ( 模糊认知映射 )

Fuzzy signed directed graphs with feedback. It model the world as a collection of classes and causal relations between classes.

The directed edge from causal concept to concept measures how much causes .

ije iC

jC iC jC

iC jCije

: Sells of computers

: Profits

iC jC

Page 6: Chapter4_Part3 Differential Heb learning &  Differential Competitive learning

2006.10.30

Fuzzy Cognitive Map of South African Politics

外国投资 矿业 雇用黑人

白人种族激进主义 工作保留法律 黑人种族联合

种族隔离 政府管理力度 民族政党支持者

1c

2C3C

4C5C 6C

7C

8C 9C

Page 7: Chapter4_Part3 Differential Heb learning &  Differential Competitive learning

2006.10.30

Causal Connection Matrix

E

1

2

3

4

5

6

7

8

9

CCCCCCCCC

0 1 1 0 0 0 0 1 10 0 1 0 0 0 0 1 00 0 0 1 0 1 0 1 10 0 0 0 0 1 1 0 10 1 1 0 0 1 1 0 00 0 0 1 0 0 1 1 00 0 0 0 1 0 0 1 00 0 0 0 0 0 1 0 00 0 0 0 1 0 0 1 0

1 2 3 4 5 6 7 8 9C C C C C C C C C

Page 8: Chapter4_Part3 Differential Heb learning &  Differential Competitive learning

2006.10.30

TAM recall process

2C 1C

We start with the foreign investment policy

Then

The arrow indicates the threshold operation with, say, ½ as the threshold value.

So zero causal input produces zero causal output. contains equals 1 because we are testing the foreign-investment policy option. Next

Next

So is a fixed point of the FCM dynamical system.

1 1 0 0 0 0 0 0 0 0C

1 0 1 1 0 0 0 0 1 1C E

1 1 1 0 0 0 0 1 1 2C

2 0 1 2 1 1 1 1 4 1C E 1 1 1 1 0 0 0 1 1 3C

3 0 1 2 1 1 0 0 4 1C E

31 1 1 1 0 0 0 1 1 C

3C

Page 9: Chapter4_Part3 Differential Heb learning &  Differential Competitive learning

2006.10.30

Strengths and weaknesses of FCM

Advantages☺Experts: 1.represent factual and evaluative concepts in an interactive framework; 2.quickly draw FCM pictures or respond to questionnaires; 3.consent or dissent to the local causal structure and perhaps the global equilibrations.☺FCM knowledge representation and inferencing structure: reduces to simple vector-matrix operations, favors integrated-circuit implementation, and allows extension to neural, statistical, or dynamical systems techniques.

Disadvantages It equally encodes the expert’s knowledge or ignorance, wisdom or prejudice. Since different experts differ in how they assign causal strengths to edges, and in which concepts they deem causally relevant, the FCM seems merely to encode its designer’s biases, and may not even encode them accurately.

Page 10: Chapter4_Part3 Differential Heb learning &  Differential Competitive learning

2006.10.30

Combination of FCMs We combined arbitrary FCM connection matrices by adding augmented(增广 )FCM matrices . We add the pointwise to yield the combined FCM matrix : Some experts may be more credible than others. We can weight each expert with a nonnegative credibility weight by multiplicatively weighting the expert’s augmented FCM matrix: Adding FCM matrices represents a simple form of causal learning.

1,..., kE E1,..., kF F iF

F

ii

F F

i

i ii

F F

Page 11: Chapter4_Part3 Differential Heb learning &  Differential Competitive learning

2006.10.30

Differential Heb Learning

Fuzzy Cognitive Maps (FCMs)Adaptive Causal InferenceKlopf’s Drive Reinforcement ModelConcomitant Variation as Statistical CovariancePulse-Coded Differential Hebbian Learning

Page 12: Chapter4_Part3 Differential Heb learning &  Differential Competitive learning

2006.10.30

Adaptive Causal Inference We infer causality between variables when we observe concomitant variation or lagged variation between them. If B changes when A changes, we suspect a causal relationship. The more correlated the changes, the more we suspect a causal relationship, or, more accurately. Time derivatives measure changes. Products of derivatives correlate changes. This leads to the simplest differential Hebbian learning law: ij ij i je e C C

Page 13: Chapter4_Part3 Differential Heb learning &  Differential Competitive learning

2006.10.30

Adaptive Causal Inference The passive decay term forces zero causality between unchanging concepts. The concomitant-variation term indicates causal increase or decrease according to joint concept movement. If and both increase or both decrease, the product of derivatives is positive, v.v. The concomitant-variation term provides a simple causal “arrow of time”.

ije

i jCC

iCjC

Page 14: Chapter4_Part3 Differential Heb learning &  Differential Competitive learning

2006.10.30

Differential Heb Learning

Fuzzy Cognitive Maps (FCMs)Adaptive Causal InferenceKlopf’s Drive Reinforcement ModelConcomitant Variation as Statistical CovariancePulse-Coded Differential Hebbian Learning

Page 15: Chapter4_Part3 Differential Heb learning &  Differential Competitive learning

2006.10.30

Klopf’s Drive Reinforcement Model Harry Klopf independently proposed the following discrete variant of differential Hebbian learning: where the synaptic difference updates the current synaptic efficacy in the first-order difference equation

ijm t

1

T

ij j j j ij i ik

m t S y t c m t k S x t k

1ij ij ijm t m t m t ijm t

Page 16: Chapter4_Part3 Differential Heb learning &  Differential Competitive learning

2006.10.30

Klopf’s Drive Reinforcement Model The term drive reinforcement arises from variables and their velocities. Klopf defines a neuronal drive as the weighted signal and a neuronal reinforcer as the weighted difference . A differentiable version of the drive-reinforcement model take the form: The synaptic magnitude amplifies the synapse’s plasticity. In particular, suppose the ijth synapse is excitatory: . Then we can derive: Implicitly the passive decay coefficient scales the term. The coefficient will usually be much smaller than unity to prevent rapid forgetting:

ij ij ij i jm m m S S

ijm0ijm

1ij ij i jm m S S

ijm

ij ij i jm m S S

ij im Sij im S

Page 17: Chapter4_Part3 Differential Heb learning &  Differential Competitive learning

2006.10.30

Klopf’s Drive Reinforcement ModelDrive-reinforcement synapses can rapidly

encode neuronal signal information. Moreover, signal velocities or directions tend to be more robust, more noise tolerant.

Unfortunately, it tend to zero as they equilibrate, and they equilibrate exponentially quickly. This holds for both excitatory and inhibitory synapses.

Page 18: Chapter4_Part3 Differential Heb learning &  Differential Competitive learning

2006.10.30

Klopf’s Drive Reinforcement Model The equilibrium condition implies that or in general. This would hold equally in a signal Hebbian model if we replaced the signal product with the magnitude -weighted product . Klopf apparently overcomes this tendency in his simulations by forbidding zero synaptic values: .

0ijm

0ij i jm S S

0ijm

0.1ijm t

i jS Sij i jm S S

Page 19: Chapter4_Part3 Differential Heb learning &  Differential Competitive learning

2006.10.30

Klopf’s Drive Reinforcement Model The simple differential Hebbian learning law equilibrates to More generally the differential Hebbian law learns an exponentially weighted average of sampled concomitant variations, since it has the solution in direct analogy to the signal-Hebbian integral equation.

ij ij i jm m S S

ij i jm S S

0

0tt s t

ij ij i jm t m e S s S s e ds

Page 20: Chapter4_Part3 Differential Heb learning &  Differential Competitive learning

2006.10.30

Differential Heb Learning

Fuzzy Cognitive Maps (FCMs)Adaptive Causal InferenceKlopf’s Drive Reinforcement ModelConcomitant Variation as Statistical CovariancePulse-Coded Differential Hebbian Learning

Page 21: Chapter4_Part3 Differential Heb learning &  Differential Competitive learning

2006.10.30

Concomitant Variation as Statistical Covariance

The very term concomitant variation resembles the term covariance. In differential Hebbian learning we interpreted variation as time change, and concomitance as conjunction or product. Alternatively we can interpret variation spatially as a statistical variance or covariance. Sejnowski has cast synaptic modification as a mean-squared optimization problem and derived a covariance-based solution. After some simplifications the optimal solution takes the form of the covariance learning law

,ij ij i i j jm m Cov S x S y

Page 22: Chapter4_Part3 Differential Heb learning &  Differential Competitive learning

2006.10.30

Concomitant Variation as Statistical Covariance

Since We can derive

The stochastic-approximation approach estimates the unknown expectation with the observed realization product

So we estimate a random process with its observed time samples

xy i j i jE S S S S

, x zCov x z E xz m m

ij ij xy i j x i y jm m E S S E S E S

ij ij i j x i y jm m S S E S E S

Page 23: Chapter4_Part3 Differential Heb learning &  Differential Competitive learning

2006.10.30

Concomitant Variation as Statistical Covariance

Suppose instead that we estimate the unknown joint-expectation term

as the observed time samples in the

integrand:

This leads to the new covariance learning law

How should a synapse estimate the unknown averages and at each time t?

xy i x i j y jE S E S S E S

xyE ,i j i x i j y jCov S S S E S S E S

ij ij i x i j y jm m S E S S E S

x iE S t y jE S t

Page 24: Chapter4_Part3 Differential Heb learning &  Differential Competitive learning

2006.10.30

Concomitant Variation as Statistical Covariance

We can lag slightly the stochastic-approximation estimate in time to make a martingale assumption. A martingale assumption estimates the immediate future as the present, or the present as the immediate past

for some time instant s arbitrarily close to t. The assumption increases in accuracy as s approaches t.

0x i x i i

i

E S t E S t S s for s t

S s

1ij ij i jm t m t S t S t

Page 25: Chapter4_Part3 Differential Heb learning &  Differential Competitive learning

2006.10.30

Concomitant Variation as Statistical Covariance

This approximation assumes that the signal processes are well-behaved: continuous, have finite variance, and are at least approximately wide-sense stationary. In an approximate sense when time averages resemble ensemble averages, differential Hebbian learning and covariance learning coincide.

Page 26: Chapter4_Part3 Differential Heb learning &  Differential Competitive learning

2006.10.30

Differential Heb Learning

Fuzzy Cognitive Maps (FCMs)Adaptive Causal InferenceKlopf’s Drive Reinforcement ModelConcomitant Variation as Statistical CovariancePulse-Coded Differential Hebbian Learning

Page 27: Chapter4_Part3 Differential Heb learning &  Differential Competitive learning

2006.10.30

Pulse-Coded Differential Hebbian LearningThe velocity-difference property for pulse-coded signal functionsThe pulse-coded differential Hebbian law replaces the signal velocities in the usual differential Hebbian law with the two differences

When no pulse are present, the pulse-coded DHL reduces to the random-signal Heb law.

ij ij i j ij

ij i j i j i j j i ij

m m S S n

m S S x y x S y S n

i i iS t x t S t j j jS t y t S t

Page 28: Chapter4_Part3 Differential Heb learning &  Differential Competitive learning

2006.10.30

Pulse-Coded Differential Hebbian Learning Replace the binary pulse functions with the bipolar pulse functions, and then suppose the pulses and the expected pulse frequencies, are pairwise independent. Then the average behavior reduces to the ensemble-averages random signal Hebbian learning law or, equivalently, the classical deterministic-signal Hebbian learning law.

ij ij i jE m E m E S E S

Page 29: Chapter4_Part3 Differential Heb learning &  Differential Competitive learning

2006.10.30

Pulse-Coded Differential Hebbian Learning In the language of estimation theory, both random-signal Heb learning and random pulse-coded differential Heb learning provide unbiased estimators of signal Heb learning. The pulse frequencies and can be interpret ergodically (time averages equaling space averages) as ensemble averages

,0i i i

i i

S t E x t x s s t

E x t x s

jSiS

Page 30: Chapter4_Part3 Differential Heb learning &  Differential Competitive learning

2006.10.30

Pulse-Coded Differential Hebbian Learning Substituting these martingale assumptions into pulse-coded DHL It suggests that random pulse-coded DHL provides a real-time stochastic approximation to covariance learning This show again how differential Heb learning and covariance learning coincide when appropriate time averages resemble ensemble averages.

ij ij i i i j j j ijm m x E x t x s y E y t y s n

,ij ij i j ijm m Cov S S n

Page 31: Chapter4_Part3 Differential Heb learning &  Differential Competitive learning

2006.10.30

Part II: Differential Competitive Learning

Learning law Learn only if change! The signal velocity is a local reinforcement mechanism. Its sign indicates whether the jth neurons are winning or losing, and its magnitude measures by how much.

ij j j i i ij ijm S y S x m n

jS

Page 32: Chapter4_Part3 Differential Heb learning &  Differential Competitive learning

2006.10.30

Differential Competitive Learning

If the velocity-difference property replaces the competitive signal velocity ,then the pulse-coded differential competitive learning law is just the difference of nondifferential competitive laws

jS

ij j j i ij ij

j i ij j i ij ij

m y S S m n

y S m S S m n

YF

Winning!

jy t jS tijm =1

Losing!

=0

0

Losing

Page 33: Chapter4_Part3 Differential Heb learning &  Differential Competitive learning

2006.10.30

Competitive signal velocity & supervised reinforcement function

Both of them use a sign change to punish misclassifying . Both of them tend to rapidly estimate unknown pattern-class centroids. The unsupervised signal velocity dose not depend on unknown class memberships, it estimates this information with instantaneous win-rate information. Even uses less information: DCL will perform comparably to SCL!

Page 34: Chapter4_Part3 Differential Heb learning &  Differential Competitive learning

2006.10.30

Computation of postsynaptic signal velocity

Velocity-difference property Nonlinear derivative reduces to the locally

available difference lies between ,except when The signal velocity at time is estimated by mere

presence or absence of the postsynaptic pulse .

jS

1 0j jS S 0j jy S 1j jand y S

0 0j jS S

jS

0 1jS

jy tt

high-speed sensory environmentsstimulus patterns shift constantly

slower, stabler pattern environments

j jy S

jS

Page 35: Chapter4_Part3 Differential Heb learning &  Differential Competitive learning

2006.10.30

Differential-competitive synaptic conjecture then states:

Synapse can physically detect the presence or absence of pulse as a change in the postsynaptic neuron’s polarization. Synapse can clearly detects the presynaptic pulse train , and thus the pulse-train’s pulse count in the most recent 30 milliseconds or so.

ix t

jy

jS t

Synapse

Incoming pulse train

Detected postsynaptic pulse

Electrochemically

Page 36: Chapter4_Part3 Differential Heb learning &  Differential Competitive learning

2006.10.30

Behavior patterns involved in animal learning

Klopf and Gluck suggest that input signal velocities provide pattern information for this.

Pulse-coded differential Hebbian learn

ing

Classical signal Hebbian learningPulse-coded differential competitive

learning

Ordinary caseMicroscopeProcess signalsstore, recognize, recall patterns

Noisy synaptic vectors can locally estimate pattern centroids in real time without supervision.

Page 37: Chapter4_Part3 Differential Heb learning &  Differential Competitive learning

2006.10.30

Differential Competitive Learning as Delta Modulation

The discrete differential competitive learning law represents a neural version of adaptive delta modulation. In communication theory, delta-modulation systems transmit consecutive sampled amplitude differences instead of the sampled amplitude values themselves. A delta-modulation system may transmit only signals, indicating local increase or decrease in the underlying sampled waveform.

1j j j j k jm k m k S y k x m k

1

Page 38: Chapter4_Part3 Differential Heb learning &  Differential Competitive learning

2006.10.30

Differential Competitive Learning as Delta Modulation

Signal difference can be approximated as the activation difference The signum operator sgn(.) behaves as a modified threshold function

It fixes the step size of the delta modulation, and a variable step size will results in adaptive delta modulation.

sgn 1

j j j

j j

S y t y t

y t y t

1 0

sgn 0 01 0

if xx if x

if x

jS jy

Page 39: Chapter4_Part3 Differential Heb learning &  Differential Competitive learning

2006.10.30

Consecutive differences, more informative than consecutive samples

We define the statistical correlation between random variables x and z, it takes values in the bipolar interval [1,-1], x and z are positively correlated if , v.v.

Let denote pulse difference

Suppose the wide-sense-stationary random sequence is zero mean, and each of them has the same finite variance.

2

2

11 ,

1

j jj j

j j

E y k y ky k y k

E y k y k

,x z

jy k

1k j jd y k y k

jykd

0

Page 40: Chapter4_Part3 Differential Heb learning &  Differential Competitive learning

2006.10.30

Consecutive differences, more informative than consecutive samples

Random sequence also has zero mean. The above properties simplifies the variance as

If consecutive samples are highly positively correlated, if the differences have less variance than the samples In the pulse-coded case, when the jth neuron wins, it emits a dense pulse train, this winning pulse frequency nay be sufficiently high to satisfy the property.

2 2

2 2 2 2

1 2 1

2 2 1k j j j jV d E y k E y k E y k y k

kd 1 ,j jy k y k

kd

1 2

Page 41: Chapter4_Part3 Differential Heb learning &  Differential Competitive learning

Thanks for your Thanks for your attention!attention!