53
1 8 th Term Neural Networks By Engr. Muhammad Imran

Neural Network

  • Upload
    buitems

  • View
    745

  • Download
    2

Embed Size (px)

Citation preview

Page 1: Neural Network

1

8th TermNeural Networks

ByEngr. Muhammad Imran

Page 2: Neural Network

Books: 1. Haykin, S., Neural Networks – A

Comprehensive Foundation, Second edition or latest, McMillan.

2. Hagan, M.T., Demuth, H.B., and Beale, M., Neural Network Design, PWS Publishing Company.

2

Page 3: Neural Network

Artificial Neural Networks (ANNs) learn by experience rather than by modeling or programming.

ANN architectures are distributed, inherently parallel and potentially real time.

They have the ability to generalize. They do not require a prior understanding

of the process or phenomenon being studied.

They can form arbitrary continuous non-linear mappings.

They are robust to noisy data. VLSI implementation is easy

3

Page 4: Neural Network

Pattern recognition including character recognition, speech recognition, face recognition, on-line signature recognition color recognition etc.

Weather forecasting, load forecasting. Intelligent routers, intelligent traffic

monitoring, intelligent filter design. Intelligent controller design Intelligent modeling.

and many more

4

Page 5: Neural Network

Conventional neural networks are black box models.

Tools for analysis and model validation are not well established.

An intelligent machine can only solve some specific problem for which it is trained.

Human brain is very complex and cannot be fully simulated with present computing power. An artificial neural network does not have capability of human brain.

Issue: What is the difference between human

brain neurons and other animal brain neurons?

5

Page 6: Neural Network

A universal definition of ANNs is not available, however, the following definition summarizes the basic features of an ANN:

Artificial Neural Networks, also called Neurocomputing or Parallel Distributed Processes (PDP) or connectionist networks or simply neural networks are interconnected assemblies of simple processing elements, called neurons, units or nodes, whose functionality is loosely based on the biological neuron. The processing ability of the network is stored in the inter-unit connection strength, or weights, obtained by a process of adaptation to, or learning from a set of training patterns.

6

Page 7: Neural Network

7

A s

imp

lifi

ed v

iew

of

a b

iolo

gica

l (re

al)

neu

ron

Page 8: Neural Network

8

Examples of some real neurons

Page 9: Neural Network

9

Image of the vertical organization of neurons in the primary visual cortex

Page 10: Neural Network

The central part of a neuron is called the soma or cell body which contains the nucleus and the protein synthesis machinery. The size of soma of a typical neuron is about 10 to 80m. Almost all the logical functions are realized in this part of a neuron.

10

Page 11: Neural Network

The dendrites represent a highly branching tree of fibers and are attached to the soma. The word dendrite has been taken from the Greek word dendro which means tree. Dendrites connect the neuron to a set of other neurons. Dendrites either receive inputs from other neurons or connect other dendrites to the synaptic outputs.

11

Page 12: Neural Network

It is a long tubular fiber which divides itself into a number of branches towards its end. Its length can be from 100 m to 1 m. The function of an axon is to transmit the generated neural activity to other neurons or to muscle fibers. In other words, it is output channel of the neuron. The point where the axon is connected to its cell body is called the Hillock zone.

12

Page 13: Neural Network

The junction at which a signal is passed from one neuron to the next is called a synapse (from the Greek verb to join). It has a button like shape with diameter around 1 m. Usually a synapse is not a physical connection (the axon and the dendrite do not touch) but there is a gap called the synaptic cleft that is normally between 200Å to 500Å (1Å = 10-10 m). The strength of synaptic connection between neurons can be chemically altered by the brain in response to favorable and unfavorable stimuli in such a way as to adapt the organism to function optimally within its environment.

13

Page 14: Neural Network

14

“The nerve fibre is clearly a signalling mechanism of limited scope. It can only transmit a succession of brief explosive waves, and the message can only be varied by changes in the frequency and in the total number of these waves. … But this limitation is really a small matter, for in the body the nervous units do not act in isolation as they do in our experiments. A sensory stimulus will usually affect a number of receptor organs, and its result will depend on the composite message in many nerve fibres.” Lord Adrian, Nobel Acceptance Speech, 1932.

Page 15: Neural Network

Single neurons are highly complex electrochemical devices

Synaptically connected networks are only part of the story

Many forms of interneuron communication now known – acting over many different spatial and temporal scales

15

Page 16: Neural Network

16

Single neuron activity

• Membrane potential is the voltage difference between a neuron and its surroundings (0 mV)

CellCell

CellCell

0 Mv

Membrane potential

Page 17: Neural Network

17

Single neuron activity •If you measure the membrane potential of a neuron and print it out on the screen, it looks like:

spike

Page 18: Neural Network

18

Single neuron activity

•A spike is generated when the membrane potential is greater than its threshold

Page 19: Neural Network

19

Abstraction•So we can forget all sub-threshold activity and concentrate on spikes (action potentials), which are the signals sent to other neurons

Spikes

Page 20: Neural Network

20

• Only spikes are important since other neurons receive them (signals)

• Neurons communicate with spikes

• Information is coded by spikes

• So if we can manage to measure the spiking time, we decipher how the brain works ….

Page 21: Neural Network

21

Again its not quite that simple

• spiking time in the cortex is random

Page 22: Neural Network

22

With identical input for the identical neuron

spike patterns are similar, but not identical

Page 23: Neural Network

23

Recording from a real neuron: membrane potential

Page 24: Neural Network

24

Single spiking time is meaninglessTo extract useful information, we have to average

to obtain the firing rate r

for a group of neurons in a local circuit where neuron codes the same information over a time window

Local circuit

=

Time window = 1 sec

r =

Hz

Page 25: Neural Network

25

So we can have a network of these local groups

w1: synaptic strength

wn

r1

rn

R f w rj j ( )

Hence we have firing rate of a group of neurons

Page 26: Neural Network

26

ri is the firing rate of input local circuit

The neurons at output local circuits receive signals in the form

The output firing rate of the output local circuit is then given by R

where f is the activation function, generally a Sigmoidal function of some sort

N

iiirw

1

)(1

N

iiirwfR

wi weight, (synaptic strength) measuring the strength of the interaction between neurons.

Page 27: Neural Network

27

Artificial Neural networks

Local circuits (average to get firing rates)

Single neuron (send out spikes)

Page 28: Neural Network

No, with present computing power is it almost impossible to simulate a full behavior of human brain.

Reasons: It is estimated that the human cerebral

cortex contains 100 billion neurons. Each neuron has as many as 1000 dendrites and, hence within the cerebral cortex there are approximately 100,000 billion synapses.

The behavior of the real nervous system is very complex and not yet fully known.

28

Page 29: Neural Network

29

w2

w1

wp

f

x1

x2

Activation function

xp

b

y

p

1iii bpwfy

Page 30: Neural Network

30

w2

w1

wp

f

x1

x2

Activation function

xp

y

w0x0

For simplicity, the threshold contribution b may be treated as an extra input to the neuron, as shown below, where x0 = 1, w0 = b.

p

0iiixwfy

In this case

Page 31: Neural Network

The activation function of a neuron, also known as transfer function, may be linear or non-linear. A particular activation function is chosen to satisfy some specification of the problem that the neuron is attempting to solve.

A variety of activation functions have been proposed. Some of these are discussed below:

31

Page 32: Neural Network

(a) The hard limit activation function: This activation function sets the output of the neuron to 0 if the function argument is less than 0, or 1 if its argument is greater than or equal to 0. (see figure below). This function is generally used to create neurons that classify inputs into two distinct classes.

32

0

1

Page 33: Neural Network

Example 1: The input to a neuron is 3, and its weight is 1.8.

(a) What is the net input to the activation function?

(b) What is the output of the neuron if it has the hard limit transfer function?

Solution:(a) Net input = u = 1.83 = 5.4(b) Neuron output = f(u) = 1.

33

Page 34: Neural Network

Example 2: Repeat example 1 if its bias is

(i) –2 (ii) –6Solution: (a) net input = u = (1.83) +

(-2) = 5.4 – 2 = 3.4 output = f(u) = 1.0 (b) net input = u = (1.83) + (-

6) = 5.4 – 6 = -0.6 output = 0.

34

Page 35: Neural Network

Example 3: Given a two input neuron with the following parameters: b = 1.4, W = [1.2 3.4] and P = [-3 5], calculate the neuron output for the hard limit transfer function.

Solution: net input to the activation function

=

Neuron output = f(u) = 1.0

35

u

8.144.15

34.32.1u

Page 36: Neural Network

(b) Symmetrical hard limit:This activation function is defined as

follows: y = -1 u < 0 y = +1 u 0Where u is the net input and y is the

output of the function (see figure below):

36

-1

0

1

y

u0

Page 37: Neural Network

Example 4: Given a two output neuron with the following parameters: b = 1.2, W = [3 2], and p = [-5 6]. Calculate the neuron output for the symmetrical hard limit activation function.

Solution:

Output = y = f(u) = -1

37

8.12.16

523u

Page 38: Neural Network

(c) Linear activation function:The output of a linear activation function is equal to its input:y = u

as shown in the following figure:

38

0

0 u

y

Page 39: Neural Network

(d) The log-sigmoid activation function:This activation function takes the input (which may have any value between plus and minus infinity) and squashes the output into the range 0 to 1, according to the expression:

39

ue1

1y

A plot of this function is given below:

0 u0

1

Page 40: Neural Network

Example 5: Repeat example 2 for a log-sigmoid activation function.

Solution:(a) net input = u = 3.4

output = f(u) = 1/(1 + e-3.4) = 0.9677

(b) net input u = -0.6 output = f(u) = 1/(1 + e0.6) =

0.3543Example 6: Repeat example 3 for a

log sigmoid activation function.Solution: output = 1

40

Page 41: Neural Network

(e) Hyperbolic Tangent Sigmoid:This activation function is shown in the following figure and is defined mathematically as follows:

41

uu

uu

ee

eey

0-1

0

1

Page 42: Neural Network

42

0

+1

-1

Saturation Limiter Gaussian function Schmitt trigger

+1

-1

and many more

Page 43: Neural Network

RULES FOR SIGNAL FLOW GRAPH

Rule 1: A signal flows along a link only in the direction defined by the arrow on the link. (see fig a: synaptic link, fig b: activation link)

Rule 2: A node signal equals the algebraic sum of all signals entering the pertinent node via the incoming links. (see fig c)

Rule 3: The signal at a node is transmitted to each outgoing link originating from the node, with the transmission being entirely independent of the transfer function of the outgoing links. (see fig d)

43

Page 44: Neural Network

The block diagram of a neuron shown in figure, provides a functional description of the various elements that constitute the model of an artificial neuron.

We can simplify the appearance by using signal flow graph.

44

Page 45: Neural Network

Feedback is said to exist in a dynamic system whenever the output of an element in the system influences in the input applied to that particular element, thereby giving rise to one or more closed paths for the transmission of signals around the system.

The Open loop architectural graph of a neuron

45

Page 46: Neural Network

46

Page 47: Neural Network

Connecting several neurons in some specific manner yields an artificial neural network.

The architecture of an ANN defines the network structure, that is the number of neurons in the network and their interconnectivity.

In a typical ANN architecture, the artificial neurons are connected in layers and they operate in parallel.

The weights or the strength of connection between the neurons are adapted during use to yield good performance.

Each ANN architecture has its own learning rule.

47

Page 48: Neural Network

48

(a) Single Layer feed forward networks:

Input layer Output layer

Page 49: Neural Network

49

(b) Multilayer Feed forward Networks

Input layer Hidden layer Output layer

Ful

ly c

onne

cted

fee

d fo

rwar

d ne

twor

k

wit

h on

e hi

dden

laye

r

Page 50: Neural Network

50

(b) Multilayer Feedforward NetworksP

arti

ally

con

nect

ed f

eedf

orw

ard

netw

ork

wit

h on

e hi

dden

laye

r

Input signal

Input layer Hidden layer Output layer

Page 51: Neural Network

51Rec

urr

ent

net

wor

k w

ith

no

hid

den

neu

ron

s

Page 52: Neural Network

52

Page 53: Neural Network

53