16
EE 538 Neural Networks Soo-Young Lee Department of Electrical Engineering / Brain Science Research Center KAIST EE538 Neural Networks Fall 2015 1-1 Lecture 1 Introduction Biological Neuron Models Course Objectives • Understanding engineering models (artificial neural networks / machine learning) of cognitive functions for “smart” machine – Feature extraction – Clustering – Separation – Classification/Recognition – Prediction – Motor control – and more EE538 Neural Networks Fall 2015 1-2 Related Topics EE538 Neural Networks Fall 2015 1-3 Rule-based Systems Learning-based Systems Artificial Intelligence Big Data/ Data Mining Machine Learning Neural Networks (Connectionist Model) Probabilistic Learning Knowledge Development Language Understanding Pattern/Image Recognition Speech Recognition Course Emphasis Connections between biological and artificial neural networks Related to computational models of brain functions (Computational Neuroscience, not to heuristic rules) Learn from hints • Term-Project Utilize neural network models to benchmark/competition problems EE538 Neural Networks 1-4

EE538-01.pdf

Embed Size (px)

Citation preview

EE 538 Neural Networks

Soo-Young LeeDepartment of Electrical Engineering /

Brain Science Research CenterKAIST

EE538 Neural Networks Fall 2015 1-1

Lecture 1 Introduction Biological Neuron Models

Course Objectives• Understanding engineering models (artificial

neural networks / machine learning) of cognitive functions for “smart” machine– Feature extraction– Clustering– Separation– Classification/Recognition– Prediction– Motor control– and more

EE538 Neural Networks Fall 2015 1-2

Related Topics

EE538 Neural Networks Fall 2015 1-3

Rule-basedSystems

Learning-based Systems

Artificial Intelligence

Big Data/Data Mining

Machine Learning

Neural Networks (Connectionist Model)

Probabilistic Learning

KnowledgeDevelopment

LanguageUnderstanding Pattern/Image

RecognitionSpeech

Recognition

Course Emphasis• Connections between biological and artificial neural

networks– Related to computational models of brain functions

(Computational Neuroscience, not to heuristic rules)– Learn from hints

• Term-Project– Utilize neural network models to benchmark/competition

problems

EE538 Neural Networks 1-4

Approaches• Lecture• Homework (20%) and Quiz (10%)• Mid-Term Exam (20%)• Course Participation (10%)• Term Project (Proposal 15%, Report 35%)

– Repeat what others had done– Add your own ideas

EE538 Neural Networks Fall 2015 1-5

Contents• Background• Course Overview• Biological Neural Networks • Artificial Neural Networks

EE538 Neural Networks 1-6

International CES• 3D• Bigger• More pixels• Faster• Smaller• Distinct color

• SMART, SMART, and SMART!

EE538 Neural Networks 1-7 EE538 Neural Networks 1-8

EE538 Neural Networks 1-9 EE538 Neural Networks 1-10

EE538 Neural Networks 1-11

Historical Sketch• Dankoon and Pygmalion• Pre-1940: von Hemholtz, Mach, Pavlov, etc.

– General theories of learning, vision, conditioning– No specific mathematical models of neuron operation

• 1940s: Hebb, McCulloch and Pitts– Mechanism for learning in biological neurons– Neural-like networks can compute any arithmetic

function• 1950s: Rosenblatt, Widrow and Hoff

– First practical networks and learning rules

EE538 Neural Networks 1-12

Historical Sketch• 1960s-1970s: Dark age

– Minsky and Papert demonstrated limitations of artificial neural networks

– Progress continues, although at a slower pace • Mid 1980s: First Renaissance

– Important new developments caused a resurgence• Late 1980s - Mid 2010s: Continuing Developments

– New algorithms and applications• Mid-2010s – Today : Second Renaissance

– Smart, Smart, and Smart applications!– Big data (in Internet and mobile networks) and cheap

multicore processors (GPUs)EE538 Neural Networks 1-13

Sen-sor

FeatureExtraction

Recogn./Tracking

SelectiveAttention

Sen-sor

FeatureExtraction

Recog./Understand

SelectiveAttention

Perception/Planning

Touch Olf. Smell

Inference

BodyControl

Individual/Group

Behavior

Learning/Memory/Language

Fusion

Pattern Recognition

Speech Recognition

Cognition & Inference

Brain-like SystemBehavior

Vision

Auditory

EE538 Neural Networks 1-14

Korean BrainTech21 Program (Brain Engineering / Neuroinformatics: 1998-2008)

Active Cognitive Development and Situation Awarenessbased on Cognitive Science and Information Technology

AudioVideo

Artificial Cognitive Systems (2009-2014)

EE538 Neural Networks 1-15

Example: Auditory Processing

Feature Extraction

Sound Localization & Speech Enhancement(Separation)

Top-Down Attention

特徵抽出

音源探知 /音聲向上

下向 注意 集中

Classification/RecognitionPrediction

EE538 Neural Networks 1-16

ICA Features from Natural Images

EE538 Neural Networks 1-17

Speech Features at Cochlea (J.H.Lee, et.al., 2000)

Time Domain

Frequency Domain

ICA on speech signals (x=c1f1+c2f2+…+cMfM)Gabor-like features both in time and frequency domain

Frame

EE538 Neural Networks 1-18

ICA-based Complex Features(T. Kim, et.al., 2005)

Time Onset/Offset

Multifrequency Tone(Timbre)

Frequency Modulation

Time Frame

Mel-

Fre

quency

Each training sample

EE538 Neural Networks 1-19

Signal Separation (T. Kim, et al.)

• Recorded microphone signals

• Separated output signals

• Recording environment– Recorded in reverberant real

conference room – Sampling rate: 8 kHz (down

sampled from 44.1 kHz)– 3 human speakers are speaking

and 1 loud speaker is playing a hip-hop music simultaneously

(.wav) (.wav)

(.wav)

(.wav) (.wav)

(.wav)

EE538 Neural Networks 1-20

Classification/Recognition• Speech Recognition

• Image Recognition

EE538 Neural Networks 1-21

Time Frame

Mel-

Fre

quency

Each training sample

Top-Down Attention• Training Patterns (32x32 pixels)

• Deformed Test Patterns

EE538 Neural Networks 1-22

Prediction• Stock price prediction

EE538 Neural Networks 1-23

Motor Control

EE538 Neural Networks 1-24

Contents• Background• Course Overview• Biological Neural Networks • Artificial Neural Networks

EE538 Neural Networks 1-25

Neural Network ModelsWe need introduce • Neuron models• Network architectures • Learning algorithms

– Mathematical models• Applications to engineering problems

– Signal/image processing and recognition– Language understanding and knowledge

developmentof artificial neural networksEE538 Neural Networks 1-26

Neural Networks

Cognitive Science

/NeuroscienceIT (EECS)

Biologically-Inspired Electronic SystemsDigital Brain (Neural Networks)

Biologically-Oriented Electronic SystemsNeuroinformatic DB, BioMedical Imaging

Bionic Life

EE538 Neural Networks 1-27 BiS554 Neural Networks

Top of

the World

EE538 Neural Networks 1-28

Research Scope

Systems

Molecules

Data Mathematical Applications/Measurement Modeling Implementation

GenProtein

MembraneNeuronCircuitsSystem

BehaviorNeural

Networks

EE538 Neural Networks 1-29

Research Modality

Neuroscience Data

Mathematical Model

Brain-like Functional Systems

Neuromorphic Chips

Analysis Software

Measurement Technology

EE538 Neural Networks 1-30

Research Tools• Mathematics

– Optimization– Linear algebra– Differential equation– Statistics / Information theory

• Computer Programming• Hardware Devices

– Silicon-based VLSI– MEMS

EE538 Neural Networks 1-31

Contents• Background• Course Overview• Biological Neural Networks• Artificial Neural Networks

EE538 Neural Networks 1-32

Brain vs. Computer• Human Brain

– about 10^10 to 10^11 neurons– about 10^3 to 10^4 synapses for each neuron– about 10^14 synapses – about 1.5 kg (2% of body weight)– about 1-10 msec time scale

• While computers just conduct programmed instructions, human brain interact with and learns from environments.

EE538 Neural Networks 1-33

Cortical Parameters

EE538 Neural Networks 1-34

Neuron andSynapses

EE538 Neural Networks 1-35

Different Types of Ion Channels

EE538 Neural Networks 1-36

Ion Channels and Concentrations

EE538 Neural Networks 1-37

ActionPotential

EE538 Neural Networks 1-38

EE538 Neural Networks Fall 2015 1-39

Contents• Background• Course Overview• Biological Neural Networks• Artificial Neural Networks

EE538 Neural Networks 1-40

Neural Network Models• Neuron model

– Integrate-and-fire– Saturating non-linearity– Rectified Linear Unit

• Network architecture– Feedforward

• Layered, Convolutive, Max-Pooling– Recurrent

• Learning algorithm– Unsupervised learning– Supervised learning– Reinforcement learning

EE538 Neural Networks 1-41

How to Conduct Researches ? • Set hypothesis, make model, and evaluate the model.

– Hypothesis: best educated guess• Knowledge on neuroscience

– Model: network architecture, mathematical equations

• Knowledge on mathematics (linear algebra, statistics, differential equations, etc.)

– Evaluation• Experiment and/or simulation• Find parameters• Performance measure and decision criteria

EE538 Neural Networks 1-42

Example: Neuron Model• Hypothesis:

– Interaction between membrane potential and ion concentration with voltage-dependent ion channels

• Model: Hodgkin-Huxley model• Evaluation

– Voltage clamp experiment– Find parameters

EE538 Neural Networks 1-43

Action Potential

EE538 Neural Networks 1-44

Hodgkin-Huxley model

EE538 Neural Networks 1-45

Basic BioPhysics• Fick’s law: diffusion

• Ohm’s law: drift

• Einstein relationship

• Equllibriumq

kTD

dxdnDJdiffusion

dxdVZnJdrift

0

dxdVZn

dxdn

qkTJJJ driftdiffusion

EE538 Neural Networks 1-46

Nernst Potential

i

on

n

V

Voi n

nqZkTdn

nqZkTdVVVE

i

o

i

o

ln1

• Nernst potential

EE538 Neural Networks 1-47

Ion Parameters

EE538 Neural Networks 1-48

Resting Potential (1)• Goldman equation for multi-ions• For one-type ions

where

nV

kTqZ

dxdnP

dxdVZn

dxdn

qkTJJJ mdriftdiffusion

D

qkTPV

dxdV m ,

dnn

qZPVkTJqZV

kTdnnPV

kTqZJ

Pdxi

o

i

o

n

n

m

m

n

nm

11,1

0

)/exp(1)/exp(,lnkTZqV

nkTZqVnPVkTqZJ

nqZPV

kTJ

nqZPV

kTJ

kTqVZ

m

omim

im

omm

EE538 Neural Networks 1-49

Resting Potential (2)• For Potassium and Chlorine ions

• Goldman equation for Potassium and Chlorine ions

• Goldman equation for Potassium, Sodium, and Chlorine ions ?

• Do we need any other ion transfer mechanism?

)/exp(1][)/exp(][

)/exp(1][)/exp(][

kTqVClkTqVCl

kTPqVJ

kTqVKkTqVK

kTPqVJ

m

omiClmCl

m

omiKmK

oCliK

iCloKm ClPKP

ClPKPq

kTV][][][][ln

EE538 Neural Networks 1-50

Hodgkin-Huxley Experiment• Voltage-clamping

EE538 Neural Networks 1-51

H-H Voltage Clamp Experiment

EE538 Neural Networks 1-52

HH Model: Refractory Period (1)

EE538 Neural Networks 1-53

HH Model: Refractory Period (2)

EE538 Neural Networks 1-54

Integrate-and-Fire Neuron• Leaky integrator

• Firing threshold

• Reset voltage

• Refractory period

j i

jij

m

ttwtI

tRItudt

tdu

)()(

)()()(

)( jitu

restj

ij utu

)(lim0

EE538 Neural Networks 1-55

Response of IF Neurons• Membrane potential with a constant current before

spikes

mte

RIuRItu /)0(11)(

EE538 Neural Networks 1-56

I-O Transfer Function

RIuRIt

r

RIuRIttt

or

tttotherwise

tttthenttifRIu

RIt

restmref

restmrefii

refii

iiirefi

restmi

ln

1

ln

ln

1

1

1

EE538 Neural Networks 1-57

Noise models of IF-neurons• Noise sources

– Stochastic threshold– Random reset– Noisy integration (noisy input)

EE538 Neural Networks 1-58

Neural Coding

j

i

EE538 Neural Networks 1-59

Population Neuron Model

EE538 Neural Networks 1-60

)(1

kk

m

jkjkjk

vy

bxwv

Activation Function

0if00 if1

k

kk v

vy

)exp(11)(

kk av

v

EE538 Neural Networks 1-61

)T/vexp()v(P

-P(v)P(v)

x

11

1 ty probabili with0)(or 1 ty probabili with1

Determinisitic Neuron

Stochastic Neuron

Network Architecture

Fully-connected Feed-forward

EE538 Neural Networks 1-62

Layered Feed-forward Convolutive

Learning Algorithms • While computers just conduct programmed

instructions, human brain learns from environments.– Supervised Learning

• Learn from teachers– Unsupervised Learning

• Learn from environments without teacher• Active (Self-searching for what need be learnt)• Need objective(s)

– Reinforcement Learning• Critique

EE538 Neural Networks 1-63

Summary• “The genetic code was cracked in the mid-20th century. The

neural code will be cracked in the mid-21st century.”– Intelligence to Machine!

Freedom to Mankind!• Neural code may be learnt by learning from data.

– Unsupervised learning– Supervised learning– Reinforcement learning

• Need define– Neuron model– Network architecture– Learning algorithm

EE538 Neural Networks 1-64