lecture1 introduction to ann

Upload: shivan-biradar

Post on 02-Jun-2018

245 views

Category:

Documents


0 download

TRANSCRIPT

  • 8/10/2019 Lecture1 Introduction to ANN

    1/70

    Introduction to ArtificialNeural Networks

    :

  • 8/10/2019 Lecture1 Introduction to ANN

    2/70

    Content Fundamental Concepts of ANNs.

    Basic Models and Learning Rules Neuron Models

    ANN structures

    Learning

    Distributed Representations

    Conclusions

  • 8/10/2019 Lecture1 Introduction to ANN

    3/70

    Introduction to ArtificialNeural Networks

    FundamentalConcepts of ANNs

  • 8/10/2019 Lecture1 Introduction to ANN

    4/70

    What is ANN? Why ANN?

    ANN Artificial Neural Networks

    To simulate human brain behavior A new generation of information

    processing system.

  • 8/10/2019 Lecture1 Introduction to ANN

    5/70

    Applications Pattern Matching Pattern Recognition Associate Memory (Content Addressable Memory) Function Approximation Learning Optimization Vector Quantization Data Clustering . . .

  • 8/10/2019 Lecture1 Introduction to ANN

    6/70

    Applications Pattern Matching Pattern Recognition Associate Memory (Content Addressable Memory) Function Approximation Learning Optimization Vector Quantization Data Clustering . . .

    Traditional Computers are inefficientat these tasks although their

    computation speed is faster.

  • 8/10/2019 Lecture1 Introduction to ANN

    7/70

    The Configuration of ANNs

    An ANN consists of a large number of

    interconnected processing elements calledneurons. A human brain consists of ~1011neurons of

    many different types.

    How ANN works? Collectivebehavior.

  • 8/10/2019 Lecture1 Introduction to ANN

    8/70

    The Biologic Neuron

  • 8/10/2019 Lecture1 Introduction to ANN

    9/70

    The Biologic Neuron

  • 8/10/2019 Lecture1 Introduction to ANN

    10/70

    The Biologic Neuron

    Excitatory or Inhibitory

  • 8/10/2019 Lecture1 Introduction to ANN

    11/70

    The Artificial Neuron

    x1

    x2

    xm

    wi1

    wi2

    wim

    yi

    f (.) a (.)

    i

  • 8/10/2019 Lecture1 Introduction to ANN

    12/70

    The Artificial Neuron

    x1

    x2

    xm

    wi1

    wi2

    wim

    yi

    f (.) a (.)

    i

    ij

    m

    j

    iji xwf 1

    )(

    )()1( fatyi

    otherwise

    ffa

    0

    01)(

  • 8/10/2019 Lecture1 Introduction to ANN

    13/70

    The Artificial Neuron

    x1

    x2

    xm

    wi1

    wi2

    wim

    yi

    f (.) a (.)

    i

    wij

    positive excitatorynegative inhibitoryzero no connection

  • 8/10/2019 Lecture1 Introduction to ANN

    14/70

    The Artificial Neuron

    x1

    x2

    xm

    wi1

    wi2

    wim

    yi

    f (.) a (.)

    i

    Proposed by McCullochand Pitts [1943]

    M-P neurons

  • 8/10/2019 Lecture1 Introduction to ANN

    15/70

    What can be done by M-P neurons?

    A hard limiter. A binary threshold unit. Hyperspace separation.

    otherwise

    fy

    xwxwf

    i

    i

    0

    0)(1

    )( 2211

    x1

    x2

    x1 x2

    y

    w1 w2 10

  • 8/10/2019 Lecture1 Introduction to ANN

    16/70

  • 8/10/2019 Lecture1 Introduction to ANN

    17/70

    Three Basic Entities of ANN Models

    Models of Neurons or PEs.

    Models of synaptic interconnections andstructures.

    Training or learning rules.

  • 8/10/2019 Lecture1 Introduction to ANN

    18/70

    Introduction to ArtificialNeural Networks

    Basic Models and Learning Rules

    Neuron Models

    ANN structures

    Learning

  • 8/10/2019 Lecture1 Introduction to ANN

    19/70

    Processing Elements

    f (.) a (.)

    i

    What integrationfunctions we mayhave?

    What activationfunctions we mayhave?

    Extensions of M-P neurons

  • 8/10/2019 Lecture1 Introduction to ANN

    20/70

    Integration Functions

    f (.) a (.)

    iij

    m

    j

    iji xwf

    2

    1

    QuadraticFunction

    i

    m

    j

    ijji wxf 1

    2)(SphericalFunction

    1 1

    j k

    m m

    i ijk j k j k i

    j k

    f w x x x x

    PolynomialFunction

    ij

    m

    j

    ijii xwnetf 1

    M-P neuron

  • 8/10/2019 Lecture1 Introduction to ANN

    21/70

    Activation Functions

    f (.) a (.)

    i

    M-P neuron: (Step function)

    otherwise

    ffa

    0

    01)(

    1

    a

    f

  • 8/10/2019 Lecture1 Introduction to ANN

    22/70

    Activation Functions

    f (.) a (.)

    i

    Hard Limiter (Threshold function)

    01

    01)sgn()(

    f

    fffa

    1

    a

    1f

  • 8/10/2019 Lecture1 Introduction to ANN

    23/70

    Activation Functions

    f (.) a (.)

    i

    Ramp function:

    00

    10

    11

    )(

    f

    ff

    f

    fa

    1

    a

    1f

  • 8/10/2019 Lecture1 Introduction to ANN

    24/70

  • 8/10/2019 Lecture1 Introduction to ANN

    25/70

    Activation Functions

    f (.) a (.)

    i

    Bipolar sigmoid function:

    11

    2)(

    fefa

    -1.5

    -1

    -0.5

    0

    0.5

    1

    1.5

    -4 -3 -2 -1 0 1 2 3 4

  • 8/10/2019 Lecture1 Introduction to ANN

    26/70

    x

    y

    Example: Activation Surfaces

    L1

    L2

    L3

    x y

    L1 L2

    L3

  • 8/10/2019 Lecture1 Introduction to ANN

    27/70

    x

    y L1

    L2

    L3

    Example: Activation Surfaces

    x1=0

    y1=0

    xy+4=0

    x y

    L1

    L2

    L3

    10

    1=1

    0 1

    2=1

    1

    1

    3= 4

  • 8/10/2019 Lecture1 Introduction to ANN

    28/70

  • 8/10/2019 Lecture1 Introduction to ANN

    29/70

    x

    y L1

    L2

    L3

    Example: Activation Surfaces

    z=1

    z=0L4

    z

    x y

    L1

    L2

    L3

  • 8/10/2019 Lecture1 Introduction to ANN

    30/70

    x

    y L1

    L2

    L3

    Example: Activation Surfaces

    z=1

    z=0L4

    z

    x y

    L1

    L2

    L3

    1

    4=2.5

    1 1

  • 8/10/2019 Lecture1 Introduction to ANN

    31/70

    Example: Activation Surfaces

    L4

    z

    x y

    L1

    L2

    L3

    M-P neuron: (Step function)

    otherwise

    ffa

    0

    01)(

    U i l i id 1

  • 8/10/2019 Lecture1 Introduction to ANN

    32/70

    Example: Activation Surfaces

    L4

    z

    x y

    L1

    L2

    L3

    =2 =3

    =5 =10

    Unipolar sigmoidfunction: fe

    fa

    1

    1)(

  • 8/10/2019 Lecture1 Introduction to ANN

    33/70

    Introduction to ArtificialNeural Networks

    Basic Models and Learning Rules

    Neuron Models

    ANN structures

    Learning

  • 8/10/2019 Lecture1 Introduction to ANN

    34/70

    ANN Structure (Connections)

  • 8/10/2019 Lecture1 Introduction to ANN

    35/70

    Single-Layer Feedforward Networks

    y1 y2 yn

    x1

    x2

    xm

    w11 w12

    w1mw21 w22

    w2m wn1

    wnmw

    n2

    . . .

  • 8/10/2019 Lecture1 Introduction to ANN

    36/70

    Multilayer Feedforward Networks

    . . .

    . . .

    . . .

    . . .

    x1 x2 xm

    y1 y2 yn

    Hidden Layer

    Input Layer

    Output Layer

  • 8/10/2019 Lecture1 Introduction to ANN

    37/70

    Multilayer Feedforward Networks

    Pattern Recognition

    Input

    Analysis

    Classification

    OutputLearning

    Where theknowledge from?

  • 8/10/2019 Lecture1 Introduction to ANN

    38/70

    Single Node with Feedback to Itself

    FeedbackLoop

  • 8/10/2019 Lecture1 Introduction to ANN

    39/70

    Single-Layer Recurrent Networks

    . . .

    x1

    x2

    xm

    y1 y2 yn

  • 8/10/2019 Lecture1 Introduction to ANN

    40/70

    Multilayer Recurrent Networks

    x1 x2 x3

    y1 y2 y3

    . . .

    . . .

  • 8/10/2019 Lecture1 Introduction to ANN

    41/70

    Introduction to ArtificialNeural Networks

    Basic Models and Learning Rules

    Neuron Models

    ANN structures

    Learning

  • 8/10/2019 Lecture1 Introduction to ANN

    42/70

    Learning Consider an ANN with nneurons and each

    with madaptive weights.

    Weight matrix:

    nmnn

    m

    m

    T

    n

    T

    T

    www

    www

    www

    21

    22221

    11211

    2

    1

    w

    w

    w

    W

  • 8/10/2019 Lecture1 Introduction to ANN

    43/70

    Learning Consider an ANN with nneurons and each

    with madaptive weights.

    Weight matrix:

    nmnn

    m

    m

    T

    n

    T

    T

    www

    www

    www

    21

    22221

    11211

    2

    1

    w

    w

    w

    W

    To Learn the weight matrix.

    How?

  • 8/10/2019 Lecture1 Introduction to ANN

    44/70

    Learning RulesSupervisedlearning

    Reinforcementlearning

    Unsupervisedlearning

  • 8/10/2019 Lecture1 Introduction to ANN

    45/70

    Supervised Learning Learning with a teacher

    Learning by examples

    Training set

    (1) (2)(1) (2 )) ( )(( , ), ( , ), , ( , ),kk d d dx xT x

    (1) (2)(1) (2 )) ( )( kk

  • 8/10/2019 Lecture1 Introduction to ANN

    46/70

    Supervised Learning

    x

    Error

    signal

    Generator

    d

    yANN

    W

    (1) (2)(1) (2 )) ( )(( , ), ( , ), , ( , ),kk d d dx xT x

  • 8/10/2019 Lecture1 Introduction to ANN

    47/70

    Reinforcement Learning

    Learning with a critic

    Learning by comments

  • 8/10/2019 Lecture1 Introduction to ANN

    48/70

    Reinforcement Learning

    x

    Critic

    signal

    Generator

    yANN

    W ReinforcementSignal

  • 8/10/2019 Lecture1 Introduction to ANN

    49/70

    Unsupervised Learning

    Self-organizing

    Clustering Form proper clusters by

    discovering the similarities anddissimilarities among objects.

  • 8/10/2019 Lecture1 Introduction to ANN

    50/70

    Unsupervised Learning

    x yANN

    W

  • 8/10/2019 Lecture1 Introduction to ANN

    51/70

    The General Weight Learning Rule

    1

    1

    m

    i ijijj

    net xw

    Input:

    Output: ( )i iy a net

    i

    .

    .

    .

    .

    .

    .

    wi1

    wi2

    wij

    wi,m-1

    x1

    x2

    xj

    xm-1

    yi

    i

  • 8/10/2019 Lecture1 Introduction to ANN

    52/70

    The General Weight Learning Rule

    1

    1

    m

    i ijijj

    net xw

    Input:

    Output: ( )i iy a net

    i

    .

    .

    .

    .

    .

    .

    wi1

    wi2

    wij

    wi,m-1

    x1

    x2

    xj

    xm-1

    yi

    i

    We want to learn the weights& bias.

  • 8/10/2019 Lecture1 Introduction to ANN

    53/70

    We want to learn the weights& bias.

    The General Weight Learning Rule

    1

    1

    m

    i ijijj

    net xw

    Input:

    i

    .

    .

    .

    .

    .

    .

    wi1

    wi2

    wij

    wi,m-1

    x1

    x2

    xj

    xm-1

    i1

    ij

    m

    i j

    j

    net xw

    Letxm = 1and wim = i.

  • 8/10/2019 Lecture1 Introduction to ANN

    54/70

    The General Weight Learning Rule

    1

    1

    m

    i ijijj

    net xw

    Input:

    i

    .

    .

    .

    .

    .

    .

    wi1

    wi2

    wij

    wi,m-1

    x1

    x2

    xj

    xm-11

    ij

    m

    i j

    j

    net xw

    Letxm = 1and wim = i.

    xm=

    1

    wim=i

    ( )TWe want

  • 8/10/2019 Lecture1 Introduction to ANN

    55/70

    The General Weight Learning Rule

    Input:

    i

    .

    .

    .

    .

    .

    .

    wi1

    wi2

    wij

    wi,m-1

    x1

    x2

    xj

    xm-1

    1 ij

    m

    i jj

    net xw

    xm=

    1

    wim=i

    yi

    wi=(wi1, wi2 ,,wim)T

    wi(t) = ?

    We wantto learn

  • 8/10/2019 Lecture1 Introduction to ANN

    56/70

  • 8/10/2019 Lecture1 Introduction to ANN

    57/70

    The General Weight Learning Rule

    wix yi

    r diLearningSignalGenerator

    ( , , )r i if dw x

  • 8/10/2019 Lecture1 Introduction to ANN

    58/70

    The General Weight Learning Rule

    wix yi

    r diLearningSignalGenerator

    )()( trti xw

    ( , , )r i if dw x

    )()(

  • 8/10/2019 Lecture1 Introduction to ANN

    59/70

    The General Weight Learning Rule

    wix yi

    r diLearningSignalGenerator

    ( ) ( )i t tr w x

    )()( trti xw

    Learning Rate

    ( , , )r i if dw x

    ( )TWe want

  • 8/10/2019 Lecture1 Introduction to ANN

    60/70

    The General Weight Learning Rule

    wi=(wi1, wi2 ,,wim)T

    W wto learn

    ( , , )r i ir f d w x( ) ( )i t tr w x

    ( 1) ( ) ( ) ( ) ( ) ( )( , , )t t t t t t i i r i if d w w w x x

    Discrete-Time Weight Modification Rule:

    Continuous-Time Weight Modification Rule:

    ( )( )i

    d tr t

    dt

    w

    x

  • 8/10/2019 Lecture1 Introduction to ANN

    61/70

    Hebbs Learning Law Hebb [1994] hypothesis that when an axonal input

    fromAtoBcauses neuron B to immediately emit

    a pulse (fire) andthis situation happensrepeatedly or persistently.

    Then, the efficacy of that axonal input, in termsof ability to help neuron B to fire in future, issomehow increased.

    Hebbs learning rule is a unsupervisedlearningrule.

  • 8/10/2019 Lecture1 Introduction to ANN

    62/70

  • 8/10/2019 Lecture1 Introduction to ANN

    63/70

    Introduction to ArtificialNeural Networks

    DistributedRepresentations

  • 8/10/2019 Lecture1 Introduction to ANN

    64/70

    Distributed Representations Distributed Representation:

    An entity is represented by a pattern ofactivity distributed over many PEs.

    Each Processing element is involved inrepresenting many different entities.

    Local Representation: Each entity is represented by one PE.

  • 8/10/2019 Lecture1 Introduction to ANN

    65/70

    ExampleP0P1P2P3P4P5P6P7P8P9P10P11P12P13P14P15

    + _ + + _ _ _ _ + + + + + _ _ _

    +_

    + +_ _ _ _

    +_

    +_

    + +_

    +

    + +_

    +_

    + +_

    +_ _

    + + + +_

    Dog

    Cat

    Bread

    Act as a content addressable memory.

  • 8/10/2019 Lecture1 Introduction to ANN

    66/70

    P0P1P2P3P4P5P6P7P8P9P10P11P12P13P14P15

    + _ + + _ _ _ _ + + + + + _ _ _

    +_

    + +_ _ _ _

    +_

    +_

    + +_

    +

    + +_

    +_

    + +_

    +_ _

    + + + +_

    Dog

    Cat

    Bread

    Advantages

    P0P1P2P3P4P5P6P7P8P9P10P11P12P13P14P15

    + + + +

    What is this?

    Act as a content addressable memory.

    M k i d ti

  • 8/10/2019 Lecture1 Introduction to ANN

    67/70

    Advantages

    P0P1P2P3P4P5P6P7P8P9P10P11P12P13P14P15

    + _ + + _ _ _ _ + + + + + _ _ _

    +_

    + +_ _ _ _

    +_

    +_

    + +_

    +

    + +_

    +_

    + +_

    +_ _

    + + + +_

    Dog

    Cat

    Bread

    Make induction easy.

    P0P1P2P3P4P5P6P7P8P9P10P11P12P13P14P15

    +_ _

    +_ _ _ _

    + + + + + +_ _

    Fido

    Dog has 4 legs? How many for Fido?

    Act as a content addressable memory.

    M k i d ti

  • 8/10/2019 Lecture1 Introduction to ANN

    68/70

    Advantages

    P0P1P2P3P4P5P6P7P8P9P10P11P12P13P14P15

    + _ + + _ _ _ _ + + + + + _ _ _

    +_

    + +_ _ _ _

    +_

    +_

    + +_

    +

    + +_

    +_

    + +_

    +_ _

    + + + +_

    Dog

    Cat

    Bread

    Make induction easy.

    Make the creation of new entities orconcept easy (without allocation ofnew hardware).

    + + _ _ _ + + _ + _ _ _ + + + _Doughnut

    Add doughnut by changing weights.

    Act as a content addressable memory.

    M k i d ti

  • 8/10/2019 Lecture1 Introduction to ANN

    69/70

    Advantages

    P0P1P2P3P4P5P6P7P8P9P10P11P12P13P14P15

    + _ + + _ _ _ _ + + + + + _ _ _

    +_

    + +_ _ _ _

    +_

    +_

    + +_

    +

    + +_

    +_

    + +_

    +_ _

    + + + +_

    Dog

    Cat

    Bread

    Make induction easy.

    Make the creation of new entities orconcept easy (without allocation ofnew hardware).

    Fault Tolerance.

    Some PEs break down dont cause problem.

  • 8/10/2019 Lecture1 Introduction to ANN

    70/70

    Disadvantages How to understand?

    How to modify?

    Learning procedures are required.