330 likes | 339 Views
This lecture explores the training and analysis of neural networks, including topics such as the perceptron algorithm, local delta training algorithms, and general definitions of neural networks. Examples of neural network structures and methods for analyzing and synthesizing neural networks are also discussed.
E N D
Nanjing University of Science & Technology Pattern Recognition:Statistical and Neural Lonnie C. Ludeman Lecture 20 Oct 26, 2005
Lecture 20 Topics 1. Perceptron Algorithm Revisited 2. Local Delta Training Algorithm for ANE 3. General Definition of Neural Networks 4. Basic Neural Network Structures-Examples 5. Analysis and Synthesis of Neural Networks
Signum Function Activation Training Algorithm(Perceptron) Review y = +1 if input vector x is from C1 y = -1 if input vector x is from C2 Weight Update Algorithm
Question How do we train an Artificial Neural Element(ANE) to do classification ??? Answer Use the Delta Training Algorithm !!!
Given an Artificial Neural Element as follows Wish to find weight vector such that training patterns are correctly classified
Given: x(p) ε { x1, x2, … , xK } d( x(p) ) = { d(x1), d(x2), … , d(xK) } Define a performance measure Ep for sample x(p) and decision d[ x(p) ] as
Derivation of Delta weight update Equation Use the gradient method to minimize Ep New Weight wk+1 in terms of previous weight wk where the Gradient is
Substituting the gradient vector into the weight update gives the GeneralLocal Delta Algorithm or rewriting gives w(p+1) = w(p) + {d[x(p)] – f(net)} f /(net)) x(p) where net = wT(p)x(p) General Local Delta Algorithm Weight Update Equation
Sometimes called the Continuous Perceptron Training Algorithm
Case 1: Local Delta Algorithm for Training an ANE with Logistic Activation Function Given: Solution:
Substituting the derivative gives the Local algorithm for the Logistic Activation function as Local Weight Update Equation for Logistic Activation Function
Case 2: Local Delta Algorithm for for Training an ANE - Hyperbolic Tangent Activation Function Given: Solution; Taking derivative of the nonlinearity and substituting into the general update equation yields the following Local Weight Update Equation for Hyperbolic Activation Function
Scale Factors for Case 2: Tanh Activation Function SF = ( d[x(p) ] –f(net) )(1 – f 2(net) ) d[x(p)]= 1 SF1 = ( 1 – f(net) )(1 – f 2(net) ) d[x(p)] = -1 SF-1 = ( -1 – f(net) )(1 – f 2(net) ) d[x(p)]= 1 d[x(p)] = -1
Scale Factors for Case 2: Tanh Activation Function (desired values = +0.9 and -0.9 )
Case 3: Local Delta Algorithm for Training an ANE - Linear Activation Function Given: Solution:Taking derivative and substituting in general update equation gives Local Weight Update Equation for Linear Activation Function ( Widrow-Hoff Training Rule )
General Global Delta Algorithm Define a performance measure ETOT for all samples xk and decisions d[ xk) ] as Using Gradient technique gives the Global Delta Algorithm as Global Weight Update Equation
Definitions A Neural Network is defined as any connection of Neural Elements. An Artificial Neural Networkis defined as any connection of Artificial Neural Elements.
Examples of Artificial Neural Networks Feed Forward Artificial Neural Networks (a) Two Layer neural Network (b) Special Three Layer Form: Hyperplane-AND-OR structure (c) General 3-Layer Feedforward structure and nomenclature Feedback Artificial Neural Networks (d) One Layer Hopfield Net (e) Two Layer Feedback
(a) Example - Two Layer Neural Network Using Signum Nonlinearity
(b) Special Hyperplane-AND-OR structure input output Layer 1 Layer 2 Layer 3 y x Hyperplanes Logical AND Logical OR
Building Block- AND μ -(n-½)
(b) Example- Hyperplanes-AND-OR Structure Hyperplanes Layer allf(·) = u(·) unit step AND Layer OR Layer
Definitions: Analysis of Neural Networks- Given a Neural Network describe the output for all inputs ( Mathematical or computer generated) Synthesis of Neural Networks- Given a list of properties and requirements build a Neural Network to satisfy the requirements ( Mathematical or computer generated)
Example:Analyze the following Neural Network -1 0 1 1 1 0 0 -1 1 Solution: Determine the output y1(2) for all (x1,x2). (Next Lecture)
Example:Synthesize a Neural Network Given the following decision regions build a neural network to perform the classification process Solution: Use Hyperplane-AND-OR Structure (Next Lecture)
Summary Lecture 20 1. Perceptron Algorithm Revisited 2. Local Delta Training Algorithms for ANE 3. General Definition of Neural Networks 4. Basic Neural Network Structures-Examples 5. Analysis and Synthesis of Neural Networks
Question How do we train an Artificial Neural Network to perform the classification problem??? Answer Not a simple answer but we will look at one way that uses the backpropagation algorithm to do the Training. Not Today, we have to wait until Friday. ☺☻☺☻☺☻☺☻☺