270 likes | 286 Views
Introduction. Course Objectives. This course gives a basic neural network architectures and learning rules.
E N D
Course Objectives This course gives a basic neural network architectures and learning rules. Emphasis is placed on the mathematical analysis of these networks, on methods of training them and on their application to practical engineering problems in such areas as pattern recognition, signal processing and control systems.
What Will Not Be Covered • Review of all architectures and learning rules • Implementation • VLSI • Optical • Parallel Computers • Biology • Psychology
Historical Sketch • Pre-1940: von Hemholtz, Mach, Pavlov, etc. • General theories of learning, vision, conditioning • No specific mathematical models of neuron operation • 1940s: Hebb, McCulloch and Pitts • Mechanism for learning in biological neurons • Neural-like networks can compute any arithmetic function • 1950s: Rosenblatt, Widrow and Hoff • First practical networks and learning rules • 1960s: Minsky and Papert • Demonstrated limitations of existing neural networks, new learning algorithms are not forthcoming, some research suspended • 1970s: Amari, Anderson, Fukushima, Grossberg, Kohonen • Progress continues, although at a slower pace • 1980s: Grossberg, Hopfield, Kohonen, Rumelhart, etc. • Important new developments cause a resurgence in the field
Applications • Aerospace • High performance aircraft autopilots, flight path simulations, aircraft control systems, autopilot enhancements, aircraft component simulations, aircraft component fault detectors • Automotive • Automobile automatic guidance systems, warranty activity analyzers • Banking • Check and other document readers, credit application evaluators • Defense • Weapon steering, target tracking, object discrimination, facial recognition, new kinds of sensors, sonar, radar and image signal processing including data compression, feature extraction and noise suppression, signal/image identification • Electronics • Code sequence prediction, integrated circuit chip layout, process control, chip failure analysis, machine vision, voice synthesis, nonlinear modeling
Applications • Financial • Real estate appraisal, loan advisor, mortgage screening, corporate bond rating, credit line use analysis, portfolio trading program, corporate financial analysis, currency price prediction • Manufacturing • Manufacturing process control, product design and analysis, process and machine diagnosis, real-time particle identification, visual quality inspection systems, beer testing, welding quality analysis, paper quality prediction, computer chip quality analysis, analysis of grinding operations, chemical product design analysis, machine maintenance analysis, project bidding, planning and management, dynamic modeling of chemical process systems • Medical • Breast cancer cell analysis, EEG and ECG analysis, prosthesis design, optimization of transplant times, hospital expense reduction, hospital quality improvement, emergency room test advisement
Applications • Robotics • Trajectory control, forklift robot, manipulator controllers, vision systems • Speech • Speech recognition, speech compression, vowel classification, text to speech synthesis • Securities • Market analysis, automatic bond rating, stock trading advisory systems • Telecommunications • Image and data compression, automated information services, real-time translation of spoken language, customer payment processing systems • Transportation • Truck brake diagnosis systems, vehicle scheduling, routing systems
Biology • Neurons respond slowly • The brain uses massively parallel computation – »1011 neurons in the brain – »104 connections per neuron
Biology The dendrites are tree-like receptivenetworks of nerve fibers that carry electrical signals into the cell body The cell body effectively sums and thresholds these incoming signals. The axonis a single long fiber that carries the signal from the cell body out to other neurons. The point of contact between an axon of one cell and a dendrite ofanother cell is called a synapse.
Neuron Model the weight “w” corresponds to the strength of a synapse the cell body is represented by the summation and the transfer function the neuron output “a”represents the signal on the axon
Single-Input Neuron Model The scalar input “p” is multiplied by “w” the scalar weight “w” to form “wp”, one of the terms that is sent to thesummer. The other input, 1, is multiplied by abias “b”and then passed to the summer. The summer output “n”, often referred to as thenet input , goes into a transfer function, which produces the scalar neuron output “a” .
Neuron ModelTransfer Functions The hard limit transfer function setsthe output of the neuron to 0 if the function argument is less than 0 or setsthe output of the neuron to 1 ifits argument is greater than or equal to 0.
Neuron ModelTransfer Functions The output of a linear transfer function is equal to its input a=n
Neuron ModelTransfer Functions This transfer function takes the input and squashes the output into the range 0 to 1, according to the expression:
Multiple-Input Neuron Model The neuron has a biasb, which is summed with the weightedinputs to form the net input n n = Wp + b the neuron output can be written as a = f (Wp + b)
Multiple-Input Neuron Model The neuron has a biasb, which is summed with the weightedinputs to form the net input n n = Wp + b the neuron output can be written as a = f (Wp + b) The first index indicates the particular neurondestination for that weight. The second index indicates the source ofthe signal fed to the neuron.
Network Architectures A Layer of Neurons Each of theR inputs is connected to each of the neurons The layer includes the weight matrixW, the summers, the bias vector b, thetransfer function boxes and the output vectora
Network Architectures A Layer of Neurons Abbreviated Notation
Network Architectures Multiple Layers of Neurons • Each layer has its own weightmatrix , its own bias vector , a netinput vector and an output vector • The number of the layer as a superscript to the names for each ofthese variables • A layer whose output is the network output is called an output layer. Theother layers are called hidden layers.
Network Architectures Multiple Layers of Neurons Abbreviated Notation
Simulation using MATLAB Neuron output=?
Simulation using MATLAB To set up this feedforward network net = newlin([1 3;1 3],1);
Simulation using MATLAB To set up this feedforward network net = newlin([1 3;1 3],1); Assignments net.IW{1,1} = [1 2]; net.b{1} = 0; P = [1 2 2 3; 2 1 3 1];
Simulation using MATLAB simulate the network A = sim(net,P) A = 5 4 8 5 To set up this feedforward network net = newlin([1 3;1 3],1); Assignments net.IW{1,1} = [1 2]; net.b{1} = 0; P = [1 2 2 3; 2 1 3 1];