50 likes | 185 Views
Back-Propagation MLP Neural Network Optimizer. ECE 539 Andrew Beckwith. Back-Propagation MLP Network Optimizer. Purpose Methods Features. Purpose. Configuring a Neural Network and its parameters is often a long and experimental process with much guess work. Let the computer do it for you.
E N D
Back-Propagation MLP Neural Network Optimizer ECE 539 Andrew Beckwith
Back-Propagation MLP Network Optimizer • Purpose • Methods • Features
Purpose • Configuring a Neural Network and its parameters is often a long and experimental process with much guess work. • Let the computer do it for you. • Design and implement a program that can test multiple network configurations with easy setup. • Allow user to modify data properly by enhancing important features and minimizing features with little importance or detrimental qualities.
Methods • Use back-propagation algorithm with momentum • To test multiple configurations, use brute force method and keep track of most successful configuration. • Only parameter user cannot control is the number of neurons per hidden layer. • Each configuration is tested with 2, 3, 5, and 10 neurons per hidden layer. The last test is a random initialization between 1 and 10 for each layer. • Use hyperbolic tangent activation function for hidden neurons and sigmoidal activation function for output neurons. One could change this in the source code if desired.
Features • Allow user to open data file, view mean and standard deviation for each feature of each class for modification purposes. • Allow user to enter ranges and number of trials for parameters such as: max epoch, epoch size, learning rate, momentum constant, and the number of hidden layers. • Allow user to set a tolerance to achieve maximum classification rate. • Allow user to view entire network – network configuration, weight values, etc.