Nova Publishers
My Account Nova Publishers Shopping Cart
HomeBooksSeriesJournalsReference CollectionseBooksInformationSalesImprintsFor Authors
  Top » Catalog » Books » Computer Science and Robotics » Horizons in Computer Science Research. Volume 7 Chapters » My Account  |  Cart Contents  |  Checkout   
Quick Find
Use keywords to find the product you are looking for.
Advanced Search
What's New? more
Advances in Animal Science and Zoology. Volume 12
Shopping Cart more
0 items
Shipping & Returns
Privacy Notice
Conditions of Use
Contact Us
Notifications more
NotificationsNotify me of updates to Parallelization of Neural Network Building and Training: An Original Decomposition Method (pp.193-223)
Tell A Friend
Tell someone you know about this product.
Parallelization of Neural Network Building and Training: An Original Decomposition Method (pp.193-223) $100.00
Authors:  (Marc Sauget, Sylvain Contassot-Vivier, Michel Salomon, IRMA/ENISYS, University of Franche-Comte, France, and others)
Since the first developments of neural networks by Pitts and McCulloch, the major
encountered problems lie in their building and learning. Indeed, there are some results
proving that a feed­forward multi­layer perceptron neural network can be used as an
universal interpolator. Unfortunately, there is neither any indication on how to build
an optimized topology, nor a method to choose the best suited learning algorithm to
train the network. Many learning algorithms give good results, like the classical back­
propagation algorithm for which various optimizations have been proposed. Some of
these optimizations change the network structure, like the Square MLP or the HPU
designs, whereas others improve the learning process, like the QuickProp or the Re­
silient back­Propagation (RPROP) algorithms. Nonetheless, these works are based on
neural networks having a static structure which have to be inferred manually accord­
ing to the user's experience. In this chapter, we present a way to adapt automatically
the neural network topology to the application context. In fact, we present an efficient
method that permits to obtain a parallel building and learning based on an original domain decomposition. This chapter describes, for both aspects, the corresponding
algorithms and gives comparative results showing the relevance of our approach. In
addition, the exploitation aspect of the obtained neural network is also addressed in
the last part. We present a multi­threaded version of our Neurad application used to
compute irradiation doses in any environment. 

Available Options:
This Item Is Currently Unavailable.
Special Focus Titles
01.Laryngeal Cancer: Risk Factors, Symptoms and Treatment
02.Breast Surgery: Indications and Techniques
03.Cursed? Biologic and Cultural Aspects of the Menstrual Cycle and Menstruation
04.The Uses of Cocoa and Cupuaçu Byproducts in Industry, Health, and Gastronomy
05.Agriculture, Food, and Food Security: Some Contemporary Global Issues
06.Fungicides: Perspectives, Resistance Management and Risk Assessment
07.Current Developments in Alexithymia - A Cognitive and Affective Deficit
08.Political Concerns and Literary Topoi in French Grand Opera
09.Abdul Aziz Said: The Mualim, The Inspiration
10.Teachers and Teaching: Global Practices, Challenges, and Prospects
11.A Comprehensive Investigation on Executive-Employee Pay Gap of Chinese Enterprises: Antecedents and Consequences
12.American Alligators: Habitats, Behaviors, and Threats

Nova Science Publishers
© Copyright 2004 - 2018

Parallelization of Neural Network Building and Training: An Original Decomposition Method (pp.193-223)