1 / 24

Evolving Neural Networks

Evolving Neural Networks. Learning and Evolution: Their secret conspiracy to take over the world. Adaptation. There are two forms of adaptation Learning Training on a set of examples. Fitting your behavior to training data. Minimizing error. Evolution Population based search.

bsnook
Download Presentation

Evolving Neural Networks

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Evolving Neural Networks Learning and Evolution: Their secret conspiracy to take over the world.

  2. Adaptation • There are two forms of adaptation • Learning • Training on a set of examples. • Fitting your behavior to training data. • Minimizing error. • Evolution • Population based search. • Random Mutation • Reproduction • Fitness Selection

  3. Combining The Two • Combining the strategies of evolutionary algorithms with learning produces “evolutionary artificial neural networks” or EANNs. • This increases the adaptability of both in a way that neither system could achieve on their own. • This also can give rise to extremely complex relationships between the two.

  4. What Can an EANN Do? • Adjust the network weights. • Learning rules. • Evolution • Build an architecture to fit the problem at hand • Does it need hidden layers? • Is the propagation delay too large? • Is the environment dynamic? • Perhaps the most lofty goal is evolving a learning rule for the network.

  5. Evolving The Weights • Why evolve the weights? What’s wrong with backpropagation? • Backpropagation is a gradient accent algorithm. These algorithms can get stuck on a local maximum or local minimum solutions. Optimal solution Local Max Initial weights

  6. Overcoming Local Min/Max • Since evolutionary algorithms are population based they have no need for gradient information. • Subsequently they are better in noisy or complex environments • Though getting stuck on local maxes can be avoided there is no guarantee that any maxima will be found. Population Samples

  7. The Permutation Problem • A gigantic problem that reflects the noisy real world solution is the permutation or competing convention problem. • Caused by a many-to-one mapping from genotype to phenotype within neural nets. • This problem kills the efficiency and effectiveness of crossover because effective parents can produce slacker offspring…..sorry mom and dad. 7 3 3 7 10 4 4 2 10 2

  8. Does Evolution Beat Backpropagation? • Absolutely! • D.L. Prados claims that GA-based training algorithms completed the tests in 3 hours 40 minutes, while networks using the generalized delta rule finished in 23 hours 40 minutes. • No way man! • H. Kitano claims that when testing a Genetic Algorithm backpropagation combination, it was at best as efficient as other backpropagation variants in small networks, but rather crappy in large networks. • It just goes to show you that the best algorithm is always problem dependent.

  9. Evolving Network Architectures • The architecture is very important to the information processing capabilities of the neural network. • Small networks without a hidden layer can’t solve problems such as XOR, that are not linearly separable. • Large networks can easily overfit a problem to match the training data, constricting their ability to generalize a problem set.

  10. Constructing the Network • Constructive algorithms take a minimal network and build up new layers nodes and connections during training. • Destructive algorithms take a maximal network and prunes unnecessary layers nodes and connections during training. • The network is evaluated based on specific performance criteria, for example, lowest training error or lowest network complexity.

  11. The Difficulties Involved • The architectural search space is infinitely large • Changes in nodes and connection can have a discontinuous effect on network performance. • Mapping from network architecture to behavior is indirect, dependent on evaluation and epistatic…don’t ask. • Similar architectures may have highly divergent performance. • Similar performance may be attained by diverse architectures.

  12. Evolving The Learning Rule • The optimal learning rules is highly dependent on the architecture of the network • Designing an optimal learning rule is very hard when little is known about the architecture, which is generally the case. • Different rules apply to different problems. • Certain rules make it easier to learn patterns and in this regard are more efficient. • Less efficient learning rules can learn exceptions to the patterns. • Which one is better? Depends on who you ask.

  13. Facing Reality • One can see the advantage of having a network that is capable of learning to learn. Combined with the ability to adjust it architecture, this sort of neural net would seem to be approaching intelligence. • The reality is that the relationship between learning and evolution is extremely complex. • Subsequently research into evolving learning rules is really in it’s infant stages.

  14. Lamarckian Evolution. • On a side note Belew McInerney & Schraudolf state that their findings suggest a reason why Lamarckian inheritance cannot be possible. • Due to issues like the permutation problem it is impossible to transcribe network behavior into a genomic encoding since it is possible that there are infinitely many encodings that will produce a phenotype.

  15. Can It Be Implemented? • Yes it can. • That’s all I got for this slide. It really seems like a waste, but at least its not paper..

  16. Go-Playing Neural Nets • Alex Lubberts and Risto Miikkulainen have created a Co-Evolving Go-playing Neural Network. • No program play go at any significant level of experience. • They decided that a parasite host relationship will foster a competitive environment for evolution.

  17. The Fight to Survive • The host • Attempt to find an optimal solution for winning on scaled down 5X5 Go board. • The parasites • Attempt to find and exploit weaknesses within the host population. • Populations are evaluated one at a time so each population takes turns being a host or a parasite.

  18. Tricks of the Trade • Competitive Fitness Sharing • Unusual or special individuals are rewarded. • If an individual whups an opponent that is very tough even if that individual lost most of it’s other games that host may still be rewarded. • Shared Sampling • To cut down on the number of games a sample set is pitted against opponents. • Hall of Fame • Old timers that have competed well are put into a steel cage match with the tyros to ensure that new generations are improving.

  19. Conclusions • They found that co-evolution did in fact increase the playing ability of their networks. • After 40 generations using their tournament style selection the co-evolved networks had nearly tripled the number of wins against a similar network that was evolved without a host parasite relationship.

  20. EANN Used to Classify Galaxies • Erick Cantu-Paz & Chandrika Kamath • Attempting to bring automation in the classification of galaxies using neural networks. • The learning algorithm must be carefully tuned to the data. • The relevant features of a classification problem may not be known before building the network or tuning the learning rule.

  21. How It Worked • Six combinations of GA and NN where compared. • The interesting part is the evolution of feature selection • GAs where to select what features are important enough that the NN needs to know them in order to classify a galaxy. • GAs consistently selected half the features and supposedly of the half they selected most of the features where relevant to classification such as symmetry measures and angles. • Two point crossover was used along with a fitness bias towards networks that learn quickly.

  22. Findings • Several of the evolved networks where competitive with human designed networks. • The best evolving feature selection. • Identifying Bent-Double Galaxies 92.99% accuracy. • Identifying Non-Bent 83.65% accuracy.

  23. Alright! He’s About to Shut Up • So in conclusion there is a lot of room for improvement in EANNs and there is a lot to explore. • So quit what your doing and build an evolving learning program. Tell your friends that you have already given it a body and soon you will make it capable of replicating itself. If they are gullible enough you might get to watch them squirm.

  24. References, The lazy way. • Richard Belew, John McInerney, Nicol Schraudolph: Evolving Networks http://www-cse.ucsd.edu/users/rik/papers/alife91/evol-net.ps • Erick Cantu-Paz C. Kamath: Evolving Neural Networks for the Classification of Galaxies http://www.llnl.gov/CASC/sapphire/pubs/147020.pdf • Xin Yao Evolving Artificial Neural Networkshttp://www.cs.bham.ac.uk/~xin/papers/published_iproc_sep99.pdf • Alex Lubberts, Risto Miikkulainen Co-Evolving a Go-Playing Neural Network http://nn.cs.utexas.edu/downloads/papers/lubberts.coevolution-gecco01.pdf

More Related