ICANN ’94


Book Description

From its early beginnings in the fifties and sixties the field of neural networks has been steadily growing. The first wave was driven by a handful of pioneers who first discovered analogies between machines and biological systems in communication, control and computing. Technological constraints held back research considerably, but gradually computers have become less expensive and more accessible and software tools inceasingly more powerful. Mathematical techniques, developed by computer-aware people, have steadily accumulated and the second wave has begun. Researchers from such diverse areas as psychology, mathematics, physics, neuroscience and engineering now work together in the neural networking field.




ICANN '94


Book Description




Self-Organizing Maps


Book Description

The Self-Organizing Map (SOM), with its variants, is the most popular artificial neural network algorithm in the unsupervised learning category. About 4000 research articles on it have appeared in the open literature, and many industrial projects use the SOM as a tool for solving hard real world problems. Many fields of science have adopted the SOM as a standard analytical tool: statistics, signal processing, control theory, financial analyses, experimental physics, chemistry and medicine. This new edition includes a survey of over 2000 contemporary studies to cover the newest results. Case examples are provided with detailed formulae, illustrations, and tables. Further, a new chapter on software tools for SOM has been included whilst other chapters have been extended and reorganised.




RAM-based Neural Networks


Book Description

RAM-based networks are a class of methods for building pattern recognition systems. Unlike other neural network methods, they learn very quickly and as a result are applicable to a wide variety of problems. This important book presents the latest work by the majority of researchers in the field of RAM-based networks.




Advances in Neural Information Processing Systems 7


Book Description

November 28-December 1, 1994, Denver, Colorado NIPS is the longest running annual meeting devoted to Neural Information Processing Systems. Drawing on such disparate domains as neuroscience, cognitive science, computer science, statistics, mathematics, engineering, and theoretical physics, the papers collected in the proceedings of NIPS7 reflect the enduring scientific and practical merit of a broad-based, inclusive approach to neural information processing. The primary focus remains the study of a wide variety of learning algorithms and architectures, for both supervised and unsupervised learning. The 139 contributions are divided into eight parts: Cognitive Science, Neuroscience, Learning Theory, Algorithms and Architectures, Implementations, Speech and Signal Processing, Visual Processing, and Applications. Topics of special interest include the analysis of recurrent nets, connections to HMMs and the EM procedure, and reinforcement- learning algorithms and the relation to dynamic programming. On the theoretical front, progress is reported in the theory of generalization, regularization, combining multiple models, and active learning. Neuroscientific studies range from the large-scale systems such as visual cortex to single-cell electrotonic structure, and work in cognitive scientific is closely tied to underlying neural constraints. There are also many novel applications such as tokamak plasma control, Glove-Talk, and hand tracking, and a variety of hardware implementations, with particular focus on analog VLSI.




Artificial Neural Nets and Genetic Algorithms


Book Description

This is the third in a series of conferences devoted primarily to the theory and applications of artificial neural networks and genetic algorithms. The first such event was held in Innsbruck, Austria, in April 1993, the second in Ales, France, in April 1995. We are pleased to host the 1997 event in the mediaeval city of Norwich, England, and to carryon the fine tradition set by its predecessors of providing a relaxed and stimulating environment for both established and emerging researchers working in these and other, related fields. This series of conferences is unique in recognising the relation between the two main themes of artificial neural networks and genetic algorithms, each having its origin in a natural process fundamental to life on earth, and each now well established as a paradigm fundamental to continuing technological development through the solution of complex, industrial, commercial and financial problems. This is well illustrated in this volume by the numerous applications of both paradigms to new and challenging problems. The third key theme of the series, therefore, is the integration of both technologies, either through the use of the genetic algorithm to construct the most effective network architecture for the problem in hand, or, more recently, the use of neural networks as approximate fitness functions for a genetic algorithm searching for good solutions in an 'incomplete' solution space, i.e. one for which the fitness is not easily established for every possible solution instance.




Handbook Of Character Recognition And Document Image Analysis


Book Description

Optical character recognition and document image analysis have become very important areas with a fast growing number of researchers in the field. This comprehensive handbook with contributions by eminent experts, presents both the theoretical and practical aspects at an introductory level wherever possible.




Hardware Implementation of Intelligent Systems


Book Description

Intelligent systems are now being used more commonly than in the past. These involve cognitive, evolving and artificial-life, robotic, and decision making systems, to name a few. Due to the tremendous speed of development, on both fundamental and technological levels, it is virtually impossible to offer an up-to-date, yet comprehensive overview of this field. Nevertheless, the need for a volume presenting recent developments and trends in this domain is huge, and the demand for such a volume is continually increasing in industrial and academic engineering 1 communities. Although there are a few volumes devoted to similar issues , none offer a comprehensive coverage of the field; moreover they risk rapidly becoming obsolete. The editors of this volume cannot pretend to fill such a large gap. However, it is the editors' intention to fill a significant part of this gap. A comprehensive coverage of the field should include topics such as neural networks, fuzzy systems, neuro-fuzzy systems, genetic algorithms, evolvable hardware, cellular automata-based systems, and various types of artificial life-system implementations, including autonomous robots. In this volume, we have focused on the first five topics listed above. The volume is composed of four parts, each part being divided into chapters, with the exception of part 4. In Part 1, the topics of "Evolvable Hardware and GAs" are addressed. In Chapter 1, "Automated Design Synthesis and Partitioning for Adaptive Reconfigurable Hardware", Ranga Vemuri and co-authors present state-of-the-art adaptive architectures, their classification, and their applications.




Advances in Self-Organising Maps


Book Description




Second-Order Methods for Neural Networks


Book Description

About This Book This book is about training methods - in particular, fast second-order training methods - for multi-layer perceptrons (MLPs). MLPs (also known as feed-forward neural networks) are the most widely-used class of neural network. Over the past decade MLPs have achieved increasing popularity among scientists, engineers and other professionals as tools for tackling a wide variety of information processing tasks. In common with all neural networks, MLPsare trained (rather than programmed) to carryout the chosen information processing function. Unfortunately, the (traditional' method for trainingMLPs- the well-knownbackpropagation method - is notoriously slow and unreliable when applied to many prac tical tasks. The development of fast and reliable training algorithms for MLPsis one of the most important areas ofresearch within the entire field of neural computing. The main purpose of this book is to bring to a wider audience a range of alternative methods for training MLPs, methods which have proved orders of magnitude faster than backpropagation when applied to many training tasks. The book also addresses the well-known (local minima' problem, and explains ways in which fast training methods can be com bined with strategies for avoiding (or escaping from) local minima. All the methods described in this book have a strong theoretical foundation, drawing on such diverse mathematical fields as classical optimisation theory, homotopic theory and stochastic approximation theory.