Computational Learning Theory and Natural Learning Systems: Making learning systems practical


Book Description

This is the fourth and final volume of papers from a series of workshops called "Computational Learning Theory and Ǹatural' Learning Systems." The purpose of the workshops was to explore the emerging intersection of theoretical learning research and natural learning systems. The workshops drew researchers from three historically distinct styles of learning research: computational learning theory, neural networks, and machine learning (a subfield of AI). Volume I of the series introduces the general focus of the workshops. Volume II looks at specific areas of interaction between theory and experiment. Volumes III and IV focus on key areas of learning systems that have developed recently. Volume III looks at the problem of "Selecting Good Models." The present volume, Volume IV, looks at ways of "Making Learning Systems Practical." The editors divide the twenty-one contributions into four sections. The first three cover critical problem areas: 1) scaling up from small problems to realistic ones with large input dimensions, 2) increasing efficiency and robustness of learning methods, and 3) developing strategies to obtain good generalization from limited or small data samples. The fourth section discusses examples of real-world learning systems. Contributors : Klaus Abraham-Fuchs, Yasuhiro Akiba, Hussein Almuallim, Arunava Banerjee, Sanjay Bhansali, Alvis Brazma, Gustavo Deco, David Garvin, Zoubin Ghahramani, Mostefa Golea, Russell Greiner, Mehdi T. Harandi, John G. Harris, Haym Hirsh, Michael I. Jordan, Shigeo Kaneda, Marjorie Klenin, Pat Langley, Yong Liu, Patrick M. Murphy, Ralph Neuneier, E.M. Oblow, Dragan Obradovic, Michael J. Pazzani, Barak A. Pearlmutter, Nageswara S.V. Rao, Peter Rayner, Stephanie Sage, Martin F. Schlang, Bernd Schurmann, Dale Schuurmans, Leon Shklar, V. Sundareswaran, Geoffrey Towell, Johann Uebler, Lucia M. Vaina, Takefumi Yamazaki, Anthony M. Zador.




Computational Learning Theory and Natural Learning Systems: Intersections between theory and experiment


Book Description

Annotation These original contributions converge on an exciting and fruitful intersection of three historically distinct areas of learning research: computational learning theory, neural networks, and symbolic machine learning. Bridging theory and practice, computer science and psychology, they consider general issues in learning systems that could provide constraints for theory and at the same time interpret theoretical results in the context of experiments with actual learning systems. In all, nineteen chapters address questions such as, What is a natural system? How should learning systems gain from prior knowledge? If prior knowledge is important, how can we quantify how important? What makes a learning problem hard? How are neural networks and symbolic machine learning approaches similar? Is there a fundamental difference in the kind of task a neural network can easily solve as opposed to those a symbolic algorithm can easily solve? Stephen J. Hanson heads the Learning Systems Department at Siemens Corporate Research and is a Visiting Member of the Research Staff and Research Collaborator at the Cognitive Science Laboratory at Princeton University. George A. Drastal is Senior Research Scientist at Siemens Corporate Research. Ronald J. Rivest is Professor of Computer Science and Associate Director of the Laboratory for Computer Science at the Massachusetts Institute of Technology.




Boosting


Book Description

An accessible introduction and essential reference for an approach to machine learning that creates highly accurate prediction rules by combining many weak and inaccurate ones. Boosting is an approach to machine learning based on the idea of creating a highly accurate predictor by combining many weak and inaccurate “rules of thumb.” A remarkably rich theory has evolved around boosting, with connections to a range of topics, including statistics, game theory, convex optimization, and information geometry. Boosting algorithms have also enjoyed practical success in such fields as biology, vision, and speech processing. At various times in its history, boosting has been perceived as mysterious, controversial, even paradoxical. This book, written by the inventors of the method, brings together, organizes, simplifies, and substantially extends two decades of research on boosting, presenting both theory and applications in a way that is accessible to readers from diverse backgrounds while also providing an authoritative reference for advanced researchers. With its introductory treatment of all material and its inclusion of exercises in every chapter, the book is appropriate for course use as well. The book begins with a general introduction to machine learning algorithms and their analysis; then explores the core theory of boosting, especially its ability to generalize; examines some of the myriad other theoretical viewpoints that help to explain and understand boosting; provides practical extensions of boosting for more complex learning problems; and finally presents a number of advanced theoretical topics. Numerous applications and practical illustrations are offered throughout.




Computational Learning Theory


Book Description

Content Description #Includes bibliographical references and index.





Book Description




ECAI 2000


Book Description




Intelligent Data Engineering and Automated Learning - IDEAL 2000. Data Mining, Financial Engineering, and Intelligent Agents


Book Description

X Table of Contents Table of Contents XI XII Table of Contents Table of Contents XIII XIV Table of Contents Table of Contents XV XVI Table of Contents K.S. Leung, L.-W. Chan, and H. Meng (Eds.): IDEAL 2000, LNCS 1983, pp. 3›8, 2000. Springer-Verlag Berlin Heidelberg 2000 4 J. Sinkkonen and S. Kaski Clustering by Similarity in an Auxiliary Space 5 6 J. Sinkkonen and S. Kaski Clustering by Similarity in an Auxiliary Space 7 0.6 1.5 0.4 1 0.2 0.5 0 0 10 100 1000 10000 10 100 1000 Mutual information (bits) Mutual information (bits) 8 J. Sinkkonen and S. Kaski 20 10 0 0.1 0.3 0.5 0.7 Mutual information (mbits) Analyses on the Generalised Lotto-Type Competitive Learning Andrew Luk St B&P Neural Investments Pty Limited, Australia Abstract, In generalised lotto-type competitive learning algorithm more than one winner exist. The winners are divided into a number of tiers (or divisions), with each tier being rewarded differently. All the losers are penalised (which can be equally or differently). In order to study the various properties of the generalised lotto-type competitive learning, a set of equations, which governs its operations, is formulated. This is then used to analyse the stability and other dynamic properties of the generalised lotto-type competitive learning.




Semi-Supervised Learning


Book Description

A comprehensive review of an area of machine learning that deals with the use of unlabeled data in classification problems: state-of-the-art algorithms, a taxonomy of the field, applications, benchmark experiments, and directions for future research. In the field of machine learning, semi-supervised learning (SSL) occupies the middle ground, between supervised learning (in which all training examples are labeled) and unsupervised learning (in which no label data are given). Interest in SSL has increased in recent years, particularly because of application domains in which unlabeled data are plentiful, such as images, text, and bioinformatics. This first comprehensive overview of SSL presents state-of-the-art algorithms, a taxonomy of the field, selected applications, benchmark experiments, and perspectives on ongoing and future research.Semi-Supervised Learning first presents the key assumptions and ideas underlying the field: smoothness, cluster or low-density separation, manifold structure, and transduction. The core of the book is the presentation of SSL methods, organized according to algorithmic strategies. After an examination of generative models, the book describes algorithms that implement the low-density separation assumption, graph-based methods, and algorithms that perform two-step learning. The book then discusses SSL applications and offers guidelines for SSL practitioners by analyzing the results of extensive benchmark experiments. Finally, the book looks at interesting directions for SSL research. The book closes with a discussion of the relationship between semi-supervised learning and transduction.




Inductive Logic Programming


Book Description

This book constitutes the refereed proceedings of the 9th International Conference on Inductive Logic Programming, ILP-99, held in Bled, Slovenia, in June 1999. The 24 revised papers presented were carefully reviewed and selected from 40 submissions. Also included are abstracts of three invited contributions. The papers address all current issues in inductive logic programming and inductive learning, from foundational and methodological issues to applications, e.g. in natural language processing, knowledge discovery, and data mining.




Inductive Logic Programming


Book Description

Wewishtothank AlfredHofmannandAnnaKramerofSpringer-Verlagfortheircooperationin publishing these proceedings. Finally, we gratefully acknowledge the nancial supportprovidedbythesponsorsofILP-99.