Adaptive Computation and Machine Learning
The goal of building systems that can adapt to their environments and learn from their experience has attracted researchers from many fields, including computer science, engineering, mathematics, physics, neuroscience, and cognitive science. Out of this research has come a wide variety of learning techniques, including methods for learning decision trees, decision rules, neural networks, statistical classifiers, and probabilistic graphical models. The researchers in these various areas have also produced several different theoretical frameworks for understanding these methods, such as computational learning theory, Bayesian learning theory, classical statistical theory, minimum description length theory, and statistical mechanics approaches. These theories provide insight into experimental results and help to guide the development of improved learning algorithms. A goal of the series is to promote the unification of the many diverse strands of machine learning research and to foster high quality research and innovative applications. This series will publish works of the highest quality that advance the understanding and practical application of machine learning and adaptive computation. Research monographs, introductory and advanced level textbooks, how-to books for practitioners will all be considered. For information on the submission of proposals and manuscripts, please contact any of the series editors above or the publisher, Marie Lee (email@example.com). The series editor is Thomas G. Dietterich, and the associate series editors are Christopher M. Bishop, David Herckerman, Michael I. Jordan, and Michael Kerns.