In this engaging book, Jerry Fodor argues against the widely held view that mental processes are largely computations, that the architecture of cognition is massively modular, and that the explanation of our innate mental structure is basically Darwinian. Although Fodor has praised the computational theory of mind as the best theory of cognition that we have got, he considers it to be only a fragment of the truth. In fact, he claims, cognitive scientists do not really know much yet about how the mind works (the book's title refers to Steve Pinker's How the Mind Works).
In this book Simon Parsons describes qualitative methods for reasoning under uncertainty, "uncertainty" being a catch-all term for various types of imperfect information. The advantage of qualitative methods is that they do not require precise numerical information. Instead, they work with abstractions such as interval values and information about how values change. The author does not invent completely new methods for reasoning under uncertainty but provides the means to create qualitative versions of existing methods.
This book is for students and researchers who have a specific interest in learning and memory and want to understand how computational models can be integrated into experimental research on the hippocampus and learning. It emphasizes the function of brain structures as they give rise to behavior, rather than the molecular or neuronal details. It also emphasizes the process of modeling, rather than the mathematical details of the models themselves.
By the mid-1980s researchers from artificial intelligence, computer science, brain and cognitive science, and psychology realized that the idea of computers as intelligent machines was inappropriate. The brain does not run "programs"; it does something entirely different. But what? Evolutionary theory says that the brain has evolved not to do mathematical proofs but to control our behavior, to ensure our survival. Researchers now agree that intelligence always manifests itself in behavior—thus it is behavior that we must understand.
The annual conference on Neural Information Processing Systems (NIPS) is the flagship conference on neural computation. The conference is interdisciplinary, with contributions in algorithms, learning theory, cognitive science, neuroscience, vision, speech and signal processing, reinforcement learning and control, implementations, and diverse applications. Only about 30 percent of the papers submitted are accepted for presentation at NIPS, so the quality is exceptionally high. These proceedings contain all of the papers that were presented at the 2000 conference.
The idea of knowledge bases lies at the heart of symbolic, or "traditional," artificial intelligence. A knowledge-based system decides how to act by running formal reasoning procedures over a body of explicitly represented knowledge—a knowledge base. The system is not programmed for specific tasks; rather, it is told what it needs to know and expected to infer the rest.
Among the many approaches to formal reasoning about programs, Dynamic Logic enjoys the singular advantage of being strongly related to classical logic. Its variants constitute natural generalizations and extensions of classical formalisms. For example, Propositional Dynamic Logic (PDL) can be described as a blend of three complementary classical ingredients: propositional calculus, modal logic, and the algebra of regular events. In First-Order Dynamic Logic (DL), the propositional calculus is replaced by classical first-order predicate calculus.
Knowledge discovery and data mining (KDD) deals with the problem of extracting interesting associations, classifiers, clusters, and other patterns from data. The emergence of network-based distributed computing environments has introduced an important new dimension to this problem—distributed sources of data. Traditional centralized KDD typically requires central aggregation of distributed data, which may not always be feasible because of limited network bandwidth, security concerns, scalability problems, and other practical issues.
The annual AAAI National Conference provides a forum for information exchange and interaction among researchers from all disciplines of AI. Contributions include theoretical, experimental, and empirical results. Topics cover principles of cognition, perception, and action; the design, application, and evaluation of AI algorithms and systems; architectures and frameworks for classes of AI systems; and analyses of tasks and domains in which intelligent systems perform.
The annual conference on Neural Information Processing System (NIPS) is the flagship conference on neural computation. It draws preeminent academic researchers from around the world and is widely considered to be a showcase conference for new developments in network algorithms and architectures. The broad range of interdisciplinary research areas represented includes computer science, neuroscience, statistics, physics, cognitive science, and many branches of engineering, including signal processing and control theory.