The globalization of the economy, increasing number of transnational organizations, and rapid changes in robotics, information, and telecommunication technologies are just a few of the factors significantly altering organizational time scales, forms, complexity, and environments. Time scales have shrunk, new organizational forms are emerging, and organizational environments are expanding and mutating at unprecedented rates. Computational modeling affords opportunities to both understand and respond to these complex changes.
foreword by Gordon Bell and afterword by H.T. Kung Although researchers have proposed many mechanisms and theories for parallel systems, only a few have actually resulted in working computing platforms. The iWarp is an experimental parallel system that was designed and built jointly by Carnegie Mellon University and Intel Corporation. The system is based on the idea of integrating a VLIW processor and a sophisticated fine-grained communication system on a single chip. This book describes the complete iWarp system, from instruction-level parallelism to final parallel applications.
Building a computer ten times more powerful than all the networked computing capability in the United States is the subject of this book by leading figures in the high performance computing community. It summarizes the near-term initiatives, including the technical and policy agendas for what could be a twenty-year effort to build a petaFLOP scale computer. (A FLOP—Floating Point OPeration—is a standard measure of computer performance and a PetaFLOP computer would perform a million billion of these operations per second.)
The thirteen chapters written expressly for this book by logicians, theoretical computer scientists, philosophers, and semanticists address, from the perspective of mathematical logic, the problems of understanding and studying the flow of information through any information-processing system.
The Handbook of Theoretical Computer Science provides professionals and students with a comprehensive overview of the main results and developments in this rapidly evolving field. Volume A covers models of computation, complexity theory, data structures, and efficient computation in many recognized subdisciplines of theoretical computer science.
At the 1900 International Congress of Mathematicians, held that year in Paris, the German mathematician David Hilbert put forth a list of 23 unsolved problems that he saw as being the greatest challenges for twentieth-century mathematics. Hilbert's 10th problem, to find a method (what we now call an algorithm) for deciding whether a Diophantine equation has an integral solution, was solved by Yuri Matiyasevich in 1970. Proving the undecidability of Hilbert's 10th problem is clearly one of the great mathematical results of the century.
Parallel Distributed Processing is transforming the field of cognitive science. Microcognition provides a clear, readable guide to this emerging paradigm from a cognitive philosopher's point of view. It explains and explores the biological basis of PDP, its psychological importance, and its philosophical relevance.
Theoretical computer science provides the foundations for understanding and exploiting the concepts and mechanisms in computing and information processing. This handbook will provide professionals and students with a comprehensive overview of the main results and developments in this rapidly evolving field. It consists of thirty-seven chapters in two volumes, all addressing core areas of theoretical computer science as it is practiced today. The material is written by leading American and European researchers, and each volume may be used independently.
Animation provides a rich environment for actively exploring algorithms. Multiple, dynamic, graphical displays of an algorithm reveal properties that might otherwise be difficult to comprehend or even remain unnoticed. This exciting new approach to the study of algorithms is taken up by Marc Brown in Algorithm Animation.