Military drones have recently been hailed as a revolutionary new technology that will forever change the conduct of war. And yet the United States and other countries have been deploying such unmanned military systems for more than a century. Written by a renowned authority in the field, this book documents the forgotten legacy of these pioneering efforts, offering the first comprehensive historical and technical accounting of unmanned air, land, sea, and underwater systems. Focusing on examples introduced during the two world wars, H. R.
In the 1960s, a team of Stanford musicians, engineers, computer scientists, and psychologists used computing in an entirely novel way: to produce and manipulate sound and create the sonic basis of new musical compositions. This group of interdisciplinary researchers at the nascent Center for Computer Research in Music and Acoustics (CCRMA, pronounced “karma”) helped to develop computer music as an academic field, invent the technologies that underlie it, and usher in the age of digital music.
The rise of the Indian information technology industry is a remarkable economic success story. Software and services exports from India amounted to less than $100 million in 1990, and today come close to $100 billion. But, as Dinesh Sharma explains in The Outsourcer, Indian IT’s success has a long prehistory; it did not begin with software support, or with American firms’ eager recruitment of cheap and plentiful programming labor, or with India’s economic liberalization of the 1990s.
The vast majority of all email sent every day is spam, a variety of idiosyncratically spelled requests to provide account information, invitations to spend money on dubious products, and pleas to send cash overseas. Most of it is caught by filters before ever reaching an in-box. Where does it come from?
In the 1930s a series of seminal works published by Alan Turing, Kurt Gödel, Alonzo Church, and others established the theoretical basis for computability. This work, advancing precise characterizations of effective, algorithmic computability, was the culmination of intensive investigations into the foundations of mathematics. In the decades since, the theory of computability has moved to the center of discussions in philosophy, computer science, and cognitive science.
Global warming skeptics often fall back on the argument that the scientific case for global warming is all model predictions, nothing but simulation; they warn us that we need to wait for real data, “sound science.” In A Vast Machine Paul Edwards has news for these skeptics: without models, there are no data. Today, no collection of signals or observations—even from satellites, which can “see” the whole planet with a single instrument—becomes global in time and space without passing through a series of data models. Everything we know about the world’s climate we know through models.
Today, women earn a relatively low percentage of computer science degrees and hold proportionately few technical computing jobs. Meanwhile, the stereotype of the male “computer geek” seems to be everywhere in popular culture. Few people know that women were a significant presence in the early decades of computing in both the United States and Britain. Indeed, programming in postwar years was considered woman’s work (perhaps in contrast to the more manly task of building the computers themselves).
This is a book about the computer revolution of the mid-twentieth century and the people who made it possible. Unlike most histories of computing, it is not a book about machines, inventors, or entrepreneurs. Instead, it tells the story of the vast but largely anonymous legions of computer specialists--programmers, systems analysts, and other software developers--who transformed the electronic computer from a scientific curiosity into the defining technology of the modern era.
The history of computing could be told as the story of hardware and software, or the story of the Internet, or the story of “smart” hand-held devices, with subplots involving IBM, Microsoft, Apple, Facebook, and Twitter. In this concise and accessible account of the invention and development of digital technology, computer historian Paul Ceruzzi offers a broader and more useful perspective.
Long ago, in 1985, personal computers came in two general categories: the friendly, childish game machine used for fun (exemplified by Atari and Commodore products); and the boring, beige adult box used for business (exemplified by products from IBM). The game machines became fascinating technical and artistic platforms that were of limited real-world utility. The IBM products were all utility, with little emphasis on aesthetics and no emphasis on fun. Into this bifurcated computing environment came the Commodore Amiga 1000.