In the era of “big data,” science is increasingly information driven, and the potential for computers to store, manage, and integrate massive amounts of data has given rise to such new disciplinary fields as biomedical informatics. Applied ontology offers a strategy for the organization of scientific information in computer-tractable form, drawing on concepts not only from computer and information science but also from linguistics, logic, and philosophy.
The field of Artificial Life (ALife) is now firmly established in the scientific world, but it has yet to achieve one of its original goals: an understanding of the emergence of life on Earth. The new field of Artificial Chemistries draws from chemistry, biology, computer science, mathematics, and other disciplines to work toward that goal. For if, as it has been argued, life emerged from primitive, prebiotic forms of self-organization, then studying models of chemical reaction systems could bring ALife closer to understanding the origins of life.
The genetic code is the Rosetta Stone by which we interpret the 3.3 billion letters of human DNA, the alphabet of life, and the discovery of the code has had an immeasurable impact on science and society. In 1968, Marshall Nirenberg, an unassuming government scientist working at the National Institutes of Health, shared the Nobel Prize for cracking the genetic code. He was the least likely man to make such an earth-shaking discovery, and yet he had gotten there before such members of the scientific elite as James Watson and Francis Crick.
The goal of structured prediction is to build machine learning models that predict relational information that itself has structure, such as being composed of multiple interrelated parts. These models, which reflect prior knowledge, task-specific relations, and constraints, are used in fields including computer vision, speech recognition, natural language processing, and computational biology. They can carry out such tasks as predicting a natural language sentence, or segmenting an image into meaningful components.
Sparse modeling is a rapidly developing area at the intersection of statistical learning and signal processing, motivated by the age-old statistical problem of selecting a small number of predictive variables in high-dimensional datasets. This collection describes key approaches in sparse modeling, focusing on its applications in fields including neuroscience, computational biology, and computer vision.
In this book, Dan Gusfield examines combinatorial algorithms to construct genealogical and exact phylogenetic networks, particularly ancestral recombination graphs (ARGs). The algorithms produce networks (or information about networks) that serve as hypotheses about the true genealogical history of observed biological sequences and can be applied to practical biological problems.
Systems techniques are integral to current research in molecular cell biology, and system-level investigations are often accompanied by mathematical models. These models serve as working hypotheses: they help us to understand and predict the behavior of complex systems. This book offers an introduction to mathematical concepts and techniques needed for the construction and interpretation of models in molecular systems biology.
The introduction of high-throughput methods has transformed biology into a data-rich science. Knowledge about biological entities and processes has traditionally been acquired by thousands of scientists through decades of experimentation and analysis. The current abundance of biomedical data is accompanied by the creation and quick dissemination of new information. Much of this information and knowledge, however, is represented only in text form--in the biomedical literature, lab notebooks, Web pages, and other sources.
Recent research in molecular biology has produced a remarkably detailed understanding of how living things operate. Becoming conversant with the intricacies of molecular biology and its extensive technical vocabulary can be a challenge, though, as introductory materials often seem more like a barrier than an invitation to the study of life.
Using the tools of information technology to understand the molecular machinery of the cell offers both challenges and opportunities to computational scientists. Over the past decade, novel algorithms have been developed both for analyzing biological data and for synthetic biology problems such as protein engineering. This book explains the algorithmic foundations and computational approaches underlying areas of structural biology including NMR (nuclear magnetic resonance); X-ray crystallography; and the design and analysis of proteins, peptides, and small molecules.