vv COMPUTER SCIENCE TECHNICAL REPORT ABSTRACTS
CMU-CS-10-113
Computer Science Department
School of Computer Science, Carnegie Mellon University



CMU-CS-10-113

Learning Generative Models for
Protein Fold Families

Siveraman Balakirshnan*, Hetunandan Kamisetty*,
Jaime G. Carbonell, Christopher James Langmead

March 2010

CMU-CS-10-113.pdf


Keywords: Structure learning, generative models, probabilistic graphical models, proteins

Statistical models of the amino acid composition of the proteins within a fold family are widely used in science and engineering. Existing techniques for learning probabilistic graphical models from multiple sequence alignments either make strong assumptions about the conditional independencies within the model (e.g., HMMs), or else use sub-optimal algorithms to learn the structure and parameters of the model. We introduce an approach to learning the topological structure and parameters of an undirected probabilistic graphical model. The learning algorithm uses block-L1 regularization and solves a convex optimization problem, thus guaranteeing a globally optimal solution at convergence. The resulting model encodes both the position-specific conservation statistics and the correlated mutation statistics between sequential and long-range pairs of residues. Our model is generative, allowing for the design of new proteins that have corresponding statistical properties to those seen in nature. We apply our approach to two widely studied protein families: the WW and the PDZ folds. We demonstrate that our model is able to capture interactions that are important in folding and allostery. Our results additionally indicate that while the network of interactions within a protein is sparse, it is richer than previously believed.

24 pages

*Contributed equally to this paper.


Return to: SCS Technical Report Collection
School of Computer Science

This page maintained by reports@cs.cmu.edu