Tutorial Abstracts
AM1. Computational analyses across the BioCyc collection of Pathway/Genome Databases
AM2. Synthetic biology with BioBrick parts
AM3. Flux-balance analysis of metabolic networks
PM1. The science behind 23andme
PM2. Next generation sequencing technologies and applications
PM3. Decoding ENCODE and other genomics information at UCSC
AM1. Computational analyses across the BioCyc collection of Pathway/Genome Databases
Peter Karp, Bioinformatics Research Group at SRI International
BioCyc is a collection of 370 Pathway/Genome Databases for many
organisms whose genomes have been completely sequenced. It is a large
and comprehensive resource for systems biology research. We expect
that many bioinformatics and computational biology researchers will be
interested in computing with BioCyc to address global biological
questions, such as studying the phylogenetic distribution and
evolution of metabolic pathways. The goal of this tutorial will be to
provide researchers with the information they need to perform global
analyses of BioCyc. The tutorial will cover the methodologies used to
create BioCyc, a description of the database schema and ontologies
that underly BioCyc, and descriptions of the APIs (in Perl, Java, and
Lisp) that are available to query BioCyc. The tutorial will also
present the Pathway Tools semantic inference layer, which is a library
of commonly used queries that we have encoded to save researchers
time.
Expected outcomes and goals: Students will learn how to perform
computational analyses across the large BioCyc collection of
Pathway/Genome Databases.
Prequisites: Basic familiarity with programming and databases, and
basic familiarity with biological concepts in genomics and metabolic
pathways.
Return to Top or Tutorial Schedule
AM2. Synthetic biology with BioBrick parts
Mackenzie Cowell and Jason Morrison, MIT
Synthetic biology is awesome! It's the application of engineering principles and process to the design and assembly of biological
systems, including a cycle of specification, design, modeling, implementation, and testing It is enabled by augmenting the genetic
engineer's toolbox (recombinant DNA, PCR, and automated sequencing) with nascent foundational technologies: automated construction,
standardization, and abstraction.
One of the first goals of this growing field is to create a robust set of freely available standard biological parts - modular
biological functions engineered to be easily combined while retaining predictable behavior - used to build novel biological devices
and systems. The production of these interchangeable biological parts depends on the successful development of standards for both
their structural and functional assembly, while their utility depends on the development of a usable abstraction hierarchy, in which
complexity can be compartmentalized and hidden, greatly accelerating the design process of devices and systems.
Already over 2000 standard biological parts exist and are available for use from the Registry of Standard Biological Parts at MIT.
Last year, the International Genetically Engineered Machine (iGEM) competition shipped over 82,000 of these biobricks(tm) standard
biological parts to more that 53 participating teams around the world. These teams used the parts to create a variety of novel
devices and systems, and to create new parts, all of which were contributed back to the community.
In this tutorial we will go over in detail the principles and practice of Synthetic Biology, as described above, and will introduce you
to some of the main community resources and participants in the field, such as the Registry of Standard Biological Parts, the iGEM
competition, the BioBricks Foundation, SynBERC, and others.
The BioBricks Foundation (BBF) is a non-profit organization devoted to enabling the development of free collections of standard
biological parts, currently by supporting an open technical standards setting process and developing a BioBricks(tm) legal scheme.
Return to Top or Tutorial Schedule
AM3. Flux-balance analysis of metabolic networks
Markus Covert, Stanford
This tutorial will demonstrate the power of linear optimization for large-scale modeling, particularly of metabolism. First, we will
discuss the need for new approaches to model large biological networks. Next, we will describe linear optimization methods, and show
how they greatly reduce the parameter problems that are associated with large-scale network modeling. We will then discuss how to
reconstruct the stoichiometric matrix for a metabolic network and show examples of how the flux balance approach can drive an experimental
discovery process.
Return to Top or Tutorial Schedule
PM1. The science behind 23andme
Serge Saxonov and Brian Naughton, 23andme
23andme is a personal genomics startup focused on creating web-based tools and content that allow consumers to access and benefit
from published scientific research related to their genetic information. 23andme currently uses an Illumina BeadChip to assay a standard
panel of 550K tagging SNPs, and a custom panel of 30K targeted SNPs.
This tutorial will cover the following:
• pros and cons of various genotyping platforms, including the design of 23andMe's custom panel of SNPs
• finding, parsing and relating new genetic association research to the public
• designing tools and algorithms to help users explore their genomes
• the potential for new research discoveries with community-driven research
Participants will learn the lay of the land of the new personal genomics field; understand how SNP chips work, and how they compare
to present- and next-generation sequencing; learn how the now-ubiquitous genome-wide association studies like Wellcome Trust's Case
Control Consortium efforts are performed, and how community-driven research could drive future discoveries.
Return to Top or Tutorial Schedule
PM2. Next generation sequencing technologies and applications
Nader Pourmand, Biomolecular Engineering, UCSC
Over the past few years, exciting advances in next-generation sequencing technology have seen the production of commercial instruments
capable of profound improvements in sequence throughput and economy. Early applications include cancer genome analysis, microbial
genomics, and the study of the Neanderthal genome. Competition in the field is increasingly intense, with start-ups vying with
established life science instrument makers to provide the gold-standard technology that will drive us towards the threshold of
the "$1000 genome.” A challenge for the Next Generation Sequencing Systems has been assembling relatively short read lengths to
overcome the complexity of nearly all genomes. We will discuss the development, chemistry and application of both the Genome Sequencer
FLX system and the SOLiD™ System.
Return to Top or Tutorial Schedule
PM3. Decoding ENCODE and other genomics information at UCSC
Jim Kent, UCSC
The Encyclopedia of DNA Elements (ENCODE) Project contains a treasure trove of information on the function of various pieces
of the genome based on a variety of modern high throughput experimental and computational methods. The project is especially
interesting for biologists studying chromatin, transcription factors, noncoding RNA, and those interested in a very high quality,
curated, collection of human coding genes. All of the ENCODE data is available at
http://genome.ucsc.edu, though some, as a
courtesy to the labs producing the data, should be considered pre-publication. The UCSC web site contains a number of useful
tools for browsing and analyzing ENCODE and other genomics data. All the data is also downloadable in dense, easy to parse,
and well documented file formats for analysis on your own computers.
This tutorial will start with an overview of the ENCODE data. This includes chromatin immunoprecipitation (ChIP) of histones
in various methylation and acylation states, and nuclease hypersensitivity experiments, and RNA expression assays on many cell
lines. Together these experiments shed much light on the chromatin state of various cell types, and the implications for transcription.
There is also ChIP data for a large number of transcription factors that control the expression of specific genes, and which together
are helpful in determining the combinatorical logic of the regulation of transcription. Other experiments aimed at understanding
transcriptional regulation include formaldehyde-assisted isolation of regulatory elements, and assays for the methylation status of
CpG rich regions. There is a sub-project, GENCODE, that aims to produce an exhaustive and high quality list of functional human
transcripts employing a variety of methods including high throughput sequencing of the ends of G-cap selected RNA, manual curation
of existing RNA data, computational predictions, and reverse-transcriptase PCR confirmation of predictions. In addition to these
genome-wide experiments the ENCODE project includes smaller, pilot projects to develop high throughput technologies to explore DNA
methylation, RNA/protein interactions and long-range DNA/DNA interactions, and new DNA/protein interaction assays both computationally
and experimentally. In all the ENCODE data is of interest to a wide variety of biomedical researchers, especially those interested in gene regulation.
The ENCODE data is stored and displayed, alongside other useful genomics data, at the UCSC Genome Informatic web site
(genome.ucsc.edu http://genome.ucsc.edu ).
In this tutorial we'll explore some of the high points of the ENCODE data as it relates
to some interesting genes using the UCSC Genome Browser. In the process we'll review the basic operations of the Genome Browser, and
introduce some new and advanced features such as sending a view in email so it can be shared with a colleague, and adding your own
tracks that can be seen alongside the genes, comparative genomics, and ENCODE data that is built into the Genome Browser. We'll then
shift to tools that are more useful for mining data from the genome as a whole rather than a particular gene. We'll use the Gene Sorter
to rapidly find and explore sets of genes that can be defined by many criteria including tissue expression levels, protein domains, and
Gene Ontology terms. We'll use the Table Browser to look at the databases underlying the genome.ucsc.edu
http://genome.ucsc.edu
site, and to combine annotation tracks to focus on particularly interesting regions of the genome - whether transcribed or regulatory
in nature. Finally we'll see how to download the data for further analysis on your own computers.
Return to Top or Tutorial Schedule