CS 6293 Advanced Topics: Translational Bioinformatics

Preview:

DESCRIPTION

CS 6293 Advanced Topics: Translational Bioinformatics. Intro & Ch2 - Data-Driven View of Disease Biology Jianhua Ruan. Road map. What is translational bioinformatics Probability and statistics background Data-driven view of disease biology Bayesian Inference - PowerPoint PPT Presentation

Citation preview

CS 6293 Advanced Topics: Translational Bioinformatics

Intro & Ch2 - Data-Driven View of Disease Biology

Jianhua Ruan

Road map

• What is translational bioinformatics

• Probability and statistics background

• Data-driven view of disease biology– Bayesian Inference– Network of functional related genes– Evaluation of network

What is translational bioinformatics?

– Advancement in biological technology (for high-throughput data collection) and computing technology (for cheap and efficient large-scale data storage, processing, and management) has shifted modern biomedical research towards integrative and translational

– Translational medical research:• the process of moving discoveries and innovations generated during

research in the laboratory, and in preclinical studies, to the development of trials and studies in humans, leading to improved diagnosis, prognosis, and treatment.

– Barriers to translating our molecular understanding into technologies that impact patients:

• understanding health market size and forces, the regulatory milieu, how to harden the technology for routine use, and how to navigate an increasingly complex intellectual property landscape

• Connecting the stuff of molecular biology to the clinical world

– The book chapters in this PLoS Comput Bio collection deals mostly with computational methodologies that likely to have an impact on clinical research / practice

Topic 1: Network-based understanding of disease mechanisms

• Chapter 2: Data-Driven View of Disease Biology

• Chapter 4: Protein Interactions and Disease

• Chapter 5: Network Biology Approach to Complex Diseases

• Chapter 15: Disease Gene Prioritization

Topic 2: drug design / discovery using computational / systems approaches

• Chapter 3: Small Molecules and Disease

• Chapter 7: Pharmacogenomics

• Chapter 17: Bioimage Informatics for Systems Pharmacology

Topics 3: Genome sequencing and disease

• Chapter 6: Structural Variation and Medical Genomics

• Chapter 12: Human Microbiome Analysis

• Chapter 14: Cancer Genome Analysis

Topic 4: Automated knowledge discovery and representation

• Chapter 8: Biological Knowledge Assembly and Interpretation

• Chapter 9: Analyses Using Disease Ontologies

• Chapter 13: Mining Electronic Health Records in the Genomics Era

• Chapter 16: Text Mining for Translational Bioinformatics

Ch2: Data-Driven View of Disease Biology

• Diverse genome-scale datasets exist– Genome sequences– Microarrays– genome-wide association studies– RNA interference screens– Proteomics databases– Databases of gene functions, pathways, chemicals, protein interactions, etc.

• Promise to provide systems level understanding of disease mechanisms– Modeling (understand)– Inference (make prediction)

• Integration is the key challenge– Experimental noise– Biological heterogeneity: e.g. source of material – cells in culture or biopsied

tissues?– Computational heterogeneity: e.g. data format – discrete or continuous?

Bayesian Inference

• Powerful tool used to make predictions based on experimental evidence

• Simple yet elegant probabilistic theories

• Easy to understand and implement

• Data-driven modeling– No explicit assumption about the underlying

biological mechanisms

Probability Basics

• Definition (informal)– Probabilities are numbers assigned to events

that indicate “how likely” it is that the event will occur when a random experiment is performed

– A probability law for a random experiment is a rule that assigns probabilities to the events in the experiment

– The sample space S of a random experiment is the set of all possible outcomes

Example

0 P(Ai) 1

P(S) = 1

Random variable

• A random variable is a function from a sample to the space of possible values of the variable– When we toss a coin, the number of times

that we see heads is a random variable– Can be discrete or continuous

• The resulting number after rolling a die• The weight of an individual

Cumulative distribution function (cdf)

• The cumulative distribution function FX(x) of a random variable X is defined as the probability of the event {X≤x}

F (x) = P(X ≤ x) for −∞ < x < +∞

Probability density function (pdf)

• The probability density function of a continuous random variable X, if it exists, is defined as the derivative of FX(x)

• For discrete random variables, the equivalent to the pdf is the probability mass function (pmf):

Probability density function vs probability

• What is the probability for somebody weighting 200lb?

• The figure shows about 0.62– What is the probability of

200.00001lb?

• The right question would be:– What’s the probability for somebody

weighting 199-201lb.

• The probability mass function is true probability– The chance to get any face is 1/6

Some common distributions

• Discrete:– Binomial– Multinomial– Geometric– Hypergeometric– Possion

• Continuous– Normal (Gaussian)– Uniform– EVD– Gamma– Beta– …

Probabilistic Calculus

• If A, B are mutually exclusive:– P(A U B) = P(A) + P(B)

• Thus: P(not(A)) = P(Ac) = 1 – P(A)

A B

Probabilistic Calculus

• P(A U B) = P(A) + P(B) – P(A ∩ B)

Conditional probability

• The joint probability of two events A and B P(A∩B), or simply P(A, B) is the probability that event A and B occur at the same time.

• The conditional probability of P(B|A) is the probability that B occurs given A occurred.

P(A | B) = P(A ∩ B) / P(B)

Example

• Roll a die– If I tell you the number is less than 4– What is the probability of an even number?

• P(d = even | d < 4) = P(d = even ∩ d < 4) / P(d < 4)• P(d = 2) / P(d = 1, 2, or 3) = (1/6) / (3/6) = 1/3

Independence

• P(A | B) = P(A ∩ B) / P(B)

=> P(A ∩ B) = P(B) * P(A | B)• A, B are independent iff

– P(A ∩ B) = P(A) * P(B) – That is, P(A) = P(A | B)

• Also implies that P(B) = P(B | A)– P(A ∩ B) = P(B) * P(A | B) = P(A) * P(B | A)

Examples

• Are P(d = even) and P(d < 4) independent?– P(d = even and d < 4) = 1/6– P(d = even) = ½– P(d < 4) = ½– ½ * ½ > 1/6

• If your die actually has 8 faces, will P(d = even) and P(d < 5) be independent?

• Are P(even in first roll) and P(even in second roll) independent?

• Playing card, are the suit and rank independent?

Bayes theorem

• P(A ∩ B) = P(B) * P(A | B) = P(A) * P(B | A)

AP

BPABP

)(

)()|( ==>

Posterior probability of B Normalizing constant

BAP )|( Prior of B

Likelihood

This is known as Bayes Theorem or Bayes Rule, and is (one of) the most useful relations in probability and statistics

Bayes Theorem is definitely the fundamental relation in Statistical Pattern Recognition

Example

• Prosecutor’s fallacy– Some crime happened– The suspect did not leave any evidence, except some

hair– The police got his DNA from his hair

• Some expert matched the DNA with that of a suspect– Expert said that both the false-positive and false

negative rates are 10-6

• Can this be used as an evidence of guilty against the suspect?

Prosecutor’s fallacy

• Prob (match | innocent) = 10-6

• Prob (no match | guilty) = 10-6

• Prob (match | guilty) = 1 - 10-6 ~ 1

• Prob (no match | innocent) = 1 - 10-6 ~ 1

• Prob (guilty | match) = ?

Prosecutor’s fallacy

P (g | m) = P (m | g) * P(g) / P (m)~ P(g) / P(m)• P(g): the probability for someone to be

guilty with no other evidence • P(m): the probability for a DNA match• How to get these two numbers?

– We don’t really care P(m)– We want to compare two models:

• P(g | m) and P(i | m)

Prosecutor’s fallacy

• P(i | m) = P(m | i) * P(i) / P(m) = 10-6 * P(i) / P(m)

• ThereforeP(i | m) / P(g | m) = 10-6 * P(i) / P(g)

• P(i) + P(g) = 1

• It is clear, therefore, that whether we can conclude the suspect is guilty depends on the prior probability P(i)

• How do you get P(i)?

Prosecutor’s fallacy

• How do you get P(i)?• Depending on what other information you have on the

suspect• Say if the suspect has no other connection with the

crime, and the overall crime rate is 10-7

• That’s a reasonable prior for P(g)• P(g) = 10-7, P(i) ~ 1• P(i | m) / P(g | m) = 10-6 * P(i) / P(g) = 10-6/10-7 = 10

Prosecutor’s fallacy

• P(i | m) / P(g | m) = 10-6/10-7 = 10• Therefore, we would say the suspect is

more likely to be innocent than guilty, given only the DNA samples

• We can also explicitly calculate P(i | m):P(m) = P(m|i)*P(i) + P(m|g)*P(g)

= 10-6 * 1 + 1 * 10-7

= 1.1 x 10-6

P(i | m) = P(m | i) * P(i) / P(m) = 1 / 1.1 = 0.91

Another example

• A test for a rare disease claims that it will report a positive result for 99.5% of people with the disease, and 99.9% of time of those without.

• The disease is present in the population at 1 in 100,000

• What is P(disease | positive test)?

• What is P(disease | negative test)?

Relation to multiple testing problem

• When searching a DNA sequence against a database, you get a high score, with a significant p-value

• P(unrelated | high score) / P(related | high score) =

P(high score | unrelated) * P(unrelated)

P(high score | related) * P(related)

• P(high score | unrelated) is much smaller than P(high score | related)

• But your database is huge, and most sequences should be unrelated, so P(unrelated) is much larger than P(related)

Likelihood ratio

Combining Diverse Data Using Bayesian Inference

• Want to calculate the probability that a gene of unknown function is involved in a disease

• Collect positive and negative genes (gold standard)

• Measure their activities under three hypothetical conditions

• Figure 1. Potential distributions of experimental results obtained for datasets collected under three different conditions.

Greene CS, Troyanskaya OG (2012) Chapter 2: Data-Driven View of Disease Biology. PLoS Comput Biol 8(12): e1002816. doi:10.1371/journal.pcbi.1002816http://www.ploscompbiol.org/article/info:doi/10.1371/journal.pcbi.1002816

Higher score in cond A and lower score in cond C => involved in disease

P (involved in disease | experimental data)?

• Table 1. A contingency table for the experimental results for Condition A.

Greene CS, Troyanskaya OG (2012) Chapter 2: Data-Driven View of Disease Biology. PLoS Comput Biol 8(12): e1002816. doi:10.1371/journal.pcbi.1002816http://www.ploscompbiol.org/article/info:doi/10.1371/journal.pcbi.1002816

• Probability that a gene i is involved in disease given the experimental results for gene i

Prior

likelihood

Normalizing factor

Prior

Combining datasets using Naïve Bayes

P(D | EB, EC) P(EB, EC | D) P(D)

P(EB | D) P(EC | D) P(D)

P(~D | EB, EC) P(EB, EC | ~D) P(~D)

P(EB | ~D) P(EC | ~D) P(~D)

P(D | EB, EC) + P(~D | EB, EC) = 1.

Define Gold Standard (training samples) for gene-gene network

• Positive examples: genes within the same biological process– Rely on expert selected Gene Ontology terms

• biological regulation • response to stimulus • cell-matrix adhesion involved in tangential migration using

cell-cell interactions • response to DNA damage stimulus • ldehyde metabolism

• Negative examples: random gene pairs– Assuming most gene pairs are not related

Building a Network of Functionally Related Genes

• P(FRij | Eij) = P(Eij | FRij) P(FRij)

• Eij: evidence (score) for a functional relationship between gene i and gene j from a particular dataset

• For some dataset, e.g., physical interaction data, obtaining Sij is trivial

• In general, Sij can be calculated using gene-wise correlation

Fisher's z-transformation

• Pearson correlation coefficient

Source: wikipedia

Z-transformation

Purpose: stabilizing variance

• Figure 4. The highest and lowest contributing datasets for the pair of APOE and PLTP are shown (http://hefalmp.princeton.edu/gene/one_specific_gene/18543?argument=21697&amp;context=0).

Greene CS, Troyanskaya OG (2012) Chapter 2: Data-Driven View of Disease Biology. PLoS Comput Biol 8(12): e1002816. doi:10.1371/journal.pcbi.1002816http://www.ploscompbiol.org/article/info:doi/10.1371/journal.pcbi.1002816

• Figure 5. The diseases that are significantly connected to APOE through the guilt by association strategy used in HEFalMp.

Greene CS, Troyanskaya OG (2012) Chapter 2: Data-Driven View of Disease Biology. PLoS Comput Biol 8(12): e1002816. doi:10.1371/journal.pcbi.1002816http://www.ploscompbiol.org/article/info:doi/10.1371/journal.pcbi.1002816

Used Fisher’s exact test

• Figure 6. The genes that are most significantly connected to Alzheimer disease genes using the HEFalMp network and OMIM disease gene annotations (http://hefalmp.princeton.edu/disease/all_genes/55?context=0).

Greene CS, Troyanskaya OG (2012) Chapter 2: Data-Driven View of Disease Biology. PLoS Comput Biol 8(12): e1002816. doi:10.1371/journal.pcbi.1002816http://www.ploscompbiol.org/article/info:doi/10.1371/journal.pcbi.1002816

Evaluating Functional Relationship Networks

• TPR vs FPR plot (ROC curve) and AUC

• Separate gold standard into training and testing

• Cross validation

• Literature evaluation

Summary

• We talked about– Prob / stats background– Bayes inference method to integrate multiple large-

scale, noisy datasets to predict• gene-disease associations• gene-gene associations

– Network useful for discovering novel gene functions and directing experimental followups

• Advantage against curated literature or analysis based on single dataset

• Limited by availability / quality of gold standard data

Recommended