30
CS 2001 ML in Bioinformatics CS 2001 Intro to research in CS Milos Hauskrecht milos@c s.pitt.edu 5329 Sennott Square Machine Learning in Bioinformatics CS 2001 ML in Bioinformatics Who am I? Milos Hauskrecht An assistant professor at CS department Secondary affiliations: ISP, CBMI and UPCI Research: AI Planning, optimization and learning in the presence of uncertainty Applications: Biomedical informatics. Modeling and control of large stochastic network systems. Decision-making for patient-management. Anomaly detection in clinical databases.

Machine Learning in Bioinformaticsmosse/courses/cs2001/lecture-milos.pdf · • Machine Learning – An area that studies methods and the design of computer programs that let us learn

  • Upload
    others

  • View
    1

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Machine Learning in Bioinformaticsmosse/courses/cs2001/lecture-milos.pdf · • Machine Learning – An area that studies methods and the design of computer programs that let us learn

1

CS 2001 ML in Bioinformatics

CS 2001 Intro to research in CS

Milos [email protected] Sennott Square

Machine Learning in Bioinformatics

CS 2001 ML in Bioinformatics

Who am I?• Milos Hauskrecht• An assistant professor at CS department• Secondary affiliations: ISP, CBMI and UPCI• Research: AI

– Planning, optimization and learning in the presence of uncertainty

• Applications:– Biomedical informatics.– Modeling and control of large stochastic network systems. – Decision-making for patient-management.– Anomaly detection in clinical databases.

Page 2: Machine Learning in Bioinformaticsmosse/courses/cs2001/lecture-milos.pdf · • Machine Learning – An area that studies methods and the design of computer programs that let us learn

2

CS 2001 ML in Bioinformatics

Bioinformatics• Bioinformatics

– Application of CS and informatics to biological and clinical sciences

• Bioinformatics is the field of science in which biology, computer science, and information technology merge to form a single discipline. The ultimate goal of the field is to enable the discovery of new biological insights as well as to create a global perspective from which unifying principles in biology can be discerned

• Machine Learning– An area that studies methods and the design of computer

programs that let us learn from past experience

CS 2001 ML in Bioinformatics

Machine Learning in Bioinformatics• Why do we need machine learning in bioinformatics?

– A large volumes of data generated in medicine• patient records with clinical data• special studies with new high-throughput data sources:

– DNA microarray data, Protemic profiles, multiplexed arrays

• Use the data to:– Learn a model that let us screen for the presence of a disease

for new patients or predicts a good therapy choice – Identification of disease biomarkers

Page 3: Machine Learning in Bioinformaticsmosse/courses/cs2001/lecture-milos.pdf · • Machine Learning – An area that studies methods and the design of computer programs that let us learn

3

CS 2001 ML in Bioinformatics

This talk• Overview of data sources

1. DNA microarray data2. SELDI-TOF-MS proteomic data profiles3. Luminex arrays

• Basics about the methods of analysis– Classification– Evaluation

CS 2001 ML in Bioinformatics

Basics

Page 4: Machine Learning in Bioinformaticsmosse/courses/cs2001/lecture-milos.pdf · • Machine Learning – An area that studies methods and the design of computer programs that let us learn

4

CS 2001 ML in Bioinformatics

Measuring RNA and Protein levels

Differences in normaland disease cells

CS 2001 ML in Bioinformatics

Measuring RNA and Protein levels

Gene-expression microarrays

MS proteomic profiling

Page 5: Machine Learning in Bioinformaticsmosse/courses/cs2001/lecture-milos.pdf · • Machine Learning – An area that studies methods and the design of computer programs that let us learn

5

CS 2001 ML in Bioinformatics

DNA microarrays• Are small, solid supports onto which the sequences or

subsequences from thousands of different genes are attached. The supports are usually glass microscope slides, or silicon chips or nylon membranes. The DNA is printed, spotted, or actually synthesized directly onto the support. The spots can be DNA, cDNA, or oligonucleotides

© 2003, Phillip Stafford and Peng Liu. Ch. 15, Microarray technology, comparison, statistical analysis and experimental design, IN: Microarray Methods and Applications: Nuts & Bolts (G. Hardiman, ed., DNA Press)

CS 2001 ML in Bioinformatics

Extracting the information from microarrays• Hybridization probing: a technique that uses fluorescently

labeled nucleic acid molecules as "mobile probes" to identify complementary molecules, sequences that are able to base-pair with one another on the microarray chip.

• A single-stranded DNA fragment is made up of four different nucleotides:– adenine (A), thymine (T), guanine (G), and cytosine (C),

that are linked end to end. – Pairs or complements: Adenine (A) – Thymine (T)

Guanine (G) -- Cytosine (C )• Two complementary sequences always find each other and

lock together or hybridize:– immobilized target DNA on the DNA microarray

hybridizes with mobile probe DNA, cDNA, or mRNA,

Page 6: Machine Learning in Bioinformaticsmosse/courses/cs2001/lecture-milos.pdf · • Machine Learning – An area that studies methods and the design of computer programs that let us learn

6

CS 2001 ML in Bioinformatics

Extracting the information from microarrays• Assume two types of cells:

– Normal cells– Cancer cells

• Extracted mRNA for each type is labeled with a different florescent dye (Cy 3 or Cy 5)

• Combine them in equal amounts

• Hybridize them on the microarray

• Scan the array: the color tells if the one cell type is over-expressed as compared to the other type

scan

R,G

CS 2001 ML in Bioinformatics

Sources of variation• The signal we obtain from a microarray is not perfect

– If we take two microarrays for the same samples we see a lot of variation

• Why?– Technical artifacts: uneven spotting– Variation in RNA purity– Different labeling efficiencies of flourescently labeled

nucleotides– Variations in expression level measurements (scan), …

• Preprocessing:– Attempts to remove unwanted sources of variation while

preserving a useful information– Eliminate systematic biases

Page 7: Machine Learning in Bioinformaticsmosse/courses/cs2001/lecture-milos.pdf · • Machine Learning – An area that studies methods and the design of computer programs that let us learn

7

CS 2001 ML in Bioinformatics

Preprocessing• Cy3 and Cy5 are incorporated into DNA with different

efficiencies– Genes that are expressed at comparable levels come with

different than 1:1 dye mixing ratio– This creates a background signal

• Example solution methods:– Global normalization: assure that expression level for the

array is 1 on average. – Housekeeping genes: genes that are known to be stable

across different conditions (e.g. beta-actin). Correct the signal using the mean of expressions at housekeeping genes.

CS 2001 ML in Bioinformatics

PreprocessingExample:• correct the slope so that 1:1 on average

Page 8: Machine Learning in Bioinformaticsmosse/courses/cs2001/lecture-milos.pdf · • Machine Learning – An area that studies methods and the design of computer programs that let us learn

8

CS 2001 ML in Bioinformatics

PreprocessingExample:• Loess normalization (locally weighted polynomial regression)Traditional analysis relies on log-ratios: M=log(R/G)

and averages: A=1/2 log(RG) as the primary data• M makes gene expression measurements relative.

M

A

M

A

Loess

CS 2001 ML in Bioinformatics

Mass Spectrometry proteomic profiling• Surface Enhanced Laser Desorption/Ionization Time of

Flight Mass Spectrometry (SELDI-TOF-MS)• A technology recently developed by Ciphergen Biosystems• A mass profile of a sample (serum, urine, cell lysates) in the

range of 0-200K Daltons• A profile is believed to provide a rich source of information

about a patient. Potentially useful for:• Detection of a disease

– Many promising results on various types of cancer• Discovery of disease biomarkers

– Identify species (protein/peptides) responsible for the differences

Page 9: Machine Learning in Bioinformaticsmosse/courses/cs2001/lecture-milos.pdf · • Machine Learning – An area that studies methods and the design of computer programs that let us learn

9

CS 2001 ML in Bioinformatics

Surface Enhanced Laser Desorption/Ionization Time of Flight Mass Spectrometry

Sample (serum, cell lysates, urine)

Mirror Lens Laser

Ion Detector

Ions with the smallest mass fly the fastest, and are detected first.

SELDI-TOF MS

Vacuum Tube

0 1 2 3 4 5 6 7 8 9 10

x 104

0

10

20

30

40

50

60

70

80

90

100

M/z

Inte

nsity

Spectral Data

Reflects the mass of proteins, peptides, nucleic acids in the sample

CS 2001 ML in Bioinformatics

Importance of SELDI-TOF MS dataSELDI profiles are believed to be useful for:• Detection of a disease

– Promising results on various types of cancer• Discovery of disease biomarkers

– Identify species (protein/peptides) responsible for the differences

-2000 0 2000 4000 6000 8000 10000 12000 14000 16000 18000 20000-20

0

20

40

60

80

100

-2000 0 2000 4000 6000 8000 10000 12000 14000 16000 18000 20000-20

0

20

40

60

80

100

-2000 0 2000 4000 6000 8000 10000 12000 14000 16000 18000 20000-20

0

20

40

60

80

100

case

control

?

Page 10: Machine Learning in Bioinformaticsmosse/courses/cs2001/lecture-milos.pdf · • Machine Learning – An area that studies methods and the design of computer programs that let us learn

10

CS 2001 ML in Bioinformatics

Challenges in MS data analysisMany sources of variation

– Sample collection, sample storage & sample processing – Instrument conditions

+ natural variation in protein expression levels among individuals

Example: profiles of two patients with pancreatic cancer

0 1000 2000 3000 4000 5000 6000 7000 80000

10

20

30

40

50

60

70

80

90

100

m/z

intens

ity

CS 2001 ML in Bioinformatics

Challenges in data analysis

Three types of systematic instrument errors– Baseline shift– Intensity measurement error– Mass inaccuracy

Example: two profiles for the same reference serum

2000 3000 4000 5000 6000 70005

10

15

20

25

30

35

Mz

Intens

ity

Page 11: Machine Learning in Bioinformaticsmosse/courses/cs2001/lecture-milos.pdf · • Machine Learning – An area that studies methods and the design of computer programs that let us learn

11

CS 2001 ML in Bioinformatics

Baseline Shift• A systematic error where intensity measurements differ from 0• Example:

2 profiles of the same pooled reference (QA/QC) serum

2000 3000 4000 5000 6000 70005

10

15

20

25

30

35

Mz

Inte

nsity

CS 2001 ML in Bioinformatics

Intensity measurements errors• Intensity measurements fluctuate randomly• Example: 2 profiles of the same pooled reference

(QA/QC) serum

• A higher intensity values come with a higher variance

7400 7600 7800 8000 8200 8400 8600 8800 90000

0.2

0.4

0.6

0.8

1

M/z

Intens

ity

Page 12: Machine Learning in Bioinformaticsmosse/courses/cs2001/lecture-milos.pdf · • Machine Learning – An area that studies methods and the design of computer programs that let us learn

12

CS 2001 ML in Bioinformatics

Mass inaccuracy

• Misalignment of readings for different m/z values. • Example:

– 2 profiles of the same pooled reference (QA/QC) serum– A clear phase shift

2200 2300 2400 2500 2600 2700 2800 2900 3000 31000

0.5

1

1.5

2

2.5

3

3.5

4

M/z

Intens

ity

Mass inaccuracy

CS 2001 ML in Bioinformatics

Profile preprocessing • Aims is to remove the noise and systematic biases in the signal

while preserving useful information• Typical preprocessing steps:

– Smoothing:• Aims to eliminate the noise in the signal

– Calibration/rescaling :• Attempts to eliminate differences in intensity among profiles

– Profile transformations (variance reduction)• Reduces the multiplicative noise

– Baseline correction:• Reduces the systematic baseline error

– Profile alignment:• Correct for mass inaccuracy

Page 13: Machine Learning in Bioinformaticsmosse/courses/cs2001/lecture-milos.pdf · • Machine Learning – An area that studies methods and the design of computer programs that let us learn

13

CS 2001 ML in Bioinformatics

Preprocessing: variance stabilization • Aims is to eliminate the dependency of the noise on the

intensity of the signal; • Example transforms used in variance stabilization:

– Log transform – Square-root transform– Cube-root transform - appears to be the best

500 1000 1500 2000 2500 3000 3500 4000 4500 50000

5

10

15

20

25

30

35

40

45

50

500 1000 1500 2000 2500 3000 3500 4000 4500 50000

0.5

1

1.5

2

2.5

3

3.5

4

4.5

5

CS 2001 ML in Bioinformatics

Preprocessing: Smoothing • Aims is to eliminate a high-frequency noise in the signal• Threat: a loss of information in the high frequency signal

1.26 1.27 1.28 1.29 1.3 1.31 1.32 1.33 1.34 1.35

x 104

7.4

7.5

7.6

7.7

7.8

7.9

8

1.26 1.27 1.28 1.29 1.3 1.31 1.32 1.33 1.34 1.35

x 104

7.4

7.5

7.6

7.7

7.8

7.9

8

Page 14: Machine Learning in Bioinformaticsmosse/courses/cs2001/lecture-milos.pdf · • Machine Learning – An area that studies methods and the design of computer programs that let us learn

14

CS 2001 ML in Bioinformatics

Preprocessing: Baseline correction

• Aims is to eliminate a systematic intensity bias by which the profile readings differ from 0

0 0.2 0.4 0.6 0.8 1 1.2 1.4 1.6 1.8 2

x 104

0

5

10

15

20

25

30

35

40

45

50

0 0.2 0.4 0.6 0.8 1 1.2 1.4 1.6 1.8 2

x 104

0

5

10

15

20

25

30

35

40

45

50

CS 2001 ML in Bioinformatics

Disease detectionHow to use the high throughput information about gene

expression levels or protein (peptide) expressions?

• Detection of a disease– Many promising results on various types of cancer

• Discovery of disease biomarkers– Identify species (protein/peptides) responsible for the

differences

• Here: the problem of building a classification model that is capable to determine with a high accuracy the presence or absence of the disease in future patients

Page 15: Machine Learning in Bioinformaticsmosse/courses/cs2001/lecture-milos.pdf · • Machine Learning – An area that studies methods and the design of computer programs that let us learn

15

CS 2001 ML in Bioinformatics

Disease detectionObjective: • Build a classification model that is capable to determine with a

high accuracy the presence or absence of the disease in future patients

Typical steps:• Data preprocessing:• Feature selection• Building of a classification model• Evaluation• Validation

CS 2001 ML in Bioinformatics

Analysis of expression profiles

DATA

Classification

Preprocessing

Testing and Evaluation

Feature Selection

Validation

Removal of systematic (instrument) biases and noise

Identify relevant relevant for building a classification model

learn a classification model

Determine the quality (accuracy) of the classification model

Check the reproducibility of detection results on independent datasets

Page 16: Machine Learning in Bioinformaticsmosse/courses/cs2001/lecture-milos.pdf · • Machine Learning – An area that studies methods and the design of computer programs that let us learn

16

CS 2001 ML in Bioinformatics

Analysis of expression profiles

DATA

Classification

Preprocessing

Testing and Evaluation

Feature Selection

Validation

CS 2001 ML in Bioinformatics

Analysis of expression profiles

DATA

Classification

Preprocessing

Testing and Evaluation

Feature Selection

Validation

Page 17: Machine Learning in Bioinformaticsmosse/courses/cs2001/lecture-milos.pdf · • Machine Learning – An area that studies methods and the design of computer programs that let us learn

17

CS 2001 ML in Bioinformatics

Detection of a disease

Objective:• Find a classification model that assigns with a high accuracy

correct disease/no-disease labels to profiles generated for individual patients

The approach:• use past data to learn a classification model• variety of machine learning approaches exist:

– Logistic regression– Linear and quadratic discriminant analysis– Classification and regression trees (CART)

Profile Classifiermodel

Class label(case or control)

CS 2001 ML in Bioinformatics

• Example: a dataset with 2 features (e.g. expression levels of two genes)

We learn the decision rule that achieves the best separation between case and control samples

Learning a classification model

case

control

Decision boundary

Page 18: Machine Learning in Bioinformaticsmosse/courses/cs2001/lecture-milos.pdf · • Machine Learning – An area that studies methods and the design of computer programs that let us learn

18

CS 2001 ML in Bioinformatics

• Logistic regression and linear SVM methods give the linear decision boundary

Learning a classification model

case

control

Decision boundary

CS 2001 ML in Bioinformatics

Non-linear boundaries

Non-linear separatorin the input space

Linear separatorin the input space

Page 19: Machine Learning in Bioinformaticsmosse/courses/cs2001/lecture-milos.pdf · • Machine Learning – An area that studies methods and the design of computer programs that let us learn

19

CS 2001 ML in Bioinformatics

Decision trees

• An alternative approach to what we have seen so far:– Partition the input space to regions– classify independently in every region

1 11

0 0

0 0

00

01

1 1

0 00 0

1

2x

1x

CS 2001 ML in Bioinformatics

Decision trees• The partitioning idea is used in the decision tree model:

– Split the space recursively according to inputs in x– Regress or classify at the bottom of the tree

03 =xx

t f01 =x 02 =x

t tf f

Example:Binary classification Binary attributes

1 0 0 1

0

1 0

321 ,, xxx}1,0{

Page 20: Machine Learning in Bioinformaticsmosse/courses/cs2001/lecture-milos.pdf · • Machine Learning – An area that studies methods and the design of computer programs that let us learn

20

CS 2001 ML in Bioinformatics

Analysis of expression profiles

DATA

Classification

Preprocessing

Testing and Evaluation

Feature Selection

Validation

CS 2001 ML in Bioinformatics

Evaluation framework • We want our classifier to generalize well to future examples• Problem: But we do not know future examples !!!• Solution: evaluate the classifier on the test set that is withheld

from the learning stage

Learn (fit)

Dataset

Training set Testing set

Evaluation

Classifiermodel

Page 21: Machine Learning in Bioinformaticsmosse/courses/cs2001/lecture-milos.pdf · • Machine Learning – An area that studies methods and the design of computer programs that let us learn

21

CS 2001 ML in Bioinformatics

Evaluation framework

Learn on the training set The model Evaluate on

the test set

case casecontrol control

Data set

Training set Test set

CS 2001 ML in Bioinformatics

Actual

Evaluation metrics

Confusion matrix: Records the percentages of examples in the testing set that fall into each group

Prediction

FN0.2

FP0.1

Control

Control

TN0.4

Case

TP0.3

Case

FNFPE +=

FNTPTPSN+

=

Sensitivity:

Specificity:

FPTNTNSP+

=

Misclassification error:

Page 22: Machine Learning in Bioinformaticsmosse/courses/cs2001/lecture-milos.pdf · • Machine Learning – An area that studies methods and the design of computer programs that let us learn

22

CS 2001 ML in Bioinformatics

Evaluation

• To limit the effect of one lucky or unlucky train/test split PDAP supports averaging through: – bootstrap– random sub-sampling– k-fold cross-validation

Classify/Evaluate

Data

TestTrain

Split randomly into 70% Train, 30% Test

Learning

Average Stats

TestTestTrainTrainExample:• random subsampling

CS 2001 ML in Bioinformatics

Validation of the predictive signal via PACE

Data

Random re-labeling

Average Stats

k-fold x-validation or random sub-sampling

Feature List

x100

• Additional assurance on the existing model: Is the signal real? Does it differ significantly from a random process?

Permutation testUse a random re-labeling of profiles’ classes and evaluate the classifiers learned using such dataIdea: Reject the null hypothesis that the achieved classification error (ACE) on the true data is obtained purely by chance

Page 23: Machine Learning in Bioinformaticsmosse/courses/cs2001/lecture-milos.pdf · • Machine Learning – An area that studies methods and the design of computer programs that let us learn

23

CS 2001 ML in Bioinformatics

Analysis of expression profiles

DATA

Classification

Preprocessing

Testing and Evaluation

Feature Selection

Validation

CS 2001 ML in Bioinformatics

Validation• Keep an independent test validation set aside

– Taken out from the original data at the beginning(the researcher holds the decoder ring)

– New data from the follow-up study– New data from different site

Data

Model

Independent validation data

Stats

Page 24: Machine Learning in Bioinformaticsmosse/courses/cs2001/lecture-milos.pdf · • Machine Learning – An area that studies methods and the design of computer programs that let us learn

24

CS 2001 ML in Bioinformatics

Analysis of expression profiles

DATA

Classification

Preprocessing

Testing and Evaluation

Feature Selection

Validation

CS 2001 ML in Bioinformatics

Dimensionality problem One important thing before we can learn a classifier

• The dimension of the profile for proteomic data is very large (~60,000 M/z measurements)

• A large number of model parameters must be to learned (estimated)

• The total number of cases in the study is relatively small (typically < 100)

• Leads to a large variance of parameter estimates (overfit)

Page 25: Machine Learning in Bioinformaticsmosse/courses/cs2001/lecture-milos.pdf · • Machine Learning – An area that studies methods and the design of computer programs that let us learn

25

CS 2001 ML in Bioinformatics

ML dimensionality reductionGoal: identify a small subset of features that achieve a good

discriminatory performanceSolution: apply ML dimensionality reduction methods and their

combinations• Feature selection

– Filter methods – Wrapper methods– Regularized classifiers

• Feature aggregation:– PCA– ICA– Clustering

• Heuristics

CS 2001 ML in Bioinformatics

Feature filtering

Solution: identify a small set of features that are good individual discriminators

Univariate scores in PDAP :– Fisher score– t-test score– Wilcoxon rank-sum score– etc.

• Univariate Scores:– can be ordered relative to each other– Are often related to some statistic (p-value)

• Features picked using a variety of criteria:– Best k features– p-value or False discovery rate (FDR) thresholds

Page 26: Machine Learning in Bioinformaticsmosse/courses/cs2001/lecture-milos.pdf · • Machine Learning – An area that studies methods and the design of computer programs that let us learn

26

CS 2001 ML in Bioinformatics

Differentially expressed features

• Is a specific profile position a good individual discriminator between case and control?

• How to measure a separation?

5890 Da

intensity

CS 2001 ML in Bioinformatics

Differentially expressed features• Example: Measures based on the t-test• Used in statistics in hypothesis testing:• H0 - Null hypothesis:

– the mean of one group (case) is not different from the mean of the other group (controls)

• H1 alternative hypothesis:

• A t-test (see if we can reject a null hypothesis):• Relies on the t statistic, that represents a normalized distance

measurement between populations

• We can compute a p-value (the significance for the null reject)

5890 Da

intensity

+

+

−+− +−=

nnti

22

)( σσµµ

)( +− = µµ

)( +− ≠ µµ

Follows Student distribution

Page 27: Machine Learning in Bioinformaticsmosse/courses/cs2001/lecture-milos.pdf · • Machine Learning – An area that studies methods and the design of computer programs that let us learn

27

CS 2001 ML in Bioinformatics

Differentially expressed features

Solution: Pick only features that are differentially are good individual discriminators for the case versus control

Scores based on univariate statistical analysis :– Fisher score and its variants– AUC (Area Under ROC Curve) – T-test score (Long & Baldi)– Wilcoxon rank-sum score– etc.

• Additional criteria may control significance levels and False discovery rate (Benjamini & Hochberg)

CS 2001 ML in Bioinformatics

Differentially expressed features• A score based on Wilcoxon rank-sum testPancreatic cancer data

1000 2000 3000 4000 5000 6000 7000 8000 9000 10000 110000

5

10

15Statgram:Wilcoxon

Mz

Sco

re

1000 2000 3000 4000 5000 6000 7000 8000 9000 10000 110000

1

2

3

Mz

Inte

sity

(cas

e)

1000 2000 3000 4000 5000 6000 7000 8000 9000 10000 110000

1

2

3

Mz

Inte

sity

(con

trol)

Page 28: Machine Learning in Bioinformaticsmosse/courses/cs2001/lecture-milos.pdf · • Machine Learning – An area that studies methods and the design of computer programs that let us learn

28

CS 2001 ML in Bioinformatics

Differentially expressed features• A score based on Wilcoxon rank-sum testPancreatic cancer data

1000 2000 3000 4000 5000 6000 7000 8000 9000 10000 110000

5

10

15Statgram:Wilcoxon

Mz

Sco

re

1000 2000 3000 4000 5000 6000 7000 8000 9000 10000 110000

1

2

3

Mz

Inte

sity

(cas

e)

1000 2000 3000 4000 5000 6000 7000 8000 9000 10000 110000

1

2

3

Mz

Inte

sity

(con

trol)

CS 2001 ML in Bioinformatics

Differentially expressed features• A score based on Wilcoxon rank-sum testPancreatic cancer data

1000 2000 3000 4000 5000 6000 7000 8000 9000 10000 110000

5

10

15Statgram:Wilcoxon

Mz

Sco

re

1000 2000 3000 4000 5000 6000 7000 8000 9000 10000 110000

1

2

3

Mz

Inte

sity

(cas

e)

1000 2000 3000 4000 5000 6000 7000 8000 9000 10000 110000

1

2

3

Mz

Inte

sity

(con

trol)

Page 29: Machine Learning in Bioinformaticsmosse/courses/cs2001/lecture-milos.pdf · • Machine Learning – An area that studies methods and the design of computer programs that let us learn

29

CS 2001 ML in Bioinformatics

Multivariate feature selectionUnivariate methods• each feature (profile position) is considered individually• May not be sufficient to capture differences in disease/control

variability

Multivariate methods• a biomarker panel (a more than one feature) is needed to define

the differences in between case/control• Subsets of features need to be considered

CS 2001 ML in Bioinformatics

Multivariate feature selectionExample of a multivariate method: • Wrapper methods• uses a classification model that is tried with different subsets of

features

• Idea:– Learn a classifier on subset of samples in the training set– Obtain a measure of accuracy on the remaining train

samples (internal test set)– Repeat 1 and 2 multiple times using cross-validation

schemes – Pick sets of features leading to the best average cross-

validation measure

Page 30: Machine Learning in Bioinformaticsmosse/courses/cs2001/lecture-milos.pdf · • Machine Learning – An area that studies methods and the design of computer programs that let us learn

30

CS 2001 ML in Bioinformatics

Wrapper methodsThe problem with wrapper methods• Combinatorial optimization • To find the best combination of k features they need to evaluate

configurations

• Recall: n is huge

Solutions: • Greedy wrappers: pick features one by one greedily• Apply heuristics to make the search more efficient

kn

CS 2001 ML in Bioinformatics

Conclusions• High-throughput data sources measure the expression of

genes, proteins, and their fragments – Potentially useful information for disease detection– Biomarker (biomarker panel) discovery

• Problems: Many sources of variability – Steps need to be taken before the profiles can be analyzed– Data collection, storage and sample processing confounding

and reproducibility is an issue• Multivariate analysis:

– More than one feature (gene, protein expression level) used to detect the disease

• Encouraging: good results on many diseases