21
GENERATING WELL- BEHAVED LEARNING CURVES: AN EMPIRICAL STUDY Gary M. Weiss Alexander Battistin Fordham University

Generating Well-Behaved Learning Curves : An Empirical Study

Embed Size (px)

DESCRIPTION

Generating Well-Behaved Learning Curves : An Empirical Study. Gary M. Weiss Alexander Battistin Fordham University. Motivation. Classification performance related to amount of training data Relationship visually represented by learning curve Performance increases steeply at first - PowerPoint PPT Presentation

Citation preview

Page 1: Generating Well-Behaved Learning Curves : An  Empirical Study

GENERATING WELL-BEHAVED LEARNING

CURVES:AN EMPIRICAL STUDY

Gary M. WeissAlexander BattistinFordham University

Page 2: Generating Well-Behaved Learning Curves : An  Empirical Study

DMIN 2014 2

Motivation

Classification performance related to amount of training data Relationship visually represented by learning

curve Performance increases steeply at first Slope begins to decrease with adequate training

data Slope approaches 0 as more data barely helps

Training data often costly Cost of collecting or labeling

“Good” learning curves can help identify optimal amount of data

7/22/2014

Page 3: Generating Well-Behaved Learning Curves : An  Empirical Study

DMIN 2014 3

Exploiting Learning Curves In practice will only have learning

curve up to current number of examples when deciding on whether to acquire more

Need to predict performance for larger sizes Can do iteratively and acquire in batches Can even use curve fitting Works best if learning curves are well

behaved Smooth and regular 7/22/2014

Page 4: Generating Well-Behaved Learning Curves : An  Empirical Study

DMIN 2014 4

Prior Work Using Learning Curves

Provost, Jensen and Oates1 evaluated progressive sampling schemes to identify point where learning curves begin to plateau

Weiss and Tian2 examined how learning curves can be used to optimize learning when performance, acquisition costs, and CPU time are considered “Because the analyses are all driven by the learning curves,

any method for improving the quality of the learning curves (i.e., smoothness, monotonicity) would improve the quality of our results, especially the effectiveness of the progressive sampling strategies.”

1 Provost, F., Jensen, D., and Oates, T. (1999) .Efficient progressive sampling. In Proc. 5th Int. Conference on knowledge Discovery & Data Mining, 23-32.

2 Weiss, G.,M., and Tian, Y. 2008. Maximizing classifier utility when there are data acquisition and modeling costs. Data Mining and Knowledge Discovery, 17(2): 253-282.

7/22/2014

Page 5: Generating Well-Behaved Learning Curves : An  Empirical Study

DMIN 2014 5

What We Do

Generate learning curves for six data sets Different classification algorithms Random sampling and cross validation

Evaluate curves Visually for smoothness and

monotonicity “Variance” of the learning curve

7/22/2014

Page 6: Generating Well-Behaved Learning Curves : An  Empirical Study

DMIN 2014 6

The Data Sets

Name # Examples

Classes

# Attributes

Adult 32,561 2 14

Coding 20,000 2 15

Blackjack 15,000 2 4

Boa1 11,000 2 68

Kr-vs-kp 3,196 2 36

Arrhythmia

452 2 279

7/22/2014

Page 7: Generating Well-Behaved Learning Curves : An  Empirical Study

DMIN 2014 7

Experiment Methodology

Sampling strategies 10-fold cross validation: 90% available for

training Random sampling: 75% available for training

Training set sizes sampled at regular 2% intervals of available data

Classification algorithms (from WEKA) J48 Decision Tree Random Forest Naïve Bayes

7/22/2014

Page 8: Generating Well-Behaved Learning Curves : An  Empirical Study

DMIN 2014 8

Results: Accuracy

Dataset J48 RandomForest

NaïveBayes

Adult 86.3 84.3 83.4

Coding 72.2 79.3 71.2

Blackjack 72.3 71.7 67.8

Boa1 54.7 56.0 58.0

Kr-vs-kp 99.4 98.7 87.8

Arrhythmia

65.4 65.2 62.0

Average 75.1 75.9 71.77/22/2014

Accuracy is not our focus, but a well behaved learning curve for a method that produces poor results is not useful. These results are for the largest training set size (no reduction)

J48 and Random Forest are competitive so we will focus on them

Page 9: Generating Well-Behaved Learning Curves : An  Empirical Study

DMIN 2014 9

Results: Variances

Dataset J48 RandomForest

NaïveBayes

Adult 0.51 0.32 0.01

Coding 9.78 17.08 0.19

Blackjack 0.36 2.81 0.01

Boa1 0.20 0.31 0.73

Kr-vs-kp 3.54 12.08 4.34

Arrhythmia

41.46 15.87 9.90

7/22/2014

Variance for a curve equals average variance in performance for each evaluated training set size. The results are for 10-fold cross validation. Naïve Bayes is best followed by J48. But Naïve Bayes had low accuracy (see previous slide)

Page 10: Generating Well-Behaved Learning Curves : An  Empirical Study

DMIN 2014 10

J48 Learning Curves(10 xval)

7/22/2014

Page 11: Generating Well-Behaved Learning Curves : An  Empirical Study

DMIN 2014 11

Random Forest Learning Curves

7/22/2014

Page 12: Generating Well-Behaved Learning Curves : An  Empirical Study

DMIN 2014 12

Naïve Bayes Learning Curves

7/22/2014

Page 13: Generating Well-Behaved Learning Curves : An  Empirical Study

DMIN 2014 13

Closer Look at J48 and RF(Adult)

7/22/2014

Page 14: Generating Well-Behaved Learning Curves : An  Empirical Study

DMIN 2014 14

A Closer Look at J48 and RF

(kr-vs-kp)

7/22/2014

Page 15: Generating Well-Behaved Learning Curves : An  Empirical Study

DMIN 2014 15

Closer Look at J48 and RF(Arrhythmia)

7/22/2014

Page 16: Generating Well-Behaved Learning Curves : An  Empirical Study

DMIN 2014 16

Now lets compare cross validation to Random Sampling, which we find generates less well behaved curves

7/22/2014

Curves used Cross validation

Page 17: Generating Well-Behaved Learning Curves : An  Empirical Study

DMIN 2014 17

J48 Learning Curves(Blackjack Data Set)

7/22/2014

Page 18: Generating Well-Behaved Learning Curves : An  Empirical Study

DMIN 2014 18

RF Learning Curves(Blackjack Data Set)

7/22/2014

Page 19: Generating Well-Behaved Learning Curves : An  Empirical Study

DMIN 2014 19

Conclusions

Introduced the notion of well-behaved learning curves and methods for evaluating this property

Naïve Bayes seemed to produce much smoother curves, but less accurate Low variance may be because they consistently

reach a plateau early J48 and Random Forest seem reasonable

Need more data sets to determine which is best Cross validation clearly generates better

curves than random sampling (less randomness?)

7/22/2014

Page 20: Generating Well-Behaved Learning Curves : An  Empirical Study

DMIN 2014 20

Future Work

Need more comprehensive evaluation Many more data sets Compare more algorithms Additional metrics

Count number of drops in performance with greater size (i.e., “blips”). Need better summary metric.

Vary number of runs. More runs almost certainly yields smoother learning curves.

Evaluate in context Ability to identify optimal learning point Ability to identify plateau (based on some

criterion)

7/22/2014

Page 21: Generating Well-Behaved Learning Curves : An  Empirical Study

DMIN 2014 21

If Interested in This AreaProvost, F., Jensen, D., and Oates, T.

(1999) .Efficient progressive sampling. In Proc. 5th Int. Conference on knowledge Discovery & Data Mining, 23-32.

Weiss, G.,M., and Tian, Y. 2008. Maximizing classifier utility when there are data acquisition and modeling costs. Data Mining and Knowledge Discovery, 17(2): 253-282.

Contact me if you want to work on expanding this paper ([email protected])

7/22/2014