17
Emily H. Wughalter, Ed.D. Measurement & Evaluation Spring 2010

Emily H. Wughalter, Ed.D. Measurement & Evaluation Spring 2010

Embed Size (px)

Citation preview

Page 1: Emily H. Wughalter, Ed.D. Measurement & Evaluation Spring 2010

Emily H. Wughalter, Ed.D.

Measurement & Evaluation

Spring 2010

Page 2: Emily H. Wughalter, Ed.D. Measurement & Evaluation Spring 2010

Reliability means the consistency in the measurement from one testing to the next trial.

Page 3: Emily H. Wughalter, Ed.D. Measurement & Evaluation Spring 2010

Validity means does the measure or the score measure what it purports (is supposed) to measure.

Page 4: Emily H. Wughalter, Ed.D. Measurement & Evaluation Spring 2010

For a measure to be valid it must be reliable; however, a measure can be reliable without it being valid.

Page 5: Emily H. Wughalter, Ed.D. Measurement & Evaluation Spring 2010

Xobserved

XTrue

XError

Page 6: Emily H. Wughalter, Ed.D. Measurement & Evaluation Spring 2010

Xobserved = Xtrue+XError

Page 7: Emily H. Wughalter, Ed.D. Measurement & Evaluation Spring 2010

Testing environment should be favorable Testing environment should be well-

organized Administrator should be competent to run a

favorable testing environment Range of talent Motivation of the performers

Page 8: Emily H. Wughalter, Ed.D. Measurement & Evaluation Spring 2010

Good day vs bad day Learning, forgetting, and fatigue Length of the test Test difficulty Ability of the test to discriminate Number of performers

Page 9: Emily H. Wughalter, Ed.D. Measurement & Evaluation Spring 2010

Nature of measurement device Selection of scoring unit Precision Errors in measurement Number of trials Recording errors

Page 10: Emily H. Wughalter, Ed.D. Measurement & Evaluation Spring 2010

Classroom management Warm-up opportunity

Page 11: Emily H. Wughalter, Ed.D. Measurement & Evaluation Spring 2010

Test-Retest (Stability Reliability) Internal Consistency

Page 12: Emily H. Wughalter, Ed.D. Measurement & Evaluation Spring 2010

Difference or change scores should be avoided because of ceiling and floor effects.

Difference scores are highly unreliable scores

Page 13: Emily H. Wughalter, Ed.D. Measurement & Evaluation Spring 2010

Objectivity means interrater reliability or consistency from one person (rater) to another.

Page 14: Emily H. Wughalter, Ed.D. Measurement & Evaluation Spring 2010

A criterion score represents the score that will be used to represent an individual’s performance The criterion score may be the mean of a series

of trials. The criterion score may be the best score for a

series of trials.

Page 15: Emily H. Wughalter, Ed.D. Measurement & Evaluation Spring 2010

When selecting a criterion measure, whether it is the best score, the mean of all the trials score, or the mean of the best 2 or 3 trials score, a researcher must determine which of the criterion measures represents the most reliable and most valid score.

Page 16: Emily H. Wughalter, Ed.D. Measurement & Evaluation Spring 2010

Inconsistent scorers Inconsistent performance Inconsistent measures Failure to develop a good set of instructions Failure to follow testing procedures

Page 17: Emily H. Wughalter, Ed.D. Measurement & Evaluation Spring 2010