Upload
polly-daniels
View
215
Download
0
Tags:
Embed Size (px)
Citation preview
1
LANGUAE TEST RELIABILITY
2
What Is Reliability?
Refer to a quality of test scores, and has to do with the consistency of measures across different time, test form, raters, and other characteristics of the measurement context.
3
Con. Test reliability is related to high variance
of the true score distribution. (person separability)
reliability is a measure of accuracy, consistency, dependability or fairness of scores resulting from administration of the particular examination.
4
The Measurement Model Observed score= True score + Error score X = T + E Observed score: a score that a test taker actually
received on a test. (Raw or Obtained score).True Score: as there is always some error in any
measurement an individual true score on a test would be his observed score minus some error.
T = X - E
5
Standard Error of Measurement The standard error of measurement (SEM)
is an estimate of error to use in interpreting an individual’s test score.
SEM = s 1 – r)
S = the standard deviation for the test
r = the reliability coefficient for the test
6
Standard Error of Measurement
For example, A test has a split-half reliability
coefficient of .96 and a standard deviation of 15 calculate the SEM for this test.
7
Standard Error of Measurement
SEM = s ( 1 – r ) =
15 ( 1-.96) = 15 .04
= 15 x .2 = 3
8
Threat to test Reliability
Sources of Error What are some of the factors that introduce error into
measurement? 1) Student Factors 2) Construction of the Items 3) Test administration- 4) Scoring 5) Length, difficulty and boundary effect of the Test 6) Regulatory Fluctuation 7)Discriminability, Speediness, and Homogeneity 8)Fluctuation in Response
9
Sources of Error
(1) Student Factors--Student fatigue, illness, or anxiety can induce error and lower reliability because they affect performance and keep a test from being a measure of their true ability or achievement.
2) Construction of the Items -- A major threat to reliable measurement is poorly worded or ambiguous questions or tricky questions.
10
Sources of Error 3) Test administration--Environmental
factors such as heat, light, noise, confusing directions, and different testing time allowed to different students can affect students' scores.
11
Sources of Error 4) Scoring – An objective test is more reliable because the test
scores reflect true differences in achievement among students and not the judgment and opinions of the scorer.
subjectivity in score or mechanical errors in scoring process may introduce inconsistency in score and produce unreliable measurement, that usually occur with in or between the rater themselves.
12
Scoring A. Intra- Rater Reliability(mark/er-mark
reliability) (Bachman, 1990) when an individual subjectively judges or rates
the adequacy of a given sample of language performance for at least two times and gives consistent results, we say that this rating have intra- rater reliability.
B. Inter-rater reliability Which refers to consistency of rating given by
different raters to a sample of language performance.
13
Sources of Error (5) length, difficulty and boundary effect of the
Test A- reliability is affected by number of item in the
test. More items in the test make a grater range of
score and grater reliability. B- A test that is either too easy or too difficult for
the class taking it will typically have low reliability. This occurs because the scores will be clustered together at either the high end or the low end of the scale, with small differences among students( boundary effect).
14
Sources of Error (6) Regulatory Fluctuation –Differences in
the clarity of instructions, the time of test administration, test administrator interaction with examinees, prevention of cheating behavior, and reporting of time remaining are all potential source of measurement error.
15
Sources of Error (7)Discriminability, Speediness, and
Homogeneity A- Discriminability: the degree to which a test or an item of the
test distinguishes among stronger and weaker test taker.
Great discriminate = Great reliability
16
Sources of Error B- Speediness: Speed test: A test in which the items are easy but
the time limits are so short that a few or non of the test takers can complete all the items. such a test aims to determining the speed of the testees to do certain task
Power test :A test in which item difficulty generally increase gradually but ample time is given to all candidate. The aim is determine how much an individual is able to do, not how rapidly.
17
Sources of Error In power test failure to allow examinees a
reasonable amount of time to complete the test will reduce the reliability.
If the test becomes more difficult as a result of the element of speedness , reliability will diminish.
C- Homogeneity We can increase reliability and reduce error by
including items of similar format and content.(e.g split half method)
18
Sources of Error (8)Fluctuation in Response
A- Response arbitrariness B- Wiseness and familiarity Response
19
Methods of Reliability Computation
The choice of the method of computation of reliability will depend on such factor as
Nature of threats to reliability present Ease of computation Nature of the test Testing situation
20
Methods of Reliability Computation
Test-Retest Method Parallel Form Method Inter – Rater Reliability Split Half Reliability KR-20 KR-21
21
Methods of Reliability Computation
1-Test-Retest Method Refer to correlation of two sets of score for
the same persons. An approach to estimating reliability in
which we administer the test twice to the group of individuals and then compute the correlation between two sets of scores .
R= r1,2
22
Methods of Reliability Computation
Test-Retest Method disadvantage: 1-time consuming. it is difficult to arrange
two testing session an preparing similar condition for the same group of examinee.
2- test effect .students may learn or memorize some question
23
Methods of Reliability Computation
2-Parallel Form Method Two tests of the same ability, and with
equal length and difficulty that are administrated to the same sample of persons.
disadvantage: constructing two parallel forms of a test
is not an easy task.
24
Methods of Reliability Parallel Form Method
Equated test: Any two sets of scores from
different test (assuming that the same trait is being tested) that have been reduced to a common scale to facilitate comparison.
25
Methods of Reliability Parallel Form Method
Random parallel tests: It has been used to described tests that have
been composed of items drawn randomly from the same population of items
ru = rA,B ru = the reliability coefficient rA,B = the correlation of form A with the
form B of the test
26
Methods of Reliability Computation
3-Inter – Rater ReliabilityEstimation based on the correlation of scores
between/among two or more raters who rate the same item, scale, or instrument .
the actual level of reliability will depend on number of raters or judges.
the more rater present in the determination of the mark, the more reliable will be the mark.
27
Intra- Rater Reliability(mark/er-mark reliability) (Bachman, 1990)
when an individual subjectively judges or rates the adequacy of a given sample of language performance for at least two times and gives consistent results, we say that this rating have intra- rater reliability.
28
Methods of Reliability Inter – Rater Reliability
there are two steps in the estimation of inter-rater reliability:
1-an average of all correlation coefficients 2-Spearman Brown Prophecy Formula
29
Methods of Reliability Computation
4-Split Half Reliability Obtained from a single administration by
dividing the tests into two comparable halves and comparing the resulting scores for each individual (split into odds and evens).
an approach to estimating the internal consistency of a test.
30
Methods of Reliability Split Half Reliability
disadvantage: a- reliability can be change according to the
manner in which the test is divided. (split into odds and evens)
b- homogeneous item. Because assuming the equality between the two halves is not always the safe assumption.( different subsection , in a test e.g. grammar , vocab, reading,…will change test homogeneity and thus reduce the test score reliability )
31
Methods of Reliability Split Half Reliability
advantage: it is more practical than other. Because: 1-no need to administer the same test
twice. 2-not necessary to develop two parallel
forms of the same test. 3-single administration will be enough .
32
Methods of Reliability Split Half Reliability
Spearman Brown Prophecy Formula.
e.g.
if the reliability coefficient of half of the test is computed to be 0.80 .what would be the reliability of the total test?
33
Methods of Reliability Split Half Reliability
it should be logically clear that the reliability of the total test will always be higher than the reliability of half of the test.
34
Methods of Reliability Computation
5-KR-20 Kuder-Richardson Formula 20 Permit
us to arrive at the same final estimate of reliability without having to compute reliability estimates for every possible split half combination.
35
Kuder-Richardson Formula 20It is based on number of item on the test = n or K difficulty of the individual items variance of the total test score = V
36
Methods of Reliability Computation
6-KR-21 Kuder-Richardson Formula 21 is a formula
that is easier to use but less accurate than KR 20.
This formula is based on the assumption that all item in the test are designs to measure a single trait.
37
Kuder-Richardson Formula 21 This formula, known as KR-21
38
Kuder-Richardson Formula 21 e.g. Suppose we gave a 50-item test and
the mean score was 43 and the variance was 25 Putting these values into KR-21.
K= 50 X= 43 V= 25
39
Kuder-Richardson Formula 21
Solving for r obtains: r= (1.02) (0.76) = 0.78 the reliability coefficient is greater than 0.70, so we can
use this test with some degree of confidence.
40
Correction for Attenuation Henning,1987 A way of holding reliability constant when
making comparison among correlation coefficient . It is made by dividing the correlation coefficient by the square root of the cross-product of reliability.
41
Correction for Attenuation
E.g If a test of composition writing correlated 0.55 with the test
of grammar usage , disattenuate this correlation, assuming that KR20 reliabilities of the tests were 0.70 for composition writing and 0.80 for grammar usage .
42
Correction for Attenuation