11
Journal of Speech and Hearing Disorders, Volume 49, 338-348, November 1984 USE AND MISUSE OF NORM-REFERENCED TESTS IN CLINICAL ASSESSMENT: A HYPOTHETICAL CASE REBECCA J. McCAULEY LINDA SWISHER University of Arizona The purposes of this paper are to discuss concepts that are fundamental to the proper use of norm-referenced tests in clinical assessment, to consider eommon errors in the use of such tests, and to suggest alternatives to norm-referenced testing for certain assessment purposes. A hypothetical client is used to illustrate the following errors: the use of age-equivalent scores as the sole summary of test results, the use of individual items to formulate therapy objectives, and the failure to consider the possible effects of measurement error when difference scores are used to assess progress or to examine patterns of impairment. Assessment, the application of numbers to a perform- ance, is a major component of clinical evaluation, a larger process that requires the integration of information from many sources. The proper use of norm-referenced tests in assessment entails a series of decisions and activities on the part of the speech-language clinician. First, the clini- cian decides whether or not a norm-referenced instru- ment is appropriate for a particular assessment purpose, and then which specific test or group of tests of the skill being assessed is psychometrically most acceptable (Buros, 1972, 1978; Darley, 1979; Kilburg, 1982; Ivauner & Lahey, 1981; McCauley & Swisher, 1984; Weiner & Hoock, 1973), and which reports norms relevant to the client. Next, the clinician administers and scores each test in accordance with the procedures outlined by the test developer. Prior to integrating test results with other sources of information during evaluation, the clinician interprets test results in light of the relevant test norms and the assessment question to arrive at a statement of the test taker's performance. Norm-referenced tests provide evidence regarding the existence of a problem. Properly used, they can suggest a need for further assessment or help document a need for the initiation or continuation of therapy. Misused, they can lead to a mistaken understanding of a client's prob- lem, to inappropriate and fruitless therapy programs, or to inaccurate conclusions regarding the efficacy of therapy. The purposes of this paper are to discuss concepts that are fundamental to the proper use of norm-referenced tests, to consider four common errors in the use of such tests, and to offer alternatives to norm-referenced testing for certain assessment purposes. CONCEPTS UNDERLYING THE USE OF NORM-REFERENCED TESTS A brief review of some concepts pertinent to the com- parison of a client's score to the test norms is included to highlight some of the characteristics of norm-referenced tests that make them susceptible to misuse. These con- eepts are related to psychometric principles that are treated more fully elsewhere (e.g., Anastasi, 1976; Salvia & Ysseldyke, 1981). Such concepts are usually discussed in relation to test construction and the psychometric evaluation of norm-referenced tests; however, an ap- preciation of them is equally vital to the proper use of such tests. Although the comparison of a test taker's score to the relevant norms may at first glance appear to be a straight- forward matter, it is complicated by the fact that tests only estimate the underlying skill or behavior they are de- signed to assess. This means that the comparison of the test taker's score and the relevant norms involves a comparison of estimated, rather than absolute, or true values. The comparison process is further complicated by the fact that individuals with similar characteristics will vary to some extent in their test performance without such variability being indicative of a problem. These two complications, then, make the comparison of an indivi- dual's score to the relevant test norms a fairly difficult enterprise. The first of these complicationsuthe fact that test scores are imperfect estimates of whatever they were designed to measure--can be more fully appreciated and responded to following a discussion of the concept of test reliability. Reliability can be defined as the consistency with which a test measures a given attribute or behavior. It can be estimated for a test by examining the consist- ency of scores obtained by individuals when they are given the same test on two occasions, an equivalent form of the test on the second occasion, or when they are tested under other, variable conditions (Anastasi, 1976, p. 103). For the clinician who wishes to compare a single test score to the norms, the imperfect reliability of even the most reliable test means that the test taker's observed score will vary from the score the test taker would receive if the test's reliability were perfect. That ideal score is referred to as the individual's true score. When evidence collected by the test designer indicates high reliability, there is probably little difference between the in- dividual's true and observed scores. When reliability is © 1984, American Speech-Language-Hearing Association 338 0022-4677/84/4904-0338501.00/0 Downloaded From: http://jshd.pubs.asha.org/ by a Ohio State University - Library User on 12/06/2015

USE AND MISUSE OF NORM-REFERENCED TESTS …u.osu.edu/cclab/files/2015/11/use-and-misuse-2crk3u3.pdf · USE AND MISUSE OF NORM-REFERENCED TESTS IN CLINICAL ASSESSMENT: A HYPOTHETICAL

  • Upload
    hahuong

  • View
    231

  • Download
    1

Embed Size (px)

Citation preview

Page 1: USE AND MISUSE OF NORM-REFERENCED TESTS …u.osu.edu/cclab/files/2015/11/use-and-misuse-2crk3u3.pdf · USE AND MISUSE OF NORM-REFERENCED TESTS IN CLINICAL ASSESSMENT: A HYPOTHETICAL

Journal of Speech and Hearing Disorders, Volume 49, 338-348, November 1984

U S E A N D M I S U S E OF N O R M - R E F E R E N C E D T E S T S IN C L I N I C A L A S S E S S M E N T : A H Y P O T H E T I C A L C A S E

REBECCA J. McCAULEY LINDA SWISHER University of Arizona

The purposes of this paper are to discuss concepts that are fundamental to the proper use of norm-referenced tests in clinical assessment, to consider eommon errors in the use of such tests, and to suggest alternatives to norm-referenced testing for certain assessment purposes. A hypothetical client is used to illustrate the following errors: the use of age-equivalent scores as the sole summary of test results, the use of individual items to formulate therapy objectives, and the failure to consider the possible effects of measurement error when difference scores are used to assess progress or to examine patterns of impairment.

Assessment , the appl ica t ion of numbers to a perform- ance, is a major componen t of cl inical evaluat ion, a larger process that requires the integrat ion of information from many sources. The p roper use of norm-refe renced tests in assessment entai ls a series of decis ions and activit ies on the part of the speech- language cl inician. First , the clini- cian dec ides whe the r or not a norm-referenced instru- men t is appropr ia te for a par t icular assessment purpose, and then which specific test or group of tests of the skill be ing assessed is psychomet r ica l ly most acceptable (Buros, 1972, 1978; Darley, 1979; Kilburg, 1982; Ivauner & Lahey, 1981; McCau ley & Swisher, 1984; Weine r & Hoock, 1973), and which reports norms re levant to the client. Next, the c l in ic ian adminis ters and scores each test in accordance with the procedures out l ined by the test deve loper . Pr ior to in tegrat ing test results with other sources of information dur ing evaluat ion, the c l inician in terprets test results in l ight of the re levant test norms and the assessment ques t ion to arrive at a s ta tement of the test taker 's performance.

Norm-re fe renced tests p rovide ev idence regarding the exis tence of a p roblem. Proper ly used, they can suggest a need for further assessment or he lp document a need for the ini t ia t ion or cont inuat ion of therapy. Misused, they can lead to a mis taken under s t and ing of a c l ient ' s prob- lem, to inappropr ia te and fruitless therapy programs, or to inaccurate conclusions regarding the efficacy of therapy. The purposes of this pape r are to discuss concepts that are fundamenta l to the proper use of norm-referenced tests, to cons ider four common errors in the use of such tests, and to offer a l ternat ives to norm-refe renced test ing for cer tain assessment purposes .

C O N C E P T S U N D E R L Y I N G T H E U S E O F N O R M - R E F E R E N C E D

T E S T S

A br i e f r ev iew of some concepts pe r t inen t to the com- par ison of a c l ient ' s score to the test norms is inc luded to h ighl igh t some of the characteris t ics of norm-referenced tests that make them suscep t ib le to misuse. These con-

eepts are re la ted to psychomet r ic pr inc ip les that are t reated more fully e l s ewhere (e.g., Anastasi , 1976; Salvia & Ysseldyke, 1981). Such concepts are usual ly d iscussed in relat ion to test construct ion and the psychometr ic evaluat ion of norm-refe renced tests; however , an ap- prec ia t ion of them is equa l ly vital to the p rope r use of such tests.

Al though the compar ison of a test taker 's score to the re levant norms may at first g lance appear to be a straight- forward matter, it is compl ica ted by the fact that tests only est imate the unde r ly ing skill or behav io r they are de- s igned to assess. This means that the compar ison of the test taker 's score and the re levan t norms involves a compar ison of es t imated, rather than absolute , or true values. The compar ison process is fur ther compl ica ted by the fact that ind iv idua ls with s imilar characterist ics will vary to some extent in thei r test per formance without such var iabi l i ty be ing indica t ive of a p roblem. These two complicat ions, then, make the compar ison of an indivi- dual ' s score to the re levan t test norms a fairly difficult enterpr ise .

The first of these c o m p l i c a t i o n s u t h e fact that test scores are imperfec t es t imates of wha tever they were des igned to m e a s u r e - - c a n be more fully apprec ia ted and r e sponded to fol lowing a d iscuss ion of the concept of test re l iabi l i ty . Rel iab i l i ty can be def ined as the consis tency with which a test measures a given at t r ibute or behavior. I t can be es t imated for a test by examining the consist- ency of scores ob ta ined by ind iv idua ls when they are given the same test on two occasions, an equ iva len t form of the test on the second occasion, or when they are tes ted under other, var iable condi t ions (Anastasi, 1976, p. 103).

For the cl inician who wishes to compare a s ingle test score to the norms, the imperfec t re l iab i l i ty of even the most re l iab le test means that the test taker 's observed score will vary from the score the test taker would receive if the test 's re l iab i l i ty were perfect . That ideal score is referred to as the ind iv idua l ' s true score. When ev idence col lec ted by the test des igner indicates high rel iabi l i ty , there is p robab ly l i t t le difference b e t w e e n the in- d iv idual ' s true and obse rved scores. When re l iabi l i ty is

© 1984, American Speech-Language-Hearing Association 338 0022-4677/84/4904-0338501.00/0

Downloaded From: http://jshd.pubs.asha.org/ by a Ohio State University - Library User on 12/06/2015Terms of Use: http://pubs.asha.org/ss/rights_and_permissions.aspx

Page 2: USE AND MISUSE OF NORM-REFERENCED TESTS …u.osu.edu/cclab/files/2015/11/use-and-misuse-2crk3u3.pdf · USE AND MISUSE OF NORM-REFERENCED TESTS IN CLINICAL ASSESSMENT: A HYPOTHETICAL

MCCAULEY & SWISHER: Norm-Referenced Tests and Assessment 339

low, however , there may be a cons iderab le d iscrepancy b e t w e e n those scores.

In eve ryday speech, it is not uncommon to hear a s t a t ement such as, " H e ' s p robab ly 50 years o l d - - g i v e or take a few years ." Wi th such a s tatement, the speaker a t tempts to convey a bes t es t imate of the a t t r ibute under cons idera t ion , as wel l as information about the confi- dence with which the es t imate is made. Comparab le qual if icat ions are necessary for the more formal es t imates p r o d u c e d by s tandard ized tests. In order to qualify the informat ion c o n v e y e d by a test score, one calculates an es t ima ted true score (the bes t est imate) and a conf idence interval (the expec ted re l iab i l i ty of that est imate) based on a statistic known as the standard error of measure- ment (SEM). The calcula t ion of the test taker 's es t imated true score and the conf idence interval a round it requires the fo l lowing p ieces of information: (a) an es t imate of the test ' s re l iabi l i ty , (b) the mean and s tandard devia t ion of scores ob ta ined by the normat ive sample to which the test taker ' s score is to be compared , and (c) the test taker 's obse rved score (Nunnally, 1978, p. 240; Salvia & Ysseldyke, 1981, pp. 86-94). When all of these p ieces of informat ion are a v a i l a b l e - - a n d often they are not (McCauley & Swisher, 1984)- - they can be used to esti- mate the s tandard error of measu remen t and, then, the conf idence interval in which one would expect to find the test taker 's t rue score. As a resul t of these calculations, one knows cons ide rab ly more about the prec is ion of the test ' s m e a s u r e m e n t than from the obse rved score alone. Thus, one can es t imate the degree of uncer ta in ty that results from the test 's imperfec t re l iabi l i ty . In addit ion, this informat ion al lows one to in terpre t test performance as a range of scores ra ther than as a specific value (American Psychological Associat ion, 1974, p. 69).

An example of the calculat ions desc r ibed above is i nc luded in the appendix . [Calculat ions that differ from those de sc r ibed by Salvia and Ysseldyke can be found in Gui l ford (1954, p. 389) and Anastasi (1976, p. 128).] In the example , a c l ien t rece ives a raw score of 75 on a test for which a r e l evan t normat ive sample rece ived a mean score of 100 with a s tandard devia t ion of 10. A re l iabi l i ty coefficient r epor ted for the same normat ive sample was .70. This less than ideal , but not uncommon, level of re l i ab i l i ty is u sed to i l lustrate the loss of prec is ion result- ing when tests of even modera te re l iab i l i ty are used. Given these p ieces of information, the test taker 's esti- ma ted true score is 82.5 and the 95% conf idence intelwal a round the true score extends fi-om 71.76 to 93.24. Thus, 95 t imes out of 100, the test taker 's true score can be expec ted to fall b e t w e e n the values of 71.76 and 93.24 and the s ingle bes t guess one can make is 82.5.

The second compl ica t ion that can mudd le the compari- son of a test score to the test norms is that indiv iduals will vary in the i r abi l i t ies , and hence in their test scores, wi thout that difference s ignal ing the exis tence of a prob- lem. At what po in t is a score different enough from the range of scores r ece ived by the normat ive sample to suggest a poss ib le p rob l em? Is a 10-point difference from the sample mean large enough? Is a 20-point difference

large enough? The concepts of s tandard devia t ion and z scores can be used to he lp answer these quest ions.

The s tandard devia t ion is a measure of the var iabi l i ty of the scores ob ta ined by ind iv idua ls in the normative sample. I t is used in conjunct ion with the mean and serves to summar ize the degree to which test scores cluster a round the mean or are scat tered over a wide range of values. The adequacy of the mean score as a descr ip t ion of a set of scores is affected by the var iabi l i ty of those scores. I f most test scores in the group lie close to the mean, the mean is a more represen ta t ive value than if most scores l ie a cons ide rab le d is tance from the mean.

Cons ide r the fol lowing l ines which r ep re sen t two pairs of scores and the i r mean values:

Pair 1: 35 41 47 Mean

Pair 2: 12 41 70 The mean for each pair of scores is 41. However , desp i te the fact that both pairs of numbers can be summar ized by that same value, the scores 35 and 47 are much more s imilar to the mean of 41 than are the scores 12 and 70. The s tandard devia t ions for these two sets of numbers reflect this difference: The first pai r has a s tandard devia- t ion of 6, whereas the second has the much larger stand- ard devia t ion of 29. Thus, the larger the s tandard devia- tion, the greater the var iabi l i ty of scores a round the mean.

Z scores are cons ide red useful because they incorpo- rate information about the s tandard deviat ion, as wel l as about the mean. For example , a score equ iva len t to the mean would have a z score of 0, and scores 1 s tandard devia t ion above or b e l o w the mean would have z scores of + 1.00 or - 1.00, respect ive ly . Z scores can be used to est imate the probabi l i ty of a g iven score 's occurrence because sets of test scores usual ly approximate a normal d is t r ibut ion around the mean. One proper ty of normal dis t r ibut ions is that a p red ic t ab le percen tage of values (test scores) will fall wi th in a given d is tance from the mean value, when the d is tance is expressed as a mul t ip le of the s tandard deviat ion. By conver t ing an ind iv idua l ' s raw test score into a z score, one can es t imate the percentage of the normat ive sample that r ece ived a test score at least that different from the mean. This permits the test user to es t imate how many scores of a par t icular magni tude should be found and, thus, how unexpec ted a given score is. For example , scores associa ted with z scores of less than - 2 (i.e., those ly ing more than two standard devia t ions be low the mean) can be expec ted to be rece ived by less than 3% of the popula t ion and are, therefore, cons ide red by some cl inicians to be indicat ive of nonnormal per formance (Ludlow, 1983).

A par t icular ly informative way in which this might be done is shown in F igure 1. In that figure, the scores used in the d iscuss ion of conf idence intervals have b e e n trans- formed to z scores, and the conf idence interval has been calculated. These values have b e e n p lo t ted along a nor- mal p robabi l i ty curve. The range of values inc luded wi thin the conf idence interval i l lustrates the dangers of taking a raw score at face value. In compar isons of a test taker 's score to a cutoff value (such as two s tandard devia t ions be low the mean or 1.65 s tandard devia t ions

Downloaded From: http://jshd.pubs.asha.org/ by a Ohio State University - Library User on 12/06/2015Terms of Use: http://pubs.asha.org/ss/rights_and_permissions.aspx

Page 3: USE AND MISUSE OF NORM-REFERENCED TESTS …u.osu.edu/cclab/files/2015/11/use-and-misuse-2crk3u3.pdf · USE AND MISUSE OF NORM-REFERENCED TESTS IN CLINICAL ASSESSMENT: A HYPOTHETICAL

340 Journal of Speech and Hearing Disorders 49 338-348 November 1984

Estimated Trm Score//~ f-- Observed Score I / / I ~ l

I

/ '

_zscores_4 -3 -2 -1 0 + L i

I

95 % Confidence Interval FIGURE t. A normal probability curve showing the 95% confi- dence interval and estimated true score (0) for an observed score (o).

+2 +3 +4

below the mean), the estimated true score and the width of the confidence interval around it should affect the test user's interpretation. Because the estimated true score takes the reliability of the test into account, it is the single best value to compare to the arbitrary cutoff value. The width of the confidence interval around it indicates how much above or below the estimated true value, and by inference how much above or below the cutoff value, the actual true score might be.

In summary, the proper interpretation of a test taker's score on a norm-referenced test requires use of informa- tion on the reliability of the test taker's score and that score's probability for the normative sample. Several of the errors in test use that will be discussed in the course of this paper arise when the importance of these pieces of information is not appreciated.

S P E C I F I C P R O B L E M S I N T H E U S E O F N O R M - R E F E R E N C E D

T E S T S

Examples of the four errors in the use of norm-referenced tests that are considered in the course of this paper are illustrated in the hypothetical case that appears in the Exhibit. The case involves a preschool child referred because of a possible language impair- ment. Although the report refers primarily to the use of language tests, it illustrates errors to be avoided when- ever a norm-referenced test of any behavior is used. In the hypothetical case, the errors are presented in the order in which they will be discussed in the following sections. This exhibit contains only the background and assessment sections of a case report. The evaluation section, in which infbrmation from different sources would be combined, is excluded from the exhibit.

AGE-EQUIVALENT SCOtlES AND THE SUMMARIZATION OF TEST RESULTS

Age-equivalent scores, such as language ages, are often reported in test manuals when the skills to be assessed change considerably with age during childhood. They are frequently used by clinicians to explain test results to nonprofessionals, such as the parents of a tested child, because it is assumed that they summarize the test results in a form that is easily understood.

In the hypothetical case report (see the Exhibit, Error #1), an age-equivalent score was used to summarize Paul's performance on test B and to support the assertion that Paul's receptive vocabulary was impaired. In spite of the frequency with which age-equivalent scores are used in this manner, there are problems associated both with the use of age-equivalent scores as a summary of test results and with the interpretation of that summary. Several of these problems will be discussed in the section that follows.

Error 1: Using Age-Equivalent Scores as Test Summaries

A sizable number of problems with the use of grade- or age-equivalent scores have been identified by various authors (Anastasi, 1976, pp. 73-76; Mehrens & Lehmann, 1980, pp. 97-99; Salvia & Ysseldyke, 1981, pp. 66-69). Generally, these problems fall into two categories: those that cause such scores to be considered "psychometri- cally crude" (Anastasi, 1976, p. 74) and those that lead to mistaken inferences.

Two psychometric problems limit the reliability and, therefore, the validity of age-equivalent scores and, thus, cause such scores to be considered "extremely vague indications of developmental level" (Salvia & Ysseldyke, 1981, p. 73). The first psychometric problem concerns the relation of age-equivalent scores and the raw scores on which they are based (Salvia & Ysseldyke, 1981, p. 67). For most tests, as age increases, similar differences in age-equivalent scores are the result of smaller and smaller differences in raw scores. Therefore, an age- equivalent score that is 6 months behind an individual's chronological age may indicate a larger difference in actual test performance for younger test takers than for older ones. This happens because individuals at higher developmental levels, who are quickly approaching adult competenee, are making fewer errors. Because of this, the 1-year delay on test B reported for Paul, who is 4 years old, may have resulted from 12 missed items; whereas a 1-year delay for a 10-year-old child taking the test might be due to only one or two missed items. This causes the reliability of age-equivalent scores to be poorer for devel- opmentally more advanced test takers.

A second psyehometrie problem with age-equivalent scores is that they are not necessarily based directly on evidence collected for children of that chronological age (Mehrens & Lehmann, 1980, pp. 97-98; Salvia & Ys-

Downloaded From: http://jshd.pubs.asha.org/ by a Ohio State University - Library User on 12/06/2015Terms of Use: http://pubs.asha.org/ss/rights_and_permissions.aspx

Page 4: USE AND MISUSE OF NORM-REFERENCED TESTS …u.osu.edu/cclab/files/2015/11/use-and-misuse-2crk3u3.pdf · USE AND MISUSE OF NORM-REFERENCED TESTS IN CLINICAL ASSESSMENT: A HYPOTHETICAL

MCCAULEY & SWISHER: Norm-Referenced Tests and Assessment 341

EXHIBIT. Sections of a hypothetical ease report. The numerals correspond to test use errors and to the num- bered sections in the text.

Client: Paul B. Parents: John & Mary Ann B. Address: l l21 Buena Vista

D.O.B.: 5/27/78 Date: 1/20/83 Clinician: R. F.

Background

Initial interview. Paul B., age 4:2, was first seen at this office on 7/23/82 for evaluation. His parents had been referred by a teacher at Paul's preschool. Both parents reported that Paul did not talk as much as his older brother and sister did at his age, and that when he talked, he used mostly two- and three-word utterances. His parents volun- teered that they considered Paul a "bright" child and that they were satisfied with his development in other areas.

Information from other sources . . . . In summary, then, the nonverbal intelligence and audiometrie assessments conducted prior to our initial evaluation indicated normal function.

Initial Speech and Language Assessment

Norm-referenced assessment. Three norm-referenced tests were administered: a test of articulation (test A), a test of single-word receptive vocabulary (test B), and a test of receptive and expressive syntax (test C).

Test A . . . . Therefore, Paul's performance on test A indicated normal test performance.

Test B. Paul received a score of 45 on test B. This 03 corresponded to a language age of 3:2, indicating a delay

m receptive vocabulary of 1 year when compared with his chronological age of 4:2 . . . .

Test C . . . . On the 36-item subtests of test C, Paul received a score of 30 on receptive subtest #1 compared with his score of only 25 on expressive subtest #2. These ® scores corresponded to z scores of -2.00 and -2.40, respectively. Thus, it appeared that Paul exhibited a greater deficit in expressing than in understanding the syntax of language . . . .

• . . Because of Paul's difficulty with the plural - s on test C, these structures were included'in the initial therapy

® objectives . . . .

Current Assessment Information

Results obtained when tests B and C were readminis- tered indicated little progress. Thus, therapy does not ® appear to have been effective for those areas.

se ldyke, 198I, p. 68). Ins tead, a given age-equiva len t score is often ca lcu la ted ind i rec t ly e i ther by in terpola t ing b e t w e e n two ages for which data are avai lable or by ext rapola t ing from ages for which data are avai lable to o lder or younger ages, for which data were not gathered. For example , Paul ' s raw score on test B probab ly fell b e t w e e n the average scores of the 3-year-olds and 4-year- olds in the normat ive sample, and he was ass igned the age-equ iva len t of 3:2 by in terpola t ion al though no chil- d ren of age 3:2 were sampled in the normat ive studies. D e p e n d e n c e on ind i rec t ly es t imated ( in terpola ted or ex- t rapolated) scores, therefore, may involve assumptions about the cont inu i ty of language d e v e l o p m e n t across chronological age that cannot be justif ied. Thus, extrapo-

la ted and in te rpo la ted age-equiva lents are l ike ly to be less stable es t imates of the test faker 's behav ior than other kinds of de r ived scores, such as z scores (Mehrens & Lehmann, 1980, p. 98).

Because of thei r psychomet r ic p rob lems , age-equiva- lent scores often serve as the basis for mis interpreta t ions . Two inferences which foster such mis in terpre ta t ions oc- cur with sufficient f requency to warrant at tent ion here. One inference is that cur rent funct ioning is accurately desc r ibed by an age-equ iva len t descr ip t ion (Mehrens & Lehmann, 1980, p. 98 Salvia & Ysseldyke, 1981, pp. 67-68). Someone faced with the hypothe t ica l ease of Paul and his age-equ iva len t score of 3:2 on a vocabulary test might infer that he functions l ike a 3-year-old in his under s t and ing of s ingle words• This inference is n o t wel l - founded, however , because there is as ye t no evi- dence that a 4 -yea r -o ld - -who has a year more exper ience with the wor ld and with language than the younger ch i ld ren on whom the normat ive data were o b t a i n e d - - approaches a language test in the same way or makes the same kinds of errors as a younger chi ld (Mehrens & Lehmann, 1980, p. 98; Salvia & Ysseldyke, 1981, p. 68; Thorndike , 1982, p. 108). Similarly, a 60-year-old suf- fering from aphas ia might rece ive an age-equ iva len t score of 10 years on the vocabulary test. I t is unl ike ly , however , that such a c l ient would make the same k ind of errors as the 10-year-old or that he would exhib i t s imilar com- munica t ion skills.

Perhaps the most per i lous inference in the interpreta- t ion of age-equ iva len t scores involves thei r use to in- dicate impa i rmen t (Anastasi, 1976, p. 76; Bloom & Lahey, 1978, p. 352). Here the inference is made that an age- equ iva len t score that is a cer tain n u m b e r of months lower than chronological age necessar i ly indicates a de lay or even an impa i rmen t for the tasks add res sed by the test. Al though there are minor differences in the way in which test deve lopers calculate age-equ iva len t scores, such scores do not take into account the normal ind iv idua l differences to be expec ted wi th in any of the age groups. Thus, they do not p rovide ev idence as to whe the r a chi ld who scores be low the average ob ta ined by his age peers d id so by an amount that represents normal variat ion or by an amount that is sufficiently unexpec t ed to suggest impairment• I f cons iderab le variat ion is conrmon for nor- mal ch i ld ren wi thin an age group, even a large age- equ iva len t de lay may not imply impairment•

A graphic i l lustrat ion may he lp make this po in t in relat ion to the age-equ iva len t score used to summarize Paul 's per formance on test B. F igure 2 shows two ideal- ized curves that r ep resen t test scores r ece ived by a large number of 3- and 4-year-olds on the test. The range of scores shown by ch i ldren in each group is ind ica ted by the width of the be l l - shaped curve. Normal variat ion resul ted in some over lap in the scores r ece ived by the two groups. P a u l - - a n d all of those 4-year-olds whose scores fall in the shaded area of the c u r v e - - w o u l d be said to have rece ived an age-equ iva len t score of 3 years or less. However , when the var iabi l i ty shown by the 4-year- olds as a group is taken into account us ing z scores, the same score may not be found to be sufficiently different

Downloaded From: http://jshd.pubs.asha.org/ by a Ohio State University - Library User on 12/06/2015Terms of Use: http://pubs.asha.org/ss/rights_and_permissions.aspx

Page 5: USE AND MISUSE OF NORM-REFERENCED TESTS …u.osu.edu/cclab/files/2015/11/use-and-misuse-2crk3u3.pdf · USE AND MISUSE OF NORM-REFERENCED TESTS IN CLINICAL ASSESSMENT: A HYPOTHETICAL

342 Journal of Speech and Hearing Disorders 49 338-348 November 1984

f i 3-year-olds / / 7 '~ / .i

/ -:~i / ..:~:~:~:}~ I ~ l ~

/ :::~::::::::.:::.:~

4-year-olds

I _zscoresfor - 4 - 3 - 2 - 4-year-olds 0 +1 +2 +3 +4

FIGURE 2. Hypothetical distributions of scores received by two groups of children. The shaded area of the distribution for 4-year-olds includes scores that would yield age-equivalent scores of 3 years or less but z scores of - 2 or better.

from the average ob ta ined by this group to suggest im- pai rment . Certainly, i f a z score be t te r than - 2 . 0 were cons ide red wi thin the normal range, almost all of the scores fall ing in the shaded area on the figure would be cons ide red normal, desp i t e their cor respondence to age- equ iva len t delays of up to a year or more. Thus, age- equ iva len t delays do not necessar i ly indicate de lay or impai rment .

In summary, age-equ iva len t scores are psychometr i - cal ly imprecise . In addit ion, they may encourage mis- in terpre ta t ion by any but the most cautious consumer, and they may be par t icular ly mis lead ing to the nonprofes- sionals to whom they are f i 'equently presented .

Suggestions Begarding the Proper Use of Age- Equivalent Scores

Summar iz ing test results with s tandard scores (e.g., z scores) or with percen t i l e ranks is r e c o m m e n d e d by the Amer ican Psychological Associat ion (1974, p. 23) as an a l ternat ive to the use of deve lopmenta l scores such as age-equ iva len t scores.

Unfortunately, the calculat ion of these s tandard scores is not always poss ib le because it requires access to the mean and s tandard devia t ion of the scores obta ined by a re levant normat ive sample. Desp i te its importance, this information is often miss ing fi'om speech and language test manuals (McCauley & Swisher, 1984). Cl inicians may wish to seek out tests that inc lude this information and to avoid tests for which only age-equiva lent scores are avai lable .

I f age-equ iva len t scores are used, awareness of their l imitat ions is an impor tant s tep toward minimiz ing errors in interpreta t ion. I t has been sugges ted that age- equ iva len t scores be used in conjunct ion with other, be t te r descr ip t ions of test results, such as z scores or pe rcen t i l e ranks (Mehrens & Lehmann, 1980, p. 99). The i r mis in te rpre ta t ion can further be avoided if age- equ iva len t scores are accompan ied by a thorough de- scr ipt ion of the reasonable inferences they support.

U)

O O

CO N

0

-1 ,

- 2 -

- 3

Subtest 1 (Receptive)

Subtest 2 (Expressive)

FIGURE 3. A hypothetical profile of Paul's performance on two subtests of test C,

P R O F I L E S AND T H E D E S C R I P T I O N OF P A T T E R N OF I M P A I R M E N T

Profiles are typica l ly const ructed by graphing two or more different test scores for the same ind iv idua l using the same unit of m e a s u r e m e n t [e.g., z scores (Salvia & Ysseldyke, 1981, p. 479)]. Because such profiles consoli- date test information, they may be used by cl inicians to facilitate visual comparisons of test scores and to extract information about an ind iv idua l ' s pat tern of impairment . The i r use somet imes involves the mis taken assumption that a seeming ly significant difference in test scores indicates a co r respond ing difference in associated be- haviors.

The discussion of Paul 's per formance on the two subtests of test C (see the Exhibi t , Error #2) is an example of the way in which test profiles are often in terpre ted . F igure 3 is a hypothe t ica l profile that could be const ructed from Paul 's per formance on test C. In the example, Paul 's lower score on subtes t 2 than on subtest 1 was cons ide red to be an indica t ion that he was more impai red on behaviors assessed by subtes t 2. However , unless steps are taken to demons t ra te that the difference be tween Paul 's two scores is p robab ly not due to chance, this in terpre ta t ion cannot be just if ied. Because such steps are re la t ive ly t ime consuming, are not commonly appreci- ated, and are not always poss ib le , they are often not taken.

Error 2: Profiles as Descriptions of Pattern of Impairment

The compl icat ions that make profile analysis t ime con- suming are c losely re la ted to the compl ica t ions accompa- nying the compar ison of an ind iv idua l ' s test score to the norms. The scores to be compared in a profile are only est imates of the ideal or true scores one would obtain if the scores were f lee from error, jus t as a s ingle score and the test norms are only es t imates of the values one would

Downloaded From: http://jshd.pubs.asha.org/ by a Ohio State University - Library User on 12/06/2015Terms of Use: http://pubs.asha.org/ss/rights_and_permissions.aspx

Page 6: USE AND MISUSE OF NORM-REFERENCED TESTS …u.osu.edu/cclab/files/2015/11/use-and-misuse-2crk3u3.pdf · USE AND MISUSE OF NORM-REFERENCED TESTS IN CLINICAL ASSESSMENT: A HYPOTHETICAL

McCAULEY & SWISHER: Norm-Referenced Tests and Assessment 343

obtain if the scores were free from error. Because error contributes to the magnitude of each score, it can affect the size of the difference between pairs of scores. There- fore, two scores within a test profile can seem quite different from one another due to measurement error rather than to real differences in the behaviors being measured.

To emphasize the possible effect of misinterpretation in profile analysis, Salvia and Ysseldyke (1981) wrote that "simplistic and unsophistieated profile analyses . . . are likely to be interpretations more of error than of real differences" (p. 499). In other words, one can jump to the wrong conclusion about an individual's relative strengths and weaknesses by assuming that all apparent differences in test scores represent real differences in behavior. Other writers (Anastasi, 1976, pp. 129-131; Mehrens & Lehmann, 1980, pp. 56-57, 101-105) make somewhat different suggestions regarding the procedures to be used in profile analysis. Their suggestions, however, share the common goal of avoiding the mistaken interpretation of error as indicative of real differences in behavior.

Suggestions Regarding the Use of P@les

Salvia and Ysseldyke (1981, pp. 483--487) discuss at length three steps to avoid simplistic profile analysis. These steps will be described as they relate to scores obtained on different tests, but similar procedures could be used to compare subtests of the same test instrument.

As the first step in a complete profile analysis, the test user determines whether the observed differences be- tween two scores, called a difference score, is likely to be reliable. That is, if the tests (or subtests) were read- ministered, would the difference between test scores tend to remain the same or would it change? Answering this question requires the calculation of the standard error of measurement of the difference (SEMdif) between the two scores to be compared. That value is comparable to the standard error of measurement for a single score and can be used to determine whether the observed difference was likely to have occurred by chance. The required calculations make use of information about the reliability of each test (subtest) and about the intercor- relation of each pair of tests (subtests) for which scores will be compared. Because the calculations are too lengthy to summarize here, the reader is encouraged to consult the detailed examples provided by Salvia and Ysseldyke (1981, pp. 487-505).

If results of the first step indicate the existence of a reliable difference, the test user then undertakes the second step in the analysis. In this step, one examines the characteristics of the normative sample used for each test (subtest) to see if differences in the characteristics of the two samples can explain that reliable difference. The following is an extreme example in which differences between the normative samples are largely responsible for a reliable difference in test scores. Suppose that Paul had taken two receptive vocabulary tests which were identical except that one had been standardized on ehil-

dren who had participated in an academic enrichment program for gifted children and the other had been standardized on children who were acquiring English as a second language. When Paul is compared to these two groups, we might expect him to perform poorly when compared to the first group of children but well when compared to the second group. Hence, a reliable, and possibly large, difference between the z scores obtained by Paul on the two tests might be found, yet the differ- ence would be due to the normative samples, not to real differences in Paul's vocabulary skills from one testing to the next.

The second step described above is not needed for profiles of subtest scores which make use of a single normative sample (Salvia & Ysseldyke, 1981, p. 504) unless there is some evidence that different children were used in the standardization of different subtests. In actual practice, one may wish to examine differences in normative samples as the first step in profile analysis in order to avoid determining the reliability of an observed difference only to find it is probably due to differences between the normative samples.

The third step in profile analysis--concluding that a difference between two scores suggests a real difference in underlying skills--is taken only if the reliable differ- ence between two test scores does not seem to be ex- plained by differences in normative samples. Many so- called profile analyses include only this last step. How- ever, this step is better regarded as a possible outcome of the procedures described above than as an automatic conclusion that can be reached through visual inspection of the profile alone.

Despite its usefulness, the test user will frequently be unable to perform a profile analysis because of the ab- sence of information on reliability or the intercorrelation of the test instruments. As a conservative, but reasonable and relatively easy alternative to the complete profile analysis discussed by Salvia and Ysseldyke, one might confine observations about patterns of impairment to observations regarding the presence or absence of impair- ment in different areas.

TEST ITEMS AND THE PLANNING OF THERAPY OBJECTIVES

Clinicians sometimes base therapy objectives on an analysis of errors a client has made on a norm-referenced test. This practice is probably motivated in part by a desire to make the most of already collected information. It is almost certainly based on the premise that missed items indicate significant and real areas of weakness.

In the Exhibit (Error #3), Paul's performance on the plural -s on test C provided the rationale for including those structures in initial therapy objectives. Unfortu- nately, the use of this kind of analysis may lead the clinician to miss important deficits while focusing ther- apy objectives on less important or even falsely identified deficits.

Downloaded From: http://jshd.pubs.asha.org/ by a Ohio State University - Library User on 12/06/2015Terms of Use: http://pubs.asha.org/ss/rights_and_permissions.aspx

Page 7: USE AND MISUSE OF NORM-REFERENCED TESTS …u.osu.edu/cclab/files/2015/11/use-and-misuse-2crk3u3.pdf · USE AND MISUSE OF NORM-REFERENCED TESTS IN CLINICAL ASSESSMENT: A HYPOTHETICAL

344 Journal of Speech and Hearing Disorders 49 338-348 November 1984

Error 3: Performance on Individual Test Items as Indications of Deficit

To begin with, norm-referenced speech and language tests do not describe all skills that might need to be addressed in therapy (Leonard, Prutting, Perozzi, & Berkley, 1978; Miller, 1983; Sehery, 1981; Snyder, 1983). Because of gaps in the skills assessed by available tests, the use of norm-referenced tests to plan therapy pre- cludes the identification of problems in many areas.

Using the results of test items to plan therapy objec- tives also presents grave difficulties for skills that are covered by existing tests. First, the relatively small number of items included in a norm-referenced test cannot adequately sample all of the specific forms and developmental levels that might be appropriate (McLean & Snyder-MeLean, 1978, p. 129). For example, language tests may include no items or only a small number of items for a given form (Owens, Haney, Giesow, Dooley, & Kelly, 1983) or a given developmental level (Leonard et al., 1978). Norm-referenced tests, after all, are designed to function in somewhat the way that headlines do in a newspaper- - they give the user the quickest possible summary of a story (e.g., Paul B. Mag Have Language Problem), but they fail to tell the story itself.

An example may emphasize the potential impact of this lack of detail on therapy planning. Suppose Paul had acquired the plural -s morpheme in some developmen- tally early contexts, but not in later ones. This pattern might not be evident from his performance on test C which would only include a few such items, almost all of which might be developmentally similar. As this example illustrates, important information about the scope of im- pairment and the nature of existing skills will usually be missing from the results of a norm-referenced test.

A second, related problem with planning therapy ob- jectives from test items is that this approach can result in an inaccurate conceptualization of the individual's im- pairment. A single error on a standardized test may have any of several causes ranging from the test taker's mo- mentary lapse in attention to lack of mastery of the skill tapped by that item. Similarly, correct answers on indi- vidual items may result from an age-appropriate mastery of the linguistic structure being probed or from a poten- tially unproductive familiarity with the structure in an isolated context. A correct answer may even result from a lueky guess for certain kinds of tests [e.g,, receptive tasks with a limited number of alternative responses (Burns, 1978)]. In short, individual errors represent only possible deficits and correctly answered items only possible eom- peteneies. Only multiple observations of similar phenom- ena cause an individual's performance on the test as a whole to be less subject to alternative explanations.

A third problem that limits the utility of norm- referenced tests in planning therapy is that the tasks included in such tests assess behaviors only within a very restricted range of communicative contexts (Bloom & Lahey, 1978, pp. 313-314; Danwitz, 1981; Muma, Lubinski, & Pierce, 1982). Restrictions on nonlinguistic

contexts and the purpose of communication can result in an incorrect characterization of a child's functional com- munication skills (Bloom & Lahey, 1978, p. 492; Launer & Lahey, 1981; Prutting, Gallagher, & Mulae, 1975).

A fourth problem with planning objectives from errors on specific test items concerns the typical scoring systems used for speech and language tests. With a few exceptions (e.g., Porch, 1967, 1974), most scoring systems fail to provide subtle descriptive clues which, in sufficient num- bers, would prove useful in therapy planning (McLean & Snyder-McLean, 1978, p. 131). Description of an indi- vidual's responses--both correct and incorrect--may help differentiate between different kinds or degrees of im- pairment. For instance, it might be useful to distinguish errors indicating response bias from those indicating the client's mastery of some aspect of the tested skill or a complete absence of the tested skill. Similarly, it might be desirable to separate correet responses that are im- mediate from those that are delayed. This information would prove valuable to the elinieian, who could in- corporate in early therapy objectives a behavior that had been assoeiated with self-corrected responses rather than one that had evoked no observed response.

When test items are translated too literally into therapy objectives, their use may also result in a fifth problem-- the invalidation of the norm-referenced test as a test of ability. For example, subsequent administrations of a vocabulary test to a child who has been drilled on items from that test will be almost useless in indieating the child's general receptive vocabulary skills.

Because of the problems noted above, the use of test items in the planning of therapy may often result in an impoverished and potentially inappropriate set of objec- tives. There is probably no circumstance in which norm- referenced test items ean profitably be used fbr this purpose.

Suggestions Regarding Tests and Therapy Objectives

As argued above, the determination of therapy objec- tives requires a more comprehensive description of a client's current level of skills than is typically obtained from a norm-referenced test. In the case of a developing child, this detailed description needs to be made with reference to stages of normal development in order to determine which skills are emerging and, thus, which are appropriate therapy objectives (McLean & Snyder- McLean, 1978, p. 253). For language deficits, a detailed description would address both the expression and com- prehension of language in both spoken and written form. At present, however, available data permit a detailed description of only some of the important components of expressive, spoken language. These data have been used to develop procedures to analyze spontaneous language samples (Bloom & Lahey, 1978, pp. 445-488) and to choose the content of criterion-referenced tests (e.g., McCarthy, 1974).

Downloaded From: http://jshd.pubs.asha.org/ by a Ohio State University - Library User on 12/06/2015Terms of Use: http://pubs.asha.org/ss/rights_and_permissions.aspx

Page 8: USE AND MISUSE OF NORM-REFERENCED TESTS …u.osu.edu/cclab/files/2015/11/use-and-misuse-2crk3u3.pdf · USE AND MISUSE OF NORM-REFERENCED TESTS IN CLINICAL ASSESSMENT: A HYPOTHETICAL

MCCAULEY & SWISHER: Norm-Referenced Tests and Assessment 345

Systematic language sampling and analysis. Although still in need of further study, language analysis (e.g., Bloom & Lahey, 1978, pp. 445-488; Lee, 1974; Lee & Canter, 1971; Tyack & Gottsleben, 1974) is an important clinical tool for the description of children's expressive language. The purpose of such procedures is to provide a detailed description of the language used by the client in relation to the normal stages of language acquisition. This detailed description oilers a fertile source of suitable therapy objectives.

Although the virtues of expressive language sampling are everywhere praised, the procedures themselves ,nay be infrequently used in some settings: Two recent sur- veys (Kilburg, cited in Kilburg, 1982; Muma, Pierce, & Muma, 1983) suggest that language sampling is currently more popular among ASHA-accredited educational insti- tutions than it is among ASHA-accredited clinics. Kilburg described the results of a national survey in which she asked speech-language pathologists in ASHA-accredited speech-language clinics what kind of language measures they preferred and which methods they actually used in the assessment of children up to 3 years of age. She noted that although most respondents preferred the analysis of spontaneous behavior, less than one third of them re- ported that they actually made use of that kind of analysis.

Muma, Pierce, and Mnma (1983) conducted a survey of ASHA-accredited educational institutions and found that language sampling and analysis procedures were the most frequently used tools in the assessment of child language among 76 programs. Future surveys will be needed to determine what accounts for the differences between these two surveys. For example, are apparent differences in the frequency of language sampling in these two settings due to differences in survey methods, in the education of the two groups of respondents, or in time constraints that might prove more compelling in the clinical setting observed by Kilburg than in the setting observed by Muma and his coauthors?

The reportedly limited use of language analysis meth- ods in clinical settings may stem from the relatively lengthy and complex procedures they require. Nonethe- less, in light of the meagre information provided by norm-referenced tests for the planning of therapy, the value of language analysis outweighs the difficulties it presents. Furthermore, recent advances in computer- assisted language analysis (e.g., Hixson, 1983; Mordecai & Palin, 1983) may dramatically reduce these practical difficulties.

Although the concepts ofreliabiiity and validity may be more easily overlooked for language analysis procedures, these concepts nonetheless apply to any behavioral mea- sure. For example, language analysis has been shown to be sensitive to the child's activity while the sample is collected (Longhurst & Grubb, 1974), as well as to the identity of the speaker's interlocnter (Olswang & Carpen- ter, 1978). Therefore, this assessment method, as others, requires additional study.

Criterion-referenced tests. Whereas norm-referenced tests are designed to compare the test taker's performance

to that of the normative sample, criterion-referenced tests are designed to determine which skills a test taker has attained along a continuum of skills judged by the test developer to be important. Instead of being used to answer questions such as "how does this child perform compared to his age peers?," criterion-referenced tests are used to answer questions such as "does this child have these skills?"

Although criterion-referenced tests represent a theo- retical alternative for planning therapy, they do not yet present a practical alternative (Schery, 1981). To date, few criterion-referenced tests have been designed to address content areas of interest to speech-language clini- cians (e.g., McCarthy, 1974). Such criterion-referenced tests have usually been developed in association with therapy programs and are limited in scope to the therapy objectives of a specific program.

At least in part, the limited number of available criterion-referenced tests is due to the fact that such tests are difficult to develop. They require a detailed descrip- tion of the content area they are meant to assess--a requirement that can lead to a cumbersome instrument that might not gain wide acceptance. From a psycho- metric point of view, they present difficulties to the test designer because of ongoing controversies about ap- propriate standardization methods [e.g., the appropriate way in which to determine reliability for such instru- ments (Carver, 1974; Millman, 1979; Proger & Mann, 1973)]. Nonetheless, because norm-referenced tests pro- vide incomplete and possibly misleading information for the formulation of language objectives, language analysis and the development of criterion-referenced tests require increased clinical and research attention.

MULTIPLE TEST SCORES AND THE ASSESSMENT OF PROGRESS

Spurred on by demands for increased accountability (e.g., Caceamo, 1973; Mowrer, 1972), some clinicians have begun using periodic administrations of norm- referenced tests to document the effects of therapy. The rationale behind this practice is presumably that test instruments that are appropriate for the identification of impairment must be equally appropriate for the measure- ment of progress. Their use seems to allow clinicians to adapt measures that they consider objective and manage- able to the troublesome task of validly assessing progress.

In the hypothetical case (see the Exhibit, Error #4), tests B and C were readministered after a period of therapy for the purpose of measuring Paul's progress in therapy. In spite of clinical observations to the contrary, no substantial change in test scores was obtained, and the clinician concluded that Paul was not making much progress. Despite its apparent efficiency, however, this practice can result in spurious conclusions regarding changes in behavior during therapy.

Downloaded From: http://jshd.pubs.asha.org/ by a Ohio State University - Library User on 12/06/2015Terms of Use: http://pubs.asha.org/ss/rights_and_permissions.aspx

Page 9: USE AND MISUSE OF NORM-REFERENCED TESTS …u.osu.edu/cclab/files/2015/11/use-and-misuse-2crk3u3.pdf · USE AND MISUSE OF NORM-REFERENCED TESTS IN CLINICAL ASSESSMENT: A HYPOTHETICAL

346 Journal of Speech and Hearing Disorders 49 338-348 November 1984

Error 4: Repeated Testing as a Means of Assessing Progress

The use of' norm-referenced tests to assess progress can resnlt in the underestimation or overestimation of change. It can result in the underestimation of change because the purpose for which they are designed is largely incompatible with the assessment of progress. To allow the comparison of a child to his peers, a test needs to examine gross, relatively stable behavior patterns (e.g., single-word vs. more complex stages of production). Norm-referenced tests are typically designed with this purpose in mind. On the other hand, to examine progress in therapy, a test needs to examine the specific behaviors that will be affected by learning within as well as be- tween major developmental stages (Salvia & Ysseldyke, 1982, p. 544; Schery, 1981). Criterion-referenced tests are more often designed for this second purpose. Because norm-referenced tests are designed to compare individu- als, they are likely to be less sensitive to changes in behavior over time than criterion-referenced tests.

Guidelines used in choosing items during the stand- ardization process differ depending upon which of these two purposes the test is intended to serve (Carver, 1974). Items that result in a broad range of' performances are more appropriate for a test designed to look at differences between individuals. On the other hand, items that result in a narrow range of performances (e.g., those passed by most individuals who have had a particular course of intervention) are more appropriate fbr tests designed to look at progress over time. Thus, norm-referenced tests, which by definition are designed to look at differences between individuals, will ahnost always be composed of itelns that can be expected to result in a broad range of performances and, therefore, will rarely be sensitive enough to behavioral change to document progress in therapy.

Besides its potential for the underestimation of change, the use of norm-referenced test scores to assess change can result in the overestimation of progress in therapy. This occurs when positive changes in test scores, called gain scores, are automatically interpreted as progress. When any positive change, regardless of magnitude, is considered significant, an individual may seem to im- prove on the test without showing improved language skills in therapy sessions or everyday talking situations.

Determining the amount of change that is noteworthy requires a detailed analysis similar to that used in profile analysis. One must remember that some amount of change in test scores can occur as a result of the test's imperfect reliability. Because gain scores can be quite unreliable (Mehrens & Lehmann, 1980, p. 57; Salvia & Ysseldyke, 1981, p. 544), even a very large difference in test scores might occur by chance, making improvement difficult to document with norm-referenced tests.

Still another reason that the use of norm-referenced tests can result in the overestimation of progress is that the client may simply "learn" the test because of re- peated administrations. Also, the clinician may intention-

ally or unintentionally teach correct responses to some items during intervention. Using a norm-referenced test no more frequently than suggested by the test developers reduces the likelihood of invalidating later test results in this way.

Thus, because norm-referenced tests were constructed to serve a different purpose and because gain scores can be particularly unreliable, norm-referenced tests do not lend themselves to use in monitoring an individual's performance over time. Their use can engender inflated illusions of success or unwarranted delusions of failure and can invalidate their future use as tests of" skill.

Suggestions Regarding Multiple Test Scores and Assessing Progress

Repeated, relatively infrequent administrations of a norm-referenced test may be a valid component of an ongoing assessment program if the intent ofreadminister- ing the test is to obtain evidence about the continued existence of an impairment rather than to obtain evidence about the amount of change observed in the test taker's behavior. When a norm-referenced test is used in this way, it can yield indirect information about an indi- vidual's progress. That is, if the test faker's test perform- anee no longer differs significantly from that of his peers, then one can infer that therapy is no longer needed.

Because the use of norm-referenced tests offers, at best, a rough and potentially inaccurate measure of behavioral change, a finer description of the effects of therapy must be found elsewhere. The assessment of behavioral change is more appropriately made with measures such as criterion-referenced tests and within-subject experi- mental designs.

Criterion-referenced tests. Criterion-referenced tests were discussed previously in regard to their use in the description of impaired functioning. They are also useful in the assessment of progress (Carver, 1974) because they can be designed to be especially sensitive to changes in behaviors for which therapy has been undertaken. As mentioned earlier, however, criterion-referenced tests are in relatively early stages of development and so present a theoretical rather than practical alternative.

Within-subject exl)erimental designs. Within-subject or single-subject experimental designs, as they are also called, represent a particularly powerful method of' as- sessing progress in therapy (Coste!lo, 1979; MeI3eynolds, 1983; McReynolds & Kearns, 1983). These designs offer several distinct advantages over other methods. First, they can permit one to determine more directly whether or not therapy is responsible for changes in behaviors occurring during therapy. Second, because they involve the continuous collection of data, they facilitate the al- teration of therapy techniques in response to client be- haviors throughout therapy.

McPteynolds (1974, p. 131) discussed four components of systematic procedures for training children's language that are particularly applicable to within-subject experi-

Downloaded From: http://jshd.pubs.asha.org/ by a Ohio State University - Library User on 12/06/2015Terms of Use: http://pubs.asha.org/ss/rights_and_permissions.aspx

Page 10: USE AND MISUSE OF NORM-REFERENCED TESTS …u.osu.edu/cclab/files/2015/11/use-and-misuse-2crk3u3.pdf · USE AND MISUSE OF NORM-REFERENCED TESTS IN CLINICAL ASSESSMENT: A HYPOTHETICAL

MCCAULEY • SWISHER: Norm-Referenced Tests and Assessment 347

mental designs: (a) the isolation and description of a behavior that will be the focus of therapy; (b) the speci- fication of stimuli, responses, and procedures used in therapy; (c) the collection of data sensitive to changes in the behavior that is the target of therapy; and (d) the assessment of the effectiveness of therapy through the use of experimental control. Although none of the four com- ponents is completely alien to speech-language clini- cians, the final component is perhaps the least familiar. The assessment of therapy effectiveness is achieved through the clinician's demonstration that (a) changes in behavior observed during therapy are reliable and (b) that the therapy procedures, stimuli, situation, and so forth have been held constant.

S U M M A R Y A N D C O N C L U S I O N S

In the body of this paper, four errors were considered that arise when norm-referenced tests are used in the following ways: the use of age-equivalent scores as the only summary of test results, the use of test profiles to compare an individual's strengths and weaknesses, the use of test items to formulate specific therapy objectives, and the use of repeated testing to assess changes in client behaviors during therapy. Only one of these practices was considered a proper use of norm-referenced test results-- the use of test profiles to describe a client's strengths and weaknesses. Yet even that use was considered acceptable only when rigorous steps are taken to rule out the inter- pretation of error as real differences in performance. For each misuse of a norm-referenced test that was consid- ered in this paper, however, alternative methods that entail different assessment tools or a somewhat different application of the norm-referenced test are available.

One factor contributing to the errors outlined in this paper may be an insufficient appreciation of certain es- sential psychometric concepts and of the proper role of norm-referenced tests in assessment. However, another is certainly the dearth of adequately developed alternatives to norm-referenced testing. In particular, criterion- referenced tests and within-subject design techniques have not received sufficient research attention. To some degree, the correct use of norm-referenced tests will depend upon the accessibility of these and other, ap- propriate assessment methods.

The proper use of norm-referenced tests as part of a clinical assessment cannot be regarded as a simple pro- tess that begins with the administration of the test. When the information being sought concerns the formulation of specific therapy objectives or the assessment of a client's behavior over time, methods other than norm-referenced testing are more appropriate. When the information being sought concerns the existence of a language impairment or the assessment of an individual's pattern of impair- ment, the correct use of norm-referenced tests begins with the choice of specific tests that are accompanied by the data required to obtain this information accurately. In short, the use of norm-referenced tests must be incorpo- rated within the broader context of clinical evaluation. In

that broader context, specific assessment questions are matched with appropriate tools, and assessment results are evaluated by the clinician who weighs the merits and limitations of each source of information.

A C K N O W L E D G M E N T S

This paper was written while R. McCauley was supported by National Institute of Neurological and Communicative Dis- orders and Stroke, National Research Service Award 5 F32 NS06390 CMS and L. Swisher was partially supported by De- partment of Education Grant G008301459. Portions of this paper were presented at the Annual Convention of the American Speech-Language-Hearing Association in Cincinnati, 1983. The authors wish to thank Anthony DeFeo, Linda Lilley, James Nation, Joseph Perozzi, and Gloriajean Wallace for their helpful comments on various versions of this manuscript. Special thanks are extended to Richard Curlee for his questions and criticisms.

R E F E R E N C E S

AMERICAN PSYCHOLOGICAL ASSOCIATION. (1974). Standards for educational and psychological tests. Washington, DC: Ameri- can Psychological Association.

ANASTASI, A. (1976). Psychological testing (4th ed.). New York: Macmillan.

BLOOM, L., & LAHEY, M. (1978). Language development and language disorders. New York: Wiley.

BURNS, E. (1978). Further comments concerning the Northwest- ern Syntax Screening Test. Journal of Speech and Hearing Disorders, 43, 414--415.

BUROS, O. (Ed.). (1972). The seventh mental measurement year- book (2 vols.). Highland Park, NJ: Gryphon Press.

BuRos, O. (Ed.). (1978). The eighth mental measurement year- book (2 vols.). Highland Park, NJ: Gryphon Press.

CACCAMO, J. (1973). Accountability--a matter of ethics? Asha, 15, 411-412.

CARVER, R. (1974). Two dimensions of tests: Psychometric and edumetrie. American Psychologist, 29, 512-518.

COSTELLO, J. M. (1979). Clinicians and researchers: A necessary dichotomy?Journal of the National Student Speech and Hear- ing Association, 7, 6-26.

DANWlTZ, M.W. (1981). Formal versus informal assessment: Fragmentation versus holism. Topics in Language Disorders, 1, 95--106.

DARLEY, F. (1979). Evaluation of appraisal techniques in speech and language pathology. Reading, MA: Addison-Wesley.

GUILFORD, J. (1954). Psychometric methods (2nd ed.). New York: McGraw-Hill.

HIXSON, P. (1983). DSS computer program. Omaha, NE: Peter Hixson.

KILBURG, H. (1982). The assessment of emerging communica- tion. Communication Disorders, 7(7), 87-101.

LAUNER, P., & LAHEY, M. (1981). Passages from the fifties to the eighties in language assessment. Topics in Language Dis- orders, 1, 11-30.

LEE, L. (1974). Developmental Sentence Analysis. Evanston, IL: Northwestern University Press.

LEE, L., & CANTER, S. (1971). Developmental Sentence Scoring: A clinical procedure for estimating syntactic developments in children's spontaneous speech. Journal of Speech and Hearing Disorders, 36, 315-341.

LEONARD, L., PRUTTING, C., PEROZZI, J., & BERKLEY, R. (1978). Nonstandardized approaches to the assessment of language behaviors. Asha, 20, 371-379.

LONGHURST, T., & GRUBB, S. (1974). A comparison of language samples collected in four situations. Language, Speech, and Hearing Services in Schools, 5, 71-78.

Downloaded From: http://jshd.pubs.asha.org/ by a Ohio State University - Library User on 12/06/2015Terms of Use: http://pubs.asha.org/ss/rights_and_permissions.aspx

Page 11: USE AND MISUSE OF NORM-REFERENCED TESTS …u.osu.edu/cclab/files/2015/11/use-and-misuse-2crk3u3.pdf · USE AND MISUSE OF NORM-REFERENCED TESTS IN CLINICAL ASSESSMENT: A HYPOTHETICAL

3 4 8 Journal of Speech and Hearing Disorders 49 3 3 8 - 3 4 8 N o v e m b e r 1 9 8 4

LUDLOW, C. (1983). Identification and assessnrent of aphasic patients for language intervention. In J. Miller, D. Yoder, & R. Schiefelbuseh (Eds.), Contemporary issues in language inter- vention. Roekville, MD: American Speech-Language-Hearing Association.

McCARTHY, D. (1974). McCarthy Scales of Children's Abilities. New York: Psychological Corporation.

McCAULEY, R., & SWISHER, L. (1984). Psychometric review of" language and articulation tests for preschool children. Journal of Speech and Hearing Disorders, 49, 34-42.

McLEAN, J., & SNYDER-MCLEAN, L. (1978). A transactional approach to early language training. Columbus, OH: Merrill.

MCREYNOLDS, L. (1974). Application of systematic procedures in cIinical environments. In L. McReynolds (Ed.), Developing systematic procedures for training children's language. (ASHA Monograph No. 18). Washington, DC: American Speech and Hearing Association.

MCI~EYNOLDS, L. (1983). Evaluating program effectiveness. In J. Miller, D. Yoder, & R. Schiefelbusch (Eds.), Contemporary issues in language intervention. Rockville, MD: American Speech-Language-Hearing Association.

MCREYNOLDS, L., & KEARNS, K. (1983). Single-subject experi- mental designs in communicative disorders. Baltimore: Uni- versity Park Press.

MEHRENS, W., & LEHMANN, I. (1980). Standardized tests in education (3rd ed.). New York: Holt, Rinehart, & Winston.

MILLER, J. (1983). Identifying children with language disorders and describing their language perfbrmance. In J. Miller, D. Yoder, & R. Schiefelbusch (Eds.), Contemporary issues in language intervention. Rockville, MD: American Speech- Language-Hearing Association.

MILLMAN, J. (1979). Reliability and validity of criterion- referenced test scores. In R. Traub (Ed.), New directions for testing and measurement: Methodological developments. San Francisco: Jossey-Bass.

MORDECAI, D., & PALIN, M. (1983). Lingquest. Napa, CA: Lingquest Software.

MOWRER, D. (t972). Accountability and speech therapy in the public schools. Asha, 14, 111-115.

MUMA, J., LUBINSKI, R., & PIERCE, S. (1982). A new" era in language assessment: Data or evidence. In N. Lass (Ed.), Speech and language: Advances in basic research and practice (Voh 7). New York: Academic Press.

MUMA, J., PIERCE, S., & MUMA, D. (1983). Language training in speech-language pathology: Substantive domains. Asha, 25, 35--40.

NUNNALLY, J. (1978). Psychometric theory (2nd ed.). New York: McGraw-Hill.

OLSWANG, L., & CARPENTER, R. (I978). Elicitor effects on the language obtained fiom young language-impaired children. Journal of Speech and Hearing Disorders, 43, 76-88.

OWENS, t:{., HANEY, M., GIESOW, V., DOOLEY, L., & KELLY, R. (1983). Language test content: A comparative study. Language, Speech, and Hearing Services in Schools, 14, 7-21.

PORCH, B. (1967). Porch Index of Communicative Ability. Palo Alto, CA: Consulting Psychologists Press.

PORCH, B. (1974). Porch Index of Communicative Ability in Children. Palo Alto, CA: Consulting Psychologists Press.

PROGER, B., & MANN, L. (1973). Criterion-referenced measure- ment: The world of gray versus black and white. Journal of Learning Disabilities, 6, 19-30.

PRUTTING, C., GALLAGHER, T., & MULAC, A. (1975). The expres- sive portion of the NSST compared to a spontaneous language sample. Journal of Speech and Hearing Disorders, 40, 40-49.

SALVIA, J., & YSSELDYKE, J. (1981). Assessment in special and remedial education (2nd ed.). Boston: Houghton-Mifflin.

SCHERY, T. (i981). Selecting assessment strategies for language- disordered children. Topics in Language Disorders, 1(3), 59-73.

SNYDER, L. (1983). From assessment to intervention: Problems and solutions. In J. Miller, D. Yoder, & R. Schiefelbusch (Eds.), Contemporary issues in language intervention.

Rockville, MD: American Speech-Language-Hearing Associa- tion.

THORNDIKE, R. (1982). Applied psychometrics. Boston: Houghton-Mifflin.

TYACK, D., & GOTTSLEBEN, 1~. (1974). Language sampling, analysis, and training: A handbook for teachers and clini- cians. Palo Alto, CA: Consulting Psychologists Press.

WEINER, P,, & HOOCK, W. (1973). The standardization of tests: Criteria and criticisms. Journal of Speech and Hearing Be- search, 16, 616-626.

Received October 25, 1983 Accepted August 14, 1984

Requests for reprints should be sent to Rebecca J. McCauley, Ph.D., Communication Sciences and Disorders, The John F. Kennedy Institute for Handicapped Children, 707 North Broad- way, Baltimore, MD 21205.

A P P E N D I X

Calculations leading to a test taker's estimated true score and the 95% confidence interval around the estimated true score. The equations used in this example and a more detailed discus- sion of this topic can be found in Salvia and Ysseldyke (1981, pp. 86--92).

E x a m p l e 1. Obtain the fbllowing intbrmation (ifavailable) from the test

manual and the test taker's corrected test form:

Test taker's score: X = 75 Mean score of relevant normative sample: M = 100 Standard deviation of scores obtained by sample: SD = I0 Reliability coefficient reported for similar sample: r = .70

2. Calculate the estimated true score (X') using the following equation:

X' = M + (r)(X - M) = 100 + (.70) (75 - 100) = 82.5

3. Calculate the 95% confidence interval using the following steps:

a. Estimate the standard error of measurement (SEM) using the following equation:

SEM = SD X/1 - r = 1 0 V 1 - - .70 - 5.48

b. Calculate the lower and upper limits of the 95% confidence interval using the following equations:

Lower limit - X' [(1.96) 1 (SEM)] = 82 .5 - [ ( 1 . 9 6 ) ( 5 . 4 8 ) ] = 71.76

Upper limit - X' + [(1.96) (SEM)] = 82.5 + [(1.96)(5.48)] = 93.24

These calculations indicate the test taker's true score can be expected to fall between the values of 71.76 and 93.24 95 times out of 100 and that the single best guess one can make as to the test taker's true score is 82.5.

JThis value is used to produce a 95% confidence interval.

Downloaded From: http://jshd.pubs.asha.org/ by a Ohio State University - Library User on 12/06/2015Terms of Use: http://pubs.asha.org/ss/rights_and_permissions.aspx