\\ 93 ! 1\"[lIIlI/fII;On lind f·Jftdrlll'y FUllt'/itlll.l m• 2 n 1.15 f 0 r •m t I 0 0.& n 0 23 .. -4 -3 -2 -1 0 Ability Figure 6.1. Hem Information Functions for Six Typical Test Items 3. When c > O. other thing!! being (,I,ual. an itcm is lesl! IIseful for assessing abilily. (This can be seen by comparing .he item information functions for Items I and 3.) 4. An item with low discriminating power is nearly useless stati,~tically in R test (see Item 6), 5. Even the most discriminating items (Items I and 4) provide less informa- lion for assessing ability in SOIllC rcgions of .he uhility cOlltinullm Ifllln do less discriminating items (Item 5). Hem 5 would he more useful for assessing abilities of middle-ability examinees than (say) either Item I or Item 4. Clearly, item information functions provide new directions for judging the utility of test items and constructing tesls. Because ilem information functions are lower, generally, when c > 0 Ihan when c =O. researchers might be tempted to consider fitting one- or two-parameter models to their test dala. Resulting item information fUllctions will be higher; however, the onc- alld two-parameter item information curves will only be useful when Ihe ICCs from which they (Ire derived fit the test data. The lise of ICC's Ihal dn 110t adequalely fit the lest dala and their corresponding item informalion curves is far from
94 FUNDAMENTALS OF ITEM RESPONSE TIIEOI{Y optimal and will give misleading results (see, for example, de Gruijtcr, 1986), Test Information Functions The information function for a test, denoted 19 a'nd derived by Birnbaum (1968, chapter 17), is given by L\" [6.41 1(9) :::: Ii (0) i= I The information provided by a test at 9 is simply the sum of the item information functions at 9, From Equation 6.4 it is clear that items contribute independently to the test information function, Thus. the contribution of individual test items can be determined without knowl- edge of the other items in the test. This feature is not available in classical test theory procedures. The contribution of test items to test reliability and item discrimination indices (e.g., point-biserial correla- tions) cannot be determined independently of the characteristics of the remaining items in the test. This is true because the test score, which is used in these calculations, is dependent on the particular selection of test items. Changing even one item will have an effect on test scores and, hence, all of the classical item and test indices will change. The amount of informat ion provided by a test at 9 is inversely related to the precision with which ahility is estimllted at that point: SE(O\" ) = ---I -- (fi.51 .[r(O) 1\\ where SE(O) is called the standard error of estimation. This resull holds whenever maximumlikclihood estimates of 0 are obtained. With knowl- edge of the test informntion at 9. a confidence hand can be found for usc in interpreting the abililY estimale (see chaplcr 3). In the framework of \"IRT, SE(9) serves Ihe same role as the standard error of Im:asuremcnt in classical measuxcment theory. It is important 10 note, however. thul the value of SE(9) varies with abilily level, whereas Ihe classical standard error of measurement does nol.
\"The stalldard error of \"0, SE(O), is the standard deviation of the aSYlllptoticully florlllal distributioll of the maximulIl likel ihood estimate of ability for II given true value of ahility O. The distribution is nonTI<11 when the test is long. Even with tests as short as In to 20 items, however, the Ilormal approxilllation is satisfactory for lIlost purpose.~ (Silll1~jjllla, 1977). The magnitude of the standard error depends, in general, on (a) the number of test items (smaller standard errors are associated with longer tests); (b) the quality of test items (in general, smaller standard errors are associated with highly discriminating items for which the correct answers cannot be obtained by guessing); and (c) the match between item difficulty and examinee ability (smaller standard errors are asso- ci.lted with tests composed of items with difficulty parameters approx- imately equal to the ability parameter of the examinee, as opposed to tests that are relatively easy or relatively difficult). The size of the standard error quickly stabilizes, so that increasing information beyond a value of (say) 25, has only a small effect on the size of errors in ability estimation (see, for example, Green, Yen, & Burket, 1989). Relative ..~rnciency Test developers are interested sometimes in comparing the infor- mation functions for two or more tests that measure the same ability. For example, a committee assigned the task of developing a national achievement test may want to compare the test information functions provided by tesls composed of different items or exercises. Comparing information functions for two or more tests can serve as an aid in test evaluation and selection (see, for example, Lord, 1977). Another exam- ple would be a school district or state department or education inter- ested in choosing a standardized achievement test. Based on prior information about student performance, the lest that provides the most information in the region of the ability scale of interest would be preferred. (Other factors, however. should be taken into account in the selection of tests, such as validity. cost, content, and test length.) The comparison of information functions is done by computing the relative efficiency of one test, compared with the other, as an estimator of anility at 0:
96 FUNDAMENTALS 01' ITEM RESI'ONSE THEORY =RE(O) 16.()J where RE(9) denotes relative efficiency and IA(O) and IIl(O) are the information funclions for Tests A and B, respectively, defined over a common ability scale, O. If, for example, IA(O) = 25.0 81\\<l)n(0) = 20.0, =then RE(9) 1.25. and it is said that at 0, Test A is functioning as if it were 25% longer than Test B. Then, Test B would need to be lengthened by 25% (by adding comparable items to those items already in the test) to yield the same precision of measurement as Test A at O. Alternatively, Test A could be shortened by 20% and still produce estimates of ability at 9 possessing the same precision as eSlimates produced by Test B. These conclusions concerning the lengthening and shorlening of tests are based on the assumption that items (or tasks) added or deleted are comparable in statistical quality to other items (or tasks) in the test. In the next chapter, two examples involving ilem and test informalion functions and relative efficiency are presented. Exercises for Chapter 6 I. II. For eadl of Ihe six ilellls given in Tahle 2.1. dclcnninc Ihe vllluc of 0 for which the infomlalioll fU\"clion is a maximum. and determine rhe maximum value of Ihe informalion. h. Which ilems would you choose 10 make lip a Iwo-ilem lesllhal will he most useful for nJllkin)\\ decisions IIhOll1 ex IIll1illec.~ 111 0 = 1.0'1 What is Ihe vlllue of the lesl informalion fUllctioll for Iht' Iwu·jtem lesl III Ihis value of 97 2. n. Show Ihal if p= then e 17,,(8 - h) p where Q =0 1 - P. Q
Illformiltioll (lml fJficirnry Flln,·ti/rs 97 h. Show Ihal Iht' cxpressiolJ given h)' E'lllIlIion 6.2 lIlay he wriUen as c. Deduce thaI for the two-parameter model, 1. hem parameters for an \"ilem hank\" made up of four items are given in Table 6.1. TAHLE 6.1 - - ,.. .....~--,-----~~---- c Itrlll \"\" 0.00 {J.OO 1 1.2.'i -0.5 0.00 2, 1.~() n.n n.oo 4 1.25 1.0 100 1.5 SlIppose it is Ilen~ssary to conslructll 1I:.'IIII1I1<1e III' of Ihree items frnlllihis bank. Compule Ihe lesl infonna!ioll function at 0 values of--2, -1,0, I, 2 for Ihe four Ihree-item leslS thaI can hc conslrllcled from Ihc b.mk. Piol the four lesl infortnnlion f\\lnclion~. Which sci of items would you IIS~ if lilt' It~sl is (ksigfl{~d n~ II mnstery le.~1 with II ('II! S(,Of(' set III 0 '\" 1.0\" Answers to Exercises for Chapter 6 I. 8. See Tahle 6.2. 1.00 I(Om .. ) I. O() TAnLE 6.2 1.10 2.34 1.'i0 0.46 Ill.'m h a (' !l\",•• = \" + -0.41 1.29 11.111 2.34 1 I.n 1.8 0.00 0.85 2 1.0 O.S 0.00 3 1.0 1.8 0.25 UH 4 --I.:'i 1.8 0.00 ~ --n.~ 1.2 0.11) 0 O.:'i 0.4 0.15
98 FUNDAMENTALS OF ITEM RESPONSE TI~[(ORY b. Since Items I llnd 2 have their ma)(imullI informatioll al (I = I, Ihese would be Ihe items of choice. lIem 2 contribules milch less Ihan Item I, and, hence, Item I may be !lufficienl. el.7at9- hI 2. a. If P = ---:-:::~.--:- - -+-e~l.1-a(-9 ~II) ,. 1+ =then Q I P :: -----'--~-~- + el.7a{O - II) Hence. I + eI.7a (9-II) = I I Q. frolll which il follows Ihal eI.7a (9-II) = .! _ I = I P Q Q -Q b. This follows direclly from E)(pression 6.2 and part a. c. For the two-parameler model. c = O. Hence. from parI b 3. See Table 6.3. Tur(/. 2. 4) Tes/ (I . .J. 4) Te.s, (2. J. 4) TABLE 6•.3 0.219 0.1117 0.0.'\\4 1.339 0.965 0';40 e Test (I. 2. 3) 2.681 1.486 2.2.'iO 1.215 1.907 2.172 -2 0.219 0.667 1.059 1.076 -1 1.361 0 2.918 1 1.738 2 0.492 The lest consisting of Items 2. 3, and 4 would be Ihe most useful since it =gives the mosl information III 0 1.0.
7 Test Construction Background The construction of achievement and aptitude tests using classical test theory techniques involves the selection of items according to their content and characteristics-item difficulty and discrimination. Items with high discriminating power are generally the most desirable. and the appropriate level of item difficulty is determined by the purpose of the test and the anticipated ability distribution of the group for whom the test is intended. / As noted in earlier chapters. classical indices are not invariant over populations that differ in ability. Hence. the success of classical item selection techniques depends on how closely the group used to deter- mine the item indices matches the populntion for whom the test is intended. When the match is poor, the item indices obtained will not be appropriate for the intended population. In many practical situations, the group for which item indices are ohtained and the group for whom the test is intended are quite different. Consider, for example, the common practice in school districts of field-testing items in the full for use in year-elld tests in Mayor June. While such a field test is useful for detecting gross flaws in the items, the item indices themselves are not likely to be very useful in test development because the ability distribution of the students tested in the filii will differ substantially from the ability distrihution of the students tested at the end of the school year. Another situation in which classical item indices are obtained for groups that may differ from the intended population is in item banking. In developing an item bank, the characteristics of the items to be stored in the bank must be determined. In practice, these items. often called \"experimental\" items, are embedded in a test and administered to a <)<)
100 FlJNIJAMIINTALS OF I rI'M RESI'ONSE TIIEORY group of examinees so that their item indices can be ohtained. If the experimental items are numerous, obviously not all can be embedded in one test. Multiple forms of the test are created, cach containing different experimental items and different forms are administered to different groups of examinees. It is usually not possihle to ensure that the different forms are administered to equivalent group\"; hence, the item indices for experimental items that were given to differcnt groups of examinees may not be comparable. If the items are banked with the assumption that the item indices nre comparable, any test constructed from the bank will not he appropriate for a given popul.ltion. Apart from the problem of noninvariant item indices, the major drawback of classical procedures for test construction is that even when a well-constructed item bank is available, items cannot he selected to yield a test that meets a fixed specificalion in terms of precision of measurement. The contribution of an item to Ihe reliahility of the test does not depend on the characteristics or the item alolle, but also on the relationship between the ilem and the other items in the test. Thus, it is not possible to isolate the contribution of an item to the reliability and, hence, to the standard error of measurement of a test. Item response theory offers a more powcrfulmelhod of item selection than does classical test theory. Item parameters are invariant, overcom- ing thl', prohlems of classical item indict'S descrihed IIhovc. In addilioll, \\ item difficulty and cXlIlilince ahility IIn~ measured 011 the salllt' scall\\ making il possihle to select items that arc most useful in certain regions of the ahililY scule; (or eXilmple, at a cUI-off st:Ort~ for separating masters Jand nonmasters. Perhaps the 1I10s1 important advanlage or IRT is that it permits the selection or items based 011 the mllollnt of information the items contribute to the total amount of information needed in the test to meet the test specifications. Sillce information is related to precision of measurement, it is possible to choose items to produce a test that hilS the desired precision of measurement at any ability level, for example, al a cul-ofr score. Basic Approach A procedure for using item information functions to build tests to meet any desired set of test specifications was outlined by Lord (1977). The procedure employs an item bank with item parameter estimates
1'1'.\\'1 ('omlrllctiml 101 available for Ihe IRT model of dlOice. with accoJllPllnying information fUlictiolls. The steps in the procedure suggested by Lord ( 1977) are as follows: I. Decide on the shape of the desired test infoml3tion function. This was termed the /arllfl informmioll fune/ion by Lord (1977 J. 2. Select hems from the item bank wilh item information flln('tiolls that will fill up hard-to-liII areas under the larget information function. 3. After ea('h item is added to the test, calculate the lest infmmation function for the selected test ilemll. 4. Continue selecting test items until the test information funclion approxi- mates the target information function to it satisfactory degree. These steps are implemented usually within a frall\\ework defined by the content specifications of the test. For a broad-range ability test, the target informatioll runction should be fairly flat, reflecting the desire tu produce a test that would provide (approximately) equally precise ability estimates over the ability scale. For a criterion-referenced test with a cut-off score to separate masters and nonmasters, the desired target information function should be highly peaked near the CUI-off score on the ahility scale, The lise of item information fUllctions allows the test developer to prndul'e a test that precisely fulfills any scI of tc~1 specificationll (assuming thllt the item bank is sufrkienlly hllge aflll cOlltains items of high quality). An example of how item information fun('tions Clln be applicli in a huge test-development projCI.:t was given hy Yen (19R3). A procedure for automating item selection to matdl <I tcst information fUllction, where constraints can he placed 011 thc resulting test to ensure content validity. desired length, and other charactcristics, has been developed recently by van der Linden and Roekkooi-Timminga (1989). Using the procedure suggested by Lord (1977) with a pool of items known to fit a particular item response model, it is possible to construct a test that \"discriminates\" well at a particular region of the ability continuum; that is, if we have a good idea of the ability level of a group of examinees, test items can be selected to maximize test information in the region of ability spanned by the examinees being measured. This selection of test items will contribute optimally to the precision with which ability parameters are estimated.
102 FIlNDAMENTALS or ITEM RESI}ONSE THEORY As an illustration of the Hhove procedure. consider an achievelllcni tesl. On achievement tests. it is common to observe lower performance on a pretest than Oil a poslles!. Knowing this. Ihe test ('ollslrllclor might select easier items for the pretest and Illore difficult items for till' posCles!. On each testing occasion. precision of measurement will be maximized in the region of Hbility where the examinees would 1I10st likely be located, Moreover, because the items on both tc.'\\ts meHsure the same ability and ability estimates do /lot depend on the IliIrticular choice of items, growth can be measured by subtracting the pretest ability estimate from the postlest ability estinHlte. Investigations of the effects of optimal item selection on the decision- making accuracy of a test when the intended cut-off scores or standards for the test are known in advance of test development were conducted by de Gruijter and Hambleton (1983) and Hambleton and de Gruijter (1983). To provide a baseline for interpreting the results, tests were constructed also by selecting test items on a random basis, Random item selection from pools of acceptable test items is a common practice in criterion-referenced testing. Error rates (probabilities of misclassifica- tion) for the test constructed by random item selection procedures were nearly double the error rates obtained with the optimally selected test items. Optimal item selection is made possible within an IRT frame- work because items. pef.'lOnS, and cut-off scores are reported on the same scale. The scenario simulated in the Hambleton and de Gruijter studies is not uncommon in testing practice. For example. in the summer of 1990, a group of educators and noneducators from the United States set standards for marginalIy basic, proficient. Hnd advanced Grade 4 stu- dents on the 1990 NAEP Mathematics Assessment. These three stan- danls were mapped onto the NAEP Reporting (Ahility) Scale \\Ising the test characteristic function defined for thc lotal pool of Grndc 4 math- ematics items. In 1991, whcn test items nre selected for the 1992 Mathematics Assessment, test items could be chosen to maximize the test information at each of the standards. In this way. more accurate information about the percentage of students in each of the four ability intervals defined by the three standards could be obtained. Similar procedures could be used for the Grades 8 and 12 NAEP Mathematics Assessments. A discussion of the process of setting target information functions and selecting items was provided by Kingston and Stocking (1986).
'1'1'.1'/ COIiJ//'l/t'liOlI un Scveral problems, however, remain In he addrcssed. One problem is Ihat lise of statistical criteria for itclll selcction alone will not ensure a l·ontcnl·valid test. Unfortunately, it is easy to ovcrclIlphllsi7.C statistical niteri'l and lIot takc into account the important role that item content plays in test development. Fuilurc to allcnd to content considerations might result in Ii charge thai the test lacks contcnt validity. Ways must be found to combine information about item contcnt and statistical criteria in the item selection process. A solution to this problem has been provided by van der Linden and Boekkooi-Timminga (1989), using linear programming lechniques. Another problem in using item information functions in test develop- ment is that high a values are likely to he overestimated and, hence, the information function may be biased. A test constructed using items with high a values is likcly to be different from the expected test (see, for example, Hambleton. Jones, & Rogers, 1990). Since the test informa- tion function will be overestim<lted, adding several extra items to the test will compensate for the overestimation. A hctter solution is to strive for large examinee samples so that accurate item parameter (~stimates can he obtained. Two examples of Ihe use of informal ion functions in the construction of tests for specific purposes lire given below. Example I: Rroad Abilities Test Suppose the test developer's inlent is to produce 11 wide-range ability test using the item hank in the Appendix. Suppose also that standnrd errors of (lIpproxilllately) U.50 would be acceptahle ill the abilily range ( ·2.00, 2.(0). with somewhlll largn (:rrors (lutsi(1e thai interval. A f\\ possihle larget information fUllction is shown in Figure 7.1. If SE(9) = n.)\", Ihen /(0) 4.\"0. To cOlIstrnctthe shortest possible test that meets the target, items with high discriminations. difficulties between -2.00 and +2.00, and low c values must be chosen. Figure 7.1 shows the target =information function (which is flat. with /(9) =4.00 hctweell 9 -2 and I) = 2) and the test information functions after selecting the best 10, 15. and 20 test items from the item bank for the desired test. Clearly, the rcsulting 20-item test fairly closdy approximates Ihe desired test. The acidition of items with difficulties nellr··2 and 2 would produce an even bctter match to the target information function.
104 FUNUAMENTALS OF ITEM RESPONSE TIIEORY 15 .~--.- .....---- .. -.. T Taroet •e 4 \"0 \".m.) -~ t ,n 3 0 •r 2- m t I 0 n 0 o 2 3 -3 -2 -1 Ability Figure 7.1. Test Information Funclions for 10, 15, and 20 Test hems Example 2: Criterion-Referenced Test Construction Suppose the test developer wishes to construct a 15-item criterion- referenced test (drawn from the item pool shown in the Appendix) to provide maximum information at the cut-off score 9 := -0.50. The resulting test information function is shown in Figure 7.2. The items selected were 2,3,5.7,14,24,27,30,32,36,47,48,71,72, and 73. (Others were possible, but these 15 ultimately were chosen.) For the purposes of comparison, a IS-item test (called the standard test) was constructed by drawing items at random (a common practice in crite- rion-referenced test development) and the test information function for this test also is shown in Figure 7.2. Figure 7.3 shows the relative efficiency of the optimal test compared with the standard test. Clearly. the optimal test provides greater measurement precision in the region of the cut-off score (9 ::: -0.50). The optimal test performs about 60% better than the standard lest in this region. The standard test would need to be lengthened from 15 10 24 test items to perform approximately as well as the optimal test. As can be seen in Figures 7.2 and 7.3, the optimal test does not perform as well as the standard test for high-ability examinees. This is
7'>.H COI'.'(rllcli(l1/ \\OS 7 8 n 5. f .. 0 r m •t 3 I 0 2 n o ____.__..__ l _ _ _ ..~ .. j ____.... '\" .L_ .._ ........_ .. L. ........_ ..L.... __ ..•. _. -3 -2 -1 0 23 Ability Figure 7.2. Test Information Functions for IS-Item Optimal and Randomly ConSlrucled Tests 1.8 - - . - -...- -..-.--.-......-.-----.--....--..- . -...-- -\"--' 1.6 1.4 ,,E 1.2 I C I 0.8 . e n c 0.6 y 0.4 0.2 o ____. __..__L._ _ _ ._L_..____ ._.L... -3 -2 -1 0 23 Ability Figure 7.3. Efficiency Function for IS-hem Optimal Versus Randomly Constructed Tests
FUNDAMENTALS Of ITEM RESPONSE TIIEORY _ _1.25---~'- e.\"e S Random I Optimal a n d a r d E 0.5 r r 0 0,25 0 -2 -1.5 -1 -0.5 o 0.5 Ability Figure 7.4. Standard Error Functions Cor 15-ltcm Optimal and Randomly COllstructed Tests due to the fact that the optimal test is composed l'lfgcly of items that discriminate near the cut-off score and does not contain many items that are suitable for high-ability examinees. The standard test includes a more heterogeneous selection of test items. In practice, the more heterogeneous the item pool and the shorter the test of interest in relation to the size of the item pool, the greater the advantages of optimal item selection over random item selection. The standard errors corresponding to the information functions for the two tests are shown in Figure 7A. Exercises for Chapter 7 I. Suppose you have an item bank containing six test items that provide information Ilt various 9 values as shown in Table 7.1. a. What is the test inCormation and corresponding standard error at 9 1.0 Cor a lest consisting oC hems 2, 3, and 6\" b. How many items like Item 5 are needed 10 obtain II standard crror or OAO at 9 = -1.01
~ 10'1 Tt'.H ('on\\/nl( lioll TAUL!<: 7.1 .--------- 0 111'111 -3 2 () 23 --,------_. -------------- .02 0.00 (110 0.20 01,\\ (UlK 0.04 2 .00 {LOO Il.05 o. I () LlO 0.25 o. 10 3 .no 0.(13 0.10 0.25 0.50 0.40 0.15 4 .15 1.25 14,\\ 0.10 0.02 o.on 0.00 5 .00 (110 0.00 (UO 0.10 0.0') 0.00 (j .no n.oo IUl2 OAIl 2,20 11.40 H.t5 2. Two tests are constructed from the item bank given in Exercise I. Test I consists of Items 2 and 3; Test 2 consists of Items I and 6. a. Compute the information provided by each test at 9 = 0,0, 1.0, and 2.0. b. Compute the efficiency of Test I relalive to Test 2 at e -= 0.0, 1.0. and 2.0. c. What does the relative efficiency analysis indicate about Test 17 d. How many items like Item 5 need to he added to Test I so that Tests I and 2 are (approximAtely) equally informative at 9 = 1.07 3. Suppose that it is desired to construct a criterion-referenced test that is optimally discriminating at e = 1.0. a. Jr the test consists of Items 4 and 5. what is the sHllldard error at 9 = -1.01 h. What is the probability that a candidate with 9 = 0.0 will fail the test when the cut-off score is set at 9 = 1.07 Answers to Exercises ror Chapter 7 I. a. 1(9 -= 1.0):: 3.8: SE(O -= 1.0) =0.51. =b. SE 0.40 requires 1(9) = 6.25. Since the information at9 =-1.0 is 0.60. II items like Item 5 are required to produce the desired test. 2. a. See Table 7.2. TA8U~ 7.2 T~st - 9 0.65 /,0 0.48 I (Ilems 2 and 3) Q.Q 2 (hems I and 6) 1.60 0.35 2,35 0.60
lOS FUNDAMENTALS OF ITEM RESPONSE THEORY h. See Table 7.3. o TABLE 7.3 00 1.0 VI Tr.<t O.till 1.35 Efficienc y (I vs. 2) c. Test I is providing about 58% as much information as Test 2 at 0 =0.0, and aboul68% as much information as Test 2 at 0 = 1.0. AIO :; 2.0, Test I is providing more infomlation than Test 2. d.4 3. a. SE(O ::: -1.0) =0.70 b. The standard error of ahility estimation at 0 = OJ) is 1.12. The-refore, the probability of this candidate failing the test (i.e., the probability of making a false-negative error) is 0.197.
8 Identification of Potentially Biased Test Items Background Perhaps Ihe most highly charged issue surrounding tesling, and cer- tainly the one of greatest importance to the public, is thaI of test fairness. Claims that tests are biased against racial or ethnic minorities have led to numerous lawsuits and calls by such organizations as Ihe NAACP for moratoria or bans on certain types of tests (Rudner. Getson. & Knight. 1980). Tests and testing practices have come under close public scrutiny, and test publishcrs lind uscrs IlIlJst dcmonstrate now that their tests are free of bias against minorities. One of the desirable features of itcm response theory is that it provides a unified framework for conceptUlI\\izing and investigating bias al the ilem level. Refore IRT procedures for investigating bias can be discussed, some clarification of lemlino\\ogy is necessary. Investigations of bias involve gathering empirical cvidence concerning the relative performances on the test item of members of the minority group of interest and members of the group that represents Ihe majority. Empirical evidence of differ- ential performance is necessary, but not sufficient, to draw the conclu- sion thai bias is present; this conclusion involves an inference that goes beyond thc data. To distinguish the empirical evidence from the conclu- sion,the lerm differential itemfunl·tioning (DIP) rather than bia.\\' is used commonly to described the empirical evidence obtained in investiga- tions of bias. Some argument existt; at; to Ihe appropriate definition of DIE A definition that has been used in recent legal settlements and legislation concerning fair testing is that \"an item shows U1F if the majority and minority groups differ in their mean performances on the item.\" The 100
110 FUNDAMENTALS OF ITEM RESPONSE TIIEORY problem with this defi1,itioll is Ihal il docs not take into an:oullt the possibility that other variables, such as a real hetween-group difference in ability, may be responsible for the difference illl,-values (see Lord, 1980). The definition of DIF accepted by psychometricians is that \"an item shows DIF if individuals having the same anility, but· from different groups, do not have the same probability of gelling the item right.\" IRT Methods for Detecting D1F Given the accepted definition of DIF, item response theory provides a natural framework for studying DIE Since the item characteristic function relates the probnhility of a correct response to the underlying ability and to the characteristics of the item. the definition of DIF may be restated operationally as follows: \"An item shows DIF if the item response functions across different subgroups are not identical. Con- verllcly. nn item does not show DIP if the it{~1II dJilfllctcristk hlllctiolis across different subgroups are identicaL\" Based on the definition given above, DIF may be investigated hy comparing the item characteristic functions of two or more groups. Item characteristic functions may be compared in several ways. The first and, intuitively. most direct approach is to compare the parameters Ihal describe the item characteristic curves. Since the ICCs are completely determined by their corresponding item parameters, two ICCs can be different only if the item parameters that describe them are different (Lord, 1980). A second approach is to compare the ICCs by evaluating the area between them (Rudner et aI., 1980). If the area between the ICCs is zero, then the ICCs coincide and, hence, no DIF is present. These two approaches for studying DIF are described in the following sections. Comparison of Item Parameters. If the parameters of two item char- acteristic functions are identical, then the functions will be identical at all points and the prohahilities of a correct response will be the same. The null hypothesis that the item response runctions are the same may be stated as
III where the sUbscript denotes the group in which the parameter estimates were obtained. Ir the hypothesis is rejected for a given item, we can conclude that DIF is present for that item. To test the null hypothesis, estimates of the item pawmeters and the variance~covariance matrices of the eSlimates are needed. Recall that when estimating item and ability parameters in each group, a scale for the parameters must be specified (chapter 3); this is done typically by standardizing either the ability estimates or the difficulty estimates in each group. As we shall see later (chapter 9), standardizing the ability estimates usually will result in different scales in each group, and the item parameter estimates will not be on a common scale. Standardizing the difficulty parameters will result in item parameter estimates that are on a common scale. After the item parameter estimates are placed on a common scale, the variance-covariance matrix of the parameter estimates in each group is computed. First, the information matrix is computed (see chapter 3) for each group alld is inverted. The variance-covariance matrices of the two groups are added then to yield the variance~-covariancc matrix of Ihe differences between the estimalcs. The statistic for testing the null hypothesis is where and L is Ihe variance-covariance matrix of the differences between the parametcr cstimates. The test slatistic is asymptotically (that is, ill large sumplcs) distributed as a chi-square with p degrees of freedom, where fJ is the numher of parameters compared. For the Ihree-p<nameter Illodel, when a, h, and c are compared for each item, IJ J; for the =two-parameter model, p 2; for the one-paramcter Jllodel.I' == I. In the case of the one-parameter Illodcl, thc expression for thl~ chi-slluare stalistic simplifies considerably; the test statistic in Ihis case is \"1.2 == hJifr \\'(h l ) + v(b2)
112 FUNDAMENTAI.S OF ITEM RESI'ONSE TIIEORY where l'(h1) and l'U'l) arc the reciprocals of the information fum:tions for the dilTiculty parameter est imatcs. Since the c parameter is ortell poorly estimated and, hcm:e, has a larj.!c slilndard error, its inclusion in the lest statistic may producc a very conservative test, Ihat is, a test Ihat is not powerful in dctecting DIF. An alternative is \\0 compare only the a and I, parameters a~lq to ignore Ihe c parameters. This approach is reasonable, since if differences exist in thc a and \" parameters across groups. then the item characteristic functions will be different, regardless of the c parameter values; without diffcrenees in the a and b parameters, apparent differences bet ween the (' parameters would be too unreliable to warrant the conclusion thatlhe ilem characteristic functions arc different (Lord, 1980). The comparison of itcm parameters as a lIleans of complIring itelll characteristic fUllctions has been criticized 011 the grounds that signifi- cant differences between thc parameters may he found when 110 practi- cal differences exist between the ICCs in the ahility range of interest. An example of item pHrameter values for two groups Ihat produce almost identical ICCs ill the ability range (~3, 3) was given by Linll, Levine, Hastings, and Wardrop (1981). The item parameters for the Iwo groups are given below: Group I: a::::: 1.8; \" 3.5; c 0.2 Group 2: a 0.5; b ::::: H); c::::: 0.2 Although significant differences exist between the parameters, the ICCs ror the two groups dirfer by less than 0.05 in the specified ability rHnge. It should be noted, however, that this item was extremely difficult for both groups and, hence, an inappropriate item for these groups. If the two ICCs were compared in the ability range for which this item is appropriate, a considerable di fference hetween the ICC's would be observed. For items of appropriate difficulty for at least one of the two groups of examinees (items with difficllity parameters in the ahility range of interest), it is 1101 possihle to obtain significant differclwes bel ween the item parameters for the two groups without a correspond- ing difference in the ICes. A more valid criticism of the comparison of ilem paramcters is that the distribution of the test statistic is known only asymptotically; furthermore, the asymplotic distribution is applicable only when item parameters are estimated in the presence or known ability parameters (Hambleton & Swaminllthan, 1985). It is not known how large the
/'1111'111;,,1/.1' Bia,ft't/ 'f1',{lllfnU 113 sample size must he in order for the asymptotic distribution to apply, and it is not known whether the asymptotic distribution applies when item and ability parameters arc estimated simultaneollsly. In addition to this problem. some evidence suggests that the chi-s4uarc statistic has a higher than expected raise-positive rate (McLaughlin & Drasgow, 19R7). Area Between Item Characteristic Clln·e.~. An alternative approach to the comparison of item characteristic functions is to compare the ICCs themselves rather than their parameters. If, after placing the parameter estimates on a common scale, the ICCs are identical, then the area between them should be zero (Rudner et a!., 19RO); when the area between ICCs is not zero, we can conclude that DIP is present. In computing the area, numerical procedures were used until recently. The numerical procedure involved (a) dividing the ahility range into k intervals, (b) constructing rectangles centered around the midpoint of each interval, (c) ohtaining the values of the ICCs (the probabilities) at the midpoint of eaeh interval, (d) taking the absolute value of the <Ii ffercnces between the probabilities, and (e) multiplying the difference hy the interval width and summing. Symbolically, this procedure may he expressed for item j as I.Ai = II',. (0) - Pj2 (0) I A 0 9~r The <Iuantity Ae is the width of the interval and is chosen to be as small as possible (e.g., 0.0 I). The values rand s indicate the abil ity range over which the area is to be calculated; the range is arbitrary and is chosen by the user. A typical choice for the ability range would be the range from three standard deviations below the lower group mean ability to three standard deviations above the upper group mean ahility. This choice ensures that the area i,~ calculated over the ahility range in which lIearly all eXlIlllillees fall. Raju (19RR) derived an exact expression ror COlllput ing the area hetweell the ICCs for the one-, two-, und three-parameter models. The expression for the three-parameter model is Area ( I
114 FUNDAMENTALS OF ITEM Rt:SI'ONSE TIIFORY For the tW<l-llnrameler model, the term involving (' disappears; for the one-parameter model, the expression reduces 10 the ahsolutc difference between the b-values for the two groups. In the expression for the area given ahove, the value of the (' paramo eter is assumed to be the same for both groups. Raju (19HH) has shown that when the (' parameters are not the same, the arell between the two curves is infinite if calculated over the entire range or' ability (-<X>, (0). For a finite range of ability, the area is finite; however, no expression has been derived for the area between ICes in 11 finite ability range, and so numerical methods must be used. Raju (1990) derived an expression for the standard error of the area statistic and suggested that the area statist it: divided hy its standard error can be taken as approximately normally distrilmted. This procedure is based on the assumption that the c parameter vallles are the sallie for the two groups and are fixed (Le., not estimated). When the c parameters for the two groups arc not the same, the signi ficance test for the area statistic cannot he carried 0111. The problem is to find a \"cut-off' vallie for the area statistk that can he used to decide whether 01 F is present. An empi ricalapproach to deterlllining a cHt-off is to divide the group with the hlrger sample size into two randomly equivalent groups, to estimate the ICCs in each group separately, and to determine the area between the estimated ICCs (Hamhleton & Rog- ers, 1(89). Since the groups arc fllndolllly cllllivait-lIt, thl' arell should be zero. Nonzero values of the area statistic are regarded as resultill~ from sampling fluctuations, and the largest area value obtained may he regarded as the largest value that may he expected in terms of sampling fluctuation. Any area value greater than this is assumed to he \"signifi- cant\" and, consequently, indicative of OTF when the majority and minority groups are compared. One disadvantage of the approach to establishing the cut-off value described above is that, as a result of halving the sample, the parameter estimates may be unstahle; consequently, the areH statist ic may not he a reliable indicator of 01F. An alternative approach is to use simulated data to establish the cut-off value (Rogers & Hambleton, 1989). In this approach, the two groups of interest are combined and parameters are estimated for the total group. The item parameter estimates for the total group and the ability parameter estimates for the majority group are used to generate a set of data of the same size as the majority group. Similarly, the item parameter estimates for the total group and the ability parameter estimates for the minority group are used to generate
1'011'111 iall\\, 1I;(I.\\('d Tr.'! /fen/.f 11.1 II sci of dala of Ihe same si:t.c as the minority group. The two sets of simulated data closciy resemble the data ror the majority alld minority groups ill lerms of sample siz.es, distrihutions of ahility, and item \\:haracteristics. The one difference is that the two sets of simulated data are based 011 the same item parameters und, hence, no DIF is present. Item and ability parameters then are estimated separately for each set of simulated data, and area statist ies are computed for each item. Since 110 011\" is present, nonzero area values are the result of sampling fluc.:tualions; as described ahove. the largest area value obtained in this comparison may be regarded as a cut-off value for lise in the comparison of ICCs for the real majority and minority groups. The empirical procedure described ahove for establishing critical vlllues Illay capitlllize on chance hecallse only one replication is per- formed. Mult iplc repl icat ions may he carried out and a cut-off vallie might he estahlished for each it(~lll; however, such a procedure would he so compnter-intcnsive as to he prohihitive. A problem cOl1lmon to the IRT approaches described ahove is that item parameters must be estimated in both groups. For proper estima- tion, a huge numher of ex.aminees with It large llhility range is needcd. In typical DIP sllulies, the numher of examinees in the minority group is usulllly small (around 3(0); furthermore, the group may have a restricted ahility range. Since item parameters will be estimated poorly in such situations, the 011' statistics may lead 10 erroneous decisions ahout the presence of DIF. Because of the problems associated with IRT methods for detecting DIF, alternative methods have heen sought. The mosl popular of the current non-IRT approaches for detecting DIF is the Mantel-llaenszel method (llolland & Thayer. 19B5). Unfortunately. this method is not sensitive to nonuniform DIF. More recently. Swarninathan and Rogers (1990) have provided a logistic regression procedure capable of detect- ing nonuniform, as well as uniform, DIF. The IRT approaches to the investigation of bias described in this chapter are illustrated using the New Mexico data set introduced in chapter 4. In this example the majority group is Anglo-American and the minority group is Native American. For the purposes of the example.
116 FUNDAMENTALS OF ITEM RESPONSE TlIEORY a random sample of 1,000 Anglo-Americans mId (I randolll saUl pie uf 1,000 Native Americans were drawn from the totHI sct of data. Three-parameter item response models were fillt~d separately to the item responses of each of the two groups. In computing the parameter estimates, the metric was fixed hy standardi7.ing the I, values. Since the two sets of data consisted of responses to the same items, slundllrdi7.ing the b values in each group automatically placed the item parameter estimates for the two groups on the same scale. Area statistics were computed for each item. Be(';Huse the (' values for the two groups were unequal for most items, the numerical method of calculating the area values was used. The 9 increment used in the calculations was 0.01. The area was calculated over the ability range from three standard deviations helow the lower group mean 0 to three standard deviations above the upper group mean 0; the resulting ability range was (-3.36, 3.55). Simulated data were used to determine the cut-off value, as desnihed earlier. Item responses were simulated for two groups for which no item had DIE To obtain parameter values for generating the item responses, the two groups were combined first and parameter estimates were computed for the total group (these parameter estimates are reported in the Appendix.). The ability estimates for the majority group and the item parameter estimates for the combined group were used then to simulate a set of data resembling the majority group; similarly, the ability estimates for the minority group and the item parameter estimates for the combined group were used to simulate a set of data resembling the minority group. Since the same item parameters were used to generate the data for the two groups, the simulated data represent the situation that has no DIE Three-parameter models were fitted separately to each set of simu- lated data. with the metric fix.ed by standardizing the b values in each group. Area statistics were computed for each item, and the largest area value obtained was used as the cut-off value in the Anglo-Native American comparison. The largest area vallie obtained in the simulated data comparison was 0.498. In comparing the item parameters for the two groups, two chi-square statistics were calculated. The first chi-square statistic, denoted as X~h' was based on only the (l and I, parameters for the two groups, while the second chi-square statistic, denoted as X~h(' was based on the 0,11, and c parameters. The second chi-square test was carried out primarily for
\"\"\"'\\ l'Olt'lII/IIlIy IJi!l.l\"·\" /I'M IINIH 117 TAIILE 11.1 tl('11\\ I'al ameler Fslilllah.'s, Area SlalislKS. alill ./. Values for Twcllly-l-'ivC Randomly ('hose\" TC~I Ilems M,I}O,ill' (;,.\"/\" Millorin' tim,,!, /)/f' Sld/H/1i J Ilrlll I), II, /'1 liz (12 ('2 Art·(/ X.2.,. • 2h Xii'\" I O.l!4(1 o ~7:'i 0.190 O.lIB IlX'J(, 0.170 0.417 5.X4 (d) 1 0.77.' 0.190 ··n.OO!! O.()()6 0.170 (UII!! 7.90 -' ().1 12 OAD 0.190 0170 n.W9· 21.1 J. 9.52 (l,W!! 0.190 -().!)5.~ fl.!!21 0.17f) 0.344 :'i.:l I 12.99 5 . U47 (I.flJ9 ().190 0.114 O17() (1.\\..11 17.110· 5.21 !! ().12.~ 0.714 0.190 n.2!!/) (),M5 !l.17() ()712· 2I.X(,\" 14.74 II (U19 1,()44 0.190 OJ01 0.170 04'!4 17.12· l'Ull\" (l,I')? 0551 n.DI OA05 29.1.1· 15.111 \" H.fll) I ().977 0.190 1.1.)99 0.170 O,2.lH J.S7 2.l07· ()Sltl 0.190 O.12!! 0.59.5 o 170 0.117 ;\",20 2.42 14 01()1I (1,52') 0.190 -ONW 0170 ().ll.I·]· II. 14 2.22 0.1')0 ().2!!6 O,~O7 n. 170 O.I()5 4.15 1),7H 16 ·0.19] IIA!!II -().I06 0.170 1.31 10 - 0 ..137 n.549 0.190 0,62!! O,X.l') 0115 O.1I4 14.74\" 4.64 11 1I.:'i 14 O.H,I!) 0.190 n.7lll () I7Il O.MI\" 11.62 1.7b 1.166 nAn 0.170 n.540· 12.0K Jfl 1.46' O.:'i!U o.l.n 1.175 O.2'JO ;,1] .111 -1.1611 0.661 11941 l.n'i4 1),1'It) 0.11:\"7 U.OQ ,11 I nIl O.4J I ().190 2.77!! 1).50!) 0.:!l5 0.56 1.050 0.190 0.140 ();X6 O.1J7 O.R!W· I.'J4 3.M 45 l.tWx 0.404 ().190 1.12!! 0.170 OSlb+ n.l; 46 04R I OMI5 O.2h; 0.;211 0.170 0.257 14.\"· .1I9 49 0.66:\\ 0.569 0.190 1.240 0.4'0 D.170 0467 :124\"+ 1642· 0.442 0.190 1545 1.201 n 170 0.942· 21.54+ ;0 (1.40') O..'l4() 0.190 0.497 (J.405 O.llO 0.M8· 1.19 0.640 0.190 1.IS4 O.4R9 0.170 (1.722· 1n.5 2 2.10 52 1.444 n.J17 0.190 0.3R7 0.531 (1.170 15.41\" 5.56 ;6 O.BS 0.190 -0.122 O.2XO 20.29· 15.07 ;7 n.2RI 0.190 -0.0(17 0.683 20.04· 60 n.,)04 0.190 0.5.14 1.223 n.S)· 15.24 64 0.245 0.562 6R ·1..198 7.1 0.567 - - _75 1.646 ...... •. ;d,.(X\" IH2 b. X'j,(M\" 16.27 \"SijtnifiC8l1f allhe 0.00 I level illustrative purposes. The significance level for each chi-square statistic was set al 0,001 to ensure that the overall Type I error rate was around 0.05. For the X~h' statistic, the critkal value was X1.001 16.27; for the xLoo,X~h statistic, the critical value was = 13.82. The item parameters for the two groups, the area statistics, and the chi-square values for 25 randomly chosen items are reported in Table 8.1. Of the 75 items analyzed altogether. the area statistic nagged 20 items as showing DIP, while tlte X~h stati... tic nagged 25 items. The X~h' statistic nagged only 9 items, which represented a subset of those
IIIl FUNDAMENTALS OF ITEM RESPONSE THEORY 0.9 0.8 P 0.7 r 0 0.8 - b a b 0.5 I I 0.4 I t 0.3 Y 0.2 0.1 0 -4 -3 -2 -1 0 1 2 3 .----. _ _ _ _ _ _ _ _.. _~_..._~b.!!'!Y______ Figure 8.1. Plot of ICCs for Majority and Minority Groups for Item 56 nagged by the X~\" statistic. As expected, the X~\"( statistic was more X;\"conservative than the and area statistics. The degree of agreement between the area and X~\" chi-square statis- tics was moderate: 77% of the items were classified in the same way (either showing DIF or not) by the two methods. The rank order correlation between the two methods was 0.71. Two examples of items flagged by both procedures are given in Figures R.1 and 8.2. These items differ in the type of DIF observed. In Figure R.I, the ICCs for the two groups are more or less parallel, differing mainly in their h parameters. This type of DIF is referred to as uniform TJIF; the difference in probabilities of success is uniform for the two groups over all ability levels. In Figure 8.2, the ICCs for the two groups cross; the probability of success is greater for the minority group thall for the majority group at the low end of the ability scale, but is greater for the majority group orat the high end of the ability scale. This tyP{~ DIF is referred to tiS nonuniform DIF, since the difference in probabilities is not uniform across ability levels. One of the advantages of IRT procedures for detecting DIF is their sensitivity to these different types of DIF; this feature is not shared by some of the popular non-IRT procedures
Potentially Biaud 1'1'.\\1 Ilnn.f 119 -~~-i'----------'-------\"\"\"\"~----=-:-r-, p r o b a b I I I t Y 0.1 o _ .• 1 o ___ ....~ j. __._.___l._ _ ..L...._••• 23 -<4 -3 -2 -1 _____._._ .______.__ ..Abl.I.I.t.L_ Illglire 8.2. Piol of Ices for Majorily and Minorily Groups ror hem 13 for detet:ting DlF (Holland & Thayer, 1988; Swaminalhan & Rogers, 1990). Of the 20 items flagged by the X~b statistic, 6 were not flagged by the area statistic; II of the items flagged by the area statistic were not flagged by the X~h statistic. Examination of the ICCs for the items inconsi~tently flagged revealed no reason ror the result. This finding demonstrates one of the problems of all methods for detecting DIF: while th~ agreement among methods is moderate, unexplainable differ- ences occur often. Summary Item response theory procedures for detecting DIF involve the com- parison of item chllfllcteristic functions for two grollps of interest. Two ways in which item characteristic functions may be compared are (a) by comparing their parameters or (b) by calculating the area between the t:urves. To compare item parameters ror two groups, a chi-square sta- tistic is computed. The statistic mayor may not include the c parameter;
120 FUNDAMENTALS 01 ITEM RESPONSE THEORY Ihe reason for 1I0t including I' is Ihal it is oneil poolly l'slimalcd lind. hence. is unreliabl(~. An advantage of till' chi-square slatislic is Ihat it has a known distrihution; a possihle disadvantage (If the procedure is that it may have a high false-positive rate. The area between lCCs can be computed using an exad expression when the c parameters for the two groups arc Ihe salny, ,a lid a signifi- cance test for the area is availahle in Ihis l'ase. When the c parameters are not the same, numerical procedures must be used to cakulate the nrea and no significance test is available. In this case an empirical \"cut-ofr' value must be obtained. This is done using either randomly equivalent samples or simulated data in which there is 110 DIE Area values are calculated for this comparison, and the largest value ohtained is used as the cut-off for the real analysis. Several other IRT approaches to the detection of DIF have not been described in this chapter. Linn and Harnisch (1981) suggested evaluat- ing the fit in the minority group of the rRT model obtained in the total group. The procedure is carried out by estimating item alld ability parameters for the combined majority-minorilY group; the item param- eter estimates and the abil ity parameter esl imates for the minority group obtained in the combined-group analysis are used to assess the fit of the model to the item response data for the minority group. If IlO DIF is found. the ICC obtained for the total group should fit the data for the minority group; if DIF is found. the parameters will not be invariant across the two groups and the model obtained for the total group will not fit the minority group. Goodness-of-fit statistics can be computed for each item to determine whether DIF is present. This procedure does not require the estimation of item parameters in the minority group (which is usually small) and, hence, overcollles some of the difficulties encountered in the two approaches described in this chapter. Another procedure, suggested by Linn et al. (1981). is to calculate the sum of squared differences between the ICCs for every observed value of O. This procedure may be modified 10 take into account the error in Ihe estimated probabilities (Shepard, Camilli, & Williams. 1985). Several comparisons of the effectiveness of IRT and non-IRT meth- ods for detecting DIF have been completed. The reader is referred to Mellenbergh (1989); Rudner et al. (1980); Shepard. Camilli. and Averill (1981); Shepard el al. (1985); and Subkoviak, Mack, lronson, and Craig ( 1984).
f'OIl'tltially Biaud Tot Items 12/ Exercises ror Chapter\" I. In an investigation of DlF. a onepararneter model was filled to the data srparlltcly for the minority 111lllll!ajority gro\"p~. Por a pankular item. thr difficulty parameter estimates amI the standard errors for the two groups were computed and are given in Tahle R.2. TABl.E 8.2 Majority Group Minority Group Difficulty Estimate: 0.34 0.89 Standard Error: 0.15 0.t6 a. Compute the variance of each difficulty estimate. b. Calculate the chi-square statistic for the difference between the di fri- cully estimates for the two groups. c. Does it appear that this item functions differentially in the two groups? 2. In carrying oul the DIP analysis on the New Me){ico data set, the ilem parameter estimates for Item 43 contained in Table 8.3 were obtained in the Anglo-American and Native American groups. TABI.E 8.3 (/ hc Group 0.93 0.90 0.2 0.42 1.112 0.2 Anglo-American Native American a. Using the formula given by Raju (19RR). calculate the areA between the ICCs for the two groups. E:otplain why this formulA can be used in Ihi!! situation. b. Using the cut-off value used in the eumple for the Anglo-American =versus comparison (cut-off value 0.468), determine if the item shows OW
122 FUNDAMENTALS OF ITEM ItESPONSE TIIF.()I{Y Answers Co I(xerciscs rllr Chnplcr R L n. For majority group: variance'\" SE2 = O. L'i1 0.0225 For minority group: variance = SE 2 =0.1 (/ = (J.()256 ·ih. For the one-parameter model, the stalistil.: siluplifies: 6.29 =c. Xt,o.oi 3.84. Since the calculated value exceeds the critical vallie, we can conclude that this item functions differentially in the two groups. I =-2. a. Area = (I 0.2) 2(0.42 0.93) x 1.7 x 0.42 x 0.93 In[l+exp 1.7x0.42xO.93x(1.82 Q.9Q1] (1.82-0.90)1 0.42 - 0.93 I 0.911== 0.8 x 1.536 In [I + exp (~1.20) J -- = 1.06 The formula can be used because the c values for the two groups are the same. b, The area value exceeds the cut-off value. We can conclude that Ihe item shows DIE
9 Test Score Equating Background The comparability of test scores across different tests measuring the same ability is an issue of considerable importance to test developers, measurement specialists, and test takers alike. If two examinees take different tests, how can their scores he compared? This question is particularly important when certification, selection, or pass-fail deci- sions must be made, since it should be a matter of indifference which test is used to make the decision. To compare the scores obtained on tests X and Y, a process of equating scores on the two tesls must be carried oul. Through this process a correspondence between the scores on X and Y is eSfablished, and the score on test X is converted to the metric of test Y. Thus, an examinee who obtains a score .x on test X has a converted score / on test Y; this score is comparable to the score y of an examinee taking test Y. In making pass-fail, selection, or certification decisions, the cUI-off y;score, Xc on test X can be converted to the score on test Y, and thil'! converted cut-off score may be ul'!ed 10 make the appropriate decision for examinees taking test Y. Classical Methods of Equating Classical methods of equating were described in detail by Angoff (1971) and Kolen (1988). In general, the methods fall into the two main categories: equipercentile equating and linear equating. Equipercentile equating is accomplished by considering the scores 011 tests X and Y to be equivalent if their respective percentile ranks in any given group are 123
124 FUNUAMENTALS OF ITEM RESI'ONSE TIIEORY equal. Strictly speaking, in order to equate scores on two lesls, Ihe lesls musl be given to Ihe same group of examinees. In praclice, Ihe process Iypically is carried oul by giving the lesls 10 randomly clJuivalelll groups of examinees. In lillear equating, it i.~ assumed that the .~core x Oil test X and the score yon lesl Yare linearly related, that is, ., y = ax + b The coefficienls II and b may be delermined using the relation.o;hips ~y = a~~ + b and where J.l~ (~y) and o~ (Oy) are means and standard deviations of the scores on tests X and Y, respectively. It follows that a = 0, and y Using the above expression, a score x may he placed on the metric of test Y. The ahove expression can he ohtained also hy equating the standard score on text X to the standard score on test Y, The assumption in linear equating is that the two test score distrihu- tions differ only with respect to their means and standard deviations. Therefore, standard scores will he equal in sllch cascs. Whcn this assumption is valid, linear equating hecomes a special case of eqlli-
fe.'/ Score f:q\"a/illg 125 perccntile equating; otherwise, it may be considered an approximation to eqlliperccntile equating. The linear equating method has many refinemenls. Procedures that take into account, for example, outliers and the unreliability of the test scores are given in Angoff (1971). Our purpose here is to descrihe hi iefly the classical equating procedures and to note the prohlems inherent in such approaches. Lord (1977, 19RO) has argued that in equating test scores, it should he a llIaller of indifference to the l'xallline(~s at l'very given ahility level whether they take test X or test Y. This notion of equity has several implications (Lord, 1977, 1980). I. Tests measuring different traits cannot he equated. 2. Raw scores on unequally reliahle tests cannot he equated (since otherwise a score from an unreliahle lesl can be equated to Ihe score on a reliable lesl, thus ohvialing the need for conslrtJ<:ling reliahle tesls!). 3. r~aw scores on tests with varying difficulty cannot he equated since Ihe II'sls will not he equally reliahle al different ahility levehi. 4. Falli!>le scores on tests X lind Y cannot he equaled unless Ihe tests are strictly parallel. 5. Perfectly reliahle tests can be equated. In addition to the above requirements of equity, two further COJ1(Ii- tions---sYlllmetry and invariance--lIIust be set for equating tests. The condition of symmetry dictates that equating should not depend on which test is used as the reference test. For example, if a regression procedure is used to determine the constants in the linear equating formula, the condition of symmetry will not he met because the regres- sion coefficients for predicting y from x are different from those for predicting x from y. The requirement of invariance is that the equating procedure he sample independent. These conditions, particularly those of equity, will usually not he met when using classical methods of equating. Theoretically, item response theory overcomes these problems. If the item response model fits the data, direct comparison of the ability parameters of two examinees who take different tests is made possihle hy the invariance property. Equat- ing of test scores is obviated in an item response theory framework; what must be ensured, however, is that the item and ahility parameter values hased on two tests are Oil a CO\"'IlIOII scale. Thus, in an item response theory framework, J('(J/il/~ rather thall equllting is lIecessary.
126 FUNDAMENTALS OF ITEM RESPONSE TIIEORY Nevertheless, because of the prevalence of the term eqllatitlN in the literature, the terms scaling and equotinN will be used interchangeably in this chapter. Scaling or Equating in Item Response TheQry According to item response theory, the ahility parameter 9 of an examinee is invariant across subsets of items. This means that, apart from measurement error, ability estimates also will be invariant across subsets of items. Hence, two examinees who respond to different sub- sets of items (or different tests) for which the item parameter values are known will have ability estimates that are on the same scale. No equating or scaling is necessary. When the item and ability estimates are unkllown, however. the situation changes. In this case, as explained in the chapter on esti- mation, 9 may be replaced by O' = (to + ~, b may be replaced by b' = (tb + (}. and 0 may be replaced hy o· =: 0 I (t without affecting the probability of a correct response. (For the one-parameter model, since a =: I. 9 need only he replaced by O· = 9 + ~, and b by b' = b + ~.) This invariance of the item response function with respect to linear transformations introduces an indeterminacy in the scale that must be removed before estimation ofparametcrs Clm proceed. One wily to remove the indeterminacy is to fix arbitrarily the scale of 9 (or b); in the two- and three-paramcter modeb, the ,,'01ll11l0n practice is to sct the mean and standard deviation of the 9 (or /J) values to be 0 lind I, respectively. For the one-parameter model, the mean of 9 (or h) is set to O. Fixing the scale on 9 or b is arbitrary and is dictated sometimes by the computer program used (in BleAL, for example, the mean of h is set to 0). This scaling must be undone when attempting to compare the item parameter values or the ability parameter values across two groups. To illustrate the procedures and principles underlying scaling, con- sider the situation in which one test is administered to two groups of examinees (as in studies of 01 \\1). Suppose also that the est imatioll of item and ability parameters is carried out s('l'mofcly for the two grouJls. During the estimation phase, it is necessary to fix the scale of the parameter estimates. The two possible wllys of fixing the scale arc (a) standardizing the item difficulty values, that is, fixing the mean and standard deviation of the difficulty values to be 0 and I, respectively; and (b) standardizing the ability values.
\\ 121 Pirst, consider the situation in which the scaling is done on the difficulty parameters. Since the sallie test is adlllinistered to the two groups, the item parameter estimates must he identical (except sampling fluctuations) if the model fits the <lata. Hence, scaling 011 the difficulty values will place the item parameter estimates and the ability estimates on the same scale. Suppose that the scaling is carried out on the ability values. Since the means and standard deviations of ability for the two groups of exami- nees usually will not be the same, standardizing on ability will result in item parameters that are on different scales. The item parameters will. nevertheless, be linearly related according to the linear relationship b\", = aba + f} where hA and 0A are the difficulty and discrimination parameter esti- mates in Group A, and bB and aB are the corresponding values in Group B. Once 0' and 11 are determined, the item parameter estimates in Group B may be placed on the same scale as the item parameter estimates for Group A. The more interesting problem is that of comparing the ability param- eters in Group A with those in Group B. Using the same relationship as nfor the II values above, all the ahility estimates Oil in Group may be placed on the same scale as those in Group A, using the linear relation- ship where e~ is the value of the parameter ell on the scale of Group A. The reverse situation to that described above is when one group of examinees takes two tests, X and Y. Since the ability paramefer of the examinees taking the two tests mllst be tIlt' salllC, st'ttill~ the IIU'UIl lind slandard deviation of the 9 to 0 and I, respectively, places the item parameters for the two tests on a {'ommon scale. If. however, the mean and standard deviation of the difficulty parameter values for each test are set to () and I, respectively, the ability parameter values in the two tests will differ by a linear transformation,
128 FUNDAMENTALS OF ITEM HESPONSE THEORY Oy ex ex + P The item parameters for tests X and Y mllst he placed on a common scalc using the following relationship: ., ax ay - ex These examples indicate that if it is necessary 10 compare examinees who take different tests, or if it is necessary to place items from differenl tests on a common scale, the equating study mllst be designed carefully. Clearly, if different groups of examinees take different tesls, IlO com- parison or equating is possible. Designs that permit \"linking\" of tests and comparison of examinees are discussed next. Linking Designs. In many situations. the interest is in placing the item parameter estimates from two or more tests on a common scale. This placement enables comparison of the difficulty levels of the tests and also facilitates the development of item banks (see Vale, 1986). The four linking designs that permit the scaling of item parameters (or their estimates) are the following: I. Single-Group Design. The two tests to be linked are given to the same group of examinees. This is a simple design, but it may be impractical to implement because testing time will be long. Moreover. practice and fatigue effects (if the two tests are administered one lifter the other) may have an effect on parameter estimation and, hence. on the linking results, 2. Equ;\"olent-Grollps Design, The two tests to be linked are given to equiv- alent but not identical groups of examinees. chosen randomly. This design is more practical and avoids practice and fatigue effects. 3. Am'hor-Tes( Design. The tests to be linked are given to two different groups of examinees. Each test has a set of common items (hal may be internal or external to the tests. This design is feasible and frequently used, and. if the anchor items are chosen properly (see, for example. Klein & Jarjoura, 1985), it avoids the problems in the single-group or equivalent- groups designs. 4. Common-Person Design. The two lests to be linked are given to two groups of examinees. with a common group of examinees taking both tests. Because the testing will he lengthy for .he common group, this design has the same drawbacks as the single-group design.
129 I In the single-group or equivalent-groups design, when~ olle group of :1 c'(illIlinces (or l'quivait'nt groups of examinees) takes the two tl'stS, the mcthods uescribed in the previous section may he lIsed to place the I itellls on the sume scale. In determining the sl:uling nlilstants in the equivalellt-groups design, matl:hed pairs of ahility vulues arc needed; this need presents It problem, because the groups I:ollsist of different examinces. One possible way to match examinees is 10 I'lmk order the examinees in the two groups and to consider examinees with the same rank to be equivalent. in the anchor test design, the parameters and, hence, their estimates (subject to sampling fluctuations) are .eluted linearly in the two tests, that is, where hyc .lOd lJxc are the difficulties of the common items embedded I in tests Y allli X, respt~ctively. Once thc I:Ollstants (J ami ~ are deter- mined. the item parameter estimates for all items in test X llIay be placed I 011 Ihe same scale as test Y. The adjusted itcm paramcter estimates for \\ the COlli mOil items ill test X will not he identical to Ihe corresponding ill~1ll parametcr estimates in test Y (hecausl' or estimation crrms) and. hence. Ihey should he averaged. or Ihe tll'signs desnibed nhove, the an!:lm.. lest desigll is the most feasihle. lIelll:c, determinalion of the sGlling nmslallis will be dis- clissed with rcference to this design. Determination of the Scalin~ Constants Thc llIelhods availahle for determining llie scaling c()nstanls fl. and r~ «,nly ~ wht'll the onc-parameter model is used) may be clllssified as follows: I Regression Method 2. Mean lind Sigma Method :I. I~ohust MCllll and Sigma Method 4. ChanlCteristic Curve Method
1]0 FUNDAMENTALS Of· ITEM RESPONSE TlII'.ORY R('~ rt' .,',\\';01/ Mt'lIwd. Ollel' pa irs of va 11I~:s 01 i lem pal \"meter l'st illlates in the two groups are obtained. a regression pl'Occdurc may he used to determine the line of hest fit through the points. The term e indicates the error in fitting the line, since 110' all the points will be exactly on the line, Here l}ye and I,xc are the item difficulty parameter estimates for the common items in tests Y 4Ind X, If common eXlIminees Me used. the eqn:1tion is where aYe and aXe are the ahility estimates for an examinee taking tests Y and X. respectively, BThe estimates &and of the regression coefficients are a1\\ = SYc and n1\\ /Jxc r ·~xc where r is the correlation coefficient hetween the estimates of the difficulty parameters ror the common items, /Jyc and \"xc are the respec- tive means, and SYc and \"'Xc lire the respective stamhlrd deviations. In the common-examinee design, these values are replaced by corresponding values ror the a estimates. The problem with the regression approach is that the condition or symmetry is not mel. This is true because the coefficients ror predicting hYe rrom hxe are different rrom those ror predicting IIX, rrom hYe and cannot be obtained by simply inverting the prediction equation P1\\ 1\\ bYe = a bxc + That is, it does not follow that 1\\ = Ilhxc bYe - --j\\--- n Thererore, the regression approach is not a suitable procedure ror determining the scaling constants.
131 Mean and SiJ:nlll Met/lOti. Since it follows that =- - Rnd Thus bYe 0. hxc + Jl and = -~ - bYe - 0. bxe Moreover, since \" i I :1 =bYe 0. \"Xc + 13 the transformation from bYe to bxc may be ohtained as =bxc bye - 13 --0.-- Hence, the symmetry requirement is satisfied by the mean and sigma method (when using the common-examinee design, the means and standard deviations of the corresponding 0 estimates arc used to deter- mine 0. and 13). Once (l and p are determined, the item parameter estimates for test X are placed on the same scale as test Y using the relationships b~ = 0. bx + P
1J2 HINIIAMENTA\\'S 01; I IFM I{ESI'ONSF 1'I11'ORY where b~ and a~ are the difficully and discrimiualion vnlUl~S of items in test X placed on the scale of test Y. The parameter estimates of the common items are averaged. since they will not he identicnl, as a result of estimation errors. For the one-parameter model. the item difficulty estimates for the common items are related as bYe = b xe + P that is, (l = I. It follows that =- -- + P h xe bYe and, hence, P = bYe - bxe Thus, the item difficulty estimates for test X are transformed by adding the difference in the mean difficulty levels of the common items. Robust Mean and Sigma Method_ In the mean and sigma method described above, no consideration was given to the fact that item parameters are estimated with varying degrees of accuracy (i.e., some difficulty estimates have larger standard errors than others). Linn et al. (1981) proposed the robust mean and sigma method to take into account the fact that the parameter estimates have different standurd errors. Each pair of values (bye;, bxci) for common item; in tests Y and X is weighted by the inverse of the larger of the variances of the two estimates. Pairs with large variances receive low weights, and pairs with small variances receive high weights. The variance of the difficulty parameter estimate is obtained by first inverting the information matrix (see chapter 3) and taking the appropriate diagonal element. For the three-parameter model the information matrix is of dimension 3 x 3, while for the one-parameter model it is of dimension I x I; that is, it has a single element.
In The steps in carrying out the robust mean and sigma method are summarized below: I. for each pair (hvc;./'xci), determine Ihe weighl. Wi as where ,,(bYei) and v(bXci ) are the variances of the estimates of the common items. 2. Scale the weights: w; = wilL• Wj JI where k is the number of common items in tests X and Y. 3. Compute the weishted estimates: ., hXci = Wi hXci 4. Determine the means Bnd standard deviations of the weighled item param- eler estimates. 5. Determine (l and p using the mellllS and sland!lrti deviations of the weighted estimates. Stocking and Lord (l983) have suggested that further improvement in determining a and Pmay be obtained if outliers are taken into account in the computation of the mean and standard deviation. The weights are made more robust by basing them on the perpendicular distances of points from the line bYe = abxc + fJ Starting with an initial value for a and ~. the process is repeated until the a and ~ values do not change. For details of this procedure. refer to Stocking and Lord (l983) or Hambleton and Swaminathan (1985).
134 fUNDAMENTALS OF ITEM RESPONSE THEORY Characteri.Hic Curve Method. The mean and sigma method (and its robust version) capitalizes on the relationship that I!ldsts between the difficulty parameters and ignores the relationship that exists between the discrimination parameters in determining the scaling constants. Haebara (1980) and Stocking and Lord (1983) have proposed the \"characteristic curve\" method, which takes into account. the informa- tion available from both the item difficulty and item discrimination parameters. The true score 'tXa of an examinee with ability 90 on the k common items in test X is « = L'tXa P(90 • bxei. 0Xci. eXei) i\", I Similarly. the true score 'tYa of an examinee with the same ability 90 on the k common items in test Y is l = L'tYa P(90 • bvci. aYe;' CYei) j= I For the set of common items, bYel = (l bxci + p and CYei :::: ('Xci The constants (l and p are chosen to minimize the function F where L~F =N 'tYlI)2 ('tXa a= I
TI',fl S(OI e f:qul1lin}l 135 alld N i!> the number of examinees. The function F is a function uf a and f\\ lind is an indication of the discrepancy between 'tXtl and 'ty\". The procedure for determining o. and I' is iterative, and details lire provided in Stocking and Lord (1983). In IIsing the anchor test design. the nUlllber of anchor items and. more important. their characteristics playa key role ill the quality of the linking. For example, if the anchor items are too easy for one group and too difficult for the other, the parameter estimates obtained in the two groups will be unstable. and the linking will be poor. It is important, therefore. that the common items be in an acceptable range of difficulty for the two groups. Empirical evidence suggests that best results are obtained if the common items are representative of the content of the two tests to be linked. In addition, it is important to ensure that the two groups of examinees are reasonably similar in their ability distributions, at least with respect to the common items. A rule of thumb for the number of anchor items is that the numher should be approximately 20% to 25% of the number of items in the tests. Other Linking and Equating I)rocedures With the andlor test design, concurrent calibration using the LOGIST computer program permits placing the item parameter eslimates and ability parameter estimates on a common scale without the need for a separate linking and scaling step. (A similar analysis with the one- parameter model can be carried out with the RIDA computer program.) The procedure is as follows: I. Treal Ihe dala all if (Nx + Ny) examinees have laken a lesl wilh (/Ix + /ly + \"8 ) ilems where n. denotes Ihe numher of allchor ilems. 2. Treal Ihe \"Y ilems to which Ihe Nx examinees did 1101 respond as \"noI reached\" items and code them as such. Similarly, code Ihe \"x items to which Ihe Ny examinees did nOI respond as \"nol reached.\" 1. Estimate Ihe item and ahility parameters. This procedure is simple to implement. Currently. lillie information exists regarding the accuracy of this procedure; further investigation of this issue is needed.
136 FUNDAMENTALS Of' ITEM RESPONSE THEORY In addition to the \"linking\" procedures dcsaihcd above, item re- sponse theory methods may be used to (a) eqnate trlle scores Oil two tests, and (b) equate two lests using \"observed score\" distrihutions generated for given levels of 9, These procedures are described in Lord (1980) and Hambleton and Swaminathan (1985). The reader is referred to these sources for more details. The steps in carrying out a linking are illustrated using two examples. In the first example. the linking procedure is illustrated in the context of developing an item bank. The second example deals with the problem of linking two tests. In both examples, the linking is carried out using an anchor test design. Example I Assume that a bank of test items that have been calibrated using the one-parameter model is availahle. The item difficulty esliulltles for the item bank are given in the Appendix. It is desirahle to add 10 the existing bank a set of 15 new, uncalibrated items. To add Ihese 15 items to the existing bank, we could lise the anchor lesl (Iesign with five anchor items chosen from the existing hank. Suppose that the II values for these five items are 1.65, 1.20, -O.RO, -1.25, and 2.50. These anchor items were chosen carefully to match the contcnt und, it was hoped, the difficulty levels of the 15 experimental items. Since the 15 items arc 1I1l1esled, however, it is difficull to assess their difficulty levels a priori. This information could be ohtained from a pilot tcst. In determining the scaling constants, the mean and sigma method is used for illustrative purposes bec,\\Use of its simplicity. The steps are as follows: I. The 20-item test (\\5 experimental and:; anch{lr items) is a(lministered 10 an appropriate sample of examinees (200 in this example). 2. An appropriate IRT model is chosen-·this IIllist he the same as .11l' IIHlclcl 011 which the existing itcm hank is used. Sinn'. for illustration. we have assumed that the itcms in the hank rit a nne· parameter 1ll()(leI, a one- pararneter model is filh:d also to the 20-itemtesl (with lIppropri\"lc dHTks 011 model-data fit) . .1. The mean diffit-ulty levd \"rlhe five IIl1chor item.~ i>y\" (from Ihe item hank. designated as tesl Y) based on Iheir known ill'llI p\"f<II11eler values is compuled; Ihe mean value is 0.66.
1~7 4. The :W·ilelll lest is calihraled. using (say) Ihe rom(lllier program BleAL. The mean of tilt' dirficullies hased on the 20 hem lest will be sel 10 lero in the estimlltioll process. The lIlean of the five '1Il~lor items that is pan of the 20·itel11 test is computed and designated as bxc• with a computed value of 0.25. 5. Since the ilem difficulties of the commoll items are related linearly according to Beta is calculated 8S bYe - hxc- (Note that u = I because the model used =is the one-parameter model.) In this example. p :::: 0.66 0.25 0041. 6. The item di fficulty estimates of the 15 experimental Hems are adjusled by =adding (bYe hXc) 0041 10 each difficulty estimate. 7. The com..!llon it~ms that are part of the experimental set are adjusted by adding (bYe \"Xc) toeach item difficulty value. Sinre the adjusted values will be different from the values fur the common ilems in Ihe item bank. the adjusted difficulty volues ore average.\\! with the difficulty values for the common items in the item bank. 8. The 15 experimental items an: on the same scale as the items in the item bank and are added to Ihe item hank. The eslimates for the common items 8re revised. These calculations are summarized in Table 9.1. The new items and their difficulty values have been added to the item bank (items 76 to 90) reported in the Appendix. Example 2 In this ex.ample, two proficiency tests, each with 15 items, were administered to samples of New Mexico high school students during two consecutive spring terms. Unfortunately, none of the test items came from the item bank in the Appendix. and. therefore, parameter estimates for all of the test items had to be obtained. It was desired to place the items in the test that was administered first on the same scale as the items in the test administered second. Hence. the test adminis- tered first is labeled as test X and the second as lest Y. An anchor lest design was used for the linking. The anchor test, with six. items. was conslructed to be representative in content of both lest X
us \"llNI)AMENTALS OF ITEM RESPONSE TIIEORY TABLE 9.1 Linking Procedure for Placing E)(perimenlal Items (Texl X) on the SlIme Scale as hems in all Hem Rank nest Y)\" Scaled Staled Test X Test Y Test X Test X Di//iculty Difficultv Oi//it'ulty lJi//irulty Item Cnmmon IterlU hx bye hx -t (hyc - hx;) , /Revi,Ud)b 1 1.29 1.65 1.70 1.67 2 0.75 1.20 I.Hi 1.18 3 -1.24 -0.80 -0.83 -0.82 4 -1.72 -1.25 -Ul -1.28 5 2.17 2.50 2.58 2.54 6 0.85 1.26 1.26 7 -1.88 -1.47 --1.47 8 -2.02 -1.61 -1.6t 9 0.19 0.60 10 0.22 0.60 II -1.116 0.63 0.63 12 -1.32 -1.4.'1 -1.4.'1 13 -1.10 -0.91 -0.91 14 0.74 -0.69 0.69 15 0.61 1.1.'1 16 0.50 1.02 1.15 t7 -0.80 0.91 1.02 18 1.70 -0.39 0.91 19 1.37 2.11 ·-0.39 20 1.55 1.78 2.11 1.96 1.78 1.-96 -...-~-~~- hxc = 0.25 =hyc 0.66 IlYe - hXc 0.41 ft. Common item~ ace in bold. b. Common item difficulties for tests X And Y have been Rve.a!!ed. and test Y. The tests were administered to 500 examinees on each occasion. In choosing an item response model, based on pilot studies, it was decided to use a three-parameter model with a fixed c value of 0.2. The item and ability parameters were estimated using the LOGIST computer program; in the estimation phase, the mean and standard deviation of e were set to be 0 and I, respectively.
139 In carrying out the linking, the mean and sigma method was used. primarily for pedagogical purposes. The mhllst mean al\\(I sigma or the characteristic curve methods are mOle appropriate hut are not used here hecallse of the nature of the comp\"tations involved. The steps ill carrying out the linking are as follows: I. Compute the mean and standard deviation of the difficulty estimates for the common items embedded in tests X and Y. 2. Determine the constants (l and p (since the three-parameter model was used). 3. Scale the difficulty estimates for test X by multiplying them by (l and adding p. 4. Average the difficulty values for the common items. S. Scale the discrimination parameter estimates for test X by dividing them by n. 6. Average the discrimination parameter values for the common items. The difficulty and discrimination parameter estimates fortest X are now on the same scale as those in test Y. The calculations are summarized in Tables 9.2 and 9.3. The constants a and f} can be used to place the ability values of the examinees taking tests X and Yon a common scale. Since 9y = aax + ~ :::: 0.95 Ox - O.IR, the mean ability or the examinees who took test X may be converted to a mean on test Y, had they taken it. enahling a comparison of the mean ahilities of the two groups even though they took different tests. For the group who took test X. the mean 9 value was set to zero in the estimation phase. Converting this mean to a mean on the scale of tcst Y. we obtain (ly = 0.95(0) - 0.18 -0.18 This implies that the difference in the mean abilities for the Iwo groups tak ing tests X and Y is -0.18; the group tuking Icsl X had a lower mean ahility than the grollp taking test Y. This inrormation could be used for academic or program evaluation purposes.
140 FUN()AMENTALS OF ITEM IUiSPONSE TIIEORY TABtE 9.2 Determination of Scaling Constants amI SCliled Difficulty for Tests X and Y· - - - - - - - - - - - - - - - - - - - _ . _.._--_._----- Tnt Y Tnt X Scalrd f)IJfi.~'lt). Itl'n! Difficliity Diffintlt.l' All Itml.< I 1.20 UO 1.20 2 1.75 -1.75 3 -0.80 2.10 -0.110 4 -1.28 2.75 5 1.35 -1.40 ··1.211 6 1.40 -1.65 1.35 7 1.20 0.60 1.40 8 0.50 1.81 1.20 9 0.72 2.20 0.50 10 -1.95 2.70 0.72 II -2.20 1.86 12 2.40 -1.95 13 1.80 -0.90 -2.20 14 1.45 -1.10 15 0.80 -2.30 2.40 16 1.10 1.80 17 1.85 0.511 1.45 18 2.30 0.92 0.80 19 -l.50 0.88 1.03 20 -1.110 1.92 1.8:1 21 0.40 2.10 2.36 22 2.52 -UI 23 1.60 -1.78 24 -1.20 25 O.4() 26 27 1.54 28 1.91 29 2.38 30 l.59 31 -1.04 32 -1.23 33 -2.37 34 0.31 35 0.69 36 0.66 1.64 \\,112 2.21 1.34 1.32 bYe = 0.39 hxc '\" 0.60 =(1 0.95 SYc 1.56 Sxc 1.65 11.= -0.18 a. Common ilems are in bold. b. Common hems .re averaged; .cal~d difficulty value. for leo! X = o.\"x + p.
141 TAULE 9.] Discrimination Values for Tests X alill ya '/\".fI y Tal X S,'\",,'\" IJ,.I<r;nr;,wlioll II,'''' /)isniml'llIt;on ni.v,.rimillati\"\" A/lII~m.vh I 1.02 1.01 121 2 1.21 ().90 .1 0.90 0.72 1.25 4 0.72 1.40 1.12 :'I 1.2:'1 n.75 0.92 6 lAO 0.62 0.52 7 1.12 1.98 R 0.75 1.90 1.62 9 0.92 1.01 0.95 10 0.62 1.22 1.911 II 0.52 O.til () 44 12 1.98 lI.til) !.6R IJ 1.90 1.95 1.On 14 1.62 1 70 n.ll:'i 1:'1 1.01 0.65 0.42 16 0.95 0.90 0.67 0.R4 11 1.23 1.15 11.79 IR 2.00 L116 1.29 19 0.611 0.55 IAl 1.111 20 0.45 0.40 I.IN 21 0.70 0.65 0.44 22 1.60 11 I.X5 24 I .I}() 25 I.ti2 26 OJ< I 27 n.62 2X 0.40 29 OM .10 {),l{O 11 075 .12 1.2.' .1.1 lSi '4 1.72 Vi l.l2 1.0._._.._--------_._ ..- - - - - - - - 042 _ _ __- - - - - - - - - - - - _.... a= 0,95 .._. .._ - - - - - - - a (\"ommnn items me in hold h Common itcUl'l: are a\"('m~(~(I; scaled d.~crimio,;uiun \\tt1hH.~s for I{\"~' X :':: flllX.
142 FUNDAMENTALS OF ITEM RESPONSE TIIEORY Summary Classical methods for equating have several shortcomings; most important, the condition of equity usually will not be met when using classical methods. Itcm response theory methods ohviate the need for equating because of the property of invariance of item and ability parameters. Because of the scaling that is needed 10 'eliminate the indeterminacy in item response models, item and ability parameters will be invariant only up to a linear transformation; that is. the item and ability parameters of the same items and same examinees will be related linearly in two groups. Oncc the linear relationship is identified, item parameter estimates and ability parameter estimates may be placed on a common scale. This procedure, known as linking or scaling. may be completed using several designs. The most important design is the anchor test design. where two tests containing a common set of items are administered to two different groups of examinees. Using the com- mon items and one of several methods. the coefficients of the linear transformation relating the item purameters for the two tests !.:an be determined. With knowledge of the linear transformation, the item and ability parameter estimates may be placed on a common scale. An excellent review of various designs for linking items to a common scale is provided by Vale (1986). Exercises ror Chapter 9 I. In DIF studies, the slime lest is administered to Iwo dirferenl groups anti the item parameters are estimaled separately. Bdore comparing the item parameters ror the two groups, Ihey mUSI he placed on the same scale. Explain how you would ensure that the ilem parameters are on a common scale. 2. Suppose that in an equating study two different tests are given to two dirrerent groups or examinees, with a common suhset or examinees taking both tests. Explain the procedure you would use to place the item and ability parameter estimates for the two tests and Ihe two groups on the same scale. 3. In Example I or chapter 9, it was assumed Ihat Ihe one·parameter model filled the data. a. Detennine the scaling conslants ror placing the experimental items on the Same scale as the items in the bank, assuming that a two-parameter model fits the data.
Search
Read the Text Version
- 1
- 2
- 3
- 4
- 5
- 6
- 7
- 8
- 9
- 10
- 11
- 12
- 13
- 14
- 15
- 16
- 17
- 18
- 19
- 20
- 21
- 22
- 23
- 24
- 25
- 26
- 27
- 28
- 29
- 30
- 31
- 32
- 33
- 34
- 35
- 36
- 37
- 38
- 39
- 40
- 41
- 42
- 43
- 44
- 45
- 46
- 47
- 48
- 49
- 50
- 51
- 52
- 53
- 54
- 55
- 56
- 57
- 58
- 59
- 60
- 61
- 62
- 63
- 64
- 65
- 66
- 67
- 68
- 69
- 70
- 71
- 72
- 73
- 74
- 75
- 76
- 77
- 78
- 79
- 80
- 81
- 82
- 83
- 84
- 85
- 86
- 87
- 88
- 89
- 90
- 91
- 92
- 93
- 94
- 95
- 96
- 97
- 98
- 99
- 100
- 101
- 102
- 103
- 104
- 105
- 106
- 107
- 108
- 109
- 110
- 111
- 112
- 113
- 114
- 115
- 116
- 117
- 118
- 119
- 120
- 121
- 122
- 123
- 124
- 125
- 126
- 127
- 128
- 129
- 130
- 131
- 132
- 133
- 134
- 135
- 136
- 137
- 138
- 139
- 140
- 141
- 142
- 143
- 144
- 145
- 146
- 147
- 148
- 149
- 150
- 151
- 152
- 153
- 154
- 155
- 156
- 157
- 158
- 159
- 160
- 161
- 162
- 163
- 164
- 165
- 166
- 167
- 168
- 169
- 170
- 171
- 172
- 173
- 174
- 175
- 176
- 177
- 178
- 179
- 180
- 181
- 182
- 183
- 184