test equating
Post on 15-Jan-2016
99 Views
Preview:
DESCRIPTION
TRANSCRIPT
Test Equating
Zhang ZhonghuaChinese University of Hong Kong
Question ?• Two sets of Standardized Test which measure the same trait:
A and B. • A and B were administrated separately to two groups of
students (Group 1 and Group 2). Group 1 students only took Test A, and Group 2 students only took Test B.
• The mean score on Test A for Group 1 is 84. And the mean score on Test B for Group2 is 80. t-test result indicated that there was a statistically significant difference between the mean score for Group 1 and Group 2 (p<0.05).
• Then, should the conclusion that the Group 1 students were better than the Group 2 students on the trait that the two tests measured be gotten?
Why Equate?
• To compare test scores of different forms of tests (Strictly speaking, Parallel tests) which measure the same latent trait
• To construct the item bank/pool
• Computerized Adaptive Testing (CAT)
What’s Equating?
• “Equating is a statistical process that is used to adjust scores on test forms so that scores on the forms can be used interchangeably. Equating adjusts for differences in difficulty among forms that are built to be similar in difficulty and content” (Kolen & Brennan, 2004).
• The two alternate test forms for equating: Same content and statistical specification
• Equity• Symmetric• Group Invariance
• Lord’s Equity Property
Examinees with a given true score would have identical observed score means, standard deviations, and distributional shapes of converted scores on Form X and scores on Form Y.
• First-order Equity Property
Examinees with a given true score have the same means converted score on Form X as they have on Form Y.
Form Y Raw Form X1 Raw Form X2 Raw
1 2 4
2 3 5
. . .
. . .
13 14 16
14 15 17
15 16 18
16 17 19
17 18 20
18 19 21
… … …
Equating Design
• Single Group
• Random Groups
• Single Group with Counterbalance
• Anchored/Common-item Nonequivalent Group
• Preequating
• Single Group
Sample Form X Form Y
G1 √ √
• Single Group with Counterbalancing
Sample Time 1 Time 2
G1 Form X Form Y
G2 Form Y Form X
• Random Groups
Sample Form X Form Y
G1 √
G2 √
• Common-item Nonequivalent Groups
Sample Form X Form Y Common Items V
G1 √ √
G2 √ √
• Preequating
Precalibrated IRT Parameter Item Bank
Items form Bank(Operational items)
New Items(Non-Operational
Items)
Equating Methods
• Based on Classical Testing Theory (CTT)
• Based on Item Response Theory (IRT)
Downloadable Equating Procedures
• Equating/Linking Programs
http://www.education.uiowa.edu/casma/EquatingLinkingPrograms.htm
• IRT Scale Transformation Programs
http://www.education.uiowa.edu/casma/IRTPrograms.htm
Equating Methods Based on CTT
• Mean Equating
• Linear Equating
• Equipercentiel equating
CTT-Mean Equating
• In mean equating, Form X is considered to differ in difficulty from Form Y by the difference of the mean scores between the two forms.
• Example:
MX=70, MY=75.
Let Form X as the base Form, Form Y as the target Form.
For the score 80 on Form Y, the Equated Score on the scale of Form X is 80-(75-70)=75.
CTT-Linear Equating
• In Linear Equating, scores that are an equal distance from their means in standard deviation units are set equal.
( ) ( )
( ) ( )
x X y Y
X Y
( ) ( )( ) ( ) ( )
( ) ( )Y
Y Yl x x Y X
X X
CTT-Equipercentile
• For a given Form X score, find the percentage of examinees earning scores at or below that Form X score.
• Find the Form Y score that has the same percentage of examinees at or below it.
• The Form X and Form Y score are considered to be equivalent.
• Example: 70% of the examinees got a score 75 or below on Form X. 70% of the examinees got a score 80 or below on Form Y. Then a Form X score of 75 would be considered to represent
the same level of achievement as a Form Y score of 80.
Equating Methods Based on IRT
• IRT Parameters Equating
• IRT Observed Score and IRT Truce Score Equating
Item Response Theory
• Take IRT Three-Parameter Model as an example,
• Item parameters: Item Discrimination, Item Difficulty, Guessing
( )
( )( , , ) (1 )
1
i i
i i
Da b
j i ii i Da bi
eP cba cc
e
0.0
0.5
1.0
150 200 250 300
Pro
bab
ility
Item 1 Item 2
Scale Score
Difficulty
Item 1
Item 2
0.0
0.5
1.0
150 200 250 300
Pro
bab
ility
Item 1 Item 2
Scale Score
Difficulty
Item 1
Item 2
0.0
0.5
1.0
150 200 250 300
Pro
bab
ility
Item 1 Item 2
Scale Score
Difficulty
Item 1
Item 2
Item Parameter Equating
• Linking Separate Calibration (Mean/Mean Method, Mean/Sigma Method, Stocking-Lord Method, Haebara Method)
• Concurrent Calibration
• Fixed Common-Precalibrated Item Parameter Method
IRT-Linking Separate Calibration
,
,
,
exp ( )
(1 )
1 exp ( )
exp ( )(1 )
1 exp ( )
Ji Ii
Jj Ij
IjJj
Jj Ij
IjIi Ij
Ij IjIj
Ii Ij
Ij Ii Ii
Ij Ij
Ij Ii Ii
Let
A B
b Ab B
aa
Ac c
Then
aD A B Ab BA
c ca
D A B Ab BA
Da bc c
Da b
IRT-Moment Methods
• Mean/Mean Method
• Mean/Sigma Method
T
B
a
a
MA
M T
B T
B
ab b
a
MB M M
M
B
T
b
b
SA
S B
B T
T
bb b
b
SB M M
S
IRT-Characteristic Curve Method
• Stocking-Lord method:
• Haebara method:
2
1 1 1
1[ ( , , ) ( , , )]
N n njT
ST i jB jB jB i jT jTi j j
aF P a b c P Ab B c
N A
2
1 1
1[ ( , , ) ( , , )]
N njT
H i jB jB jB i jT jTi j
aF P a b c P Ab B c
N A
Example
• Take Form Y as the base test , Form X as the target Test
• Item 1 on Form X: Item Difficulty is 1.0; Item Discrimination is 1.896; Guessing is 0.18
• Equated item parameters for Item 1 on Form X onto the scale of Base Form Y can be computed as follows,
Stocking-Lord Haebara Mean/Mean Mean/Sigma
B -0.057 -0.063 -0.087 0.028
A 0.948 0.942 0.943 0.770
( ) ( )1 0.948 1.0 0.057=0.891Y
bXeq Ab B
( )1( )
1.8962.0
0.948Xa
Y
aeq
A
( ) ( )1 0.18cY Xeq c
IRT- Concurrent Calibration
• Concurrent calibration method involves estimating item and ability parameters simultaneously on a single computer run. In the procedure, the items that are not taken by one group of subjects are taken as not reached or missing data and the item parameters for all items on the two test forms are simultaneously estimated. This one estimation run makes the item parameters for all items from the two test forms put on the same scale (Kim & Hanson, 2002; Kim & Cohen, 1998).
• Example
Concurrent Calibration for Replication 16>COMMENTSHorizontal EquatingConcurrent Calibration for Replication 16>GLOBAL NPARM=3,DFNAME='D:\RESEARCH\REP16\CONH-16\CONH-16.DAT',SAVE;>SAVE PARM='D:\RESEARCH\REP16\CONH-16\CONH-16.PAR';>LENGTH NITEMS=140;>INPUT NTOTAL=80,SAMPLE=2000,NALT=4,NIDCH=4,FORMS=2;(4X,4A1,6X,I1,1X,80A1)>FORM1 LENGTH=80,ITEMS=(1(1)80);>FORM2 LENGTH=80,ITEMS=(1(1)20,81(1)140);>TEST ITEMS=(1(1)140),LINK=(1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0);>CALIB CYCLES=20;>SCORE;
IRT-Fixed Common-Item Parameters
• This procedure combines the features of concurrent calibration and linking separate calibration methods. In the method, the item parameters for the two test forms are estimated separately. What differs from linking separate calibration is that the common item parameters from the target test will be fixed at the estimated values from the base test.
• Example
Fixed Common Item Parameters for Replication 16>COMMENTSFCIP for Replication 16Target Test Form B with N (0,1)>GLOBAL NPARM=3,DFNAME='D:\RESEARCH\REP16\FIXV-16\B11-16.DAT',SAVE;>SAVE PARM='D:\RESEARCH\REP16\FIXV-16\FIXV-16.PAR';>LENGTH NITEMS=(80);>INPUT NTOTAL=80,SAMPLE=1000,NALT=4,NIDCH=4;(4A1,1X,80A1)>TEST ITEMS=(1(1)80);>CALIB TPRIOR,SPRIOR,GPRIOR,READPRI,CYCLES=20;>PRIORSTMU=(-0.639,1.041,1.701,0.482,-1.144,-0.023,0.616,1.133,0.668,0.577,-0.257,0.029,0.904,0.232,1.602,1.642,0.537,-0.228,1.439,0.517,0.0(0)60), TSIGMA=(0.001,0.001,0.001,0.001,0.001,0.001,0.001,0.001,0.001,0.001,0.001,0.001,0.001,0.001,0.001,0.001,0.001,0.001,0.001,0.001,2.0(0)60), SMU=(-0.688,0.011,-0.810,0.614,-0.811,-0.445,-0.142,-0.387,0.292,-0.449,0.040,-0.522,0.080,0.660,0.301,0.408,-0.689,-0.079,0.294,-0.174,0.0(0)60), SSIGMA=(0.001,0.001,0.001,0.001,0.001,0.001,0.001,0.001,0.001,0.001,0.001,0.001,0.001,0.001,0.001,0.001,0.001,0.001,0.001,0.001,0.5(0)60), ALPHA=(2882.990,1329.080,3540.010,4092.470,2694.080,2652.900,2314.660,2532.500,2336.870,3725.870,2364.700,2545.460,2358.110,2307.760,3583.990,3117.190,2569.460,1817.030,1057.210,2544.350,6(0)60), BETA=(7119.010,8672.920,6461.990,5909.530,7307.920,7349.100,7687.340,7469.500,7665.130,6276.130,7637.300,7456.540,7643.890,7694.240,6418.010,6884.810,7432.540,8184.970,8944.790,7457.650,16(0)60); >SCORE;
Comparison of Different Equating Methods
• No agreements have been gotten • Methods based on CTT can be used to equate tests.
Methods based on IRT are essential to construct item bank/pool.
• Among the methods based on IRT, some researches indicated that Concurrent Calibration Method could produce more accurate equating results than that of Linking Separate Calibration Method and FCIP method.
Thank You Very Much!
top related