Abstract
Traditionally, error in equating observed scores on two versions of a test is defined as the difference between the transformations that equate the quantiles of their distributions in the sample and population of test takers. But it is argued that if the goal of equating is to adjust the scores of test takers on one version of the test to make them indistinguishable from those on another, equating error should be defined as the degree to which the equated scores realize this goal. Two equivalent definitions of equating error based on this criterion are formulated. It is shown how these definitions allow one to evaluate such key quantities as the bias and mean squared error of any equating method if the tests fit a unidimensional response model. Several alternative applications of the ideas for the case in which the tests do not fit a unidimensional response model are discussed. Index terms: bias in equating; local equating; equating error; equipercentile equating; IRT observed-score equating; mean squared error of equating; observed-score equating.
Original language | Undefined |
---|---|
Pages (from-to) | 355-378 |
Number of pages | 22 |
Journal | Applied psychological measurement |
Volume | 30 |
Issue number | 5 |
DOIs | |
Publication status | Published - 2006 |
Keywords
- IR-60068
- METIS-235076