%0 Journal Article
%J Personality and Individual Differences
%D 2010
%T Detection of aberrant item score patterns in computerized adaptive testing: An empirical example using the CUSUM
%A Egberink, I. J. L.
%A Meijer, R. R.
%A Veldkamp, B. P.
%A Schakel, L.
%A Smid, N. G.
%K CAT
%K computerized adaptive testing
%K CUSUM approach
%K person Fit
%X The scalability of individual trait scores on a computerized adaptive test (CAT) was assessed through investigating the consistency of individual item score patterns. A sample of N = 428 persons completed a personality CAT as part of a career development procedure. To detect inconsistent item score patterns, we used a cumulative sum (CUSUM) procedure. Combined information from the CUSUM, other personality measures, and interviews showed that similar estimated trait values may have a different interpretation.Implications for computer-based assessment are discussed.
%B Personality and Individual Differences
%V 48
%P 921-925
%@ 01918869
%G eng
%0 Journal Article
%J Journal of Educational Measurement
%D 2004
%T Using patterns of summed scores in paper-and-pencil tests and computer-adaptive tests to detect misfitting item score patterns
%A Meijer, R. R.
%K Computer Assisted Testing
%K Item Response Theory
%K person Fit
%K Test Scores
%X Two new methods have been proposed to determine unexpected sum scores on subtests (testlets) both for paper-and-pencil tests and computer adaptive tests. A method based on a conservative bound using the hypergeometric distribution, denoted ρ, was compared with a method where the probability for each score combination was calculated using a highest density region (HDR). Furthermore, these methods were compared with the standardized log-likelihood statistic with and without a correction for the estimated latent trait value (denoted as l-super(*)-sub(z) and l-sub(z), respectively). Data were simulated on the basis of the one-parameter logistic model, and both parametric and nonparametric logistic regression was used to obtain estimates of the latent trait. Results showed that it is important to take the trait level into account when comparing subtest scores. In a nonparametric item response theory (IRT) context, on adapted version of the HDR method was a powerful alterative to ρ. In a parametric IRT context, results showed that l-super(*)-sub(z) had the highest power when the data were simulated conditionally on the estimated latent trait level. (PsycINFO Database Record (c) 2005 APA ) (journal abstract)
%B Journal of Educational Measurement
%V 41
%P 119-136
%G eng
%0 Journal Article
%J Journal of Educational Measurement
%D 2002
%T Outlier detection in high-stakes certification testing
%A Meijer, R. R.
%K Adaptive Testing
%K computerized adaptive testing
%K Educational Measurement
%K Goodness of Fit
%K Item Analysis (Statistical)
%K Item Response Theory
%K person Fit
%K Statistical Estimation
%K Statistical Power
%K Test Scores
%X Discusses recent developments of person-fit analysis in computerized adaptive testing (CAT). Methods from statistical process control are presented that have been proposed to classify an item score pattern as fitting or misfitting the underlying item response theory model in CAT Most person-fit research in CAT is restricted to simulated data. In this study, empirical data from a certification test were used. Alternatives are discussed to generate norms so that bounds can be determined to classify an item score pattern as fitting or misfitting. Using bounds determined from a sample of a high-stakes certification test, the empirical analysis showed that different types of misfit can be distinguished Further applications using statistical process control methods to detect misfitting item score patterns are discussed. (PsycINFO Database Record (c) 2005 APA )
%B Journal of Educational Measurement
%V 39
%P 219-233
%G eng
%0 Journal Article
%J Psychologie Française
%D 2001
%T Nouveaux développements dans le domaine du testing informatisé [New developments in the area of computerized testing]
%A Meijer, R. R.
%A Grégoire, J.
%K Adaptive Testing
%K Computer Applications
%K Computer Assisted
%K Diagnosis
%K Psychological Assessment computerized adaptive testing
%X L'usage de l'évaluation assistée par ordinateur s'est fortement développé depuis la première formulation de ses principes de base dans les années soixante et soixante-dix. Cet article offre une introduction aux derniers développements dans le domaine de l'évaluation assistée par ordinateur, en particulier celui du testing adaptative informatisée (TAI). L'estimation de l'aptitude, la sélection des items et le développement d'une base d'items dans le cas du TAI sont discutés. De plus, des exemples d'utilisations innovantes de l'ordinateur dans des systèmes intégrés de testing et de testing via Internet sont présentés. L'article se termine par quelques illustrations de nouvelles applications du testing informatisé et des suggestions pour des recherches futures.Discusses the latest developments in computerized psychological assessment, with emphasis on computerized adaptive testing (CAT). Ability estimation, item selection, and item pool development in CAT are described. Examples of some innovative approaches to CAT are presented. (PsycINFO Database Record (c) 2005 APA )
%B Psychologie Française
%V 46
%P 221-230
%G eng
%0 Journal Article
%J Applied Psychological Measurement
%D 1999
%T Computerized Adaptive Testing: Overview and Introduction
%A Meijer, R. R.
%A Nering, M. L.
%K computerized adaptive testing
%X Use of computerized adaptive testing (CAT) has increased substantially since it was first formulated in the 1970s. This paper provides an overview of CAT and introduces the contributions to this Special Issue. The elements of CAT discussed here include item selection procedures, estimation of the latent trait, item exposure, measurement precision, and item bank development. Some topics for future research are also presented.
%B Applied Psychological Measurement
%V 23
%P 187-94
%G eng
%M EJ596304