00576nas a2200109 4500008004200000245016600042210006900208100002400277700002000301700001900321856012600340 In Press d 00aMeasurement efficiency for fixed-precision multidimensional computerized adaptive tests: Comparing health measurement and educational testing using example banks0 aMeasurement efficiency for fixedprecision multidimensional compu1 aPaap, Muirne, C. S.1 aBorn, Sebastian1 aBraeken, Johan uhttp://www.iacat.org/measurement-efficiency-fixed-precision-multidimensional-computerized-adaptive-tests-comparing-health01729nas a2200145 4500008003900000245016600039210006900205300001000274490000700284520118400291100002401475700002001499700001901519856004501538 2019 d00aMeasurement Efficiency for Fixed-Precision Multidimensional Computerized Adaptive Tests: Comparing Health Measurement and Educational Testing Using Example Banks0 aMeasurement Efficiency for FixedPrecision Multidimensional Compu a68-830 v433 aIt is currently not entirely clear to what degree the research on multidimensional computerized adaptive testing (CAT) conducted in the field of educational testing can be generalized to fields such as health assessment, where CAT design factors differ considerably from those typically used in educational testing. In this study, the impact of a number of important design factors on CAT performance is systematically evaluated, using realistic example item banks for two main scenarios: health assessment (polytomous items, small to medium item bank sizes, high discrimination parameters) and educational testing (dichotomous items, large item banks, small- to medium-sized discrimination parameters). Measurement efficiency is evaluated for both between-item multidimensional CATs and separate unidimensional CATs for each latent dimension. In this study, we focus on fixed-precision (variable-length) CATs because it is both feasible and desirable in health settings, but so far most research regarding CAT has focused on fixed-length testing. This study shows that the benefits associated with fixed-precision multidimensional CAT hold under a wide variety of circumstances.1 aPaap, Muirne, C. S.1 aBorn, Sebastian1 aBraeken, Johan uhttps://doi.org/10.1177/014662161876571901267nas a2200157 4500008003900000245009900039210006900138300001200207490000700219520076400226100001800990700001501008700001801023700002301041856004501064 2018 d00aA Continuous a-Stratification Index for Item Exposure Control in Computerized Adaptive Testing0 aContinuous aStratification Index for Item Exposure Control in Co a523-5370 v423 aThe method of a-stratification aims to reduce item overexposure in computerized adaptive testing, as items that are administered at very high rates may threaten the validity of test scores. In existing methods of a-stratification, the item bank is partitioned into a fixed number of nonoverlapping strata according to the items’a, or discrimination, parameters. This article introduces a continuous a-stratification index which incorporates exposure control into the item selection index itself and thus eliminates the need for fixed discrete strata. The new continuous a-stratification index is compared with existing stratification methods via simulation studies in terms of ability estimation bias, mean squared error, and control of item exposure rates.1 aHuebner, Alan1 aWang, Chun1 aDaly, Bridget1 aPinkelman, Colleen uhttps://doi.org/10.1177/014662161875828900564nas a2200145 4500008003900000245008100039210006900120100002400189700002200213700001600235700001900251700002300270700002500293856010000318 2018 d00aMeasuring patient-reported outcomes adaptively: Multidimensionality matters!0 aMeasuring patientreported outcomes adaptively Multidimensionalit1 aPaap, Muirne, C. S.1 aKroeze, Karel, A.1 aGlas, C A W1 aTerwee, C., B.1 avan der Palen, Job1 aVeldkamp, Bernard, P uhttp://www.iacat.org/measuring-patient-reported-outcomes-adaptively-multidimensionality-matters00949nas a2200169 4500008003900000022001400039245011500053210006900168260000800237300001600245490000700261520040300268100001700671700002400688700002100712856004600733 2018 d a1573-264900aSome recommendations for developing multidimensional computerized adaptive tests for patient-reported outcomes0 aSome recommendations for developing multidimensional computerize cApr a1055–10630 v273 aMultidimensional item response theory and computerized adaptive testing (CAT) are increasingly used in mental health, quality of life (QoL), and patient-reported outcome measurement. Although multidimensional assessment techniques hold promises, they are more challenging in their application than unidimensional ones. The authors comment on minimal standards when developing multidimensional CATs.1 aSmits, Niels1 aPaap, Muirne, C. S.1 aBöhnke, Jan, R. uhttps://doi.org/10.1007/s11136-018-1821-802709nas a2200157 4500008004100000245007100041210006900112260005500181520213800236653000802374653002002382653001402402100002002416700002602436856008902462 2017 eng d00aAnalysis of CAT Precision Depending on Parameters of the Item Pool0 aAnalysis of CAT Precision Depending on Parameters of the Item Po aNiigata, JapanbNiigata Seiryo Universityc08/20173 a
The purpose of this research project is to analyze the measurement precision of a latent variable depending on parameters of the item pool. The influence of the following factors is analyzed:
Factor A – range of variation of items in the pool. This factor varies on three levels with the following ranges in logits: a1 – [-3.0; +3.0], a2 - [-4.0; +4.0], a3 - [-5.0; +5.0].
Factor B – number of items in the pool. The factor varies on six levels with the following number of items for every factor: b1 - 128, b2 - 256, b3 – 512, b4 - 1024, b5 – 2048, b6 – 4096. The items are evenly distributed in each of the variation ranges.
Factor C – examinees’ proficiency varies at 30 levels (c1, c2, …, c30), which are evenly distributed in the range [-3.0; +3.0] logit.
The investigation was based on a simulation experiment within the framework of the theory of latent variables.
Response Y is the precision of measurement of examinees’ proficiency, which is calculated as the difference between the true levels of examinees’ proficiency and estimates obtained by means of adaptive testing. Three factor ANOVA was used for data processing.
The following results were obtained:
1. Factor A is significant. Ceteris paribus, the greater the range of variation of items in the pool, the higher the estimation precision is.
2. Factor B is significant. Ceteris paribus, the greater the number of items in the pool, the higher the estimation precision is.
3. Factor C is statistically insignificant at level α = .05. It means that the precision of estimation of examinees’ proficiency is the same within the range of their variation.
4. The only significant interaction among all interactions is AB. The significance of this interaction is explained by the fact that increasing the number of items in the pool decreases the effect of the range of variation of items in the pool.
10aCAT10aItem parameters10aPrecision1 aMaslak, Anatoly1 aPozdniakov, Stanislav uhttps://drive.google.com/file/d/1Bwe58kOQRgCSbB8x6OdZTDK4OIm3LQI3/view?usp=drive_web01179nas a2200157 4500008003900000245008400039210006900123300001200192490000700204520068700211100002000898700001600918700001800934700002200952856004700974 2017 d00aThe Development of MST Test Information for the Prediction of Test Performances0 aDevelopment of MST Test Information for the Prediction of Test P a570-5860 v773 aThe current study proposes novel methods to predict multistage testing (MST) performance without conducting simulations. This method, called MST test information, is based on analytic derivation of standard errors of ability estimates across theta levels. We compared standard errors derived analytically to the simulation results to demonstrate the validity of the proposed method in both measurement precision and classification accuracy. The results indicate that the MST test information effectively predicted the performance of MST. In addition, the results of the current study highlighted the relationship among the test construction, MST design factors, and MST performance.1 aPark, Ryoungsun1 aKim, Jiseon1 aChung, Hyewon1 aDodd, Barbara, G. uhttp://dx.doi.org/10.1177/001316441666296000592nas a2200169 4500008003900000022001400039245011200053210006900165300001600234490000700250100002400257700002200281700002500303700002300328700002500351856004600376 2017 d a1573-264900aItem usage in a multidimensional computerized adaptive test (MCAT) measuring health-related quality of life0 aItem usage in a multidimensional computerized adaptive test MCAT a2909–29180 v261 aPaap, Muirne, C. S.1 aKroeze, Karel, A.1 aTerwee, Caroline, B.1 avan der Palen, Job1 aVeldkamp, Bernard, P uhttps://doi.org/10.1007/s11136-017-1624-303808nas a2200169 4500008004100000245008600041210006900127260005500196520318900251653000903440653000803449653002503457100002503482700001703507700001703524856009703541 2017 eng d00aNew Results on Bias in Estimates due to Discontinue Rules in Intelligence Testing0 aNew Results on Bias in Estimates due to Discontinue Rules in Int aNiigata, JapanbNiigata Seiryo Universityc08/20173 aThe presentation provides new results on a form of adaptive testing that is used frequently in intelligence testing. In these tests, items are presented in order of increasing difficulty, and the presentation of items is adaptive in the sense that each subtest session is discontinued once a test taker produces a certain number of incorrect responses in sequence. The subsequent (not observed) responses are commonly scored as wrong for that subtest, even though the test taker has not seen these. Discontinuation rules allow a certain form of adaptiveness both in paper-based and computerbased testing, and help reducing testing time.
Two lines of research that are relevant are studies that directly assess the impact of discontinuation rules, and studies that more broadly look at the impact of scoring rules on test results with a large number of not administered or not reached items. He & Wolf (2012) compared different ability estimation methods for this type of discontinuation rule adaptation of test length in a simulation study. However, to our knowledge there has been no rigorous analytical study of the underlying distributional changes of the response variables under discontinuation rules. It is important to point out that the results obtained by He & Wolf (2012) agree with results presented by, for example, DeAyala, Plake & Impara (2001) as well as Rose, von Davier & Xu (2010) and Rose, von Davier & Nagengast (2016) in that ability estimates are biased most when scoring the not observed responses as wrong. Discontinuation rules combined with scoring the non-administered items as wrong is used operationally in several major intelligence tests, so more research is needed in order to improve this particular type of adaptiveness in the testing practice.
The presentation extends existing research on adaptiveness by discontinue-rules in intelligence tests in multiple ways: First, a rigorous analytical study of the distributional properties of discontinue-rule scored items is presented. Second, an extended simulation is presented that includes additional alternative scoring rules as well as bias-corrected ability estimators that may be suitable to improve results for discontinue-rule scored intelligence tests.
References: DeAyala, R. J., Plake, B. S., & Impara, J. C. (2001). The impact of omitted responses on the accuracy of ability estimation in item response theory. Journal of Educational Measurement, 38, 213-234.
He, W. & Wolfe, E. W. (2012). Treatment of Not-Administered Items on Individually Administered Intelligence Tests. Educational and Psychological Measurement, Vol 72, Issue 5, pp. 808 – 826. DOI: 10.1177/0013164412441937
Rose, N., von Davier, M., & Xu, X. (2010). Modeling non-ignorable missing data with item response theory (IRT; ETS RR-10-11). Princeton, NJ: Educational Testing Service.
Rose, N., von Davier, M., & Nagengast, B. (2016) Modeling omitted and not-reached items in irt models. Psychometrika. doi:10.1007/s11336-016-9544-7
10aBias10aCAT10aIntelligence Testing1 avon Davier, Matthias1 aCho, Youngmi1 aPan, Tianshu uhttp://www.iacat.org/new-results-bias-estimates-due-discontinue-rules-intelligence-testing-000491nas a2200133 4500008003900000022001400039245010800053210006900161300000700230490000600237100002500243700002400268856006500292 2017 d a2504-284X00aRobust Automated Test Assembly for Testlet-Based Tests: An Illustration with Analytical Reasoning Items0 aRobust Automated Test Assembly for TestletBased Tests An Illustr a630 v21 aVeldkamp, Bernard, P1 aPaap, Muirne, C. S. uhttps://www.frontiersin.org/article/10.3389/feduc.2017.0006301780nas a2200145 4500008003900000245008400039210006900123300001200192490000700204520131700211100001601528700002301544700001401567856005301581 2015 d00aa-Stratified Computerized Adaptive Testing in the Presence of Calibration Error0 aaStratified Computerized Adaptive Testing in the Presence of Cal a260-2830 v753 aa-Stratified computerized adaptive testing with b-blocking (AST), as an alternative to the widely used maximum Fisher information (MFI) item selection method, can effectively balance item pool usage while providing accurate latent trait estimates in computerized adaptive testing (CAT). However, previous comparisons of these methods have treated item parameter estimates as if they are the true population parameter values. Consequently, capitalization on chance may occur. In this article, we examined the performance of the AST method under more realistic conditions where item parameter estimates instead of true parameter values are used in the CAT. Its performance was compared against that of the MFI method when the latter is used in conjunction with Sympson–Hetter or randomesque exposure control. Results indicate that the MFI method, even when combined with exposure control, is susceptible to capitalization on chance. This is particularly true when the calibration sample size is small. On the other hand, AST is more robust to capitalization on chance. Consistent with previous investigations using true item parameter values, AST yields much more balanced item pool usage, with a small loss in the precision of latent trait estimates. The loss is negligible when the test is as long as 40 items.1 aCheng, Ying1 aPatton, Jeffrey, M1 aShao, Can uhttp://epm.sagepub.com/content/75/2/260.abstract01425nas a2200157 4500008003900000245009200039210006900131300001200200490000700212520092000219100002001139700001601159700001801175700002101193856005301214 2014 d00aEnhancing Pool Utilization in Constructing the Multistage Test Using Mixed-Format Tests0 aEnhancing Pool Utilization in Constructing the Multistage Test U a268-2800 v383 aThis study investigated a new pool utilization method of constructing multistage tests (MST) using the mixed-format test based on the generalized partial credit model (GPCM). MST simulations of a classification test were performed to evaluate the MST design. A linear programming (LP) model was applied to perform MST reassemblies based on the initial MST construction. Three subsequent MST reassemblies were performed. For each reassembly, three test unit replacement ratios (TRRs; 0.22, 0.44, and 0.66) were investigated. The conditions of the three passing rates (30%, 50%, and 70%) were also considered in the classification testing. The results demonstrated that various MST reassembly conditions increased the overall pool utilization rates, while maintaining the desired MST construction. All MST testing conditions performed equally well in terms of the precision of the classification decision.
1 aPark, Ryoungsun1 aKim, Jiseon1 aChung, Hyewon1 aDodd, Barbara, G uhttp://apm.sagepub.com/content/38/4/268.abstract01548nas a2200145 4500008003900000245008200039210006900121300001200190490000700202520108300209100001501292700002001307700002201327856005301349 2014 d00aUsing Multidimensional CAT to Administer a Short, Yet Precise, Screening Test0 aUsing Multidimensional CAT to Administer a Short Yet Precise Scr a614-6310 v383 aMultidimensional computerized adaptive testing (MCAT) provides a mechanism by which the simultaneous goals of accurate prediction and minimal testing time for a screening test could both be met. This article demonstrates the use of MCAT to administer a screening test for the Computerized Adaptive Testing–Armed Services Vocational Aptitude Battery (CAT-ASVAB) under a variety of manipulated conditions. CAT-ASVAB is a test battery administered via unidimensional CAT (UCAT) that is used to qualify applicants for entry into the U.S. military and assign them to jobs. The primary research question being evaluated is whether the use of MCAT to administer a screening test can lead to significant reductions in testing time from the full-length selection test, without significant losses in score precision. Different stopping rules, item selection methods, content constraints, time constraints, and population distributions for the MCAT administration are evaluated through simulation, and compared with results from a regular full-length UCAT administration.
1 aYao, Lihua1 aPommerich, Mary1 aSegall, Daniel, O uhttp://apm.sagepub.com/content/38/8/614.abstract01409nas a2200157 4500008003900000245009300039210006900132300001000201490000700211520091300218100002301131700001501154700001701169700001301186856005201199 2013 d00aThe Influence of Item Calibration Error on Variable-Length Computerized Adaptive Testing0 aInfluence of Item Calibration Error on VariableLength Computeriz a24-400 v373 aVariable-length computerized adaptive testing (VL-CAT) allows both items and test length to be “tailored” to examinees, thereby achieving the measurement goal (e.g., scoring precision or classification) with as few items as possible. Several popular test termination rules depend on the standard error of the ability estimate, which in turn depends on the item parameter values. However, items are chosen on the basis of their parameter estimates, and capitalization on chance may occur. In this article, the authors investigated the effects of capitalization on chance on test length and classification accuracy in several VL-CAT simulations. The results confirm that capitalization on chance occurs in VL-CAT and has complex effects on test length, ability estimation, and classification accuracy. These results have important implications for the design and implementation of VL-CATs.
1 aPatton, Jeffrey, M1 aYing Cheng1 aYuan, Ke-Hai1 aDiao, Qi uhttp://apm.sagepub.com/content/37/1/24.abstract01220nas a2200133 4500008003900000022001400039245003600053210003600089300001400125490000700139520088200146100001701028856004101045 2013 d a1745-398400aLongitudinal Multistage Testing0 aLongitudinal Multistage Testing a447–4680 v503 aThis article introduces longitudinal multistage testing (lMST), a special form of multistage testing (MST), as a method for adaptive testing in longitudinal large-scale studies. In lMST designs, test forms of different difficulty levels are used, whereas the values on a pretest determine the routing to these test forms. Since lMST allows for testing in paper and pencil mode, lMST may represent an alternative to conventional testing (CT) in assessments for which other adaptive testing designs are not applicable. In this article the performance of lMST is compared to CT in terms of test targeting as well as bias and efficiency of ability and change estimates. Using a simulation study, the effect of the stability of ability across waves, the difficulty level of the different test forms, and the number of link items between the test forms were investigated.
1 aPohl, Steffi uhttp://dx.doi.org/10.1111/jedm.1202800541nas a2200181 4500008004500000245006300045210006300108300001400171490000700185100002300192700002000215700002200235700001700257700001600274700001800290700002100308856003000329 2012 Engldsh 00aDevelopment of a computerized adaptive test for depression0 aDevelopment of a computerized adaptive test for depression a1105-11120 v691 aGibbons, Robert, D1 aWeiss, David, J1 aPilkonis, Paul, A1 aFrank, Ellen1 aMoore, Tara1 aKim, Jong Bae1 aKupfer, David, J uWWW.ARCHGENPSYCHIATRY.COM01421nas a2200157 4500008003900000245008000039210006900119300001200188490000700200520092800207100001601135700001801151700002101169700002001190856005301210 2012 d00aPanel Design Variations in the Multistage Test Using the Mixed-Format Tests0 aPanel Design Variations in the Multistage Test Using the MixedFo a574-5880 v723 aThis study compared various panel designs of the multistage test (MST) using mixed-format tests in the context of classification testing. Simulations varied the design of the first-stage module. The first stage was constructed according to three levels of test information functions (TIFs) with three different TIF centers. Additional computerized adaptive test (CAT) conditions provided baseline comparisons. Three passing rate conditions were also included. The various MST conditions using mixed-format tests were constructed properly and performed well. When the levels of TIFs at the first stage were higher, the simulations produced a greater number of correct classifications. CAT with the randomesque-10 procedure yielded comparable results to the MST with increased levels of TIFs. Finally, all MST conditions achieved better test security results compared with CAT’s maximum information conditions.
1 aKim, Jiseon1 aChung, Hyewon1 aDodd, Barbara, G1 aPark, Ryoungsun uhttp://epm.sagepub.com/content/72/4/574.abstract02592nas a2200313 4500008004100000020004600041245011200087210006900199250001500268260001900283300000700302490000600309520160100315100001901916700001801935700001701953700001801970700001701988700001402005700001602019700001502035700001402050700002602064700001602090700001302106700001502119700001902134856012502153 2011 Eng d a1477-7525 (Electronic)1477-7525 (Linking)00aCross-cultural development of an item list for computer-adaptive testing of fatigue in oncological patients0 aCrosscultural development of an item list for computeradaptive t a2011/03/31 cMarch 29, 2011 a100 v93 aABSTRACT: INTRODUCTION: Within an ongoing project of the EORTC Quality of Life Group, we are developing computerized adaptive test (CAT) measures for the QLQ-C30 scales. These new CAT measures are conceptualised to reflect the same constructs as the QLQ-C30 scales. Accordingly, the Fatigue-CAT is intended to capture physical and general fatigue. METHODS: The EORTC approach to CAT development comprises four phases (literature search, operationalisation, pre-testing, and field testing). Phases I-III are described in detail in this paper. A literature search for fatigue items was performed in major medical databases. After refinement through several expert panels, the remaining items were used as the basis for adapting items and/or formulating new items fitting the EORTC item style. To obtain feedback from patients with cancer, these English items were translated into Danish, French, German, and Spanish and tested in the respective countries. RESULTS: Based on the literature search a list containing 588 items was generated. After a comprehensive item selection procedure focusing on content, redundancy, item clarity and item difficulty a list of 44 fatigue items was generated. Patient interviews (n=52) resulted in 12 revisions of wording and translations. DISCUSSION: The item list developed in phases I-III will be further investigated within a field-testing phase (IV) to examine psychometric characteristics and to fit an item response theory model. The Fatigue CAT based on this item bank will provide scores that are backward-compatible to the original QLQ-C30 fatigue scale.1 aGiesinger, J M1 aPetersen, M A1 aGroenvold, M1 aAaronson, N K1 aArraras, J I1 aConroy, T1 aGamper, E M1 aKemmler, G1 aKing, M T1 aOberguggenberger, A S1 aVelikova, G1 aYoung, T1 aHolzner, B1 aEortc-Qlg, E O uhttp://www.iacat.org/content/cross-cultural-development-item-list-computer-adaptive-testing-fatigue-oncological-patients00653nas a2200157 4500008004100000020001400041245015700055210006900212100001800281700001400299700001500313700001600328700001500344700001300359856012300372 2011 eng d a1073-191100aItem banks for measuring emotional distress from the Patient-Reported Outcomes Measurement Information System (PROMIS®): depression, anxiety, and anger0 aItem banks for measuring emotional distress from the PatientRepo1 aPilkonis, P A1 aChoi, S W1 aReise, S P1 aStover, A M1 aRiley, W T1 aCella, D uhttp://www.iacat.org/content/item-banks-measuring-emotional-distress-patient-reported-outcomes-measurement-information03049nas a2200241 4500008004100000020004100041245017000082210007100252250001500323300001000338490000700348520214300355653001402498653006602512653001102578653001302589100001402602700001202616700001402628700001502642700001702657856013302674 2010 spa d a0214-9915 (Print)0214-9915 (Linking)00aDeterioro de parámetros de los ítems en tests adaptativos informatizados: estudio con eCAT [Item parameter drift in computerized adaptive testing: Study with eCAT]0 aDeterioro de parámetros de los ítems en tests adaptativos inform a2010/04/29 a340-70 v223 aEn el presente trabajo se muestra el análisis realizado sobre un Test Adaptativo Informatizado (TAI) diseñado para la evaluación del nivel de inglés, denominado eCAT, con el objetivo de estudiar el deterioro de parámetros (parameter drift) producido desde la calibración inicial del banco de ítems. Se ha comparado la calibración original desarrollada para la puesta en servicio del TAI (N= 3224) y la calibración actual obtenida con las aplicaciones reales del TAI (N= 7254). Se ha analizado el Funcionamiento Diferencial de los Ítems (FDI) en función de los parámetros utilizados y se ha simulado el impacto que sobre el nivel de rasgo estimado tiene la variación en los parámetros. Los resultados muestran que se produce especialmente un deterioro de los parámetros a y c, que hay unimportante número de ítems del banco para los que existe FDI y que la variación de los parámetros produce un impacto moderado en la estimación de θ de los evaluados con nivel de inglés alto. Se concluye que los parámetros de los ítems se han deteriorado y deben ser actualizados.Item parameter drift in computerized adaptive testing: Study with eCAT. This study describes the parameter drift analysis conducted on eCAT (a Computerized Adaptive Test to assess the written English level of Spanish speakers). The original calibration of the item bank (N = 3224) was compared to a new calibration obtained from the data provided by most eCAT operative administrations (N =7254). A Differential Item Functioning (DIF) study was conducted between the original and the new calibrations. The impact that the new parameters have on the trait level estimates was obtained by simulation. Results show that parameter drift is found especially for a and c parameters, an important number of bank items show DIF, and the parameter change has a moderate impact on high-level-English θ estimates. It is then recommended to replace the original estimates by the new set. by the new set.
10a*Software10aEducational Measurement/*methods/*statistics & numerical data10aHumans10aLanguage1 aAbad, F J1 aOlea, J1 aAguado, D1 aPonsoda, V1 aBarrada, J R uhttp://www.iacat.org/content/deterioro-de-par%C3%A1metros-de-los-%C3%ADtems-en-tests-adaptativos-informatizados-estudio-con-ecat03103nas a2200445 4500008004100000020004100041245012000082210006900202250001500271260001000286300001100296490000700307520175400314653003802068653002102106653001002127653000902137653002202146653002802168653003302196653001102229653001102240653000902251653001602260653001802276653001902294653003102313653003102344653001602375100001602391700001002407700001402417700001502431700001402446700001502460700001802475700002402493700001802517856012202535 2010 eng d a0161-8105 (Print)0161-8105 (Linking)00aDevelopment and validation of patient-reported outcome measures for sleep disturbance and sleep-related impairments0 aDevelopment and validation of patientreported outcome measures f a2010/06/17 cJun 1 a781-920 v333 aSTUDY OBJECTIVES: To develop an archive of self-report questions assessing sleep disturbance and sleep-related impairments (SRI), to develop item banks from this archive, and to validate and calibrate the item banks using classic validation techniques and item response theory analyses in a sample of clinical and community participants. DESIGN: Cross-sectional self-report study. SETTING: Academic medical center and participant homes. PARTICIPANTS: One thousand nine hundred ninety-three adults recruited from an Internet polling sample and 259 adults recruited from medical, psychiatric, and sleep clinics. INTERVENTIONS: None. MEASUREMENTS AND RESULTS: This study was part of PROMIS (Patient-Reported Outcomes Information System), a National Institutes of Health Roadmap initiative. Self-report item banks were developed through an iterative process of literature searches, collecting and sorting items, expert content review, qualitative patient research, and pilot testing. Internal consistency, convergent validity, and exploratory and confirmatory factor analysis were examined in the resulting item banks. Factor analyses identified 2 preliminary item banks, sleep disturbance and SRI. Item response theory analyses and expert content review narrowed the item banks to 27 and 16 items, respectively. Validity of the item banks was supported by moderate to high correlations with existing scales and by significant differences in sleep disturbance and SRI scores between participants with and without sleep disorders. CONCLUSIONS: The PROMIS sleep disturbance and SRI item banks have excellent measurement properties and may prove to be useful for assessing general aspects of sleep and SRI with various groups of patients and interventions.10a*Outcome Assessment (Health Care)10a*Self Disclosure10aAdult10aAged10aAged, 80 and over10aCross-Sectional Studies10aFactor Analysis, Statistical10aFemale10aHumans10aMale10aMiddle Aged10aPsychometrics10aQuestionnaires10aReproducibility of Results10aSleep Disorders/*diagnosis10aYoung Adult1 aBuysse, D J1 aYu, L1 aMoul, D E1 aGermain, A1 aStover, A1 aDodds, N E1 aJohnston, K L1 aShablesky-Cade, M A1 aPilkonis, P A uhttp://www.iacat.org/content/development-and-validation-patient-reported-outcome-measures-sleep-disturbance-and-sleep02223nas a2200289 4500008004100000020004600041245010800087210006900195250001500264300001200279490000700291520131000298100001801608700001701626700001801643700001401661700001401675700001801689700001401707700001701721700001501738700001301753700001401766700001601780700001301796856012401809 2010 Eng d a1573-2649 (Electronic)0962-9343 (Linking)00aDevelopment of computerized adaptive testing (CAT) for the EORTC QLQ-C30 physical functioning dimension0 aDevelopment of computerized adaptive testing CAT for the EORTC Q a2010/10/26 a479-4900 v203 aPURPOSE: Computerized adaptive test (CAT) methods, based on item response theory (IRT), enable a patient-reported outcome instrument to be adapted to the individual patient while maintaining direct comparability of scores. The EORTC Quality of Life Group is developing a CAT version of the widely used EORTC QLQ-C30. We present the development and psychometric validation of the item pool for the first of the scales, physical functioning (PF). METHODS: Initial developments (including literature search and patient and expert evaluations) resulted in 56 candidate items. Responses to these items were collected from 1,176 patients with cancer from Denmark, France, Germany, Italy, Taiwan, and the United Kingdom. The items were evaluated with regard to psychometric properties. RESULTS: Evaluations showed that 31 of the items could be included in a unidimensional IRT model with acceptable fit and good content coverage, although the pool may lack items at the upper extreme (good PF). There were several findings of significant differential item functioning (DIF). However, the DIF findings appeared to have little impact on the PF estimation. CONCLUSIONS: We have established an item pool for CAT measurement of PF and believe that this CAT instrument will clearly improve the EORTC measurement of PF.1 aPetersen, M A1 aGroenvold, M1 aAaronson, N K1 aChie, W C1 aConroy, T1 aCostantini, A1 aFayers, P1 aHelbostad, J1 aHolzner, B1 aKaasa, S1 aSinger, S1 aVelikova, G1 aYoung, T uhttp://www.iacat.org/content/development-computerized-adaptive-testing-cat-eortc-qlq-c30-physical-functioning-dimension00605nas a2200157 4500008004100000245011500041210006900156300001400225490001000239100001200249700001500261700001800276700001400294700001300308856012600321 2010 eng d00aEfficiency of static and computer adaptive short forms compared to full-length measures of depressive symptoms0 aEfficiency of static and computer adaptive short forms compared a125–1360 v19(1)1 aChoi, S1 aReise, S P1 aPilkonis, P A1 aHays, R D1 aCella, D uhttp://www.iacat.org/content/efficiency-static-and-computer-adaptive-short-forms-compared-full-length-measures-depressive00414nas a2200133 4500008004100000245004600041210004600087300001200133100001800145700001600163700001300179700001700192856007100209 2010 eng d00aInnovative Items for Computerized Testing0 aInnovative Items for Computerized Testing a215-2301 aParshall, C G1 aHarmes, J C1 aDavey, T1 aPashley, P J uhttp://www.iacat.org/content/innovative-items-computerized-testing00449nas a2200121 4500008004100000245006200041210006200103260002300165300000900188100002300197700001700220856009000237 2010 eng d00aItem Selection and Ability Estimation in Adaptive Testing0 aItem Selection and Ability Estimation in Adaptive Testing aNew YorkbSpringer a3-301 avan der Linden, WJ1 aPashley, P J uhttp://www.iacat.org/content/item-selection-and-ability-estimation-adaptive-testing-001871nas a2200157 4500008003900000245008900039210006900128300001200197490000700209520135600216100002501572700001601597700002101613700002601634856005301660 2010 d00aA Method for the Comparison of Item Selection Rules in Computerized Adaptive Testing0 aMethod for the Comparison of Item Selection Rules in Computerize a438-4520 v343 aIn a typical study comparing the relative efficiency of two item selection rules in computerized adaptive testing, the common result is that they simultaneously differ in accuracy and security, making it difficult to reach a conclusion on which is the more appropriate rule. This study proposes a strategy to conduct a global comparison of two or more selection rules. A plot showing the performance of each selection rule for several maximum exposure rates is obtained and the whole plot is compared with other rule plots. The strategy was applied in a simulation study with fixed-length CATs for the comparison of six item selection rules: the point Fisher information, Fisher information weighted by likelihood, Kullback-Leibler weighted by likelihood, maximum information stratification with blocking, progressive and proportional methods. Our results show that there is no optimal rule for any overlap value or root mean square error (RMSE). The fact that a rule, for a given level of overlap, has lower RMSE than another does not imply that this pattern holds for another overlap rate. A fair comparison of the rules requires extensive manipulation of the maximum exposure rates. The best methods were the Kullback-Leibler weighted by likelihood, the proportional method, and the maximum information stratification method with blocking.
1 aBarrada, Juan Ramón1 aOlea, Julio1 aPonsoda, Vicente1 aAbad, Francisco José uhttp://apm.sagepub.com/content/34/6/438.abstract00571nas a2200109 4500008004100000245010000041210006900141260009700210100001400307700001700321856012300338 2009 eng d00aAn approach to implementing adaptive testing using item response theory both offline and online0 aapproach to implementing adaptive testing using item response th aD. J. Weiss (Ed.), Proceedings of the 2009 GMAC Conference on Computerized Adaptive Testing.1 aPadaki, M1 aNatarajan, V uhttp://www.iacat.org/content/approach-implementing-adaptive-testing-using-item-response-theory-both-offline-and-online00616nas a2200121 4500008004100000245012600041210006900167260009700236100001100333700001700344700001400361856011900375 2009 eng d00aAn evaluation of a new procedure for computing information functions for Bayesian scores from computerized adaptive tests0 aevaluation of a new procedure for computing information function aD. J. Weiss (Ed.), Proceedings of the 2009 GMAC Conference on Computerized Adaptive Testing.1 aIto, K1 aPommerich, M1 aSegall, D uhttp://www.iacat.org/content/evaluation-new-procedure-computing-information-functions-bayesian-scores-computerized00516nas a2200145 4500008004100000245008100041210006900122300000900191490000600200100001700206700001200223700001500235700001400250856010600264 2009 eng d00aItem selection rules in computerized adaptive testing: Accuracy and security0 aItem selection rules in computerized adaptive testing Accuracy a a7-170 v51 aBarrada, J R1 aOlea, J1 aPonsoda, V1 aAbad, F J uhttp://www.iacat.org/content/item-selection-rules-computerized-adaptive-testing-accuracy-and-security02048nas a2200133 4500008004100000245006100041210005500102260010000157520152600257100001701783700001601800700001601816856008201832 2009 eng d00aThe nine lives of CAT-ASVAB: Innovations and revelations0 anine lives of CATASVAB Innovations and revelations aIn D. J. Weiss (Ed.), Proceedings of the 2009 GMAC Conference on Computerized Adaptive Testing.3 aThe Armed Services Vocational Aptitude Battery (ASVAB) is administered annually to more than one million military applicants and high school students. ASVAB scores are used to determine enlistment eligibility, assign applicants to military occupational specialties, and aid students in career exploration. The ASVAB is administered as both a paper-and-pencil (P&P) test and a computerized adaptive test (CAT). CAT-ASVAB holds the distinction of being the first large-scale adaptive test battery to be administered in a high-stakes setting. Approximately two-thirds of military applicants currently take CAT-ASVAB; long-term plans are to replace P&P-ASVAB with CAT-ASVAB at all test sites. Given CAT-ASVAB’s pedigree—approximately 20 years in development and 20 years in operational administration—much can be learned from revisiting some of the major highlights of CATASVAB history. This paper traces the progression of CAT-ASVAB through nine major phases of development including: research and evelopment of the CAT-ASVAB prototype, the initial development of psychometric procedures and item pools, initial and full-scale operational implementation, the introduction of new item pools, the introduction of Windows administration, the introduction of Internet administration, and research and development of the next generation CATASVAB. A background and history is provided for each phase, including discussions of major research and operational issues, innovative approaches and practices, and lessons learned.1 aPommerich, M1 aSegall, D O1 aMoreno, K E uhttp://www.iacat.org/content/nine-lives-cat-asvab-innovations-and-revelations00587nas a2200133 4500008004100000245008400041210006900125260009700194100001700291700001200308700001500320700001400335856010400349 2009 eng d00aTest overlap rate and item exposure rate as indicators of test security in CATs0 aTest overlap rate and item exposure rate as indicators of test s aD. J. Weiss (Ed.), Proceedings of the 2009 GMAC Conference on Computerized Adaptive Testing.1 aBarrada, J R1 aOlea, J1 aPonsoda, V1 aAbad, F J uhttp://www.iacat.org/content/test-overlap-rate-and-item-exposure-rate-indicators-test-security-cats01224nas a2200133 4500008003900000245008800039210006900127260000900196300001400205520072100219100002000940700002000960856011000980 2008 d00aAn Automated Decision System for Computer Adaptive Testing Using Genetic Algorithms0 aAutomated Decision System for Computer Adaptive Testing Using Ge bIEEE a655–6603 aThis paper proposes an approach to solve the triangle decision tree problem for computer adaptive testing (CAT) using genetic algorithms (GAs). In this approach, item response theory (IRT) parameters composed of discrimination, difficulty, and guess are firstly obtained and stored in an item bank. Then a fitness function, which is based on IRT parameters, of GAs for obtaining an optimal solution is set up. Finally, the GAs is applied to the parameters of the item bank so that an optimal decision tree is generated. Based on a six-level triangle-decision tree for examination items, the experimental results show that the optimal decision tree can be generated correctly when compared with the standard patterns.1 aPhankokkruad, M1 aWoraratpanya, K uhttp://www.iacat.org/content/automated-decision-system-computer-adaptive-testing-using-genetic-algorithms01728nas a2200181 4500008004100000245011200041210006900153300001100222490000700233520107200240653003401312653001701346653001901363100001601382700001801398700001501416856011501431 2008 eng d00aThe D-optimality item selection criterion in the early stage of CAT: A study with the graded response model0 aDoptimality item selection criterion in the early stage of CAT A a88-1100 v333 aDuring the early stage of computerized adaptive testing (CAT), item selection criteria based on Fisher’s information often produce less stable latent trait estimates than the Kullback-Leibler global information criterion. Robustness against early stage instability has been reported for the D-optimality criterion in a polytomous CAT with the Nominal Response Model and is shown herein to be reproducible for the Graded Response Model. For comparative purposes, the A-optimality and the global information criteria are also applied. Their item selection is investigated as a function of test progression and item bank composition. The results indicate how the selection of specific item parameters underlies the criteria performances evaluated via accuracy and precision of estimation. In addition, the criteria item exposure rates are compared, without the use of any exposure controlling measure. On the account of stability, precision, accuracy, numerical simplicity, and less evidently, item exposure rate, the D-optimality criterion can be recommended for CAT.10acomputerized adaptive testing10aD optimality10aitem selection1 aPassos, V L1 aBerger, M P F1 aTan, F E S uhttp://www.iacat.org/content/d-optimality-item-selection-criterion-early-stage-cat-study-graded-response-model00546nas a2200145 4500008004100000245009900041210006900140300001200209490000700221100001700228700001200245700001500257700001400272856011400286 2008 eng d00aIncorporating randomness in the Fisher information for improving item-exposure control in CATs0 aIncorporating randomness in the Fisher information for improving a493-5130 v611 aBarrada, J R1 aOlea, J1 aPonsoda, V1 aAbad, F J uhttp://www.iacat.org/content/incorporating-randomness-fisher-information-improving-item-exposure-control-cats01442nas a2200145 4500008003900000022001400039245007100053210006900124300001400193490000700207520098500214100002001199700002201219856005501241 2008 d a1745-398400aLocal Dependence in an Operational CAT: Diagnosis and Implications0 aLocal Dependence in an Operational CAT Diagnosis and Implication a201–2230 v453 aThe accuracy of CAT scores can be negatively affected by local dependence if the CAT utilizes parameters that are misspecified due to the presence of local dependence and/or fails to control for local dependence in responses during the administration stage. This article evaluates the existence and effect of local dependence in a test of Mathematics Knowledge. Diagnostic tools were first used to evaluate the existence of local dependence in items that were calibrated under a 3PL model. A simulation study was then used to evaluate the effect of local dependence on the precision of examinee CAT scores when the 3PL model was used for selection and scoring. The diagnostic evaluation showed strong evidence for local dependence. The simulation suggested that local dependence in parameters had a minimal effect on CAT score precision, while local dependence in responses had a substantial effect on score precision, depending on the degree of local dependence present.
1 aPommerich, Mary1 aSegall, Daniel, O uhttp://dx.doi.org/10.1111/j.1745-3984.2008.00061.x03428nas a2200385 4500008004100000020004100041245010600082210006900188250001500257260001200272300001000284490000700294520220300301653002702504653001502531653001002546653002102556653002402577653002802601653003802629653001102667653001102678653001102689653003902700653000902739653002402748653003102772653004002803100001802843700001502861700001302876700001702889700001402906856012202920 2008 eng d a0271-6798 (Print)0271-6798 (Linking)00aMeasuring physical functioning in children with spinal impairments with computerized adaptive testing0 aMeasuring physical functioning in children with spinal impairmen a2008/03/26 cApr-May a330-50 v283 aBACKGROUND: The purpose of this study was to assess the utility of measuring current physical functioning status of children with scoliosis and kyphosis by applying computerized adaptive testing (CAT) methods. Computerized adaptive testing uses a computer interface to administer the most optimal items based on previous responses, reducing the number of items needed to obtain a scoring estimate. METHODS: This was a prospective study of 77 subjects (0.6-19.8 years) who were seen by a spine surgeon during a routine clinic visit for progress spine deformity. Using a multidimensional version of the Pediatric Evaluation of Disability Inventory CAT program (PEDI-MCAT), we evaluated content range, accuracy and efficiency, known-group validity, concurrent validity with the Pediatric Outcomes Data Collection Instrument, and test-retest reliability in a subsample (n = 16) within a 2-week interval. RESULTS: We found the PEDI-MCAT to have sufficient item coverage in both self-care and mobility content for this sample, although most patients tended to score at the higher ends of both scales. Both the accuracy of PEDI-MCAT scores as compared with a fixed format of the PEDI (r = 0.98 for both mobility and self-care) and test-retest reliability were very high [self-care: intraclass correlation (3,1) = 0.98, mobility: intraclass correlation (3,1) = 0.99]. The PEDI-MCAT took an average of 2.9 minutes for the parents to complete. The PEDI-MCAT detected expected differences between patient groups, and scores on the PEDI-MCAT correlated in expected directions with scores from the Pediatric Outcomes Data Collection Instrument domains. CONCLUSIONS: Use of the PEDI-MCAT to assess the physical functioning status, as perceived by parents of children with complex spinal impairments, seems to be feasible and achieves accurate and efficient estimates of self-care and mobility function. Additional item development will be needed at the higher functioning end of the scale to avoid ceiling effects for older children. LEVEL OF EVIDENCE: This is a level II prospective study designed to establish the utility of computer adaptive testing as an evaluation method in a busy pediatric spine practice.10a*Disability Evaluation10aAdolescent10aChild10aChild, Preschool10aComputer Simulation10aCross-Sectional Studies10aDisabled Children/*rehabilitation10aFemale10aHumans10aInfant10aKyphosis/*diagnosis/rehabilitation10aMale10aProspective Studies10aReproducibility of Results10aScoliosis/*diagnosis/rehabilitation1 aMulcahey, M J1 aHaley, S M1 aDuffy, T1 aPengsheng, N1 aBetz, R R uhttp://www.iacat.org/content/measuring-physical-functioning-children-spinal-impairments-computerized-adaptive-testing00460nas a2200121 4500008003900000245009100039210007600130300001200206490000600218100002800224700002100252856006500273 2007 d00aA “Rearrangement Procedure†For Scoring Adaptive Tests with Review Options0 a“Rearrangement Procedure†For Scoring Adaptive Tests with Rev a387-4070 v71 aPapanastasiou, Elena, C1 aReckase, Mark, D uhttp://www.tandfonline.com/doi/abs/10.1080/1530505070163226200495nas a2200109 4500008004100000245006400041210006400105260009700169100001800266700001600284856008500300 2007 eng d00aDesigning templates based on a taxonomy of innovative items0 aDesigning templates based on a taxonomy of innovative items aD. J. Weiss (Ed.). Proceedings of the 2007 GMAC Conference on Computerized Adaptive Testing.1 aParshall, C G1 aHarmes, J C uhttp://www.iacat.org/content/designing-templates-based-taxonomy-innovative-items01798nas a2200217 4500008004100000020004600041245017800087210006900265260002500334300001200359490000700371520094900378653001201327653004301339653001801382653000901400100001701409700001501426700001501441856012401456 2007 eng d a1062-7197 (Print); 1532-6977 (Electronic)00aThe effect of including pretest items in an operational computerized adaptive test: Do different ability examinees spend different amounts of time on embedded pretest items?0 aeffect of including pretest items in an operational computerized bLawrence Erlbaum: US a161-1730 v123 aThe purpose of this study was to examine the effect of pretest items on response time in an operational, fixed-length, time-limited computerized adaptive test (CAT). These pretest items are embedded within the CAT, but unlike the operational items, are not tailored to the examinee's ability level. If examinees with higher ability levels need less time to complete these items than do their counterparts with lower ability levels, they will have more time to devote to the operational test questions. Data were from a graduate admissions test that was administered worldwide. Data from both quantitative and verbal sections of the test were considered. For the verbal section, examinees in the lower ability groups spent systematically more time on their pretest items than did those in the higher ability groups, though for the quantitative section the differences were less clear. (PsycINFO Database Record (c) 2007 APA, all rights reserved)10aability10aoperational computerized adaptive test10apretest items10atime1 aFerdous, A A1 aPlake, B S1 aChang, S-R uhttp://www.iacat.org/content/effect-including-pretest-items-operational-computerized-adaptive-test-do-different-ability02156nas a2200133 4500008004100000245012000041210006900161300000900230490000600239520162500245653001701870100001701887856011801904 2007 eng d00aThe effect of using item parameters calibrated from paper administrations in computer adaptive test administrations0 aeffect of using item parameters calibrated from paper administra a1-290 v53 aComputer administered tests are becoming increasingly prevalent as computer technology becomes more readily available on a large scale. For testing programs that utilize both computer and paper administrations, mode effects are problematic in that they can result in examinee scores that are artificially inflated or deflated. As such, researchers have engaged in extensive studies of whether scores differ across paper and computer presentations of the same tests. The research generally seems to indicate that the more complicated it is to present or take a test on computer, the greater the possibility of mode effects. In a computer adaptive test, mode effects may be a particular concern if items are calibrated using item responses obtained from one administration mode (i.e., paper), and those parameters are then used operationally in a different administration mode (i.e., computer). This paper studies the suitability of using parameters calibrated from a paper administration for item selection and scoring in a computer adaptive administration, for two tests with lengthy passages that required navigation in the computer administration. The results showed that the use of paper calibrated parameters versus computer calibrated parameters in computer adaptive administrations had small to moderate effects on the reliability of examinee scores, at fairly short test lengths. This effect was generally diminished for longer test lengths. However, the results suggest that in some cases, some loss in reliability might be inevitable if paper-calibrated parameters are used in computer adaptive administrations.10aMode effects1 aPommerich, M uhttp://www.iacat.org/content/effect-using-item-parameters-calibrated-paper-administrations-computer-adaptive-test02153nas a2200109 4500008003900000245012000039210006900159490000600228520167200234100001701906856012001923 2007 d00aThe Effect of Using Item Parameters Calibrated from Paper Administrations in Computer Adaptive Test Administrations0 aEffect of Using Item Parameters Calibrated from Paper Administra0 v53 aComputer administered tests are becoming increasingly prevalent as computer technology becomes more readily available on a large scale. For testing programs that utilize both computer and paper administrations, mode effects are problematic in that they can
result in examinee scores that are artificially inflated or deflated. As such, researchers have engaged in extensive studies of whether scores differ across paper and computer presentations of the same tests. The research generally seems to indicate that the more
complicated it is to present or take a test on computer, the greater the possibility of mode effects. In a computer adaptive test, mode effects may be a particular concern if items are calibrated using item responses obtained from one administration mode (i.e., paper), and those parameters are then used operationally in a different administration mode (i.e., computer). This paper studies the suitability of using parameters calibrated from a paper administration for item selection and scoring in a computer adaptive administration, for two tests with lengthy passages that required navigation in the computer administration. The results showed that the use of paper calibrated parameters versus computer calibrated parameters in computer adaptive administrations had small to
moderate effects on the reliability of examinee scores, at fairly short test lengths. This effect was generally diminished for longer test lengths. However, the results suggest that in some cases, some loss in reliability might be inevitable if paper-calibrated parameters
are used in computer adaptive administrations.
The standard error of the maximum likelihood ability estimator is commonly estimated by evaluating the test information function at an examinee's current maximum likelihood estimate (a point estimate) of ability. Because the test information function evaluated at the point estimate may differ from the test information function evaluated at an examinee's true ability value, the estimated standard error may be biased under certain conditions. This is of particular concern in adaptive testing because the height of the test information function is expected to be higher at the current estimate of ability than at the actual value of ability. This article proposes using the posterior-weighted test information function in computing the standard error of the maximum likelihood ability estimator for adaptive test sessions. A simulation study showed that the proposed approach provides standard error estimates that are less biased and more efficient than those provided by the traditional point estimate approach.
1 aPenfield, Randall, D uhttp://epm.sagepub.com/content/67/6/958.abstract02212nas a2200229 4500008004100000020004600041245008500087210006900172260004500241300001000286490000600296520140300302653003401705653002301739653002601762653001701788653002601805100001701831700001201848700001501860856010701875 2007 eng d a1614-1881 (Print); 1614-2241 (Electronic)00aMethods for restricting maximum exposure rate in computerized adaptative testing0 aMethods for restricting maximum exposure rate in computerized ad bHogrefe & Huber Publishers GmbH: Germany a14-230 v33 aThe Sympson-Hetter (1985) method provides a means of controlling maximum exposure rate of items in Computerized Adaptive Testing. Through a series of simulations, control parameters are set that mark the probability of administration of an item on being selected. This method presents two main problems: it requires a long computation time for calculating the parameters and the maximum exposure rate is slightly above the fixed limit. Van der Linden (2003) presented two alternatives which appear to solve both of the problems. The impact of these methods in the measurement accuracy has not been tested yet. We show how these methods over-restrict the exposure of some highly discriminating items and, thus, the accuracy is decreased. It also shown that, when the desired maximum exposure rate is near the minimum possible value, these methods offer an empirical maximum exposure rate clearly above the goal. A new method, based on the initial estimation of the probability of administration and the probability of selection of the items with the restricted method (Revuelta & Ponsoda, 1998), is presented in this paper. It can be used with the Sympson-Hetter method and with the two van der Linden's methods. This option, when used with Sympson-Hetter, speeds the convergence of the control parameters without decreasing the accuracy. (PsycINFO Database Record (c) 2007 APA, all rights reserved)10acomputerized adaptive testing10aitem bank security10aitem exposure control10aoverlap rate10aSympson-Hetter method1 aBarrada, J R1 aOlea, J1 aPonsoda, V uhttp://www.iacat.org/content/methods-restricting-maximum-exposure-rate-computerized-adaptative-testing00371nas a2200097 4500008004100000245003400041210003200075260009700107100001200204856005700216 2007 eng d00aA new delivery system for CAT0 anew delivery system for CAT aD. J. Weiss (Ed.). Proceedings of the 2007 GMAC Conference on Computerized Adaptive Testing.1 aPark, J uhttp://www.iacat.org/content/new-delivery-system-cat00506nas a2200121 4500008004100000245006000041210006000101260009700161100001900258700000700277700001500284856008500299 2007 eng d00aPartial order knowledge structures for CAT applications0 aPartial order knowledge structures for CAT applications aD. J. Weiss (Ed.), Proceedings of the 2007 GMAC Conference on Computerized Adaptive Testing.1 aDesmarais, M C1 aPu1 aBlais, J-G uhttp://www.iacat.org/content/partial-order-knowledge-structures-cat-applications01688nas a2200217 4500008004100000020002200041245008000063210006900143260002600212300001200238490000700250520095600257653003401213653002701247653002301274653002901297100001601326700001801342700001301360856009701373 2007 eng d a0146-6216 (Print)00aTest design optimization in CAT early stage with the nominal response model0 aTest design optimization in CAT early stage with the nominal res bSage Publications: US a213-2320 v313 aThe early stage of computerized adaptive testing (CAT) refers to the phase of the trait estimation during the administration of only a few items. This phase can be characterized by bias and instability of estimation. In this study, an item selection criterion is introduced in an attempt to lessen this instability: the D-optimality criterion. A polytomous unconstrained CAT simulation is carried out to evaluate this criterion's performance under different test premises. The simulation shows that the extent of early stage instability depends primarily on the quality of the item pool information and its size and secondarily on the item selection criteria. The efficiency of the D-optimality criterion is similar to the efficiency of other known item selection criteria. Yet, it often yields estimates that, at the beginning of CAT, display a more robust performance against instability. (PsycINFO Database Record (c) 2007 APA, all rights reserved)10acomputerized adaptive testing10anominal response model10arobust performance10atest design optimization1 aPassos, V L1 aBerger, M P F1 aTan, F E uhttp://www.iacat.org/content/test-design-optimization-cat-early-stage-nominal-response-model01637nas a2200217 4500008004100000020004600041245008900087210006900176260002500245300000900270490000700279520083400286653001901120653002801139653003001167653003301197653002101230653003501251100001801286856011501304 2006 eng d a0895-7347 (Print); 1532-4818 (Electronic)00aApplying Bayesian item selection approaches to adaptive tests using polytomous items0 aApplying Bayesian item selection approaches to adaptive tests us bLawrence Erlbaum: US a1-200 v193 aThis study applied the maximum expected information (MEI) and the maximum posterior- weighted information (MPI) approaches of computer adaptive testing item selection to the case of a test using polytomous items following the partial credit model. The MEI and MPI approaches are described. A simulation study compared the efficiency of ability estimation using the MEI and MPI approaches to the traditional maximal item information (MII) approach. The results of the simulation study indicated that the MEI and MPI approaches led to a superior efficiency of ability estimation compared with the MII approach. The superiority of the MEI and MPI approaches over the MII approach was greatest when the bank contained items having a relatively peaked information function. (PsycINFO Database Record (c) 2007 APA, all rights reserved)10aadaptive tests10aBayesian item selection10acomputer adaptive testing10amaximum expected information10apolytomous items10aposterior weighted information1 aPenfield, R D uhttp://www.iacat.org/content/applying-bayesian-item-selection-approaches-adaptive-tests-using-polytomous-items00452nas a2200109 4500008004100000245004200041210004200083260011500125100001300240700001800253856007100271 2006 eng d00aDesigning computerized adaptive tests0 aDesigning computerized adaptive tests aS.M. Downing and T. M. Haladyna (Eds.), Handbook of test development. New Jersey: Lawrence Erlbaum Associates.1 aDavey, T1 aPitoniak, M J uhttp://www.iacat.org/content/designing-computerized-adaptive-tests01613nas a2200145 4500008003900000245011900039210006900158300001200227490000700239520110300246100002001349700002001369700002501389856005301414 2006 d00aEffects of Estimation Bias on Multiple-Category Classification With an IRT-Based Adaptive Classification Procedure0 aEffects of Estimation Bias on MultipleCategory Classification Wi a545-5640 v663 aThe effects of five ability estimators, that is, maximum likelihood estimator, weighted likelihood estimator, maximum a posteriori, expected a posteriori, and Owen's sequential estimator, on the performances of the item response theory–based adaptive classification procedure on multiple categories were studied via simulations. The following results were found. (a) The Bayesian estimators were more likely to misclassify examinees into an inward category because of their inward biases, when a fixed start value of zero was assigned to every examinee. (b) When moderately accurate start values were available, however, Bayesian estimators produced classifications that were slightly more accurate than was the maximum likelihood estimator or weighted likelihood estimator. Expected a posteriori was the procedure that produced the most accurate results among the three Bayesian methods. (c) All five estimators produced equivalent efficiencies in terms of number of items required, which was 50 or more items except for abilities that were less than -2.00 or greater than 2.00.
1 aYang, Xiangdong1 aPoggio, John, C1 aGlasnapp, Douglas, R uhttp://epm.sagepub.com/content/66/4/545.abstract01330nas a2200169 4500008004100000020002700041245010300068210006900171250001500240300000600255490000600261520073400267100001301001700001401014700001401028856011801042 2006 eng d a1975-5937 (Electronic)00aEstimation of an examinee's ability in the web-based computerized adaptive testing program IRT-CAT0 aEstimation of an examinees ability in the webbased computerized a2006/01/01 a40 v33 aWe developed a program to estimate an examinee s ability in order to provide freely available access to a web-based computerized adaptive testing (CAT) program. We used PHP and Java Script as the program languages, PostgresSQL as the database management system on an Apache web server and Linux as the operating system. A system which allows for user input and searching within inputted items and creates tests was constructed. We performed an ability estimation on each test based on a Rasch model and 2- or 3-parametric logistic models. Our system provides an algorithm for a web-based CAT, replacing previous personal computer-based ones, and makes it possible to estimate an examinee's ability immediately at the end of test.1 aLee, Y H1 aPark, J H1 aPark, I Y uhttp://www.iacat.org/content/estimation-examinees-ability-web-based-computerized-adaptive-testing-program-irt-cat03383nas a2200205 4500008004100000020002200041245009700063210006900160260002500229300001200254490000700266520266000273653003402933653002802967100001502995700001903010700001703029700001403046856011703060 2006 eng d a0439-755X (Print)00a[Item Selection Strategies of Computerized Adaptive Testing based on Graded Response Model.]0 aItem Selection Strategies of Computerized Adaptive Testing based bScience Press: China a461-4670 v383 aItem selection strategy (ISS) is an important component of Computerized Adaptive Testing (CAT). Its performance directly affects the security, efficiency and precision of the test. Thus, ISS becomes one of the central issues in CATs based on the Graded Response Model (GRM). It is well known that the goal of IIS is to administer the next unused item remaining in the item bank that best fits the examinees current ability estimate. In dichotomous IRT models, every item has only one difficulty parameter and the item whose difficulty matches the examinee's current ability estimate is considered to be the best fitting item. However, in GRM, each item has more than two ordered categories and has no single value to represent the item difficulty. Consequently, some researchers have used to employ the average or the median difficulty value across categories as the difficulty estimate for the item. Using the average value and the median value in effect introduced two corresponding ISSs. In this study, we used computer simulation compare four ISSs based on GRM. We also discussed the effect of "shadow pool" on the uniformity of pool usage as well as the influence of different item parameter distributions and different ability estimation methods on the evaluation criteria of CAT. In the simulation process, Monte Carlo method was adopted to simulate the entire CAT process; 1,000 examinees drawn from standard normal distribution and four 1,000-sized item pools of different item parameter distributions were also simulated. The assumption of the simulation is that a polytomous item is comprised of six ordered categories. In addition, ability estimates were derived using two methods. They were expected a posteriori Bayesian (EAP) and maximum likelihood estimation (MLE). In MLE, the Newton-Raphson iteration method and the Fisher Score iteration method were employed, respectively, to solve the likelihood equation. Moreover, the CAT process was simulated with each examinee 30 times to eliminate random error. The IISs were evaluated by four indices usually used in CAT from four aspects--the accuracy of ability estimation, the stability of IIS, the usage of item pool, and the test efficiency. Simulation results showed adequate evaluation of the ISS that matched the estimate of an examinee's current trait level with the difficulty values across categories. Setting "shadow pool" in ISS was able to improve the uniformity of pool utilization. Finally, different distributions of the item parameter and different ability estimation methods affected the evaluation indices of CAT. (PsycINFO Database Record (c) 2007 APA, all rights reserved)10acomputerized adaptive testing10aitem selection strategy1 aPing, Chen1 aShuliang, Ding1 aHaijing, Lin1 aJie, Zhou uhttp://www.iacat.org/content/item-selection-strategies-computerized-adaptive-testing-based-graded-response-model02405nas a2200337 4500008004100000020002200041245010800063210006900171250001500240260000800255300001100263490000700274520139300281653002101674653002101695653001001716653001101726653001801737653001101755653000901766653001601775653003001791653002801821100001801849700001701867700001801884700001401902700001701916700001701933856011701950 2006 eng d a0962-9343 (Print)00aMultidimensional computerized adaptive testing of the EORTC QLQ-C30: basic developments and evaluations0 aMultidimensional computerized adaptive testing of the EORTC QLQC a2006/03/21 cApr a315-290 v153 aOBJECTIVE: Self-report questionnaires are widely used to measure health-related quality of life (HRQOL). Ideally, such questionnaires should be adapted to the individual patient and at the same time scores should be directly comparable across patients. This may be achieved using computerized adaptive testing (CAT). Usually, CAT is carried out for a single domain at a time. However, many HRQOL domains are highly correlated. Multidimensional CAT may utilize these correlations to improve measurement efficiency. We investigated the possible advantages and difficulties of multidimensional CAT. STUDY DESIGN AND SETTING: We evaluated multidimensional CAT of three scales from the EORTC QLQ-C30: the physical functioning, emotional functioning, and fatigue scales. Analyses utilised a database with 2958 European cancer patients. RESULTS: It was possible to obtain scores for the three domains with five to seven items administered using multidimensional CAT that were very close to the scores obtained using all 12 items and with no or little loss of measurement precision. CONCLUSION: The findings suggest that multidimensional CAT may significantly improve measurement precision and efficiency and encourage further research into multidimensional CAT. Particularly, the estimation of the model underlying the multidimensional CAT and the conceptual aspects need further investigations.10a*Quality of Life10a*Self Disclosure10aAdult10aFemale10aHealth Status10aHumans10aMale10aMiddle Aged10aQuestionnaires/*standards10aUser-Computer Interface1 aPetersen, M A1 aGroenvold, M1 aAaronson, N K1 aFayers, P1 aSprangers, M1 aBjorner, J B uhttp://www.iacat.org/content/multidimensional-computerized-adaptive-testing-eortc-qlq-c30-basic-developments-and02251nas a2200301 4500008004500000020001400045245009800059210007100157300001200228490000700240520127200247653003201519653002601551653002801577653001801605653002501623653001601648653001201664653001501676653001801691653001601709653001801725653002701743653001101770100001901781700001601800856013301816 2006 Spandsh a0212-972800aTécnicas para detectar patrones de respuesta atípicos [Aberrant patterns detection methods]0 aTécnicas para detectar patrones de respuesta atípicos Aberrant p a143-1540 v223 aLa identificación de patrones de respuesta atípicos es de gran utilidad para la construcción de tests y de bancos de ítems con propiedades psicométricas así como para el análisis de validez de los mismos. En este trabajo de revisión se han recogido los más relevantes y novedosos métodos de ajuste de personas que se han elaborado dentro de cada uno de los principales ámbitos de trabajo de la Psicometría: el escalograma de Guttman, la Teoría Clásica de Tests (TCT), la Teoría de la Generalizabilidad (TG), la Teoría de Respuesta al Ítem (TRI), los Modelos de Respuesta al Ítem No Paramétricos (MRINP), los Modelos de Clase Latente de Orden Restringido (MCL-OR) y el Análisis de Estructura de Covarianzas (AEC).Aberrant patterns detection has a great usefulness in order to make tests and item banks with psychometric characteristics and validity analysis of tests and items. The most relevant and newest person-fit methods have been reviewed. All of them have been made in each one of main areas of Psychometry: Guttman's scalogram, Classical Test Theory (CTT), Generalizability Theory (GT), Item Response Theory (IRT), Non-parametric Response Models (NPRM), Order-Restricted Latent Class Models (OR-LCM) and Covariance Structure Analysis (CSA).10aaberrant patterns detection10aClassical Test Theory10ageneralizability theory10aItem Response10aItem Response Theory10aMathematics10amethods10aperson-fit10aPsychometrics10apsychometry10aTest Validity10atest validity analysis10aTheory1 aNúñez, R M N1 aPina, J A L uhttp://www.iacat.org/content/t%C3%A9cnicas-para-detectar-patrones-de-respuesta-at%C3%ADpicos-aberrant-patterns-detection-methods02242nas a2200205 4500008004100000020004600041245009500087210006900182260002700251300001200278490000700290520149400297653002701791653003001818653001701848653002501865100001901890700001001909856011701919 2005 eng d a1560-4292 (Print); 1560-4306 (Electronic)00aA Bayesian student model without hidden nodes and its comparison with item response theory0 aBayesian student model without hidden nodes and its comparison w bIOS Press: Netherlands a291-3230 v153 aThe Bayesian framework offers a number of techniques for inferring an individual's knowledge state from evidence of mastery of concepts or skills. A typical application where such a technique can be useful is Computer Adaptive Testing (CAT). A Bayesian modeling scheme, POKS, is proposed and compared to the traditional Item Response Theory (IRT), which has been the prevalent CAT approach for the last three decades. POKS is based on the theory of knowledge spaces and constructs item-to-item graph structures without hidden nodes. It aims to offer an effective knowledge assessment method with an efficient algorithm for learning the graph structure from data. We review the different Bayesian approaches to modeling student ability assessment and discuss how POKS relates to them. The performance of POKS is compared to the IRT two parameter logistic model. Experimental results over a 34 item Unix test and a 160 item French language test show that both approaches can classify examinees as master or non-master effectively and efficiently, with relatively comparable performance. However, more significant differences are found in favor of POKS for a second task that consists in predicting individual question item outcome. Implications of these results for adaptive testing and student modeling are discussed, as well as the limitations and advantages of POKS, namely the issue of integrating concepts into its structure. (PsycINFO Database Record (c) 2007 APA, all rights reserved)10aBayesian Student Model10acomputer adaptive testing10ahidden nodes10aItem Response Theory1 aDesmarais, M C1 aPu, X uhttp://www.iacat.org/content/bayesian-student-model-without-hidden-nodes-and-its-comparison-item-response-theory00685nas a2200121 4500008004100000245017000041210006900211260011000280100001800390700002000408700001600428856011900444 2005 eng d00aComputerizing statewide assessments in Minnesota: A report on the feasibility of converting the Minnesota Comprehensive Assessments to a computerized adaptive format0 aComputerizing statewide assessments in Minnesota A report on the bOffice of Educational Accountability, College of Education and Human Development, University of Minnesota1 aPeterson, K A1 aL., Davison., M1 aHjelseth, L uhttp://www.iacat.org/content/computerizing-statewide-assessments-minnesota-report-feasibility-converting-minnesota01743nas a2200229 4500008004100000245008300041210006900124300001100193490000700204520103000211653003401241100001301275700001401288700001501302700001701317700001501334700001501349700001401364700001301378700001301391856010901404 2005 eng d00aAn item response theory-based pain item bank can enhance measurement precision0 aitem response theorybased pain item bank can enhance measurement a278-880 v303 aCancer-related pain is often under-recognized and undertreated. This is partly due to the lack of appropriate assessments, which need to be comprehensive and precise yet easily integrated into clinics. Computerized adaptive testing (CAT) can enable precise-yet-brief assessments by only selecting the most informative items from a calibrated item bank. The purpose of this study was to create such a bank. The sample included 400 cancer patients who were asked to complete 61 pain-related items. Data were analyzed using factor analysis and the Rasch model. The final bank consisted of 43 items which satisfied the measurement requirement of factor analysis and the Rasch model, demonstrated high internal consistency and reasonable item-total correlations, and discriminated patients with differing degrees of pain. We conclude that this bank demonstrates good psychometric properties, is sensitive to pain reported by patients, and can be used as the foundation for a CAT pain-testing platform for use in clinical practice.10acomputerized adaptive testing1 aLai, J-S1 aDineen, K1 aReeve, B B1 aVon Roenn, J1 aShervin, D1 aMcGuire, M1 aBode, R K1 aPaice, J1 aCella, D uhttp://www.iacat.org/content/item-response-theory-based-pain-item-bank-can-enhance-measurement-precision02902nas a2200409 4500008004100000245012300041210006900164260000800233300001000241490000700251520159700258653004701855653001001902653000901912653001901921653003101940653002601971653001101997653002902008653001102037653000902048653001602057653003902073653001402112653002502126653002702151653003002178653003202208653002802240653002202268100001502290700001602305700001702321700001602338700001502354856012302369 2005 eng d00aMeasuring physical function in patients with complex medical and postsurgical conditions: a computer adaptive approach0 aMeasuring physical function in patients with complex medical and cOct a741-80 v843 aOBJECTIVE: To examine whether the range of disability in the medically complex and postsurgical populations receiving rehabilitation is adequately sampled by the new Activity Measure--Post-Acute Care (AM-PAC), and to assess whether computer adaptive testing (CAT) can derive valid patient scores using fewer questions. DESIGN: Observational study of 158 subjects (mean age 67.2 yrs) receiving skilled rehabilitation services in inpatient (acute rehabilitation hospitals, skilled nursing facility units) and community (home health services, outpatient departments) settings for recent-onset or worsening disability from medical (excluding neurological) and surgical (excluding orthopedic) conditions. Measures were interviewer-administered activity questions (all patients) and physical functioning portion of the SF-36 (outpatients) and standardized chart items (11 Functional Independence Measure (FIM), 19 Standardized Outcome and Assessment Information Set (OASIS) items, and 22 Minimum Data Set (MDS) items). Rasch modeling analyzed all data and the relationship between person ability estimates and average item difficulty. CAT assessed the ability to derive accurate patient scores using a sample of questions. RESULTS: The 163-item activity item pool covered the range of physical movement and personal and instrumental activities. CAT analysis showed comparable scores between estimates using 10 items or the total item pool. CONCLUSION: The AM-PAC can assess a broad range of function in patients with complex medical illness. CAT achieves valid patient scores using fewer questions.10aActivities of Daily Living/*classification10aAdult10aAged10aCohort Studies10aContinuity of Patient Care10aDisability Evaluation10aFemale10aHealth Services Research10aHumans10aMale10aMiddle Aged10aPostoperative Care/*rehabilitation10aPrognosis10aRecovery of Function10aRehabilitation Centers10aRehabilitation/*standards10aSensitivity and Specificity10aSickness Impact Profile10aTreatment Outcome1 aSiebens, H1 aAndres, P L1 aPengsheng, N1 aCoster, W J1 aHaley, S M uhttp://www.iacat.org/content/measuring-physical-function-patients-complex-medical-and-postsurgical-conditions-computer00511nas a2200157 4500008004100000020001000041245004300051210004300094260002600137653003000163653002800193653003400221653001700255100001200272856006900284 2005 eng d a05-0500aRecent trends in comparability studies0 aRecent trends in comparability studies bPearsoncAugust, 200510acomputer adaptive testing10aComputerized assessment10adifferential item functioning10aMode effects1 aPaek, P uhttp://www.iacat.org/content/recent-trends-comparability-studies10201nas a2200553 4500008004100000245016300041210006900204300001400273490000700287520862500294653001808919653003208937100001608969700001608985700001409001700001609015700001409031700001509045700001609060700001409076700001509090700001509105700001409120700001709134700001709151700001709168700001609185700001709201700001609218700001509234700001609249700001709265700002309282700001609305700002209321700001609343700001609359700001709375700001309392700001509405700001309420700001609433700001209449700001409461700001609475700002009491700001209511856012409523 2005 eng d00aToward efficient and comprehensive measurement of the alcohol problems continuum in college students: The Brief Young Adult Alcohol Consequences Questionnaire0 aToward efficient and comprehensive measurement of the alcohol pr a1180-11890 v293 aBackground: Although a number of measures of alcohol problems in college students have been studied, the psychometric development and validation of these scales have been limited, for the most part, to methods based on classical test theory. In this study, we conducted analyses based on item response theory to select a set of items for measuring the alcohol problem severity continuum in college students that balances comprehensiveness and efficiency and is free from significant gender bias., Method: We conducted Rasch model analyses of responses to the 48-item Young Adult Alcohol Consequences Questionnaire by 164 male and 176 female college students who drank on at least a weekly basis. An iterative process using item fit statistics, item severities, item discrimination parameters, model residuals, and analysis of differential item functioning by gender was used to pare the items down to those that best fit a Rasch model and that were most efficient in discriminating among levels of alcohol problems in the sample., Results: The process of iterative Rasch model analyses resulted in a final 24-item scale with the data fitting the unidimensional Rasch model very well. The scale showed excellent distributional properties, had items adequately matched to the severity of alcohol problems in the sample, covered a full range of problem severity, and appeared highly efficient in retaining all of the meaningful variance captured by the original set of 48 items., Conclusions: The use of Rasch model analyses to inform item selection produced a final scale that, in both its comprehensiveness and its efficiency, should be a useful tool for researchers studying alcohol problems in college students. To aid interpretation of raw scores, examples of the types of alcohol problems that are likely to be experienced across a range of selected scores are provided., (C)2005Research Society on AlcoholismAn important, sometimes controversial feature of all psychological phenomena is whether they are categorical or dimensional. A conceptual and psychometric framework is described for distinguishing whether the latent structure behind manifest categories (e.g., psychiatric diagnoses, attitude groups, or stages of development) is category-like or dimension-like. Being dimension-like requires (a) within-category heterogeneity and (b) between-category quantitative differences. Being category-like requires (a) within-category homogeneity and (b) between-category qualitative differences. The relation between this classification and abrupt versus smooth differences is discussed. Hybrid structures are possible. Being category-like is itself a matter of degree; the authors offer a formalized framework to determine this degree. Empirical applications to personality disorders, attitudes toward capital punishment, and stages of cognitive development illustrate the approach., (C) 2005 by the American Psychological AssociationThe authors conducted Rasch model ( G. Rasch, 1960) analyses of items from the Young Adult Alcohol Problems Screening Test (YAAPST; S. C. Hurlbut & K. J. Sher, 1992) to examine the relative severity and ordering of alcohol problems in 806 college students. Items appeared to measure a single dimension of alcohol problem severity, covering a broad range of the latent continuum. Items fit the Rasch model well, with less severe symptoms reliably preceding more severe symptoms in a potential progression toward increasing levels of problem severity. However, certain items did not index problem severity consistently across demographic subgroups. A shortened, alternative version of the YAAPST is proposed, and a norm table is provided that allows for a linking of total YAAPST scores to expected symptom expression., (C) 2004 by the American Psychological AssociationA didactic on latent growth curve modeling for ordinal outcomes is presented. The conceptual aspects of modeling growth with ordinal variables and the notion of threshold invariance are illustrated graphically using a hypothetical example. The ordinal growth model is described in terms of 3 nested models: (a) multivariate normality of the underlying continuous latent variables (yt) and its relationship with the observed ordinal response pattern (Yt), (b) threshold invariance over time, and (c) growth model for the continuous latent variable on a common scale. Algebraic implications of the model restrictions are derived, and practical aspects of fitting ordinal growth models are discussed with the help of an empirical example and Mx script ( M. C. Neale, S. M. Boker, G. Xie, & H. H. Maes, 1999). The necessary conditions for the identification of growth models with ordinal data and the methodological implications of the model of threshold invariance are discussed., (C) 2004 by the American Psychological AssociationRecent research points toward the viability of conceptualizing alcohol problems as arrayed along a continuum. Nevertheless, modern statistical techniques designed to scale multiple problems along a continuum (latent trait modeling; LTM) have rarely been applied to alcohol problems. This study applies LTM methods to data on 110 problems reported during in-person interviews of 1,348 middle-aged men (mean age = 43) from the general population. The results revealed a continuum of severity linking the 110 problems, ranging from heavy and abusive drinking, through tolerance and withdrawal, to serious complications of alcoholism. These results indicate that alcohol problems can be arrayed along a dimension of severity and emphasize the relevance of LTM to informing the conceptualization and assessment of alcohol problems., (C) 2004 by the American Psychological AssociationItem response theory (IRT) is supplanting classical test theory as the basis for measures development. This study demonstrated the utility of IRT for evaluating DSM-IV diagnostic criteria. Data on alcohol, cannabis, and cocaine symptoms from 372 adult clinical participants interviewed with the Composite International Diagnostic Interview-Expanded Substance Abuse Module (CIDI-SAM) were analyzed with Mplus ( B. Muthen & L. Muthen, 1998) and MULTILOG ( D. Thissen, 1991) software. Tolerance and legal problems criteria were dropped because of poor fit with a unidimensional model. Item response curves, test information curves, and testing of variously constrained models suggested that DSM-IV criteria in the CIDI-SAM discriminate between only impaired and less impaired cases and may not be useful to scale case severity. IRT can be used to study the construct validity of DSM-IV diagnoses and to identify diagnostic criteria with poor performance., (C) 2004 by the American Psychological AssociationThis study examined the psychometric characteristics of an index of substance use involvement using item response theory. The sample consisted of 292 men and 140 women who qualified for a Diagnostic and Statistical Manual of Mental Disorders (3rd ed., rev.; American Psychiatric Association, 1987) substance use disorder (SUD) diagnosis and 293 men and 445 women who did not qualify for a SUD diagnosis. The results indicated that men had a higher probability of endorsing substance use compared with women. The index significantly predicted health, psychiatric, and psychosocial disturbances as well as level of substance use behavior and severity of SUD after a 2-year follow-up. Finally, this index is a reliable and useful prognostic indicator of the risk for SUD and the medical and psychosocial sequelae of drug consumption., (C) 2002 by the American Psychological AssociationComparability, validity, and impact of loss of information of a computerized adaptive administration of the Minnesota Multiphasic Personality Inventory-2 (MMPI-2) were assessed in a sample of 140 Veterans Affairs hospital patients. The countdown method ( Butcher, Keller, & Bacon, 1985) was used to adaptively administer Scales L (Lie) and F (Frequency), the 10 clinical scales, and the 15 content scales. Participants completed the MMPI-2 twice, in 1 of 2 conditions: computerized conventional test-retest, or computerized conventional-computerized adaptive. Mean profiles and test-retest correlations across modalities were comparable. Correlations between MMPI-2 scales and criterion measures supported the validity of the countdown method, although some attenuation of validity was suggested for certain health-related items. Loss of information incurred with this mode of adaptive testing has minimal impact on test validity. Item and time savings were substantial., (C) 1999 by the American Psychological Association10aPsychometrics10aSubstance-Related Disorders1 aKahler, C W1 aStrong, D R1 aRead, J P1 aDe Boeck, P1 aWilson, M1 aActon, G S1 aPalfai, T P1 aWood, M D1 aMehta, P D1 aNeale, M C1 aFlay, B R1 aConklin, C A1 aClayton, R R1 aTiffany, S T1 aShiffman, S1 aKrueger, R F1 aNichol, P E1 aHicks, B M1 aMarkon, K E1 aPatrick, C J1 aIacono, William, G1 aMcGue, Matt1 aLangenbucher, J W1 aLabouvie, E1 aMartin, C S1 aSanjuan, P M1 aBavly, L1 aKirisci, L1 aChung, T1 aVanyukov, M1 aDunn, M1 aTarter, R1 aHandel, R W1 aBen-Porath, Y S1 aWatt, M uhttp://www.iacat.org/content/toward-efficient-and-comprehensive-measurement-alcohol-problems-continuum-college-students01668nas a2200229 4500008004100000245005500041210005300096300000800149490000700157520102900164653002101193653001201214653003001226653001801256653000901274100001701283700001201300700001501312700001601327700001401343856008101357 2004 eng d00aAssisted self-adapted testing: A comparative study0 aAssisted selfadapted testing A comparative study a2-90 v203 aA new type of self-adapted test (S-AT), called Assisted Self-Adapted Test (AS-AT), is presented. It differs from an ordinary S-AT in that prior to selecting the difficulty category, the computer advises examinees on their best difficulty category choice, based on their previous performance. Three tests (computerized adaptive test, AS-AT, and S-AT) were compared regarding both their psychometric (precision and efficiency) and psychological (anxiety) characteristics. Tests were applied in an actual assessment situation, in which test scores determined 20% of term grades. A sample of 173 high school students participated. Neither differences in posttest anxiety nor ability were obtained. Concerning precision, AS-AT was as precise as CAT, and both revealed more precision than S-AT. It was concluded that AS-AT acted as a CAT concerning precision. Some hints, but not conclusive support, of the psychological similarity between AS-AT and S-AT was also found. (PsycINFO Database Record (c) 2005 APA ) (journal abstract)10aAdaptive Testing10aAnxiety10aComputer Assisted Testing10aPsychometrics10aTest1 aHontangas, P1 aOlea, J1 aPonsoda, V1 aRevuelta, J1 aWise, S L uhttp://www.iacat.org/content/assisted-self-adapted-testing-comparative-study02593nas a2200469 4500008004100000245007200041210006900113300001000182490000600192520108600198653002501284653001001309653001501319653002101334653002201355653005901377653007001436653003301506653001101539653001101550653001301561653000901574653002701583653002201610653005501632653001901687653001501706653006601721653001801787653003701805653004101842653003001883653001301913100001501926700001301941700001801954700001501972700001401987700001402001700001302015856009502028 2004 eng d00aComputerized adaptive measurement of depression: A simulation study0 aComputerized adaptive measurement of depression A simulation stu a13-230 v43 aBackground: Efficient, accurate instruments for measuring depression are increasingly importantin clinical practice. We developed a computerized adaptive version of the Beck DepressionInventory (BDI). We examined its efficiency and its usefulness in identifying Major DepressiveEpisodes (MDE) and in measuring depression severity.Methods: Subjects were 744 participants in research studies in which each subject completed boththe BDI and the SCID. In addition, 285 patients completed the Hamilton Depression Rating Scale.Results: The adaptive BDI had an AUC as an indicator of a SCID diagnosis of MDE of 88%,equivalent to the full BDI. The adaptive BDI asked fewer questions than the full BDI (5.6 versus 21items). The adaptive latent depression score correlated r = .92 with the BDI total score and thelatent depression score correlated more highly with the Hamilton (r = .74) than the BDI total scoredid (r = .70).Conclusions: Adaptive testing for depression may provide greatly increased efficiency withoutloss of accuracy in identifying MDE or in measuring depression severity.10a*Computer Simulation10aAdult10aAlgorithms10aArea Under Curve10aComparative Study10aDepressive Disorder/*diagnosis/epidemiology/psychology10aDiagnosis, Computer-Assisted/*methods/statistics & numerical data10aFactor Analysis, Statistical10aFemale10aHumans10aInternet10aMale10aMass Screening/methods10aPatient Selection10aPersonality Inventory/*statistics & numerical data10aPilot Projects10aPrevalence10aPsychiatric Status Rating Scales/*statistics & numerical data10aPsychometrics10aResearch Support, Non-U.S. Gov't10aResearch Support, U.S. Gov't, P.H.S.10aSeverity of Illness Index10aSoftware1 aGardner, W1 aShear, K1 aKelleher, K J1 aPajer, K A1 aMammen, O1 aBuysse, D1 aFrank, E uhttp://www.iacat.org/content/computerized-adaptive-measurement-depression-simulation-study01251nas a2200157 4500008003900000245006400039210006300103300001200166490000700178520076400185100002500949700002200974700002200996700002201018856005301040 2004 d00aComputerized Adaptive Testing With Multiple-Form Structures0 aComputerized Adaptive Testing With MultipleForm Structures a147-1640 v283 aA multiple-form structure (MFS) is an orderedcollection or network of testlets (i.e., sets of items).An examinee’s progression through the networkof testlets is dictated by the correctness of anexaminee’s answers, thereby adapting the test tohis or her trait level. The collection of pathsthrough the network yields the set of all possibletest forms, allowing test specialists the opportunityto review them before they are administered. Also,limiting the exposure of an individual MFS to aspecific period of time can enhance test security.This article provides an overview of methods thathave been developed to generate parallel MFSs.The approach is applied to the assembly of anexperimental computerized Law School Admission Test (LSAT).
1 aArmstrong, Ronald, D1 aJones, Douglas, H1 aKoppel, Nicole, B1 aPashley, Peter, J uhttp://apm.sagepub.com/content/28/3/147.abstract01543nas a2200229 4500008004100000020002200041245006400063210006300127260002600190300001200216490000700228520081800235653003401053653003001087653002801117653001301145100001901158700001501177700001601192700001701208856008801225 2004 eng d a0146-6216 (Print)00aComputerized adaptive testing with multiple-form structures0 aComputerized adaptive testing with multipleform structures bSage Publications: US a147-1640 v283 aA multiple-form structure (MFS) is an ordered collection or network of testlets (i.e., sets of items). An examinee's progression through the network of testlets is dictated by the correctness of an examinee's answers, thereby adapting the test to his or her trait level. The collection of paths through the network yields the set of all possible test forms, allowing test specialists the opportunity to review them before they are administered. Also, limiting the exposure of an individual MFS to a specific period of time can enhance test security. This article provides an overview of methods that have been developed to generate parallel MFSs. The approach is applied to the assembly of an experimental computerized Law School Admission Test (LSAT). (PsycINFO Database Record (c) 2007 APA, all rights reserved)10acomputerized adaptive testing10aLaw School Admission Test10amultiple-form structure10atestlets1 aArmstrong, R D1 aJones, D H1 aKoppel, N B1 aPashley, P J uhttp://www.iacat.org/content/computerized-adaptive-testing-multiple-form-structures01908nas a2200217 4500008004100000020004600041245010100087210006900188260002600257300001200283490000700295520111100302653002401413653003201437653001301469653003801482100001701520700001501537700001401552856012401566 2004 eng d a0021-9762 (Print); 1097-4679 (Electronic)00aComputers in clinical assessment: Historical developments, present status, and future challenges0 aComputers in clinical assessment Historical developments present bJohn Wiley & Sons: US a331-3450 v603 aComputerized testing methods have long been regarded as a potentially powerful asset for providing psychological assessment services. Ever since computers were first introduced and adapted to the field of assessment psychology in the 1950s, they have been a valuable aid for scoring, data processing, and even interpretation of test results. The history and status of computer-based personality and neuropsychological tests are discussed in this article. Several pertinent issues involved in providing test interpretation by computer are highlighted. Advances in computer-based test use, such as computerized adaptive testing, are described and problems noted. Today, there is great interest in expanding the availability of psychological assessment applications on the Internet. Although these applications show great promise, there are a number of problems associated with providing psychological tests on the Internet that need to be addressed by psychologists before the Internet can become a major medium for psychological service delivery. (PsycINFO Database Record (c) 2007 APA, all rights reserved)10aclinical assessment10acomputerized testing method10aInternet10apsychological assessment services1 aButcher, J N1 aPerry, J L1 aHahn, J A uhttp://www.iacat.org/content/computers-clinical-assessment-historical-developments-present-status-and-future-challenges00585nas a2200133 4500008004100000245010800041210006900149260003200218100002400250700001700274700001800291700001600309856012600325 2004 eng d00aA learning environment for english for academic purposes based on adaptive tests and task-based systems0 alearning environment for english for academic purposes based on b Springer Berlin Heidelberg1 aPITON-GONÇALVES, J1 aALUISIO, S M1 aMENDONCA, L H1 aNOVAES, O O uhttp://www.iacat.org/content/learning-environment-english-academic-purposes-based-adaptive-tests-and-task-based-systems-000717nas a2200205 4500008004100000020002200041245010800063210006900171260003300240300000900273490000900282100002400291700002300315700002700338700002900365700002100394700002300415700002300438856005000461 2004 eng d a978-3-540-22948-300aA Learning Environment for English for Academic Purposes Based on Adaptive Tests and Task-Based Systems0 aLearning Environment for English for Academic Purposes Based on bSpringer Berlin / Heidelberg a1-110 v32201 aGonçalves, Jean, P1 aAluisio, Sandra, M1 aOliveira, Leandro, H M1 aOliveira Jr., Osvaldo, N1 aLester, James, C1 aVicari, Rosa Maria1 aParaguaçu, Fábio uhttp://dx.doi.org/10.1007/978-3-540-30139-4_100541nas a2200181 4500008004100000245005000041210004800091300001000139490000700149653003400156100001400190700001500204700001500219700001400234700002600248700001300274856007200287 2004 eng d00aSiette: a web-based tool for adaptive testing0 aSiette a webbased tool for adaptive testing a29-610 v1410acomputerized adaptive testing1 aConejo, R1 aGuzmán, E1 aMillán, E1 aTrella, M1 aPérez-De-La-Cruz, JL1 aRíos, A uhttp://www.iacat.org/content/siette-web-based-tool-adaptive-testing00550nas a2200121 4500008004100000245006500041210006500106260011000171100001700281700001900298700002000317856009100337 2003 eng d00aAdaptive exploration of assessment results under uncertainty0 aAdaptive exploration of assessment results under uncertainty aProceedings 3rd IEEE ternational Conference on Advanced Learning Technologies, ICALT '03, 460-461, 2003.1 aLamboudis, D1 aEconomides, AA1 aPapastergiou, A uhttp://www.iacat.org/content/adaptive-exploration-assessment-results-under-uncertainty00483nas a2200109 4500008004100000245009300041210006900134260001500203100001700218700001600235856012200251 2003 eng d00aCalibrating CAT pools and online pretest items using marginal maximum likelihood methods0 aCalibrating CAT pools and online pretest items using marginal ma aChicago IL1 aPommerich, M1 aSegall, D O uhttp://www.iacat.org/content/calibrating-cat-pools-and-online-pretest-items-using-marginal-maximum-likelihood-methods00509nas a2200133 4500008004100000245008600041210006900127260001500196100001300211700001500224700001500239700001400254856010700268 2003 eng d00aComparison of multi-stage tests with computer adaptive and paper and pencil tests0 aComparison of multistage tests with computer adaptive and paper aChicago IL1 aRotou, O1 aPatsula, L1 aSteffen, M1 aRizavi, S uhttp://www.iacat.org/content/comparison-multi-stage-tests-computer-adaptive-and-paper-and-pencil-tests00405nas a2200109 4500008004100000245003400041210003400075260009400109100001500203700001200218856006500230 2003 eng d00aComputerized adaptive testing0 aComputerized adaptive testing aR. Fernández-Ballesteros (Ed.): Encyclopaedia of Psychological Assessment. London: Sage.1 aPonsoda, V1 aOlea, J uhttp://www.iacat.org/content/computerized-adaptive-testing-000545nas a2200145 4500008004100000245009500041210006900136260001500205100001000220700001600230700001400246700001300260700001500273856011100288 2003 eng d00aEvaluating computerized adaptive testing design for the MCAT with realistic simulated data0 aEvaluating computerized adaptive testing design for the MCAT wit aChicago IL1 aLu, Y1 aPitoniak, M1 aRizavi, S1 aWay, W D1 aSteffan, M uhttp://www.iacat.org/content/evaluating-computerized-adaptive-testing-design-mcat-realistic-simulated-data00524nas a2200109 4500008004100000245014200041210006900183260001500252100001300267700001300280856012100293 2003 eng d00aEvaluating the comparability of English- and French-speaking examinees on a science achievement test administered using two-stage testing0 aEvaluating the comparability of English and Frenchspeaking exami aChicago IL1 aPuhan, G1 aGierl, M uhttp://www.iacat.org/content/evaluating-comparability-english-and-french-speaking-examinees-science-achievement-test01816nas a2200253 4500008004100000245013100041210006900172300001000241490000600251520095700257653001501214653002901229653002501258653001501283653002001298653001101318653003101329100001401360700001601374700001401390700001401404700002101418856012301439 2003 eng d00aAn examination of exposure control and content balancing restrictions on item selection in CATs using the partial credit model0 aexamination of exposure control and content balancing restrictio a24-420 v43 aThe purpose of the present investigation was to systematically examine the effectiveness of the Sympson-Hetter technique and rotated content balancing relative to no exposure control and no content rotation conditions in a computerized adaptive testing system (CAT) based on the partial credit model. A series of simulated fixed and variable length CATs were run using two data sets generated to multiple content areas for three sizes of item pools. The 2 (exposure control) X 2 (content rotation) X 2 (test length) X 3 (item pool size) X 2 (data sets) yielded a total of 48 conditions. Results show that while both procedures can be used with no deleterious effect on measurement precision, the gains in exposure control, pool utilization, and item overlap appear quite modest. Difficulties involved with setting the exposure control parameters in small item pools make questionable the utility of the Sympson-Hetter technique with similar item pools.10a*Computers10a*Educational Measurement10a*Models, Theoretical10aAutomation10aDecision Making10aHumans10aReproducibility of Results1 aDavis, LL1 aPastor, D A1 aDodd, B G1 aChiang, C1 aFitzpatrick, S J uhttp://www.iacat.org/content/examination-exposure-control-and-content-balancing-restrictions-item-selection-cats-using00452nas a2200145 4500008004100000245005100041210005100092260001500143100001500158700001400173700001200187700001500199700001500214856007700229 2003 eng d00aMaintaining scale in computer adaptive testing0 aMaintaining scale in computer adaptive testing aChicago IL1 aSmith, R L1 aRizavi, S1 aPaez, R1 aDamiano, M1 aHerbert, E uhttp://www.iacat.org/content/maintaining-scale-computer-adaptive-testing00516nas a2200121 4500008004100000245009800041210006900139260001500208100001600223700001800239700001700257856012000274 2003 eng d00aRecalibration of IRT item parameters in CAT: Sparse data matrices and missing data treatments0 aRecalibration of IRT item parameters in CAT Sparse data matrices aChicago IL1 aHarmes, J C1 aParshall, C G1 aKromrey, J D uhttp://www.iacat.org/content/recalibration-irt-item-parameters-cat-sparse-data-matrices-and-missing-data-treatments00437nas a2200109 4500008004100000245006900041210006700110260002700177100001200204700001500216856009600231 2003 eng d00aTests adaptativos informatizados (Computerized adaptive testing)0 aTests adaptativos informatizados Computerized adaptive testing aMadrid: UNED Ediciones1 aOlea, J1 aPonsoda, V uhttp://www.iacat.org/content/tests-adaptativos-informatizados-computerized-adaptive-testing01982nas a2200289 4500008004100000245007600041210006900117260000800186300001200194490000700206520114900213653002401362653001301386653002101399653002001420653001501440653001001455653001001465653001101475653001101486653000901497653002401506653002601530100001601556700001501572856010501587 2002 eng d00aAssessing tobacco beliefs among youth using item response theory models0 aAssessing tobacco beliefs among youth using item response theory cNov aS21-S390 v683 aSuccessful intervention research programs to prevent adolescent smoking require well-chosen, psychometrically sound instruments for assessing smoking prevalence and attitudes. Twelve thousand eight hundred and ten adolescents were surveyed about their smoking beliefs as part of the Teenage Attitudes and Practices Survey project, a prospective cohort study of predictors of smoking initiation among US adolescents. Item response theory (IRT) methods are used to frame a discussion of questions that a researcher might ask when selecting an optimal item set. IRT methods are especially useful for choosing items during instrument development, trait scoring, evaluating item functioning across groups, and creating optimal item subsets for use in specialized applications such as computerized adaptive testing. Data analytic steps for IRT modeling are reviewed for evaluating item quality and differential item functioning across subgroups of gender, age, and smoking status. Implications and challenges in the use of these methods for tobacco onset research and for assessing the developmental trajectories of smoking among youth are discussed.10a*Attitude to Health10a*Culture10a*Health Behavior10a*Questionnaires10aAdolescent10aAdult10aChild10aFemale10aHumans10aMale10aModels, Statistical10aSmoking/*epidemiology1 aPanter, A T1 aReeve, B B uhttp://www.iacat.org/content/assessing-tobacco-beliefs-among-youth-using-item-response-theory-models02085nas a2200229 4500008004100000245012900041210006900170300001200239490000700251520124000258653001801498653002101516653004501537653003001582653001801612653002501630653002601655100001601681700001401697700001901711856012501730 2002 eng d00aA comparison of item selection techniques and exposure control mechanisms in CATs using the generalized partial credit model0 acomparison of item selection techniques and exposure control mec a147-1630 v263 aThe use of more performance items in large-scale testing has led to an increase in the research investigating the use of polytomously scored items in computer adaptive testing (CAT). Because this research has to be complemented with information pertaining to exposure control, the present research investigated the impact of using five different exposure control algorithms in two sized item pools calibrated using the generalized partial credit model. The results of the simulation study indicated that the a-stratified design, in comparison to a no-exposure control condition, could be used to reduce item exposure and overlap, increase pool utilization, and only minorly degrade measurement precision. Use of the more restrictive exposure control algorithms, such as the Sympson-Hetter and conditional Sympson-Hetter, controlled exposure to a greater extent but at the cost of measurement precision. Because convergence of the exposure control parameters was problematic for some of the more restrictive exposure control algorithms, use of the more simplistic exposure control mechanisms, particularly when the test length to item pool size ratio is large, is recommended. (PsycINFO Database Record (c) 2005 APA ) (journal abstract)10a(Statistical)10aAdaptive Testing10aAlgorithms computerized adaptive testing10aComputer Assisted Testing10aItem Analysis10aItem Response Theory10aMathematical Modeling1 aPastor, D A1 aDodd, B G1 aChang, Hua-Hua uhttp://www.iacat.org/content/comparison-item-selection-techniques-and-exposure-control-mechanisms-cats-using-generalized00524nas a2200121 4500008004100000245005100041210005100092260013700143100001800280700001300298700001500311856007600326 2002 eng d00aInnovative item types for computerized testing0 aInnovative item types for computerized testing aIn W. J. van der Linden and C. A. W. Glas (Eds.), Computerized adaptive testing: Theory and practice. Norwell MA: Kluwer (in press).1 aParshall, C G1 aDavey, T1 aPashley, P uhttp://www.iacat.org/content/innovative-item-types-computerized-testing03062nas a2200325 4500008004100000020004100041245008100082210006900163250001500232260000800247300001100255490000700266520201300273653001502286653001002301653004002311653005702351653003302408653001102441653001102452653001802463653000902481653002802490653001202518653005502530100001502585700001802600700001502618856010302633 2002 eng d a0025-7079 (Print)0025-7079 (Linking)00aMultidimensional adaptive testing for mental health problems in primary care0 aMultidimensional adaptive testing for mental health problems in a2002/09/10 cSep a812-230 v403 aOBJECTIVES: Efficient and accurate instruments for assessing child psychopathology are increasingly important in clinical practice and research. For example, screening in primary care settings can identify children and adolescents with disorders that may otherwise go undetected. However, primary care offices are notorious for the brevity of visits and screening must not burden patients or staff with long questionnaires. One solution is to shorten assessment instruments, but dropping questions typically makes an instrument less accurate. An alternative is adaptive testing, in which a computer selects the items to be asked of a patient based on the patient's previous responses. This research used a simulation to test a child mental health screen based on this technology. RESEARCH DESIGN: Using half of a large sample of data, a computerized version was developed of the Pediatric Symptom Checklist (PSC), a parental-report psychosocial problem screen. With the unused data, a simulation was conducted to determine whether the Adaptive PSC can reproduce the results of the full PSC with greater efficiency. SUBJECTS: PSCs were completed by parents on 21,150 children seen in a national sample of primary care practices. RESULTS: Four latent psychosocial problem dimensions were identified through factor analysis: internalizing problems, externalizing problems, attention problems, and school problems. A simulated adaptive test measuring these traits asked an average of 11.6 questions per patient, and asked five or fewer questions for 49% of the sample. There was high agreement between the adaptive test and the full (35-item) PSC: only 1.3% of screening decisions were discordant (kappa = 0.93). This agreement was higher than that obtained using a comparable length (12-item) short-form PSC (3.2% of decisions discordant; kappa = 0.84). CONCLUSIONS: Multidimensional adaptive testing may be an accurate and efficient technology for screening for mental health problems in primary care settings.10aAdolescent10aChild10aChild Behavior Disorders/*diagnosis10aChild Health Services/*organization & administration10aFactor Analysis, Statistical10aFemale10aHumans10aLinear Models10aMale10aMass Screening/*methods10aParents10aPrimary Health Care/*organization & administration1 aGardner, W1 aKelleher, K J1 aPajer, K A uhttp://www.iacat.org/content/multidimensional-adaptive-testing-mental-health-problems-primary-care00458nas a2200097 4500008004100000245008900041210006900130260001900199100002300218856011900241 2002 eng d00aA “rearrangement procedure” for administering adaptive tests with review options0 arearrangement procedure for administering adaptive tests with re aNew Orleans LA1 aPapanastasiou, E C uhttp://www.iacat.org/content/%E2%80%9Crearrangement-procedure%E2%80%9D-administering-adaptive-tests-review-options00568nas a2200133 4500008004100000245012500041210006900166260001900235100001400254700001800268700001600286700001200302856012000314 2002 eng d00aThe robustness of the unidimensional 3PL IRT model when applied to two-dimensional data in computerized adaptive testing0 arobustness of the unidimensional 3PL IRT model when applied to t aNew Orleans LA1 aZhao, J C1 aMcMorris, R F1 aPruzek, R M1 aChen, R uhttp://www.iacat.org/content/robustness-unidimensional-3pl-irt-model-when-applied-two-dimensional-data-computerized00372nas a2200133 4500008004100000245003800041210003600079300001200115490000700127100001400134700001500148700001200163856006300175 2002 eng d00aSelf-adapted testing: An overview0 aSelfadapted testing An overview a107-1220 v121 aWise, S L1 aPonsoda, V1 aOlea, J uhttp://www.iacat.org/content/self-adapted-testing-overview00454nas a2200133 4500008004100000245005900041210005900100260001900159100001500178700001400193700001200207700001300219856008800232 2002 eng d00aUpdated item parameter estimates using sparse CAT data0 aUpdated item parameter estimates using sparse CAT data aNew Orleans LA1 aSmith, R L1 aRizavi, S1 aPaez, R1 aRotou, O uhttp://www.iacat.org/content/updated-item-parameter-estimates-using-sparse-cat-data01149nas a2200205 4500008004100000245008200041210006900123260005600192520043100248653002100679653003000700653001600730653001600746653001800762100001500780700001700795700001700812700001400829856010000843 2002 eng d00aThe work ahead: A psychometric infrastructure for computerized adaptive tests0 awork ahead A psychometric infrastructure for computerized adapti aMahwah, N.J. USAbLawrence Erlbaum Associates, Inc.3 a(From the chapter) Considers the past and future of computerized adaptive tests and computer-based tests and looks at issues and challenges confronting a testing program as it implements and operates a computer-based test. Recommendations for testing programs from The National Council of Measurement in Education Ad Hoc Committee on Computerized Adaptive Test Disclosure are appended. (PsycINFO Database Record (c) 2005 APA )10aAdaptive Testing10aComputer Assisted Testing10aEducational10aMeasurement10aPsychometrics1 aDrasgow, F1 aPotenza, M P1 aFreemer, J J1 aWard, W C uhttp://www.iacat.org/content/work-ahead-psychometric-infrastructure-computerized-adaptive-tests01617nas a2200193 4500008004100000245008600041210006900127300001200196490000700208520094500215653002101160653003001181653004101211653000901252653001701261100001601278700001701294856011201311 2001 eng d00aDifferences between self-adapted and computerized adaptive tests: A meta-analysis0 aDifferences between selfadapted and computerized adaptive tests a235-2470 v383 aSelf-adapted testing has been described as a variation of computerized adaptive testing that reduces test anxiety and thereby enhances test performance. The purpose of this study was to gain a better understanding of these proposed effects of self-adapted tests (SATs); meta-analysis procedures were used to estimate differences between SATs and computerized adaptive tests (CATs) in proficiency estimates and post-test anxiety levels across studies in which these two types of tests have been compared. After controlling for measurement error the results showed that SATs yielded proficiency estimates that were 0.12 standard deviation units higher and post-test anxiety levels that were 0.19 standard deviation units lower than those yielded by CATs. The authors speculate about possible reasons for these differences and discuss advantages and disadvantages of using SATs in operational settings. (PsycINFO Database Record (c) 2005 APA )10aAdaptive Testing10aComputer Assisted Testing10aScores computerized adaptive testing10aTest10aTest Anxiety1 aPitkin, A K1 aVispoel, W P uhttp://www.iacat.org/content/differences-between-self-adapted-and-computerized-adaptive-tests-meta-analysis00478nas a2200109 4500008004100000245009900041210006900140260001500209100001400224700001700238856011300255 2001 eng d00aAn examination of item review on a CAT using the specific information item selection algorithm0 aexamination of item review on a CAT using the specific informati aSeattle WA1 aBowles, R1 aPommerich, M uhttp://www.iacat.org/content/examination-item-review-cat-using-specific-information-item-selection-algorithm00480nas a2200109 4500008004100000245009900041210006900140260001500209100001400224700001700238856011500255 2001 eng d00aAn examination of item review on a CAT using the specific information item selection algorithm0 aexamination of item review on a CAT using the specific informati aSeattle WA1 aBowles, R1 aPommerich, M uhttp://www.iacat.org/content/examination-item-review-cat-using-specific-information-item-selection-algorithm-000583nas a2200121 4500008004100000245015600041210006900197260002600266100001500292700001500307700002000322856011900342 2001 eng d00aItem and passage selection algorithm simulations for a computerized adaptive version of the verbal section of the Medical College Admission Test (MCAT)0 aItem and passage selection algorithm simulations for a computeri aMCAT Monograph Series1 aSmith, R W1 aPlake, B S1 ade Ayala, R. J. uhttp://www.iacat.org/content/item-and-passage-selection-algorithm-simulations-computerized-adaptive-version-verbal00587nas a2200133 4500008004100000245013700041210006900178260001500247100001800262700001700280700001600297700001700313856012300330 2001 eng d00aNearest neighbors, simple strata, and probabilistic parameters: An empirical comparison of methods for item exposure control in CATs0 aNearest neighbors simple strata and probabilistic parameters An aSeattle WA1 aParshall, C G1 aKromrey, J D1 aHarmes, J C1 aSentovich, C uhttp://www.iacat.org/content/nearest-neighbors-simple-strata-and-probabilistic-parameters-empirical-comparison-methods00609nas a2200121 4500008004100000245017700041210006900218260002700287100001600314700001700330700001800347856012200365 2001 eng d00aOnline item parameter recalibration: Application of missing data treatments to overcome the effects of sparse data conditions in a computerized adaptive version of the MCAT0 aOnline item parameter recalibration Application of missing data aUnpublished manuscript1 aHarmes, J C1 aKromrey, J D1 aParshall, C G uhttp://www.iacat.org/content/online-item-parameter-recalibration-application-missing-data-treatments-overcome-effects00513nam a2200097 4500008004100000245009700041210006900138260006500207100002300272856012000295 2001 eng d00aA rearrangement procedure for administering adaptive tests when review options are permitted0 arearrangement procedure for administering adaptive tests when re aUnpublished doctoral dissertation, Michigan State University1 aPapanastasiou, E C uhttp://www.iacat.org/content/rearrangement-procedure-administering-adaptive-tests-when-review-options-are-permitted01360nas a2200205 4500008004100000245016700041210007000208300001000278490000700288520051600295653003000811653003100841653004800872100001800920700001900938700002600957700002600983700001401009856013101023 2001 eng d00aRequerimientos, aplicaciones e investigación en tests adaptativos informatizados [Requirements, applications, and investigation in computerized adaptive testing]0 aRequerimientos aplicaciones e investigación en tests adaptativos a11-280 v193 aSummarizes the main requirements and applications of computerized adaptive testing (CAT) with emphasis on the differences between CAT and conventional computerized tests. Psychometric properties of estimations based on CAT, item selection strategies, and implementation software are described. Results of CAT studies in Spanish-speaking samples are described. Implications for developing a CAT measuring the English vocabulary of Spanish-speaking students are discussed. (PsycINFO Database Record (c) 2005 APA )10aComputer Assisted Testing10aEnglish as Second Language10aPsychometrics computerized adaptive testing1 aOlea Díaz, J1 aPonsoda Gil, V1 aRevuelta Menéndez, J1 aHontangas Beltrán, P1 aAbad, F J uhttp://www.iacat.org/content/requerimientos-aplicaciones-e-investigaci%C3%B3n-en-tests-adaptativos-informatizados-requirements00525nas a2200109 4500008004100000245014700041210006900188300001300257490000900270100001600279856012000295 2001 eng d00aTest anxiety and test performance: Comparing paper-based and computer-adaptive versions of the Graduate Record Examinations (GRE) General test0 aTest anxiety and test performance Comparing paperbased and compu a249-273.0 v24 1 aPowers, D E uhttp://www.iacat.org/content/test-anxiety-and-test-performance-comparing-paper-based-and-computer-adaptive-versions00722nas a2200145 4500008004100000245020100041210006900242260005700311100001700368700001700385700001400402700002000416700001600436856012400452 2001 eng d00aTesting via the Internet: A literature review and analysis of issues for Department of Defense Internet testing of the Armed Services Vocational Aptitude Battery (ASVAB) in high schools (FR-01-12)0 aTesting via the Internet A literature review and analysis of iss aAlexandria VA: Human Resources Research Organization1 aMcBride, J R1 aPaddock, A F1 aWise, L L1 aStrickland, W J1 aWaters, B K uhttp://www.iacat.org/content/testing-internet-literature-review-and-analysis-issues-department-defense-internet-testing00450nas a2200145 4500008004100000245005800041210005400099300000900153490000700162100001700169700001500186700001200201700001400213856007700227 2000 eng d00aThe choice of item difficulty in self adapted testing0 achoice of item difficulty in self adapted testing a3-120 v161 aHontangas, P1 aPonsoda, V1 aOlea, J1 aWise, S L uhttp://www.iacat.org/content/choice-item-difficulty-self-adapted-testing00562nas a2200133 4500008004100000245012400041210006900165260001900234100001300253700001200266700001400278700001500292856012100307 2000 eng d00aClassification accuracy and test security for a computerized adaptive mastery test calibrated with different IRT models0 aClassification accuracy and test security for a computerized ada aNew Orleans LA1 aRobin, F1 aXing, D1 aScrams, D1 aPotenza, M uhttp://www.iacat.org/content/classification-accuracy-and-test-security-computerized-adaptive-mastery-test-calibrated02654nas a2200133 4500008004100000245007300041210006900114300000900183490000700192520217300199653003402372100001702406856009702423 2000 eng d00aA comparison of computerized adaptive testing and multistage testing0 acomparison of computerized adaptive testing and multistage testi a58290 v603 aThere is considerable evidence to show that computerized-adaptive testing (CAT) and multi-stage testing (MST) are viable frameworks for testing. With many testing organizations looking to move towards CAT or MST, it is important to know what framework is superior in different situations and at what cost in terms of measurement. What was needed is a comparison of the different testing procedures under various realistic testing conditions. This dissertation addressed the important problem of the increase or decrease in accuracy of ability estimation in using MST rather than CAT. The purpose of this study was to compare the accuracy of ability estimates produced by MST and CAT while keeping some variables fixed and varying others. A simulation study was conducted to investigate the effects of several factors on the accuracy of ability estimation using different CAT and MST designs. The factors that were manipulated are the number of stages, the number of subtests per stage, and the number of items per subtest. Kept constant were test length, distribution of subtest information, method of determining cut-points on subtests, amount of overlap between subtests, and method of scoring total test. The primary question of interest was, given a fixed test length, how many stages and many subtests per stage should there be to maximize measurement precision? Furthermore, how many items should there be in each subtest? Should there be more in the routing test or should there be more in the higher stage tests? Results showed that, in general, increasing the number of stages from two to three decreased the amount of errors in ability estimation. Increasing the number of subtests from three to five increased the accuracy of ability estimates as well as the efficiency of the MST designs relative to the P&P and CAT designs at most ability levels (-.75 to 2.25). Finally, at most ability levels (-.75 to 2.25), varying the number of items per stage had little effect on either the resulting accuracy of ability estimates or the relative efficiency of the MST designs to the P&P and CAT designs. (PsycINFO Database Record (c) 2003 APA, all rights reserved).10acomputerized adaptive testing1 aPatsula, L N uhttp://www.iacat.org/content/comparison-computerized-adaptive-testing-and-multistage-testing00518nas a2200109 4500008004100000245012100041210006900162260001900231100001700250700001600267856012500283 2000 eng d00aDetecting test-takers who have memorized items in computerized-adaptive testing and muti-stage testing: A comparison0 aDetecting testtakers who have memorized items in computerizedada aNew Orleans LA1 aPatsula, L N1 aMcLeod, L D uhttp://www.iacat.org/content/detecting-test-takers-who-have-memorized-items-computerized-adaptive-testing-and-muti-stage00609nas a2200145 4500008004100000245013100041210006900172260002000241100001400261700001600275700001400291700001400305700001900319856012500338 2000 eng d00aAn examination of exposure control and content balancing restrictions on item selection in CATs using the partial credit model0 aexamination of exposure control and content balancing restrictio aNew Orleans, LA1 aDavis, LL1 aPastor, D A1 aDodd, B G1 aChiang, C1 aFitzpatrick, S uhttp://www.iacat.org/content/examination-exposure-control-and-content-balancing-restrictions-item-selection-cats-using-000448nas a2200109 4500008004100000245007600041210006900117260002800186100001700214700001400231856009300245 2000 eng d00aFrom simulation to application: Examinees react to computerized testing0 aFrom simulation to application Examinees react to computerized t aNew Orleans, April 20001 aPommerich, M1 aBurden, T uhttp://www.iacat.org/content/simulation-application-examinees-react-computerized-testing00529nas a2200133 4500008004100000245010200041210006900143300001000212490000700222100001600229700001600245700001700261856011700278 2000 eng d00aItem exposure control in computer-adaptive testing: The use of freezing to augment stratification0 aItem exposure control in computeradaptive testing The use of fre a28-520 v401 aParshall, C1 aHarmes, J C1 aKromrey, J D uhttp://www.iacat.org/content/item-exposure-control-computer-adaptive-testing-use-freezing-augment-stratification00485nas a2200121 4500008004100000245006200041210006200103260005900165300001100224100002300235700001700258856008800275 2000 eng d00aItem selection and ability estimation in adaptive testing0 aItem selection and ability estimation in adaptive testing aDordrecht, The NetherlandsbKluwer Academic Publishers a1–251 avan der Linden, WJ1 aPashley, P J uhttp://www.iacat.org/content/item-selection-and-ability-estimation-adaptive-testing00692nas a2200169 4500008004100000020001400041245015800055210006900213300001200282490000600294653003400300100001700334700001500351700001200366700001400378856013000392 2000 eng d a1575-910500aLos tests adaptativos informatizados en la frontera del siglo XXI: Una revisión [Computerized adaptive tests at the turn of the 21st century: A review]0 aLos tests adaptativos informatizados en la frontera del siglo XX a183-2160 v210acomputerized adaptive testing1 aHontangas, P1 aPonsoda, V1 aOlea, J1 aAbad, F J uhttp://www.iacat.org/content/los-tests-adaptativos-informatizados-en-la-frontera-del-siglo-xxi-una-revisi%C3%B3n-computerized00867nas a2200145 4500008004100000245006600041210006600107300001200173490000700185520036100192653002100553653004400574100001500618856008800633 2000 eng d00aOverview of the computerized adaptive testing special section0 aOverview of the computerized adaptive testing special section a115-1200 v213 aThis paper provides an overview of the five papers included in the Psicologica special section on computerized adaptive testing. A short introduction to this topic is presented as well. The main results, the links between the five papers and the general research topic to which they are more related are also shown. (PsycINFO Database Record (c) 2005 APA )10aAdaptive Testing10aComputers computerized adaptive testing1 aPonsoda, V uhttp://www.iacat.org/content/overview-computerized-adaptive-testing-special-section00560nas a2200121 4500008004100000245013100041210006900172260001900241100001800260700001700278700001700295856012600312 2000 eng d00aSufficient simplicity or comprehensive complexity? A comparison of probabilitic and stratification methods of exposure control0 aSufficient simplicity or comprehensive complexity A comparison o aNew Orleans LA1 aParshall, C G1 aKromrey, J D1 aHogarty, K Y uhttp://www.iacat.org/content/sufficient-simplicity-or-comprehensive-complexity-comparison-probabilitic-and-stratification00478nas a2200097 4500008004100000245009400041210006900135260004400204100001800248856011400266 2000 eng d00aTestlet-based Designs for Computer-Based Testing in a Certification and Licensure Setting0 aTestletbased Designs for ComputerBased Testing in a Certificatio aJersey City, NJ: AICPA Technical Report1 aPitoniak, M J uhttp://www.iacat.org/content/testlet-based-designs-computer-based-testing-certification-and-licensure-setting00546nas a2200133 4500008004100000245005900041210005900100260009900159100001400258700001400272700002700286700001400313856008500327 2000 eng d00aUsing Bayesian Networks in Computerized Adaptive Tests0 aUsing Bayesian Networks in Computerized Adaptive Tests aM. Ortega and J. Bravo (Eds.),Computers and Education in the 21st Century. Kluwer, pp. 217228.1 aMillan, E1 aTrella, M1 aPerez-de-la-Cruz, J -L1 aConejo, R uhttp://www.iacat.org/content/using-bayesian-networks-computerized-adaptive-tests00494nas a2200109 4500008004100000245010000041210006900141260002000210100001700230700001800247856011900265 1999 eng d00aA comparative study of ability estimates from computer-adaptive testing and multi-stage testing0 acomparative study of ability estimates from computeradaptive tes aMontreal Canada1 aPatsula, L N1 aHambleton, RK uhttp://www.iacat.org/content/comparative-study-ability-estimates-computer-adaptive-testing-and-multi-stage-testing00475nam a2200097 4500008004100000245007400041210006900115260007800184100001700262856009800279 1999 eng d00aA comparison of computerized-adaptive testing and multi-stage testing0 acomparison of computerizedadaptive testing and multistage testin aUnpublished doctoral dissertation, University of Massachusetts at Amherst1 aPatsula, L N uhttp://www.iacat.org/content/comparison-computerized-adaptive-testing-and-multi-stage-testing00492nas a2200133 4500008004100000245007400041210006700115260001300182100001600195700001300211700001300224700001400237856010700251 1999 eng d00aComputerized testing – Issues and applications (Mini-course manual)0 aComputerized testing Issues and applications Minicourse manual aMontreal1 aParshall, C1 aDavey, T1 aSpray, J1 aKalohn, J uhttp://www.iacat.org/content/computerized-testing-%E2%80%93-issues-and-applications-mini-course-manual00569nas a2200145 4500008004100000245010600041210006900147300001200216490000700228100001500235700001200250700001900262700001600281856012600297 1999 eng d00aThe effects of test difficulty manipulation in computerized adaptive testing and self-adapted testing0 aeffects of test difficulty manipulation in computerized adaptive a167-1840 v121 aPonsoda, V1 aOlea, J1 aRodriguez, M S1 aRevuelta, J uhttp://www.iacat.org/content/effects-test-difficulty-manipulation-computerized-adaptive-testing-and-self-adapted-testin-200485nas a2200121 4500008004100000245008600041210006900127260001600196100001600212700001500228700001500243856010500258 1999 eng d00aItem exposure in adaptive tests: An empirical investigation of control strategies0 aItem exposure in adaptive tests An empirical investigation of co aLawrence KS1 aParshall, C1 aHogarty, K1 aKromrey, J uhttp://www.iacat.org/content/item-exposure-adaptive-tests-empirical-investigation-control-strategies00524nas a2200133 4500008004100000245009300041210006900134260002100203100001600224700001400240700001400254700001400268856010800282 1999 eng d00aPerformance of the Sympson-Hetter exposure control algorithm with a polytomous item bank0 aPerformance of the SympsonHetter exposure control algorithm with aMontreal, Canada1 aPastor, D A1 aChiang, C1 aDodd, B G1 aYockey, R uhttp://www.iacat.org/content/performance-sympson-hetter-exposure-control-algorithm-polytomous-item-bank00390nas a2200121 4500008004100000245004400041210004400085260002100129100001300150700001700163700001800180856007000198 1999 eng d00aPretesting alongside an operational CAT0 aPretesting alongside an operational CAT aMontreal, Canada1 aDavey, T1 aPommerich, M1 aThompson, D T uhttp://www.iacat.org/content/pretesting-alongside-operational-cat00534nas a2200097 4500008004100000245014000041210006900181260004600250100001600296856012400312 1999 eng d00aTest anxiety and test performance: Comparing paper-based and computer-adaptive versions of the GRE General Test (Research Report 99-15)0 aTest anxiety and test performance Comparing paperbased and compu aPrinceton NJ: Educational Testing Service1 aPowers, D E uhttp://www.iacat.org/content/test-anxiety-and-test-performance-comparing-paper-based-and-computer-adaptive-versions-gre00522nas a2200121 4500008004100000245010600041210006900147260002100216100001200237700001500249700001700264856011900281 1999 eng d00aTests informatizados: Fundamentos y aplicaciones (Computerized testing: Fundamentals and applications0 aTests informatizados Fundamentos y aplicaciones Computerized tes aMadrid: Pirmide.1 aOlea, J1 aPonsoda, V1 aPrieto, Eds. uhttp://www.iacat.org/content/tests-informatizados-fundamentos-y-aplicaciones-computerized-testing-fundamentals-and00687nas a2200169 4500008004100000020003000041245009600071210006900167260006400236100001900300700001700319700002100336700001300357700001700370700001500387856011500402 1998 eng d aETS Research Report 98-3800aComparability of paper-and-pencil and computer adaptive test scores on the GRE General Test0 aComparability of paperandpencil and computer adaptive test score aPrinceton, N.J.bEducational Testing ServicescAugust, 19981 aSchaeffer, G A1 aBridgeman, B1 aGolub-Smith, M L1 aLewis, C1 aPotenza, M T1 aSteffen, M uhttp://www.iacat.org/content/comparability-paper-and-pencil-and-computer-adaptive-test-scores-gre-general-test00729nas a2200157 4500008004100000245018900041210006900230260004700299100001700346700001700363700002100380700001300401700001700414700001500431856012500446 1998 eng d00aComparability of paper-and-pencil and computer adaptive test scores on the GRE General Test (GRE Board Professional Report No 95-08P; Educational Testing Service Research Report 98-38)0 aComparability of paperandpencil and computer adaptive test score aPrinceton, NJ: Educational Testing Service1 aSchaeffer, G1 aBridgeman, B1 aGolub-Smith, M L1 aLewis, C1 aPotenza, M T1 aSteffen, M uhttp://www.iacat.org/content/comparability-paper-and-pencil-and-computer-adaptive-test-scores-gre-general-test-gre-board00466nas a2200121 4500008004100000245008300041210006900124300001200193490000700205100001300212700001500225856010400240 1998 eng d00aA comparison of item exposure control methods in computerized adaptive testing0 acomparison of item exposure control methods in computerized adap a311-3270 v351 aRevuelta1 aPonsoda, V uhttp://www.iacat.org/content/comparison-item-exposure-control-methods-computerized-adaptive-testing00419nas a2200097 4500008004100000245007600041210006900117260001700186100001800203856010000221 1998 eng d00aItem development and pretesting in a computer-based testing environment0 aItem development and pretesting in a computerbased testing envir aPhiladelphia1 aParshall, C G uhttp://www.iacat.org/content/item-development-and-pretesting-computer-based-testing-environment01298nas a2200157 4500008004100000245007500041210006900116300001000185490000700195520076100202653003400963100001800997700001401015700001701029856009401046 1998 eng d00aSimulating the use of disclosed items in computerized adaptive testing0 aSimulating the use of disclosed items in computerized adaptive t a48-680 v353 aRegular use of questions previously made available to the public (i.e., disclosed items) may provide one way to meet the requirement for large numbers of questions in a continuous testing environment, that is, an environment in which testing is offered at test taker convenience throughout the year rather than on a few prespecified test dates. First it must be shown that such use has effects on test scores small enough to be acceptable. In this study simulations are used to explore the use of disclosed items under a worst-case scenario which assumes that disclosed items are always answered correctly. Some item pool and test designs were identified in which the use of disclosed items produces effects on test scores that may be viewed as negligible.10acomputerized adaptive testing1 aStocking, M L1 aWard, W C1 aPotenza, M T uhttp://www.iacat.org/content/simulating-use-disclosed-items-computerized-adaptive-testing00430nas a2200121 4500008004100000245005900041210005900100260001800159100001800177700001300195700001600208856008400224 1998 eng d00aTest development exposure control for adaptive testing0 aTest development exposure control for adaptive testing aSan Diego, CA1 aParshall, C G1 aDavey, T1 aNering, M L uhttp://www.iacat.org/content/test-development-exposure-control-adaptive-testing00521nas a2200121 4500008004100000245010200041210006900143260001800212100001500230700001800245700001600263856012000279 1997 eng d00aCalibration of CAT items administered online for classification: Assumption of local independence0 aCalibration of CAT items administered online for classification aGatlinburg TN1 aSpray, J A1 aParshall, C G1 aHuang, C -H uhttp://www.iacat.org/content/calibration-cat-items-administered-online-classification-assumption-local-independence02024nas a2200289 4500008004100000245012700041210006900168300001300237490000800250520105800258653003401316653003301350653002301383653001501406653001001421653002601431653001001457653001501467653001801482653003101500100001501531700001601546700001401562700001701576700001601593856012501609 1997 eng d00aA computerized adaptive testing system for speech discrimination measurement: The Speech Sound Pattern Discrimination Test0 acomputerized adaptive testing system for speech discrimination m a2289-2980 v1013 aA computerized, adaptive test-delivery system for the measurement of speech discrimination, the Speech Sound Pattern Discrimination Test, is described and evaluated. Using a modified discrimination task, the testing system draws on a pool of 130 items spanning a broad range of difficulty to estimate an examinee's location along an underlying continuum of speech processing ability, yet does not require the examinee to possess a high level of English language proficiency. The system is driven by a mathematical measurement model which selects only test items which are appropriate in difficulty level for a given examinee, thereby individualizing the testing experience. Test items were administered to a sample of young deaf adults, and the adaptive testing system evaluated in terms of respondents' sensory and perceptual capabilities, acoustic and phonetic dimensions of speech, and theories of speech perception. Data obtained in this study support the validity, reliability, and efficiency of this test as a measure of speech processing ability.10a*Diagnosis, Computer-Assisted10a*Speech Discrimination Tests10a*Speech Perception10aAdolescent10aAdult10aAudiometry, Pure-Tone10aHuman10aMiddle Age10aPsychometrics10aReproducibility of Results1 aBochner, J1 aGarrison, W1 aPalmer, L1 aMacKenzie, D1 aBraveman, A uhttp://www.iacat.org/content/computerized-adaptive-testing-system-speech-discrimination-measurement-speech-sound-pattern00390nas a2200121 4500008004100000245005000041210005000091300001000141490000600151100001700157700001800174856007600192 1997 eng d00aFlawed items in computerized adaptive testing0 aFlawed items in computerized adaptive testing a79-960 v41 aPotenza, M T1 aStocking, M L uhttp://www.iacat.org/content/flawed-items-computerized-adaptive-testing00512nas a2200145 4500008004100000245008100041210006900122300001200191490000700203100001500210700001400225700001200239700001600251856009900267 1997 eng d00aAn investigation of self-adapted testing in a Spanish high school population0 ainvestigation of selfadapted testing in a Spanish high school po a210-2210 v571 aPonsoda, V1 aWise, S L1 aOlea, J1 aRevuelta, J uhttp://www.iacat.org/content/investigation-self-adapted-testing-spanish-high-school-population00449nas a2200109 4500008004100000245008000041210006900121260001600190100001700206700001500223856010100238 1997 eng d00aMaintaining item and test security in a CAT environment: A simulation study0 aMaintaining item and test security in a CAT environment A simula aChicago IL)1 aPatsula, L N1 aSteffen, M uhttp://www.iacat.org/content/maintaining-item-and-test-security-cat-environment-simulation-study00329nas a2200097 4500008004100000245004800041210004500089260001500134100001500149856006700164 1997 eng d00aAn overview of the LSAC CAT research agenda0 aoverview of the LSAC CAT research agenda aChicago IL1 aPashley, P uhttp://www.iacat.org/content/overview-lsac-cat-research-agenda00411nas a2200133 4500008004100000245005400041210004900095300001000144490000700154100001400161700001400175700001500189856007300204 1997 eng d00aThe role of item feedback in self-adapted testing0 arole of item feedback in selfadapted testing a85-980 v571 aRoos, L L1 aWise, S L1 aPlake, B S uhttp://www.iacat.org/content/role-item-feedback-self-adapted-testing00542nas a2200121 4500008004100000245009900041210006900140260004600209100001800255700001400273700001700287856011600304 1997 eng d00aSimulating the use of disclosed items in computerized adaptive testing (Research Report 97-10)0 aSimulating the use of disclosed items in computerized adaptive t aPrinceton NJ: Educational Testing Service1 aStocking, M L1 aWard, W C1 aPotenza, M T uhttp://www.iacat.org/content/simulating-use-disclosed-items-computerized-adaptive-testing-research-report-97-1000542nas a2200121 4500008004500000245012600045210007100171300000800242490000600250100001600256700001500272856013300287 1997 Spandsh 00aUna solución a la estimatión inicial en los tests adaptivos informatizados [A solution to initial estimation in CATs.] 0 aUna solución a la estimatión inicial en los tests adaptivos info a1-60 v21 aRevuelta, J1 aPonsoda, V uhttp://www.iacat.org/content/una-soluci%C3%B3n-la-estimati%C3%B3n-inicial-en-los-tests-adaptivos-informatizados-solution-initial00553nas a2200121 4500008004500000245015200045210006900197300001200266490000700278100001600285700001500301856011500316 1996 Spandsh 00aMetodos sencillos para el control de las tasas de exposicion en tests adaptativos informatizados [Simple methods for item exposure control in CATs]0 aMetodos sencillos para el control de las tasas de exposicion en a161-1720 v171 aRevuelta, J1 aPonsoda, V uhttp://www.iacat.org/content/metodos-sencillos-para-el-control-de-las-tasas-de-exposicion-en-tests-adaptativos00637nas a2200145 4500008004500000245018700045210006900232300001000301490000700311100001300318700001500331700001600346700001400362856011500376 1996 Spandsh 00aPropiedades psicometricas du un test adaptivo informatizado do vocabulario ingles [Psychometric properties of a computerized adaptive tests for the measurement of English vocabulary]0 aPropiedades psicometricas du un test adaptivo informatizado do v a61-730 v551 aOlea., J1 aPonsoda, V1 aRevuelta, J1 aBelchi, J uhttp://www.iacat.org/content/propiedades-psicometricas-du-un-test-adaptivo-informatizado-do-vocabulario-ingles00437nas a2200109 4500008004100000245006800041210006600109260003000175100001200205700001500217856009500232 1996 eng d00aTest adaptativos informatizados [Computerized adaptive testing]0 aTest adaptativos informatizados Computerized adaptive testing aMadrid, UNEDbUniversitas1 aOlea, J1 aPonsoda, V uhttp://www.iacat.org/content/test-adaptativos-informatizados-computerized-adaptive-testing00480nas a2200109 4500008004100000245009400041210006900135260002100204100001300225700001800238856011400256 1995 eng d00aNew algorithms for item selection and exposure control with computerized adaptive testing0 aNew algorithms for item selection and exposure control with comp aSan Francisco CA1 aDavey, T1 aParshall, C G uhttp://www.iacat.org/content/new-algorithms-item-selection-and-exposure-control-computerized-adaptive-testing00589nas a2200121 4500008004500000245017100045210006900216260001800285100001200303700001500315700001400330856012300344 1995 Spandsh 00aTests adaptivos y autoadaptados informatizados: Effects en la ansiedad y en la pecision de las estimaciones [SATs and CATS: Effects on enxiety and estimate precision]0 aTests adaptivos y autoadaptados informatizados Effects en la ans aMurcia, Spain1 aOlea, J1 aPonsoda, V1 aWise, S L uhttp://www.iacat.org/content/tests-adaptivos-y-autoadaptados-informatizados-effects-en-la-ansiedad-y-en-la-pecision-de00484nas a2200133 4500008004100000245008100041210006900122300001200191490000700203100001500210700000900225700001600234856010000250 1994 eng d00aADTEST: A computer-adaptive tests based on the maximum information principle0 aADTEST A computeradaptive tests based on the maximum information a680-6860 v541 aPonsoda, V1 aOlea1 aRevuelta, J uhttp://www.iacat.org/content/adtest-computer-adaptive-tests-based-maximum-information-principle00534nas a2200121 4500008004100000245011500041210006900156260001900225100001400244700001400258700001500272856012500287 1994 eng d00aComparing computerized adaptive and self-adapted tests: The influence of examinee achievement locus of control0 aComparing computerized adaptive and selfadapted tests The influe aNew Orleans LA1 aWise, S L1 aRoos, L L1 aPlake, B S uhttp://www.iacat.org/content/comparing-computerized-adaptive-and-self-adapted-tests-influence-examinee-achievement-locus00523nas a2200121 4500008004100000245008900041210006900130260004600199100001000245700001300255700001700268856011600285 1994 eng d00aComputerized mastery testing using fuzzy set decision theory (Research Report 94-37)0 aComputerized mastery testing using fuzzy set decision theory Res aPrinceton NJ: Educational Testing Service1 aDu, Y1 aLewis, C1 aPashley, P J uhttp://www.iacat.org/content/computerized-mastery-testing-using-fuzzy-set-decision-theory-research-report-94-3700474nam a2200097 4500008004100000245008100041210006900122260007400191100001500265856009600280 1994 eng d00aThe exploration of an alternative method for scoring computer adaptive tests0 aexploration of an alternative method for scoring computer adapti aUnpublished doctoral dissertation, Lincoln NE: University of Nebraska1 aPotenza, M uhttp://www.iacat.org/content/exploration-alternative-method-scoring-computer-adaptive-tests00415nas a2200109 4500008004100000245007100041210006700112300001000179490000700189100001600196856009300212 1994 eng d00aThe psychological impacts of computerized adaptive testing methods0 apsychological impacts of computerized adaptive testing methods a41-470 v341 aPowell, Z E uhttp://www.iacat.org/content/psychological-impacts-computerized-adaptive-testing-methods00530nas a2200145 4500008004100000245008600041210006900127300001000196490000600206100001400212700001400226700001000240700002700250856010700277 1994 eng d00aThe relationship between examinee anxiety and preference for self-adapted testing0 arelationship between examinee anxiety and preference for selfada a81-910 v71 aWise, S L1 aRoos, L L1 aPlake1 aNebelsick-Gullett, L J uhttp://www.iacat.org/content/relationship-between-examinee-anxiety-and-preference-self-adapted-testing00468nas a2200133 4500008004100000245007100041210006700112300001200179490000700191100001800198700001500216700001600231856008700247 1993 eng d00aThe application of an automated item selection method to real data0 aapplication of an automated item selection method to real data a167-1760 v171 aStocking, M L1 aSwanson, L1 aPearlman, M uhttp://www.iacat.org/content/application-automated-item-selection-method-real-data00457nas a2200133 4500008004100000245006500041210006500106300001200171490000600183100001000189700001300199700001700212856009400229 1993 eng d00aComputerized mastery testing using fuzzy set decision theory0 aComputerized mastery testing using fuzzy set decision theory a181-1930 v61 aDu, Y1 aLewis, C1 aPashley, P J uhttp://www.iacat.org/content/computerized-mastery-testing-using-fuzzy-set-decision-theory00476nas a2200109 4500008004100000245009300041210006900134260001600203100001100219700001500230856012100245 1993 eng d00aMonte Carlo simulation comparison of two-stage testing and computerized adaptive testing0 aMonte Carlo simulation comparison of twostage testing and comput aAtlanta, GA1 aKim, H1 aPlake, B S uhttp://www.iacat.org/content/monte-carlo-simulation-comparison-two-stage-testing-and-computerized-adaptive-testing-000513nas a2200145 4500008004100000245007700041210006900118300001200187490000700199100001400206700001500220700001700235700001400252856010100266 1992 eng d00aA comparison of self-adapted and computerized adaptive achievement tests0 acomparison of selfadapted and computerized adaptive achievement a329-3390 v291 aWise, S L1 aPlake, S S1 aJohnson, P L1 aRoos, S L uhttp://www.iacat.org/content/comparison-self-adapted-and-computerized-adaptive-achievement-tests00464nas a2200121 4500008004100000245007600041210006900117260001800186100001400204700001500218700001400233856009500247 1992 eng d00aThe effects of feedback in computerized adaptive and self-adapted tests0 aeffects of feedback in computerized adaptive and selfadapted tes aSan Francisco1 aRoos, L L1 aPlake, B S1 aWise, S L uhttp://www.iacat.org/content/effects-feedback-computerized-adaptive-and-self-adapted-tests00499nas a2200121 4500008004100000245008200041210006900123300000900192490000700201653003400208100002400242856011100266 1992 eng d00aTest anxiety and test performance under computerized adaptive testing methods0 aTest anxiety and test performance under computerized adaptive te a25180 v5210acomputerized adaptive testing1 aPowell, Zen-Hsiu, E uhttp://www.iacat.org/content/test-anxiety-and-test-performance-under-computerized-adaptive-testing-methods00440nas a2200097 4500008004100000245008200041210006900123260002100192100001600213856011300229 1992 eng d00aTest anxiety and test performance under computerized adaptive testing methods0 aTest anxiety and test performance under computerized adaptive te aSan Francisco CA1 aPowell, Z E uhttp://www.iacat.org/content/test-anxiety-and-test-performance-under-computerized-adaptive-testing-methods-000547nas a2200121 4500008004100000245010100041210006900142260004600211100001800257700001500275700001600290856011900306 1991 eng d00aAutomatic item selection (AIS) methods in the ETS testing environment (Research Memorandum 91-5)0 aAutomatic item selection AIS methods in the ETS testing environm aPrinceton NJ: Educational Testing Service1 aStocking, M L1 aSwanson, L1 aPearlman, M uhttp://www.iacat.org/content/automatic-item-selection-ais-methods-ets-testing-environment-research-memorandum-91-500599nas a2200121 4500008004100000245007100041210006800112260014900180100001800329700001400347700001900361856009700380 1991 eng d00aComputerized adaptive testing: Theory, applications, and standards0 aComputerized adaptive testing Theory applications and standards aR. K. Hambleton and J. N. Zaal (Eds.), Advances in educational and psychological testing: Theory and Applications (pp. 341-366). Boston: Kluwer.1 aHambleton, RK1 aZaal, J N1 aPieters, J P M uhttp://www.iacat.org/content/computerized-adaptive-testing-theory-applications-and-standards00499nas a2200145 4500008004100000245007200041210006900113300001000182490000600192100001700198700001400215700001400229700001500243856009500258 1991 eng d00aCorrelates of examinee item choice behavior in self-adapted testing0 aCorrelates of examinee item choice behavior in selfadapted testi a25-280 v41 aJohnson, J L1 aRoos, L L1 aWise, S L1 aPlake, B S uhttp://www.iacat.org/content/correlates-examinee-item-choice-behavior-self-adapted-testing00738nas a2200169 4500008004100000020000900041245012400050210006900174260008700243653003400330653001500364653001800379100001600397700001900413700001700432856011900449 1991 eng d aR-1100aPatterns of alcohol and drug use among federal offenders as assessed by the Computerized Lifestyle Screening Instrument0 aPatterns of alcohol and drug use among federal offenders as asse aOttawa, ON. CanadabResearch and Statistics Branch, Correctional Service of Canada10acomputerized adaptive testing10adrug abuse10asubstance use1 aRobinson, D1 aPorporino, F J1 aMillson, W A uhttp://www.iacat.org/content/patterns-alcohol-and-drug-use-among-federal-offenders-assessed-computerized-lifestyle00491nas a2200109 4500008004100000245009500041210006900136260002700205100001600232700001700248856011600265 1990 eng d00aA comparison of Rasch and three-parameter logistic models in computerized adaptive testing0 acomparison of Rasch and threeparameter logistic models in comput aUnpublished manuscript1 aParker, S B1 aMcBride, J R uhttp://www.iacat.org/content/comparison-rasch-and-three-parameter-logistic-models-computerized-adaptive-testing00377nas a2200109 4500008004100000245005500041210005400096300001000150490000600160100001800166856008300184 1990 eng d00aSoftware review: MicroCAT Testing System Version 30 aSoftware review MicroCAT Testing System Version 3 a82-880 v71 aPatience, W M uhttp://www.iacat.org/content/software-review-microcat-testing-system-version-300464nas a2200121 4500008004100000245007400041210006900115490001900184100001500203700001400218700001400232856009600246 1989 eng d00aEXSPRT: An expert systems approach to computer-based adaptive testing0 aEXSPRT An expert systems approach to computerbased adaptive test0 vSan Francisco.1 aFrick, T W1 aPlew, G T1 aLuk, H -K uhttp://www.iacat.org/content/exsprt-expert-systems-approach-computer-based-adaptive-testing00510nas a2200133 4500008004100000245009200041210006900133300001200202490000700214100001400221700001500235700001200250856011400262 1989 eng d00aProviding item feedback in computer-based tests: Effects of initial success and failure0 aProviding item feedback in computerbased tests Effects of initia a479-4860 v491 aWise, S L1 aPlake, B S1 aet. al. uhttp://www.iacat.org/content/providing-item-feedback-computer-based-tests-effects-initial-success-and-failure00456nas a2200133 4500008004100000245006600041210006600107300001200173490000700185100000900192700001400201700002200215856008500237 1988 eng d00aItem pool maintenance in the presence of item parameter drift0 aItem pool maintenance in the presence of item parameter drift a275-2850 v251 aBock1 aMuraki, E1 aPfeiffenberger, W uhttp://www.iacat.org/content/item-pool-maintenance-presence-item-parameter-drift00623nas a2200133 4500008004100000245012600041210006900167260006700236100001900303700001400322700001600336700001500352856012200367 1985 eng d00aArmed Services Vocational Aptitude Battery: Development of an adaptive item pool (AFHLR-TR-85-19; Technical Rep No 85-19)0 aArmed Services Vocational Aptitude Battery Development of an ada aBrooks Air Force Base TX: Air Force Human Resources Laboratory1 aPrestwood, J S1 aVale, C D1 aMassey, R H1 aWelsh, J R uhttp://www.iacat.org/content/armed-services-vocational-aptitude-battery-development-adaptive-item-pool-afhlr-tr-85-1900540nas a2200133 4500008004100000245007700041210006900118260006500187100001200252700001400264700001500278700001500293856009800308 1984 eng d00aMicrocomputer network for computerized adaptive testing (CAT) (TR-84-33)0 aMicrocomputer network for computerized adaptive testing CAT TR84 aSan Diego CA: Navy Personnel Research and Development Center1 aQuan, B1 aPark, T A1 aSandahl, G1 aWolfe, J H uhttp://www.iacat.org/content/microcomputer-network-computerized-adaptive-testing-cat-tr-84-3300616nas a2200133 4500008004100000245009600041210006900137260009700206100001400303700001600317700001800333700001400351856011700365 1980 eng d00aEffects of computerized adaptive testing on Black and White students (Research Report 79-2)0 aEffects of computerized adaptive testing on Black and White stud aMinneapolis: University of Minnesota, Department of Psychology, Psychometric Methods Program1 aPine, S M1 aChurch, A T1 aGialluca, K A1 aWeiss, DJ uhttp://www.iacat.org/content/effects-computerized-adaptive-testing-black-and-white-students-research-report-79-200516nas a2200109 4500008004100000245012400041210006900165260001700234100001800251700001700269856012000286 1980 eng d00aEffects of program parameters and item pool characteristics on the bias of a three-parameter tailored testing procedure0 aEffects of program parameters and item pool characteristics on t aBoston MA, U1 aPatience, W M1 aReckase, M D uhttp://www.iacat.org/content/effects-program-parameters-and-item-pool-characteristics-bias-three-parameter-tailored00506nas a2200109 4500008004100000245005900041210005900100260013300159100001700292700001300309856007400322 1980 eng d00aIndividualized testing on the basis of the Rasch model0 aIndividualized testing on the basis of the Rasch model aIn J. Th. Van der Kamp, W. F. Langerak, and D. N. M. de Gruijter (Eds.). Psychometrics for educational debates. New York: Wiley.1 aFischer, G H1 aPendl, P uhttp://www.iacat.org/content/individualized-testing-basis-rasch-model00483nas a2200121 4500008004100000245007800041210006900119300002500188490001600213100001300229700001700242856010200259 1980 eng d00aOperational characteristics of a one-parameter tailored testing procedure0 aOperational characteristics of a oneparameter tailored testing p a10, 66 (Ms No. 2104)0 vAugust 19801 aPatience1 aReckase, M D uhttp://www.iacat.org/content/operational-characteristics-one-parameter-tailored-testing-procedure00533nas a2200109 4500008004100000245013600041210006900177260001800246100001800264700001700282856012400299 1979 eng d00aOperational characteristics of a Rasch model tailored testing procedure when program parameters and item pool attributes are varied0 aOperational characteristics of a Rasch model tailored testing pr aSan Francisco1 aPatience, W M1 aReckase, M D uhttp://www.iacat.org/content/operational-characteristics-rasch-model-tailored-testing-procedure-when-program-parameters00568nas a2200109 4500008004100000245010400041210006900145260009700214100001400311700001400325856011900339 1978 eng d00aA comparison of the fairness of adaptive and conventional testing strategies (Research Report 78-1)0 acomparison of the fairness of adaptive and conventional testing aMinneapolis: University of Minnesota, Department of Psychology, Psychometric Methods Program1 aPine, S M1 aWeiss, DJ uhttp://www.iacat.org/content/comparison-fairness-adaptive-and-conventional-testing-strategies-research-report-78-100567nas a2200109 4500008003900000245005900039210005800098260018500156100001700341700001600358856008300374 1977 d00aAdaptive Branching in a Multi-Content Achievement Test0 aAdaptive Branching in a MultiContent Achievement Test aD. J. Weiss (Ed.), Proceedings of the 1977 Computerized Adaptive Testing Conference. Minneapolis MN: University of Minnesota, Department of Psychology, Psychometric Methods Program1 aPennell, R J1 aHarris, D A uhttp://www.iacat.org/content/adaptive-branching-multi-content-achievement-test00457nas a2200133 4500008004100000245006300041210006300104300001200167490000600179100001700185700001700202700001800219856008600237 1977 eng d00aApplication of tailored testing to achievement measurement0 aApplication of tailored testing to achievement measurement a158-1610 v91 aEnglish, R A1 aReckase, M D1 aPatience, W M uhttp://www.iacat.org/content/application-tailored-testing-achievement-measurement00360nas a2200109 4500008004100000245005000041210005000091300001200141490000600153100001800159856007300177 1977 eng d00aDescription of components in tailored testing0 aDescription of components in tailored testing a153-1570 v91 aPatience, W M uhttp://www.iacat.org/content/description-components-tailored-testing00653nas a2200097 4500008003900000245012300039210006900162260018600231100001900417856011900436 1977 d00aEffects of Knowledge of Results and Varying Proportion Correct on Ability Test Performance and Psychological Variables0 aEffects of Knowledge of Results and Varying Proportion Correct o aD. J. Weiss (Ed.), Proceedings of the 1977 Computerized Adaptive Testing Conference. Minneapolis MN: University of Minnesota, Department of Psychology, Psychometric Methods Program.1 aPrestwood, J S uhttp://www.iacat.org/content/effects-knowledge-results-and-varying-proportion-correct-ability-test-performance-and00501nas a2200097 4500008003900000245004700039210004700086260018600133100001400319856007000333 1977 d00aReduction of Test Bias by Adaptive Testing0 aReduction of Test Bias by Adaptive Testing aD. J. Weiss (Ed.), Proceedings of the 1977 Computerized Adaptive Testing Conference. Minneapolis MN: University of Minnesota, Department of Psychology, Psychometric Methods Program.1 aPine, S M uhttp://www.iacat.org/content/reduction-test-bias-adaptive-testing00536nas a2200109 4500008004100000245004600041210004600087260018600133100001400319700001800333856007500351 1977 eng d00aStudent attitudes toward tailored testing0 aStudent attitudes toward tailored testing aD. J. Weiss (Ed.), Proceedings of the 1977 Computerized Adaptive Testing Conference. Minneapolis MN: University of Minnesota, Department of Psychology, Psychometric Methods Program.1 aKoch, W R1 aPatience, W M uhttp://www.iacat.org/content/student-attitudes-toward-tailored-testing00360nas a2200109 4500008004100000245004600041210004400087260002400131100001100155700001400166856007000180 1974 eng d00aA Bayesian approach in sequential testing0 aBayesian approach in sequential testing aChicago ILc04/19741 aHsu, T1 aPingel, K uhttp://www.iacat.org/content/bayesian-approach-sequential-testing00470nas a2200109 4500008004100000245010000041210006900141300001200210490000600222100001500228856011700243 1968 eng d00aMethodological determination of the PEST (parameter estimation by sequential testing) procedure0 aMethodological determination of the PEST parameter estimation by a285-2890 v31 aPollack, I uhttp://www.iacat.org/content/methodological-determination-pest-parameter-estimation-sequential-testing-procedure00397nam a2200097 4500008004100000245005400041210005100095260006500146100001800211856007000229 1962 eng d00aAn evaluation of the sequential method of testing0 aevaluation of the sequential method of testing aUnpublished doctoral dissertation, Michigan State University1 aPaterson, J J uhttp://www.iacat.org/content/evaluation-sequential-method-testing00487nas a2200133 4500008004000000245006900040210006500109260006300174100001200237700001500249700001500264700001600279856005800295 0 engd00aMicrocomputer network for computerized adaptive testing (CAT) 0 aMicrocomputer network for computer ized adaptive testing CAT bSan Diego: Navy Personnel Research and Development Center.1 aQuan, B1 aPark, T.A.1 aSandahl, G1 aWolfe, J.H. uhttps://apps.dtic.mil/dtic/tr/fulltext/u2/a140256.pdf