TY - JOUR T1 - Optimizing cognitive ability measurement with multidimensional computer adaptive testing JF - International Journal of Testing Y1 - In Press A1 - Makransky, G. A1 - Glas, C. A. W. ER - TY - JOUR T1 - The Optimal Item Pool Design in Multistage Computerized Adaptive Tests With the p-Optimality Method JF - Educational and Psychological Measurement Y1 - 2020 A1 - Lihong Yang A1 - Mark D. Reckase AB - The present study extended the p-optimality method to the multistage computerized adaptive test (MST) context in developing optimal item pools to support different MST panel designs under different test configurations. Using the Rasch model, simulated optimal item pools were generated with and without practical constraints of exposure control. A total number of 72 simulated optimal item pools were generated and evaluated by an overall sample and conditional sample using various statistical measures. Results showed that the optimal item pools built with the p-optimality method provide sufficient measurement accuracy under all simulated MST panel designs. Exposure control affected the item pool size, but not the item distributions and item pool characteristics. This study demonstrated that the p-optimality method can adapt to MST item pool design, facilitate the MST assembly process, and improve its scoring accuracy. VL - 80 UR - https://doi.org/10.1177/0013164419901292 ER - TY - JOUR T1 - On-the-Fly Constraint-Controlled Assembly Methods for Multistage Adaptive Testing for Cognitive Diagnosis JF - Journal of Educational Measurement Y1 - 2018 A1 - Liu, Shuchang A1 - Cai, Yan A1 - Tu, Dongbo AB - Abstract This study applied the mode of on-the-fly assembled multistage adaptive testing to cognitive diagnosis (CD-OMST). Several and several module assembly methods for CD-OMST were proposed and compared in terms of measurement precision, test security, and constrain management. The module assembly methods in the study included the maximum priority index method (MPI), the revised maximum priority index (RMPI), the weighted deviation model (WDM), and the two revised Monte Carlo methods (R1-MC, R2-MC). Simulation results showed that on the whole the CD-OMST performs well in that it not only has acceptable attribute pattern correct classification rates but also satisfies both statistical and nonstatistical constraints; the RMPI method was generally better than the MPI method, the R2-MC method was generally better than the R1-MC method, and the two revised Monte Carlo methods performed best in terms of test security and constraint management, whereas the RMPI and WDM methods worked best in terms of measurement precision. The study is not only expected to provide information about how to combine MST and CD using an on-the-fly method and how do these assembled methods in CD-OMST perform relative to each other but also offer guidance for practitioners to assemble modules in CD-OMST with both statistical and nonstatistical constraints. VL - 55 UR - https://onlinelibrary.wiley.com/doi/abs/10.1111/jedm.12194 ER - TY - JOUR T1 - Online Calibration of Polytomous Items Under the Generalized Partial Credit Model JF - Applied Psychological Measurement Y1 - 2016 A1 - Zheng, Yi AB - Online calibration is a technology-enhanced architecture for item calibration in computerized adaptive tests (CATs). Many CATs are administered continuously over a long term and rely on large item banks. To ensure test validity, these item banks need to be frequently replenished with new items, and these new items need to be pretested before being used operationally. Online calibration dynamically embeds pretest items in operational tests and calibrates their parameters as response data are gradually obtained through the continuous test administration. This study extends existing formulas, procedures, and algorithms for dichotomous item response theory models to the generalized partial credit model, a popular model for items scored in more than two categories. A simulation study was conducted to investigate the developed algorithms and procedures under a variety of conditions, including two estimation algorithms, three pretest item selection methods, three seeding locations, two numbers of score categories, and three calibration sample sizes. Results demonstrated acceptable estimation accuracy of the two estimation algorithms in some of the simulated conditions. A variety of findings were also revealed for the interacted effects of included factors, and recommendations were made respectively. VL - 40 UR - http://apm.sagepub.com/content/40/6/434.abstract ER - TY - JOUR T1 - Optimal Reassembly of Shadow Tests in CAT JF - Applied Psychological Measurement Y1 - 2016 A1 - Choi, Seung W. A1 - Moellering, Karin T. A1 - Li, Jie A1 - van der Linden, Wim J. AB - Even in the age of abundant and fast computing resources, concurrency requirements for large-scale online testing programs still put an uninterrupted delivery of computer-adaptive tests at risk. In this study, to increase the concurrency for operational programs that use the shadow-test approach to adaptive testing, we explored various strategies aiming for reducing the number of reassembled shadow tests without compromising the measurement quality. Strategies requiring fixed intervals between reassemblies, a certain minimal change in the interim ability estimate since the last assembly before triggering a reassembly, and a hybrid of the two strategies yielded substantial reductions in the number of reassemblies without degradation in the measurement accuracy. The strategies effectively prevented unnecessary reassemblies due to adapting to the noise in the early test stages. They also highlighted the practicality of the shadow-test approach by minimizing the computational load involved in its use of mixed-integer programming. VL - 40 UR - http://apm.sagepub.com/content/40/7/469.abstract ER - TY - JOUR T1 - Online Item Calibration for Q-Matrix in CD-CAT JF - Applied Psychological Measurement Y1 - 2015 A1 - Chen, Yunxiao A1 - Liu, Jingchen A1 - Ying, Zhiliang AB -

Item replenishment is important for maintaining a large-scale item bank. In this article, the authors consider calibrating new items based on pre-calibrated operational items under the deterministic inputs, noisy-and-gate model, the specification of which includes the so-called -matrix, as well as the slipping and guessing parameters. Making use of the maximum likelihood and Bayesian estimators for the latent knowledge states, the authors propose two methods for the calibration. These methods are applicable to both traditional paper–pencil–based tests, for which the selection of operational items is prefixed, and computerized adaptive tests, for which the selection of operational items is sequential and random. Extensive simulations are done to assess and to compare the performance of these approaches. Extensions to other diagnostic classification models are also discussed.

VL - 39 UR - http://apm.sagepub.com/content/39/1/5.abstract ER - TY - JOUR T1 - On-the-Fly Assembled Multistage Adaptive Testing JF - Applied Psychological Measurement Y1 - 2015 A1 - Zheng, Yi A1 - Chang, Hua-Hua AB -

Recently, multistage testing (MST) has been adopted by several important large-scale testing programs and become popular among practitioners and researchers. Stemming from the decades of history of computerized adaptive testing (CAT), the rapidly growing MST alleviates several major problems of earlier CAT applications. Nevertheless, MST is only one among all possible solutions to these problems. This article presents a new adaptive testing design, “on-the-fly assembled multistage adaptive testing” (OMST), which combines the benefits of CAT and MST and offsets their limitations. Moreover, OMST also provides some unique advantages over both CAT and MST. A simulation study was conducted to compare OMST with MST and CAT, and the results demonstrated the promising features of OMST. Finally, the “Discussion” section provides suggestions on possible future adaptive testing designs based on the OMST framework, which could provide great flexibility for adaptive tests in the digital future and open an avenue for all types of hybrid designs based on the different needs of specific tests.

VL - 39 UR - http://apm.sagepub.com/content/39/2/104.abstract ER - TY - CONF T1 - Optimal Calibration Designs for Computerized Adaptive Testing T2 - Annual Conference of the International Association for Computerized Adaptive Testing Y1 - 2011 A1 - Angela Verschoor KW - balanced block design KW - CAT KW - item calibration KW - optimization KW - Rasch AB -

Optimaztion

How can we exploit the advantages of Balanced Block Design while keeping the logistics manageable?

Homogeneous Designs: Overlap between test booklets as regular as possible

Conclusions:

JF - Annual Conference of the International Association for Computerized Adaptive Testing ER - TY - JOUR T1 - Online calibration via variable length computerized adaptive testing JF - Psychometrika Y1 - 2010 A1 - Chang, Y. I. A1 - Lu, H. Y. AB - Item calibration is an essential issue in modern item response theory based psychological or educational testing. Due to the popularity of computerized adaptive testing, methods to efficiently calibrate new items have become more important than that in the time when paper and pencil test administration is the norm. There are many calibration processes being proposed and discussed from both theoretical and practical perspectives. Among them, the online calibration may be one of the most cost effective processes. In this paper, under a variable length computerized adaptive testing scenario, we integrate the methods of adaptive design, sequential estimation, and measurement error models to solve online item calibration problems. The proposed sequential estimate of item parameters is shown to be strongly consistent and asymptotically normally distributed with a prechosen accuracy. Numerical results show that the proposed method is very promising in terms of both estimation accuracy and efficiency. The results of using calibrated items to estimate the latent trait levels are also reported. VL - 75 SN - 0033-3123 ER - TY - CHAP T1 - Obtaining reliable diagnostic information through constrained CAT Y1 - 2009 A1 - Wang, C. A1 - Chang, Hua-Hua A1 - Douglas, J. CY - D. J. Weiss (Ed.), Proceedings of the 2009 GMAC Conference on Computerized Adaptive Testing. N1 - {PDF File, 252 KB} ER - TY - CHAP T1 - Optimizing item exposure control algorithms for polytomous computerized adaptive tests with restricted item banks Y1 - 2009 A1 - Chajewski, M. A1 - Lewis, C. CY - D. J. Weiss (Ed.), Proceedings of the 2009 GMAC Conference on Computerized Adaptive Testing. N1 - {PDF File, 923 KB} ER - TY - JOUR T1 - Optimal and nonoptimal computer-based test designs for making pass-fail decisions JF - Applied Measurement in Education Y1 - 2006 A1 - Hambleton, R. K. A1 - Xing, D. KW - adaptive test KW - credentialing exams KW - Decision Making KW - Educational Measurement KW - multistage tests KW - optimal computer-based test designs KW - test form AB - Now that many credentialing exams are being routinely administered by computer, new computer-based test designs, along with item response theory models, are being aggressively researched to identify specific designs that can increase the decision consistency and accuracy of pass-fail decisions. The purpose of this study was to investigate the impact of optimal and nonoptimal multistage test (MST) designs, linear parallel-form test designs (LPFT), and computer adaptive test (CAT) designs on the decision consistency and accuracy of pass-fail decisions. Realistic testing situations matching those of one of the large credentialing agencies were simulated to increase the generalizability of the findings. The conclusions were clear: (a) With the LPFTs, matching test information functions (TIFs) to the mean of the proficiency distribution produced slightly better results than matching them to the passing score; (b) all of the test designs worked better than test construction using random selection of items, subject to content constraints only; (c) CAT performed better than the other test designs; and (d) if matching a TIP to the passing score, the MST design produced a bit better results than the LPFT design. If an argument for the MST design is to be made, it can be made on the basis of slight improvements over the LPFT design and better expected item bank utilization, candidate preference, and the potential for improved diagnostic feedback, compared with the feedback that is possible with fixed linear test forms. (PsycINFO Database Record (c) 2007 APA, all rights reserved) PB - Lawrence Erlbaum: US VL - 19 SN - 0895-7347 (Print); 1532-4818 (Electronic) ER - TY - JOUR T1 - Optimal Testing With Easy or Difficult Items in Computerized Adaptive Testing JF - Applied Psychological Measurement Y1 - 2006 A1 - Theo Eggen A1 - Verschoor, Angela J. AB -

Computerized adaptive tests (CATs) are individualized tests that, from a measurement point of view, are optimal for each individual, possibly under some practical conditions. In the present study, it is shown that maximum information item selection in CATs using an item bank that is calibrated with the one or the two-parameter logistic model results in each individual answering about 50% of the items correctly. Two item selection procedures giving easier (or more difficult) tests for students are presented and evaluated. Item selection on probability points of items yields good results only with the one-parameter logistic model and not with the two-parameter logistic model. An alternative selection procedure, based on maximum information at a shifted ability level, gives satisfactory results with both models. Index terms: computerized adaptive testing, item selection, item response theory

VL - 30 UR - http://apm.sagepub.com/content/30/5/379.abstract ER - TY - JOUR T1 - Optimal testing with easy or difficult items in computerized adaptive testing JF - Applied Psychological Measurement Y1 - 2006 A1 - Theo Eggen A1 - Verschoor, Angela J. KW - computer adaptive tests KW - individualized tests KW - Item Response Theory KW - item selection KW - Measurement AB - Computerized adaptive tests (CATs) are individualized tests that, from a measurement point of view, are optimal for each individual, possibly under some practical conditions. In the present study, it is shown that maximum information item selection in CATs using an item bank that is calibrated with the one- or the two-parameter logistic model results in each individual answering about 50% of the items correctly. Two item selection procedures giving easier (or more difficult) tests for students are presented and evaluated. Item selection on probability points of items yields good results only with the one-parameter logistic model and not with the two-parameter logistic model. An alternative selection procedure, based on maximum information at a shifted ability level, gives satisfactory results with both models. (PsycINFO Database Record (c) 2007 APA, all rights reserved) PB - Sage Publications: US VL - 30 SN - 0146-6216 (Print) ER - TY - JOUR T1 - Optimal Testlet Pool Assembly for Multistage Testing Designs JF - Applied Psychological Measurement Y1 - 2006 A1 - Ariel, Adelaide A1 - Veldkamp, Bernard P. A1 - Breithaupt, Krista AB -

Computerized multistage testing (MST) designs require sets of test questions (testlets) to be assembled to meet strict, often competing criteria. Rules that govern testlet assembly may dictate the number of questions on a particular subject or may describe desirable statistical properties for the test, such as measurement precision. In an MST design, testlets of differing difficulty levels must be created. Statistical properties for assembly of the testlets can be expressed using item response theory (IRT) parameters. The testlet test information function (TIF) value can be maximized at a specific point on the IRT ability scale. In practical MST designs, parallel versions of testlets are needed, so sets of testlets with equivalent properties are built according to equivalent specifications. In this project, the authors study the use of a mathematical programming technique to simultaneously assemble testlets to ensure equivalence and fairness to candidates who may be administered different testlets.

VL - 30 UR - http://apm.sagepub.com/content/30/3/204.abstract ER - TY - JOUR T1 - Overview of quantitative measurement methods. Equivalence, invariance, and differential item functioning in health applications JF - Medical Care Y1 - 2006 A1 - Teresi, J. A. KW - *Cross-Cultural Comparison KW - Data Interpretation, Statistical KW - Factor Analysis, Statistical KW - Guidelines as Topic KW - Humans KW - Models, Statistical KW - Psychometrics/*methods KW - Statistics as Topic/*methods KW - Statistics, Nonparametric AB - BACKGROUND: Reviewed in this article are issues relating to the study of invariance and differential item functioning (DIF). The aim of factor analyses and DIF, in the context of invariance testing, is the examination of group differences in item response conditional on an estimate of disability. Discussed are parameters and statistics that are not invariant and cannot be compared validly in crosscultural studies with varying distributions of disability in contrast to those that can be compared (if the model assumptions are met) because they are produced by models such as linear and nonlinear regression. OBJECTIVES: The purpose of this overview is to provide an integrated approach to the quantitative methods used in this special issue to examine measurement equivalence. The methods include classical test theory (CTT), factor analytic, and parametric and nonparametric approaches to DIF detection. Also included in the quantitative section is a discussion of item banking and computerized adaptive testing (CAT). METHODS: Factorial invariance and the articles discussing this topic are introduced. A brief overview of the DIF methods presented in the quantitative section of the special issue is provided together with a discussion of ways in which DIF analyses and examination of invariance using factor models may be complementary. CONCLUSIONS: Although factor analytic and DIF detection methods share features, they provide unique information and can be viewed as complementary in informing about measurement equivalence. VL - 44 SN - 0025-7079 (Print)0025-7079 (Linking) N1 - Teresi, Jeanne AAG15294/AG/NIA NIH HHS/United StatesResearch Support, N.I.H., ExtramuralResearch Support, Non-U.S. Gov'tReviewUnited StatesMedical careMed Care. 2006 Nov;44(11 Suppl 3):S39-49. ER - TY - ABST T1 - Optimal testing with easy items in computerized adaptive testing (Measurement and Research Department Report 2004-2) Y1 - 2004 A1 - Theo Eggen A1 - Verschoor, A. J. CY - Arnhem, The Netherlands: Cito Group ER - TY - CONF T1 - Online calibration and scale stability of a CAT program T2 - Paper presented at the annual meeting of the American Educational Research Association Y1 - 2003 A1 - Guo, F. A1 - Wang, G. JF - Paper presented at the annual meeting of the American Educational Research Association CY - Chicago IL N1 - {PDF file, 274 KB} ER - TY - JOUR T1 - An optimal design approach to criterion-referenced computerized testing JF - Journal of Educational Measurement Y1 - 2003 A1 - Wiberg, M. VL - 28 ER - TY - JOUR T1 - Optimal stratification of item pools in α-stratified computerized adaptive testing JF - Applied Psychological Measurement Y1 - 2003 A1 - Chang, Hua-Hua A1 - van der Linden, W. J. KW - Adaptive Testing KW - Computer Assisted Testing KW - Item Content (Test) KW - Item Response Theory KW - Mathematical Modeling KW - Test Construction computerized adaptive testing AB - A method based on 0-1 linear programming (LP) is presented to stratify an item pool optimally for use in α-stratified adaptive testing. Because the 0-1 LP model belongs to the subclass of models with a network flow structure, efficient solutions are possible. The method is applied to a previous item pool from the computerized adaptive testing (CAT) version of the Graduate Record Exams (GRE) Quantitative Test. The results indicate that the new method performs well in practical situations. It improves item exposure control, reduces the mean squared error in the θ estimates, and increases test reliability. (PsycINFO Database Record (c) 2005 APA ) (journal abstract) VL - 27 ER - TY - CONF T1 - Optimal testing with easy items in computerized adaptive testing T2 - Paper presented at the conference of the International Association for Educational Assessment Y1 - 2003 A1 - Theo Eggen A1 - Verschoor, A. JF - Paper presented at the conference of the International Association for Educational Assessment CY - Manchester UK ER - TY - CONF T1 - Optimum number of strata in the a-stratified adaptive testing design T2 - Paper presented at the annual meeting of the American Educational Research Association Y1 - 2002 A1 - Wen, J.-B. A1 - Chang, Hua-Hua A1 - Hau, K-T. JF - Paper presented at the annual meeting of the American Educational Research Association CY - New Orleans LA N1 - {PDF file, 114 KB} ER - TY - JOUR T1 - Outlier detection in high-stakes certification testing JF - Journal of Educational Measurement Y1 - 2002 A1 - Meijer, R. R. KW - Adaptive Testing KW - computerized adaptive testing KW - Educational Measurement KW - Goodness of Fit KW - Item Analysis (Statistical) KW - Item Response Theory KW - person Fit KW - Statistical Estimation KW - Statistical Power KW - Test Scores AB - Discusses recent developments of person-fit analysis in computerized adaptive testing (CAT). Methods from statistical process control are presented that have been proposed to classify an item score pattern as fitting or misfitting the underlying item response theory model in CAT Most person-fit research in CAT is restricted to simulated data. In this study, empirical data from a certification test were used. Alternatives are discussed to generate norms so that bounds can be determined to classify an item score pattern as fitting or misfitting. Using bounds determined from a sample of a high-stakes certification test, the empirical analysis showed that different types of misfit can be distinguished Further applications using statistical process control methods to detect misfitting item score patterns are discussed. (PsycINFO Database Record (c) 2005 APA ) VL - 39 ER - TY - CONF T1 - On-line Calibration Using PARSCALE Item Specific Prior Method: Changing Test Population and Sample Size T2 - Paper presented at National Council on Measurement in Education Annual Meeting Y1 - 2001 A1 - Guo, F. A1 - Stone, E. A1 - Cruz, D. JF - Paper presented at National Council on Measurement in Education Annual Meeting CY - Seattle, Washington ER - TY - ABST T1 - Online item parameter recalibration: Application of missing data treatments to overcome the effects of sparse data conditions in a computerized adaptive version of the MCAT Y1 - 2001 A1 - Harmes, J. C. A1 - Kromrey, J. D. A1 - Parshall, C. G. CY - Unpublished manuscript N1 - {PDF file, 406 KB} ER - TY - JOUR T1 - Outlier measures and norming methods for computerized adaptive tests JF - Journal of Educational and Behavioral Statistics Y1 - 2001 A1 - Bradlow, E. T. A1 - Weiss, R. E. KW - Adaptive Testing KW - Computer Assisted Testing KW - Statistical Analysis KW - Test Norms AB - Notes that the problem of identifying outliers has 2 important aspects: the choice of outlier measures and the method to assess the degree of outlyingness (norming) of those measures. Several classes of measures for identifying outliers in Computerized Adaptive Tests (CATs) are introduced. Some of these measures are constructed to take advantage of CATs' sequential choice of items; other measures are taken directly from paper and pencil (P&P) tests and are used for baseline comparisons. Assessing the degree of outlyingness of CAT responses, however, can not be applied directly from P&P tests because stopping rules associated with CATs yield examinee responses of varying lengths. Standard outlier measures are highly correlated with the varying lengths which makes comparison across examinees impossible. Therefore, 4 methods are presented and compared which map outlier statistics to a familiar probability scale (a p value). The methods are explored in the context of CAT data from a 1995 Nationally Administered Computerized Examination (NACE). (PsycINFO Database Record (c) 2005 APA ) VL - 26 ER - TY - ABST T1 - Overexposure and underexposure of items in computerized adaptive testing (Measurement and Research Department Reports 2001-1) Y1 - 2001 A1 - Theo Eggen CY - Arnhem, The Netherlands: CITO Groep N1 - {PDF file, 276 KB} ER - TY - ABST T1 - Optimal stratification of item pools in a-stratified computerized adaptive testing (Research Report 00-07) Y1 - 2000 A1 - van der Linden, W. J. CY - Enschede, The Netherlands: University of Twente, Faculty of Educational Science and Technology, Department of Measurement and Data Analysis ER - TY - JOUR T1 - Overview of the computerized adaptive testing special section JF - Psicológica Y1 - 2000 A1 - Ponsoda, V. KW - Adaptive Testing KW - Computers computerized adaptive testing AB - This paper provides an overview of the five papers included in the Psicologica special section on computerized adaptive testing. A short introduction to this topic is presented as well. The main results, the links between the five papers and the general research topic to which they are more related are also shown. (PsycINFO Database Record (c) 2005 APA ) VL - 21 ER - TY - CONF T1 - On-the-fly adaptive tests: An application of generative modeling to quantitative reasoning T2 - Symposium presented at the annual meeting of the National Council on Measurement in Education Y1 - 1999 A1 - Bejar, I. I. JF - Symposium presented at the annual meeting of the National Council on Measurement in Education CY - Montreal, Canada ER - TY - JOUR T1 - Optimal design for item calibration in computerized adaptive testing JF - Dissertation Abstracts International: Section B: the Sciences & Engineering Y1 - 1999 A1 - Buyske, S. G. KW - computerized adaptive testing AB - Item Response Theory is the psychometric model used for standardized tests such as the Graduate Record Examination. A test-taker's response to an item is modelled as a binary response with success probability depending on parameters for both the test-taker and the item. Two popular models are the two-parameter logistic (2PL) model and the three-parameter logistic (3PL) model. For the 2PL model, the logit of the probability of a correct response equals ai(theta j-bi), where ai and bi are item parameters, while thetaj is the test-taker's parameter, known as "proficiency." The 3PL model adds a nonzero left asymptote to model random response behavior by low theta test-takers. Assigning scores to students requires accurate estimation of theta s, while accurate estimation of theta s requires accurate estimation of the item parameters. The operational implementation of Item Response Theory, particularly following the advent of computerized adaptive testing, generally involves handling these two estimation problems separately. This dissertation addresses the optimal design for item parameter estimation. Most current designs calibrate items with a sample drawn from the overall test-taking population. For 2PL models a sequential design based on the D-optimality criterion has been proposed, while no 3PL design is in the literature. In this dissertation, we design the calibration with the ultimate use of the items in mind, namely to estimate test-takers' proficiency parameters. For both the 2PL and 3PL models, this criterion leads to a locally L-optimal design criterion, named the Minimal Information Loss criterion. In turn, this criterion and the General Equivalence Theorem give a two point design for the 2PL model and a three point design for the 3PL model. A sequential implementation of this optimal design is presented. For the 2PL model, this design is almost 55% more efficient than the simple random sample approach, and 12% more efficient than the locally D-optimal design. For the 3PL model, the proposed design is 34% more efficient than the simple random sample approach. (PsycINFO Database Record (c) 2003 APA, all rights reserved). VL - 59 ER - TY - JOUR T1 - Optimal design of item pools for computerized adaptive testing JF - Applied Psychological Measurement Y1 - 1998 A1 - Stocking, M. L. A1 - Swanson, L. VL - 22 ER - TY - JOUR T1 - Optimal sequential rules for computer-based instruction JF - Journal of Educational Computing Research Y1 - 1998 A1 - Vos, H. J. VL - 19(2) ER - TY - JOUR T1 - Optimal test assembly of psychological and educational tests JF - Applied Psychological Measurement Y1 - 1998 A1 - van der Linden, W. J. VL - 22 ER - TY - JOUR T1 - On-line performance assessment using rating scales JF - Journal of Outcomes Measurement Y1 - 1997 A1 - Stahl, J. A1 - Shumway, R. A1 - Bergstrom, B. A1 - Fisher, A. KW - *Outcome Assessment (Health Care) KW - *Rehabilitation KW - *Software KW - *Task Performance and Analysis KW - Activities of Daily Living KW - Humans KW - Microcomputers KW - Psychometrics KW - Psychomotor Performance AB - The purpose of this paper is to report on the development of the on-line performance assessment instrument--the Assessment of Motor and Process Skills (AMPS). Issues that will be addressed in the paper include: (a) the establishment of the scoring rubric and its implementation in an extended Rasch model, (b) training of raters, (c) validation of the scoring rubric and procedures for monitoring the internal consistency of raters, and (d) technological implementation of the assessment instrument in a computerized program. VL - 1 N1 - 1090-655X (Print)Journal Article ER - TY - BOOK T1 - Optimization methods in computerized adaptive testing Y1 - 1997 A1 - Cordova, M. J. CY - Unpublished doctoral dissertation, Rutgers University, New Brunswick NJ ER - TY - CONF T1 - Overview of practical issues in a CAT program T2 - Paper presented at the annual meeting of the National Council on Measurement in Education Y1 - 1997 A1 - Wise, S. L. JF - Paper presented at the annual meeting of the National Council on Measurement in Education CY - Chicago IL N1 - [ERIC ED 408 330] ER - TY - CONF T1 - An overview of the LSAC CAT research agenda T2 - Paper presented at the annual meeting of the National Council on Measurement in Education Y1 - 1997 A1 - Pashley, P. JF - Paper presented at the annual meeting of the National Council on Measurement in Education CY - Chicago IL ER - TY - CONF T1 - Overview of the USMLE Step 2 computerized field test T2 - Paper presented at the annual meeting of the National Council on Measurement in Education Y1 - 1997 A1 - Luecht, RM A1 - Nungester, R. J. JF - Paper presented at the annual meeting of the National Council on Measurement in Education CY - Chicago IL ER - TY - ABST T1 - Optimal design of item pools for computerized adaptive testing (Research Report 96-34) Y1 - 1996 A1 - Stocking, M. L. A1 - Swanson, L. CY - Princeton NJ: Educational Testing Service ER - TY - JOUR T1 - Operational Characteristics of Adaptive Testing Procedures Using the Graded Response Model JF - Applied Psychological Measurement Y1 - 1989 A1 - Dodd, B. G. A1 - Koch, W. R. A1 - De Ayala, R. J. VL - 13 IS - 2 ER - TY - JOUR T1 - Operational characteristics of adaptive testing procedures using the graded response model JF - Applied Psychological Measurement Y1 - 1989 A1 - Dodd, B. G. A1 - Koch, W. R. A1 - De Ayala, R. J., VL - 13 ER - TY - CONF T1 - Operational characteristics of adaptive testing procedures using partial credit scoring T2 - Paper presented at the annual meeting of the American Educational Research Association Y1 - 1986 A1 - Koch, W. R. A1 - Dodd. B. G. JF - Paper presented at the annual meeting of the American Educational Research Association CY - San Francisco CA N1 - #KO86-01 ER - TY - JOUR T1 - Optimal item difficulty for the three-parameter normal ogive model JF - Psychometrika Y1 - 1981 A1 - Wolfe, J. H. VL - 46 ER - TY - JOUR T1 - Operational characteristics of a one-parameter tailored testing procedure JF - Catalog of Selected Documents in Psychology Y1 - 1980 A1 - Patience, W. M., A1 - Reckase, M. D. VL - August 1980 N1 - No. 2104). ER - TY - CONF T1 - Operational characteristics of a Rasch model tailored testing procedure when program parameters and item pool attributes are varied T2 - Paper presented at the annual meeting of the National Council on Measurement in Education Y1 - 1979 A1 - Patience, W. M. A1 - Reckase, M. D. JF - Paper presented at the annual meeting of the National Council on Measurement in Education CY - San Francisco ER - TY - CHAP T1 - Operational Considerations in Implementing Tailored Testing Y1 - 1977 A1 - Segal, H. CY - D. J. Weiss (Ed.), Proceedings of the 1977 Computerized Adaptive Testing Conference. Minneapolis MN: University of Minnesota, Department of Psychology, Psychometric Methods Program. ER - TY - CHAP T1 - Opening remarks Y1 - 1976 A1 - Gorham, W. A. CY - W. H. Gorham (Chair), Computers and testing: Steps toward the inevitable conquest (PS 76-1). Symposium presented at the 83rd annual convention of the APA, Chicago IL. Washington DC: U.S. Civil Service Commission, Personnel Research and Developement Center ER - TY - ABST T1 - An overview of tailored testing (unpublished manuscript) Y1 - 1973 A1 - Olivier, P. CY - Florida State University, Program of Educational Evaluation and Research Design ER -