%0 Book Section %B Elements of Adaptive Testing %D 2010 %T Constrained Adaptive Testing with Shadow Tests %A van der Linden, W. J. %B Elements of Adaptive Testing %P 31-56 %G eng %& 2 %R 10.1007/978-0-387-85461-8 %0 Book Section %B Elements of Adaptive Testing %D 2010 %T Designing Item Pools for Adaptive Testing %A Veldkamp, B. P. %A van der Linden, W. J. %B Elements of Adaptive Testing %P 231-245 %G eng %& 12 %R 10.1007/978-0-387-85461-8 %0 Book %D 2010 %T Elements of Adaptive Testing %A van der Linden, W. J. %A Glas, C. A. W. %I Springer %C New York %P 437 %G eng %R 10.1007/978-0-387-85461-8 %0 Book Section %B Elements of Adaptive Testing %D 2010 %T Estimation of the Parameters in an Item-Cloning Model for Adaptive Testing %A Glas, C. A. W. %A van der Linden, W. J. %A Geerlings, H. %B Elements of Adaptive Testing %P 289-314 %G eng %& 15 %R 10.1007/978-0-387-85461-8 %0 Book Section %B Elements of Adaptive Testing %D 2010 %T Item Selection and Ability Estimation in Adaptive Testing %A van der Linden, W. J. %A Pashley, P. J. %B Elements of Adaptive Testing %I Springer %C New York %P 3-30 %G eng %& 1 %R 10.1007/978-0-387-85461-8 %0 Journal Article %J British Journal of Mathematical and Statistical Psychology %D 2010 %T Marginal likelihood inference for a model for item responses and response times %A Glas, C. A. W. %A van der Linden, W. J. %X

Marginal maximum-likelihood procedures for parameter estimation and testing the fit of a hierarchical model for speed and accuracy on test items are presented. The model is a composition of two first-level models for dichotomous responses and response times along with multivariate normal models for their item and person parameters. It is shown how the item parameters can easily be estimated using Fisher's identity. To test the fit of the model, Lagrange multiplier tests of the assumptions of subpopulation invariance of the item parameters (i.e., no differential item functioning), the shape of the response functions, and three different types of conditional independence were derived. Simulation studies were used to show the feasibility of the estimation and testing procedures and to estimate the power and Type I error rate of the latter. In addition, the procedures were applied to an empirical data set from a computerized adaptive test of language comprehension.

%B British Journal of Mathematical and Statistical Psychology %7 2010/01/30 %V 63 %P 603-26 %@ 0007-1102 (Print)0007-1102 (Linking) %G eng %M 20109271 %0 Book Section %B Elements of Adaptive Testing %D 2010 %T Multidimensional Adaptive Testing with Kullback–Leibler Information Item Selection %A Mulder, J. %A van der Linden, W. J. %B Elements of Adaptive Testing %P 77-102 %G eng %& 4 %R 10.1007/978-0-387-85461-8 %0 Book Section %B Elements of Adaptive Testing %D 2010 %T Sequencing an Adaptive Test Battery %A van der Linden, W. J. %B Elements of Adaptive Testing %G eng %& 5 %R 10.1007/978-0-387-85461-8 %0 Journal Article %J Psychometrika %D 2009 %T Multidimensional Adaptive Testing with Optimal Design Criteria for Item Selection %A Mulder, J. %A van der Linden, W. J. %X Several criteria from the optimal design literature are examined for use with item selection in multidimensional adaptive testing. In particular, it is examined what criteria are appropriate for adaptive testing in which all abilities are intentional, some should be considered as a nuisance, or the interest is in the testing of a composite of the abilities. Both the theoretical analyses and the studies of simulated data in this paper suggest that the criteria of A-optimality and D-optimality lead to the most accurate estimates when all abilities are intentional, with the former slightly outperforming the latter. The criterion of E-optimality showed occasional erratic behavior for this case of adaptive testing, and its use is not recommended. If some of the abilities are nuisances, application of the criterion of A(s)-optimality (or D(s)-optimality), which focuses on the subset of intentional abilities is recommended. For the measurement of a linear combination of abilities, the criterion of c-optimality yielded the best results. The preferences of each of these criteria for items with specific patterns of parameter values was also assessed. It was found that the criteria differed mainly in their preferences of items with different patterns of values for their discrimination parameters. %B Psychometrika %7 2010/02/02 %V 74 %P 273-296 %8 Jun %@ 0033-3123 (Print)0033-3123 (Linking) %G Eng %M 20119511 %2 2813188 %0 Journal Article %J Zeitschrift für Psychologie / Journal of Psychology %D 2008 %T Adaptive models of psychological testing %A van der Linden, W. J. %B Zeitschrift für Psychologie / Journal of Psychology %V 216(1) %P 3–11 %G eng %0 Journal Article %J Zeitschrift für Psychologie / Journal of Psychology %D 2008 %T Adaptive Models of Psychological Testing %A van der Linden, W. J. %B Zeitschrift für Psychologie / Journal of Psychology %V 216 %P 1-2 %N 1 %R 10.1027/0044-3409.216.1.49 %0 Journal Article %J Zeitschrift für Psychologie %D 2008 %T Some new developments in adaptive testing technology %A van der Linden, W. J. %K computerized adaptive testing %X

In an ironic twist of history, modern psychological testing has returned to an adaptive format quite common when testing was not yet standardized. Important stimuli to the renewed interest in adaptive testing have been the development of item-response theory in psychometrics, which models the responses on test items using separate parameters for the items and test takers, and the use of computers in test administration, which enables us to estimate the parameter for a test taker and select the items in real time. This article reviews a selection from the latest developments in the technology of adaptive testing, such as constrained adaptive item selection, adaptive testing using rule-based item generation, multidimensional adaptive testing, adaptive use of test batteries, and the use of response times in adaptive testing.

%B Zeitschrift für Psychologie %V 216 %P 3-11 %G eng %0 Journal Article %J Journal of Educational and Behavioral Statistics %D 2008 %T Using response times for item selection in adaptive testing %A van der Linden, W. J. %B Journal of Educational and Behavioral Statistics %V 33 %P 5–20 %G eng %0 Book Section %D 2007 %T The shadow-test approach: A universal framework for implementing adaptive testing %A van der Linden, W. J. %C D. J. Weiss (Ed.), Proceedings of the 2007 GMAC Conference on Computerized Adaptive Testing. %G eng %0 Book Section %D 2007 %T Statistical aspects of adaptive testing %A van der Linden, W. J. %A Glas, C. A. W. %C C. R. Rao and S. Sinharay (Eds.), Handbook of statistics (Vol. 27: Psychometrics) (pp. 801838). Amsterdam: North-Holland. %G eng %0 Journal Article %J Journal of Educational and Behavioral Statistics %D 2006 %T Assembling a computerized adaptive testing item pool as a set of linear tests %A van der Linden, W. J. %A Ariel, A. %A Veldkamp, B. P. %K Algorithms %K computerized adaptive testing %K item pool %K linear tests %K mathematical models %K statistics %K Test Construction %K Test Items %X Test-item writing efforts typically results in item pools with an undesirable correlational structure between the content attributes of the items and their statistical information. If such pools are used in computerized adaptive testing (CAT), the algorithm may be forced to select items with less than optimal information, that violate the content constraints, and/or have unfavorable exposure rates. Although at first sight somewhat counterintuitive, it is shown that if the CAT pool is assembled as a set of linear test forms, undesirable correlations can be broken down effectively. It is proposed to assemble such pools using a mixed integer programming model with constraints that guarantee that each test meets all content specifications and an objective function that requires them to have maximal information at a well-chosen set of ability values. An empirical example with a previous master pool from the Law School Admission Test (LSAT) yielded a CAT with nearly uniform bias and mean-squared error functions for the ability estimator and item-exposure rates that satisfied the target for all items in the pool. %B Journal of Educational and Behavioral Statistics %I Sage Publications: US %V 31 %P 81-99 %@ 1076-9986 (Print) %G eng %M 2007-08137-004 %0 Journal Article %J Applied Psychological Measurement %D 2006 %T Equating scores from adaptive to linear tests %A van der Linden, W. J. %K computerized adaptive testing %K equipercentile equating %K local equating %K score reporting %K test characteristic function %X Two local methods for observed-score equating are applied to the problem of equating an adaptive test to a linear test. In an empirical study, the methods were evaluated against a method based on the test characteristic function (TCF) of the linear test and traditional equipercentile equating applied to the ability estimates on the adaptive test for a population of test takers. The two local methods were generally best. Surprisingly, the TCF method performed slightly worse than the equipercentile method. Both methods showed strong bias and uniformly large inaccuracy, but the TCF method suffered from extra error due to the lower asymptote of the test characteristic function. It is argued that the worse performances of the two methods are a consequence of the fact that they use a single equating transformation for an entire population of test takers and therefore have to compromise between the individual score distributions. %B Applied Psychological Measurement %I Sage Publications: US %V 30 %P 493-508 %@ 0146-6216 (Print) %G eng %M 2006-20197-003 %0 Journal Article %J Journal of Educational Measurement %D 2005 %T A comparison of item-selection methods for adaptive tests with content constraints %A van der Linden, W. J. %B Journal of Educational Measurement %V 42 %P 283-302 %G eng %0 Journal Article %J Journal of Educational Measurement %D 2005 %T A comparison of item-selection methods for adaptive tests with content constraints %A van der Linden, W. J. %K Adaptive Testing %K Algorithms %K content constraints %K item selection method %K shadow test approach %K spiraling method %K weighted deviations method %X In test assembly, a fundamental difference exists between algorithms that select a test sequentially or simultaneously. Sequential assembly allows us to optimize an objective function at the examinee's ability estimate, such as the test information function in computerized adaptive testing. But it leads to the non-trivial problem of how to realize a set of content constraints on the test—a problem more naturally solved by a simultaneous item-selection method. Three main item-selection methods in adaptive testing offer solutions to this dilemma. The spiraling method moves item selection across categories of items in the pool proportionally to the numbers needed from them. Item selection by the weighted-deviations method (WDM) and the shadow test approach (STA) is based on projections of the future consequences of selecting an item. These two methods differ in that the former calculates a projection of a weighted sum of the attributes of the eventual test and the latter a projection of the test itself. The pros and cons of these methods are analyzed. An empirical comparison between the WDM and STA was conducted for an adaptive version of the Law School Admission Test (LSAT), which showed equally good item-exposure rates but violations of some of the constraints and larger bias and inaccuracy of the ability estimator for the WDM. %B Journal of Educational Measurement %I Blackwell Publishing: United Kingdom %V 42 %P 283-302 %@ 0022-0655 (Print) %G eng %M 2005-10716-004 %0 Generic %D 2005 %T Constraining item exposure in computerized adaptive testing with shadow tests %A van der Linden, W. J. %A Veldkamp, B. P. %C Law School Admission Council Computerized Testing Report 02-03 %G eng %0 Generic %D 2005 %T Implementing content constraints in alpha-stratified adaptive testing using a shadow test approach %A van der Linden, W. J. %A Chang, Hua-Hua %C Law School Admission Council, Computerized Testing Report 01-09 %G eng %0 Journal Article %J Journal of Educational and Behavioral Statistics %D 2004 %T Constraining item exposure in computerized adaptive testing with shadow tests %A van der Linden, W. J. %A Veldkamp, B. P. %K computerized adaptive testing %K item exposure control %K item ineligibility constraints %K Probability %K shadow tests %X Item-exposure control in computerized adaptive testing is implemented by imposing item-ineligibility constraints on the assembly process of the shadow tests. The method resembles Sympson and Hetter’s (1985) method of item-exposure control in that the decisions to impose the constraints are probabilistic. The method does not, however, require time-consuming simulation studies to set values for control parameters before the operational use of the test. Instead, it can set the probabilities of item ineligibility adaptively during the test using the actual item-exposure rates. An empirical study using an item pool from the Law School Admission Test showed that application of the method yielded perfect control of the item-exposure rates and had negligible impact on the bias and mean-squared error functions of the ability estimator. %B Journal of Educational and Behavioral Statistics %I American Educational Research Assn: US %V 29 %P 273-291 %@ 1076-9986 (Print) %G eng %M 2006-01687-001 %0 Journal Article %J Journal of Educational Measurement %D 2004 %T Constructing rotating item pools for constrained adaptive testing %A Ariel, A. %A Veldkamp, B. P. %A van der Linden, W. J. %K computerized adaptive tests %K constrained adaptive testing %K item exposure %K rotating item pools %X Preventing items in adaptive testing from being over- or underexposed is one of the main problems in computerized adaptive testing. Though the problem of overexposed items can be solved using a probabilistic item-exposure control method, such methods are unable to deal with the problem of underexposed items. Using a system of rotating item pools, on the other hand, is a method that potentially solves both problems. In this method, a master pool is divided into (possibly overlapping) smaller item pools, which are required to have similar distributions of content and statistical attributes. These pools are rotated among the testing sites to realize desirable exposure rates for the items. A test assembly model, motivated by Gulliksen's matched random subtests method, was explored to help solve the problem of dividing a master pool into a set of smaller pools. Different methods to solve the model are proposed. An item pool from the Law School Admission Test was used to evaluate the performances of computerized adaptive tests from systems of rotating item pools constructed using these methods. (PsycINFO Database Record (c) 2007 APA, all rights reserved) %B Journal of Educational Measurement %I Blackwell Publishing: United Kingdom %V 41 %P 345-359 %@ 0022-0655 (Print) %G eng %M 2004-21596-004 %0 Conference Paper %B Paper presented at the annual meeting of the National Council on Measurement in Education %D 2004 %T A sequential Bayesian procedure for item calibration in multistage testing %A van der Linden, W. J. %A Alan D Mead %B Paper presented at the annual meeting of the National Council on Measurement in Education %C San Diego CA %G eng %0 Journal Article %J Applied Psychological Measurement %D 2003 %T Alpha-stratified adaptive testing with large numbers of content constraints %A van der Linden, W. J. %A Chang, Hua-Hua %B Applied Psychological Measurement %V 27 %P 107-120 %G eng %0 Book Section %D 2003 %T Bayesian checks on outlying response times in computerized adaptive testing %A van der Linden, W. J. %C H. Yanai, A. Okada, K. Shigemasu, Y. Kano, Y. and J. J. Meulman, (Eds.), New developments in psychometrics (pp. 215-222). New York: Springer-Verlag. %G eng %0 Journal Article %J Applied Psychological Measurement %D 2003 %T Computerized adaptive testing with item cloning %A Glas, C. A. W. %A van der Linden, W. J. %K computerized adaptive testing %X (from the journal abstract) To increase the number of items available for adaptive testing and reduce the cost of item writing, the use of techniques of item cloning has been proposed. An important consequence of item cloning is possible variability between the item parameters. To deal with this variability, a multilevel item response (IRT) model is presented which allows for differences between the distributions of item parameters of families of item clones. A marginal maximum likelihood and a Bayesian procedure for estimating the hyperparameters are presented. In addition, an item-selection procedure for computerized adaptive testing with item cloning is presented which has the following two stages: First, a family of item clones is selected to be optimal at the estimate of the person parameter. Second, an item is randomly selected from the family for administration. Results from simulation studies based on an item pool from the Law School Admission Test (LSAT) illustrate the accuracy of these item pool calibration and adaptive testing procedures. (PsycINFO Database Record (c) 2003 APA, all rights reserved). %B Applied Psychological Measurement %V 27 %P 247-261 %G eng %0 Conference Paper %B Paper presented at the Annual meeting of the National Council on Measurement in Education %D 2003 %T Constraining item exposure in computerized adaptive testing with shadow tests %A van der Linden, W. J. %A Veldkamp, B. P. %B Paper presented at the Annual meeting of the National Council on Measurement in Education %C Chicago IL %G eng %0 Conference Paper %B Paper presented at the Annual meeting of the National Council on Measurement in Education %D 2003 %T Constructing rotating item pools for constrained adaptive testing %A Ariel, A. %A Veldkamp, B. %A van der Linden, W. J. %B Paper presented at the Annual meeting of the National Council on Measurement in Education %C Chicago IL %G eng %0 Conference Paper %D 2003 %T Controlling item exposure and item eligibility in computerized adaptive testing %A van der Linden, W. J. %A Veldkamp, B. P. %G eng %0 Conference Paper %B Paper presented at the Annual meeting of the National Council on Measurement in Education %D 2003 %T Implementing an alternative to Sympson-Hetter item-exposure control in constrained adaptive testing %A Veldkamp, B. P. %A van der Linden, W. J. %B Paper presented at the Annual meeting of the National Council on Measurement in Education %C Chicago IL %G eng %0 Journal Article %J Applied Psychological Measurement %D 2003 %T Implementing content constraints in alpha-stratified adaptive testing using a shadow test approach %A van der Linden, W. J. %A Chang, Hua-Hua %B Applied Psychological Measurement %V 27 %P 107-120 %G eng %0 Journal Article %J Applied Psychological Measurement %D 2003 %T Optimal stratification of item pools in α-stratified computerized adaptive testing %A Chang, Hua-Hua %A van der Linden, W. J. %K Adaptive Testing %K Computer Assisted Testing %K Item Content (Test) %K Item Response Theory %K Mathematical Modeling %K Test Construction computerized adaptive testing %X A method based on 0-1 linear programming (LP) is presented to stratify an item pool optimally for use in α-stratified adaptive testing. Because the 0-1 LP model belongs to the subclass of models with a network flow structure, efficient solutions are possible. The method is applied to a previous item pool from the computerized adaptive testing (CAT) version of the Graduate Record Exams (GRE) Quantitative Test. The results indicate that the new method performs well in practical situations. It improves item exposure control, reduces the mean squared error in the θ estimates, and increases test reliability. (PsycINFO Database Record (c) 2005 APA ) (journal abstract) %B Applied Psychological Measurement %V 27 %P 262-274 %G eng %0 Generic %D 2003 %T A sequential Bayes procedure for item calibration in multi-stage testing %A van der Linden, W. J. %A Alan D Mead %C Manuscript in preparation %G eng %0 Journal Article %J Journal of Educational and Behavioral Statistics %D 2003 %T Some alternatives to Sympson-Hetter item-exposure control in computerized adaptive testing %A van der Linden, W. J. %K Adaptive Testing %K Computer Assisted Testing %K Test Items computerized adaptive testing %X TheHetter and Sympson (1997; 1985) method is a method of probabilistic item-exposure control in computerized adaptive testing. Setting its control parameters to admissible values requires an iterative process of computer simulations that has been found to be time consuming, particularly if the parameters have to be set conditional on a realistic set of values for the examinees’ ability parameter. Formal properties of the method are identified that help us explain why this iterative process can be slow and does not guarantee admissibility. In addition, some alternatives to the SH method are introduced. The behavior of these alternatives was estimated for an adaptive test from an item pool from the Law School Admission Test (LSAT). Two of the alternatives showed attractive behavior and converged smoothly to admissibility for all items in a relatively small number of iteration steps. %B Journal of Educational and Behavioral Statistics %V 28 %P 249-265 %G eng %0 Journal Article %J Psychometrika %D 2003 %T Using response times to detect aberrant responses in computerized adaptive testing %A van der Linden, W. J. %A van Krimpen-Stoop, E. M. L. A. %K Adaptive Testing %K Behavior %K Computer Assisted Testing %K computerized adaptive testing %K Models %K person Fit %K Prediction %K Reaction Time %X A lognormal model for response times is used to check response times for aberrances in examinee behavior on computerized adaptive tests. Both classical procedures and Bayesian posterior predictive checks are presented. For a fixed examinee, responses and response times are independent; checks based on response times offer thus information independent of the results of checks on response patterns. Empirical examples of the use of classical and Bayesian checks for detecting two different types of aberrances in response times are presented. The detection rates for the Bayesian checks outperformed those for the classical checks, but at the cost of higher false-alarm rates. A guideline for the choice between the two types of checks is offered. %B Psychometrika %V 68 %P 251-265 %G eng %0 Generic %D 2002 %T Constraining item exposure in computerized adaptive testing with shadow tests (Research Report No. 02-06) %A van der Linden, W. J. %A Veldkamp, B. P. %C University of Twente, The Netherlands %G eng %0 Report %D 2002 %T Mathematical-programming approaches to test item pool design %A Veldkamp, B. P. %A van der Linden, W. J. %A Ariel, A. %K Adaptive Testing %K Computer Assisted %K Computer Programming %K Educational Measurement %K Item Response Theory %K Mathematics %K Psychometrics %K Statistical Rotation computerized adaptive testing %K Test Items %K Testing %X (From the chapter) This paper presents an approach to item pool design that has the potential to improve on the quality of current item pools in educational and psychological testing and hence to increase both measurement precision and validity. The approach consists of the application of mathematical programming techniques to calculate optimal blueprints for item pools. These blueprints can be used to guide the item-writing process. Three different types of design problems are discussed, namely for item pools for linear tests, item pools computerized adaptive testing (CAT), and systems of rotating item pools for CAT. The paper concludes with an empirical example of the problem of designing a system of rotating item pools for CAT. %I University of Twente, Faculty of Educational Science and Technology %C Twente, The Netherlands %P 93-108 %@ 02-09 %G eng %0 Generic %D 2002 %T Modifications of the Sympson-Hetter method for item-exposure control in computerized adaptive testing %A van der Linden, W. J. %C Manuscript submitted for publication %G eng %0 Journal Article %J Psychometrika %D 2002 %T Multidimensional adaptive testing with constraints on test content %A Veldkamp, B. P. %A van der Linden, W. J. %X The case of adaptive testing under a multidimensional response model with large numbers of constraints on the content of the test is addressed. The items in the test are selected using a shadow test approach. The 0–1 linear programming model that assembles the shadow tests maximizes posterior expected Kullback-Leibler information in the test. The procedure is illustrated for five different cases of multidimensionality. These cases differ in (a) the numbers of ability dimensions that are intentional or should be considered as ldquonuisance dimensionsrdquo and (b) whether the test should or should not display a simple structure with respect to the intentional ability dimensions. %B Psychometrika %V 67 %P 575-588 %G eng %0 Journal Article %J Applied Psychological Measurement %D 2001 %T Computerized adaptive testing with equated number-correct scoring %A van der Linden, W. J. %X A constrained computerized adaptive testing (CAT) algorithm is presented that can be used to equate CAT number-correct (NC) scores to a reference test. As a result, the CAT NC scores also are equated across administrations. The constraints are derived from van der Linden & Luecht’s (1998) set of conditions on item response functions that guarantees identical observed NC score distributions on two test forms. An item bank from the Law School Admission Test was used to compare the results of the algorithm with those for equipercentile observed-score equating, as well as the prediction of NC scores on a reference test using its test response function. The effects of the constraints on the statistical properties of the θ estimator in CAT were examined. %B Applied Psychological Measurement %V 25 %P 343-355 %G eng %0 Generic %D 2001 %T Implementing content constraints in a-stratified adaptive testing using a shadow test approach (Research Report 01-001) %A Chang, Hua-Hua %A van der Linden, W. J. %C University of Twente, Department of Educational Measurement and Data Analysis %G eng %0 Conference Paper %B Paper presented at the Annual Meeting of the National Council on Measurement in Education %D 2001 %T Modeling variability in item parameters in CAT %A Glas, C. A. W. %A van der Linden, W. J. %B Paper presented at the Annual Meeting of the National Council on Measurement in Education %C Seattle WA %G eng %0 Conference Paper %B Paper presented at the annual meeting of the National Council on Measurement in Education %D 2001 %T Using response times to detect aberrant behavior in computerized adaptive testing %A van der Linden, W. J. %A van Krimpen-Stoop, E. M. L. A. %B Paper presented at the annual meeting of the National Council on Measurement in Education %C Seattle WA %G eng %0 Journal Article %J Applied Measurement in Education %D 2000 %T Capitalization on item calibration error in adaptive testing %A van der Linden, W. J. %A Glas, C. A. W. %K computerized adaptive testing %X (from the journal abstract) In adaptive testing, item selection is sequentially optimized during the test. Because the optimization takes place over a pool of items calibrated with estimation error, capitalization on chance is likely to occur. How serious the consequences of this phenomenon are depends not only on the distribution of the estimation errors in the pool or the conditional ratio of the test length to the pool size given ability, but may also depend on the structure of the item selection criterion used. A simulation study demonstrated a dramatic impact of capitalization on estimation errors on ability estimation. Four different strategies to minimize the likelihood of capitalization on error in computerized adaptive testing are discussed. %B Applied Measurement in Education %V 13 %P 35-53 %G eng %0 Book %D 2000 %T Computerized adaptive testing: Theory and practice %A van der Linden, W. J. %A Glas, C. A. W. %I Kluwer Academic Publishers %C Dordrecht, The Netherlands %G eng %0 Book Section %D 2000 %T Constrained adaptive testing with shadow tests %A van der Linden, W. J. %C W. J. van der Linden and C. A. W. Glas (eds.), Computerized adaptive testing: Theory and practice (pp.27-52). Norwell MA: Kluwer. %G eng %0 Book Section %D 2000 %T Cross-validating item parameter estimation in adaptive testing %A van der Linden, W. J. %A Glas, C. A. W. %C A. Boorsma, M. A. J. van Duijn, and T. A. B. Snijders (Eds.) (pp. 205-219), Essays on item response theory. New York: Springer. %G eng %0 Book Section %B Computerized adaptive testing: Theory and practice %D 2000 %T Designing item pools for computerized adaptive testing %A Veldkamp, B. P. %A van der Linden, W. J. %B Computerized adaptive testing: Theory and practice %I Kluwer Academic Publishers %C Dendrecht, The Netherlands %P 149–162 %G eng %0 Journal Article %J Applied Psychological Measurement %D 2000 %T An integer programming approach to item bank design %A van der Linden, W. J. %A Veldkamp, B. P. %A Reese, L. M. %K Aptitude Measures %K Item Analysis (Test) %K Item Response Theory %K Test Construction %K Test Items %X An integer programming approach to item bank design is presented that can be used to calculate an optimal blueprint for an item bank, in order to support an existing testing program. The results are optimal in that they minimize the effort involved in producing the items as revealed by current item writing patterns. Also presented is an adaptation of the models, which can be used as a set of monitoring tools in item bank management. The approach is demonstrated empirically for an item bank that was designed for the Law School Admission Test. %B Applied Psychological Measurement %V 24 %P 139-150 %G eng %0 Book Section %B Computerized adaptive testing: Theory and practice %D 2000 %T Item selection and ability estimation in adaptive testing %A van der Linden, W. J. %A Pashley, P. J. %B Computerized adaptive testing: Theory and practice %I Kluwer Academic Publishers %C Dordrecht, The Netherlands %P 1–25 %G eng %0 Generic %D 2000 %T Multidimensional adaptive testing with constraints on test content (Research Report 00-11) %A Veldkamp, B. P. %A van der Linden, W. J. %C Enschede, The Netherlands: University of Twente, Faculty of Educational Science and Technology, Department of Measurement and Data Analysis %G eng %0 Generic %D 2000 %T Optimal stratification of item pools in a-stratified computerized adaptive testing (Research Report 00-07) %A van der Linden, W. J. %C Enschede, The Netherlands: University of Twente, Faculty of Educational Science and Technology, Department of Measurement and Data Analysis %G eng %0 Generic %D 2000 %T Using response times to detect aberrant behavior in computerized adaptive testing (Research Report 00-09) %A van der Linden, W. J. %A van Krimpen-Stoop, E. M. L. A. %C Enschede, The Netherlands: University of Twente, Faculty of Educational Science and Technology, Department of Measurement and Data Analysis %G eng %0 Generic %D 1999 %T Adaptive testing with equated number-correct scoring (Research Report 99-02) %A van der Linden, W. J. %C Enschede, The Netherlands: University of Twente, Faculty of Educational Science and Technology, Department of Measurement and Data Analysis %G eng %0 Generic %D 1999 %T Designing item pools for computerized adaptive testing (Research Report 99-03 ) %A Veldkamp, B. P. %A van der Linden, W. J. %C Enschede, The Netherlands: University of Twente, Faculty of Educational Science and Technology, Department of Measurement and Data Analysis %G eng %0 Journal Article %J Applied Psychological Measurement %D 1999 %T Empirical initialization of the trait estimator in adaptive testing %A van der Linden, W. J. %B Applied Psychological Measurement %V 23 %P 21-29 %G eng %0 Book Section %D 1999 %T Het ontwerpen van adaptieve examens [Designing adaptive tests] %A van der Linden, W. J. %C J. M Pieters, Tj. Plomp, and L.E. Odenthal (Eds.), Twintig jaar Toegepaste Onderwijskunde: Een kaleidoscopisch overzicht van Twents onderwijskundig onderzoek (pp. 249-267). Enschede: Twente University Press. %G eng %0 Journal Article %J Journal of Educational and Behavioral Statistics %D 1999 %T Multidimensional adaptive testing with a minimum error-variance criterion %A van der Linden, W. J. %K computerized adaptive testing %X Adaptive testing under a multidimensional logistic response model is addressed. An algorithm is proposed that minimizes the (asymptotic) variance of the maximum-likelihood estimator of a linear combination of abilities of interest. The criterion results in a closed-form expression that is easy to evaluate. In addition, it is shown how the algorithm can be modified if the interest is in a test with a "simple ability structure". The statistical properties of the adaptive ML estimator are demonstrated for a two-dimensional item pool with several linear combinations of the abilities. %B Journal of Educational and Behavioral Statistics %V 24 %P 398-412 %G eng %M EJ607470 %0 Journal Article %J Applied Psychological Measurement %D 1999 %T Using response-time constraints to control for differential speededness in computerized adaptive testing %A van der Linden, W. J. %A Scrams, D. J. %A Schnipke, D. L. %K computerized adaptive testing %X An item-selection algorithm is proposed for neutralizing the differential effects of time limits on computerized adaptive test scores. The method is based on a statistical model for distributions of examinees’ response times on items in a bank that is updated each time an item is administered. Predictions from the model are used as constraints in a 0-1 linear programming model for constrained adaptive testing that maximizes the accuracy of the trait estimator. The method is demonstrated empirically using an item bank from the Armed Services Vocational Aptitude Battery. %B Applied Psychological Measurement %V 23 %P 195-210 %G eng %0 Journal Article %J Psychometrika %D 1998 %T Bayesian item selection criteria for adaptive testing %A van der Linden, W. J. %B Psychometrika %V 63 %P 201-216 %G eng %0 Generic %D 1998 %T Capitalization on item calibration error in adaptive testing (Research Report 98-07) %A van der Linden, W. J. %A Glas, C. A. W. %C Enschede, The Netherlands: University of Twente, Faculty of Educational Science and Technology, Department of Measurement and Data Analysis %G eng %0 Journal Article %J Applied Psychological Measurement %D 1998 %T A model for optimal constrained adaptive testing %A van der Linden, W. J. %A Reese, L. M. %K computerized adaptive testing %X A model for constrained computerized adaptive testing is proposed in which the information in the test at the trait level (0) estimate is maximized subject to a number of possible constraints on the content of the test. At each item-selection step, a full test is assembled to have maximum information at the current 0 estimate, fixing the items already administered. Then the item with maximum in-formation is selected. All test assembly is optimal because a linear programming (LP) model is used that automatically updates to allow for the attributes of the items already administered and the new value of the 0 estimator. The LP model also guarantees that each adaptive test always meets the entire set of constraints. A simulation study using a bank of 753 items from the Law School Admission Test showed that the 0 estimator for adaptive tests of realistic lengths did not suffer any loss of efficiency from the presence of 433 constraints on the item selection process. %B Applied Psychological Measurement %V 22 %P 259-270 %G eng %0 Journal Article %J Applied Psychological Measurement %D 1998 %T Optimal test assembly of psychological and educational tests %A van der Linden, W. J. %B Applied Psychological Measurement %V 22 %P 195-211 %G eng %0 Journal Article %J Psychometrika %D 1998 %T Stochastic order in dichotomous item response models for fixed, adaptive, and multidimensional tests %A van der Linden, W. J. %B Psychometrika %V 63 %P 211-226 %G eng %0 Generic %D 1998 %T Using response-time constraints to control for differential speededness in adaptive testing (Research Report 98-06) %A van der Linden, W. J. %A Scrams, D. J. %A Schnipke, D. L. %C Enschede, The Netherlands: University of Twente, Faculty of Educational Science and Technology, Department of Measurement and Data Analysis %G eng %0 Conference Paper %B Paper presented at the annual meeting of the National Council on Measurement in Education %D 1997 %T Detection of aberrant response patterns in CAT %A van der Linden, W. J. %B Paper presented at the annual meeting of the National Council on Measurement in Education %C Chicago IL %G eng %0 Conference Paper %B Paper presented at the annual meeting of the American Educational Research Association %D 1997 %T Multidimensional adaptive testing with a minimum error-variance criterion %A van der Linden, W. J. %B Paper presented at the annual meeting of the American Educational Research Association %C Chicago %G eng %0 Generic %D 1997 %T Multidimensional adaptive testing with a minimum error-variance criterion (Research Report 97-03) %A van der Linden, W. J. %C Enschede, The Netherlands: University of Twente, Department of Educational Measurement and Data Analysis %G eng %0 Journal Article %J Psychometrika %D 1996 %T Bayesian item selection criteria for adaptive testing %A van der Linden, W. J. %B Psychometrika %V 63 %P 201-216 %G eng %0 Generic %D 1996 %T Bayesian item selection criteria for adaptive testing (Research Report 96-01) %A van der Linden, W. J. %C Twente, The Netherlands: Department of Educational Measurement and Data Analysis %G eng %0 Conference Paper %B Paper presented at the Annual Meeting of the Psychometric Society %D 1995 %T Bayesian item selection in adaptive testing %A van der Linden, W. J. %B Paper presented at the Annual Meeting of the Psychometric Society %C Minneapolis MN %G eng %0 Journal Article %J International Journal of Educational Research %D 1989 %T Some procedures for computerized ability testing %A van der Linden, W. J. %A Zwarts, M. A. %B International Journal of Educational Research %V 13(2) %P 175-187 %G eng