%0 Journal Article %J Applied Psychological Measurement %D 2019 %T Adaptive Testing With a Hierarchical Item Response Theory Model %A Wenhao Wang %A Neal Kingston %X The hierarchical item response theory (H-IRT) model is very flexible and allows a general factor and subfactors within an overall structure of two or more levels. When an H-IRT model with a large number of dimensions is used for an adaptive test, the computational burden associated with interim scoring and selection of subsequent items is heavy. An alternative approach for any high-dimension adaptive test is to reduce dimensionality for interim scoring and item selection and then revert to full dimensionality for final score reporting, thereby significantly reducing the computational burden. This study compared the accuracy and efficiency of final scoring for multidimensional, local multidimensional, and unidimensional item selection and interim scoring methods, using both simulated and real item pools. The simulation study was conducted under 10 conditions (i.e., five test lengths and two H-IRT models) with a simulated sample of 10,000 students. The study with the real item pool was conducted using item parameters from an actual 45-item adaptive test with a simulated sample of 10,000 students. Results indicate that the theta estimations provided by the local multidimensional and unidimensional item selection and interim scoring methods were relatively as accurate as the theta estimation provided by the multidimensional item selection and interim scoring method, especially during the real item pool study. In addition, the multidimensional method required the longest computation time and the unidimensional method required the shortest computation time. %B Applied Psychological Measurement %V 43 %P 51-67 %U https://doi.org/10.1177/0146621618765714 %R 10.1177/0146621618765714 %0 Journal Article %J Applied Psychological Measurement %D 2019 %T Application of Dimension Reduction to CAT Item Selection Under the Bifactor Model %A Xiuzhen Mao %A Jiahui Zhang %A Tao Xin %X Multidimensional computerized adaptive testing (MCAT) based on the bifactor model is suitable for tests with multidimensional bifactor measurement structures. Several item selection methods that proved to be more advantageous than the maximum Fisher information method are not practical for bifactor MCAT due to time-consuming computations resulting from high dimensionality. To make them applicable in bifactor MCAT, dimension reduction is applied to four item selection methods, which are the posterior-weighted Fisher D-optimality (PDO) and three non-Fisher information-based methods—posterior expected Kullback–Leibler information (PKL), continuous entropy (CE), and mutual information (MI). They were compared with the Bayesian D-optimality (BDO) method in terms of estimation precision. When both the general and group factors are the measurement objectives, BDO, PDO, CE, and MI perform equally well and better than PKL. When the group factors represent nuisance dimensions, MI and CE perform the best in estimating the general factor, followed by the BDO, PDO, and PKL. How the bifactor pattern and test length affect estimation accuracy was also discussed. %B Applied Psychological Measurement %V 43 %P 419-434 %U https://doi.org/10.1177/0146621618813086 %R 10.1177/0146621618813086 %0 Journal Article %J Journal of Computerized Adaptive Testing %D 2018 %T Adaptive Item Selection Under Matroid Constraints %A Daniel Bengs %A Ulf Brefeld %A Ulf Kröhne %B Journal of Computerized Adaptive Testing %V 6 %P 15-36 %G English %U http://www.iacat.org/jcat/index.php/jcat/article/view/64/32 %N 2 %R 10.7333/1808-0602015 %0 Conference Paper %B IACAT 2017 Conference %D 2017 %T Adapting Linear Models for Optimal Test Design to More Complex Test Specifications %A Maxim Morin %K Complex Test Specifications %K Linear Models %K Optimal Test Design %X

Combinatorial optimization (CO) has proven to be a very helpful approach for addressing test assembly issues and for providing solutions. Furthermore, CO has been applied for several test designs, including: (1) for the development of linear test forms; (2) for computerized adaptive testing and; (3) for multistage testing. In his seminal work, van der Linden (2006) laid out the basis for using linear models for simultaneously assembling exams and item pools in a variety of conditions: (1) for single tests and multiple tests; (2) with item sets, etc. However, for some testing programs, the number and complexity of test specifications can grow rapidly. Consequently, the mathematical representation of the test assembly problem goes beyond most approaches reported either in van der Linden’s book or in the majority of other publications related to test assembly. In this presentation, we extend van der Linden’s framework by including the concept of blocks for test specifications. We modify the usual mathematical notation of a test assembly problem by including this concept and we show how it can be applied to various test designs. Finally, we will demonstrate an implementation of this approach in a stand-alone software, called the ATASolver.

Session Video

%B IACAT 2017 Conference %I Niigata Seiryo University %C Niigata, Japan %8 08/2017 %G eng %0 Conference Paper %B IACAT 2017 Conference %D 2017 %T Adaptive Item and Feedback Selection in Personalized Learning with a Network Approach %A Nikky van Buuren %A Hendrik Straat %A Theo Eggen %A Jean-Paul Fox %K feedback selection %K item selection %K network approach %K personalized learning %X

Personalized learning is a term used to describe educational systems that adapt student-specific curriculum sequencing, pacing, and presentation based on their unique backgrounds, knowledge, preferences, interests, and learning goals. (Chen, 2008; Netcoh, 2016). The technological approach to personalized learning provides data-driven models to incorporate these adaptations automatically. Examples of applications include online learning systems, educational games, and revision-aid systems. In this study we introduce Bayesian networks as a methodology to implement an adaptive framework within a personalized learning environment. Existing ideas from Computerized Adaptive Testing (CAT) with Item Response Theory (IRT), where choices about content provision are based on maximizing information, are related to the goals of personalized learning environments. Personalized learning entails other goals besides efficient ability estimation by maximizing information, such as an adaptive configuration of preferences and feedback to the student. These considerations will be discussed and their application in networks will be illustrated.

Adaptivity in Personalized Learning.In standard CAT’s there is a focus on selecting items that provide maximum information about the ability of an individual at a certain point in time (Van der Linden & Glas, 2000). When learning is the main goal of testing, alternative adaptive item selection methods were explored by Eggen (2012). The adaptive choices made in personalized learning applications require additional adaptivity with respect to the following aspects; the moment of feedback, the kind of feedback, and the possibility for students to actively influence the learning process.

Bayesian Networks and Personalized Learning.Personalized learning aims at constructing a framework to incorporate all the aspects mentioned above. Therefore, the goal of this framework is not only to focus on retrieving ability estimates by choosing items on maximum information, but also to construct a framework that allows for these other factors to play a role. Plajner and Vomlel (2016) have already applied Bayesian Networks to adaptive testing, selecting items with help of entropy reduction. Almond et al. (2015) provide a reference work on Bayesian Networks in Educational Assessment. Both acknowledge the potential of the method in terms of features such as modularity options to build finer-grained models. IRT does not allow to model sub-skills very easily and to gather information on fine-grained level, due to its dependency on the assumption of generally one underlying trait. The local independence assumption in IRT implies being interested in mainly the student’s overall ability on the subject of interest. When the goal is to improve student’s learning, we are not just interested in efficiently coming to their test score on a global subject. One wants a model that is able to map educational problems and talents in detail over the whole educational program, while allowing for dependency between items. The moment in time can influence topics to be better mastered than others, and this is exactly what we can to get out of a model. The possibility to model flexible structures, estimate abilities on a very detailed level for sub-skills and to easily incorporate other variables such as feedback in Bayesian Networks makes it a very promising method for making adaptive choices in personalized learning. It is shown in this research how item and feedback selection can be performed with help of the promising Bayesian Networks. A student involvement possibility is also introduced and evaluated.

References

Almond, R. G., Mislevy, R. J., Steinberg, L. S., Yan, D., & Williamson, D. M. (2015). Bayesian Networks in Educational Assessment. Test. New York: Springer Science+Business Media. http://doi.org/10.1007/978-0-387-98138-3

Eggen, T.J.H.M. (2012) Computerized Adaptive Testing Item Selection in Computerized Adaptive Learning Systems. In: Eggen. TJHM & Veldkamp, BP.. (Eds). Psychometrics in Practice at RCEC. Enschede: RCEC

Netcoh, S. (2016, March). “What Do You Mean by ‘Personalized Learning?’. Croscutting Conversations in Education – Research, Reflections & Practice. Blogpost.

Plajner, M., & Vomlel, J. (2016). Student Skill Models in Adaptive Testing. In Proceedings of the Eighth International Conference on Probabilistic Graphical Models (pp. 403-414).

Van der Linden, W. J., & Glas, C. A. (2000). Computerized adaptive testing: Theory and practice. Dordrecht: Kluwer Academic Publishers.

Session Video

%B IACAT 2017 Conference %I Niigata Seiryo University %C Niigata, Japan %8 08/2017 %G eng %0 Conference Paper %B IACAT 2017 Conference %D 2017 %T Adaptivity in a Diagnostic Educational Test %A Sanneke Schouwstra %K CAT %K Diagnostic tests %K Education %X

During the past five years a diagnostic educational test for three subjects (writing Dutch, writing English and math) has been developed in the Netherlands. The test informs students and their teachers about the students’ strengths and weaknesses in such a manner that the learning process can be adjusted to their personal needs. It is a computer-based assessment for students in five different educational tracks midway secondary education that can yield diagnoses of many sub-skills. One of the main challenges at the outset of the development was to devise a way to deliver many diagnoses within a reasonably testing time. The answer to this challenge was to make the DET adaptive.

In this presentation we will discuss first how the adaptivity is shaped towards the purpose of the Diagnostic Educational Test. The adaptive design, particularly working with item blocks, will be discussed as well as the implemented adaptive rules. We will also show a simulation of different adaptive paths of students and some empirical information on the paths students took through the test

Session Video

%B IACAT 2017 Conference %I Niigata Seiryo University %C Niigata, Japan %8 08/2017 %G eng %0 Conference Paper %B IACAT 2017 Conference %D 2017 %T Analysis of CAT Precision Depending on Parameters of the Item Pool %A Anatoly Maslak %A Stanislav Pozdniakov %K CAT %K Item parameters %K Precision %X

The purpose of this research project is to analyze the measurement precision of a latent variable depending on parameters of the item pool. The influence of the following factors is analyzed:

Factor A – range of variation of items in the pool. This factor varies on three levels with the following ranges in logits: a1 – [-3.0; +3.0], a2 - [-4.0; +4.0], a3 - [-5.0; +5.0].

Factor B – number of items in the pool. The factor varies on six levels with the following number of items for every factor: b1 - 128, b2 - 256, b3 – 512, b4 - 1024, b5 – 2048, b6 – 4096. The items are evenly distributed in each of the variation ranges.

Factor C – examinees’ proficiency varies at 30 levels (c1, c2, …, c30), which are evenly distributed in the range [-3.0; +3.0] logit.

The investigation was based on a simulation experiment within the framework of the theory of latent variables.

Response Y is the precision of measurement of examinees’ proficiency, which is calculated as the difference between the true levels of examinees’ proficiency and estimates obtained by means of adaptive testing. Three factor ANOVA was used for data processing.

The following results were obtained:

1. Factor A is significant. Ceteris paribus, the greater the range of variation of items in the pool, the higher the estimation precision is.

2. Factor B is significant. Ceteris paribus, the greater the number of items in the pool, the higher the estimation precision is.

3. Factor C is statistically insignificant at level α = .05. It means that the precision of estimation of examinees’ proficiency is the same within the range of their variation.

4. The only significant interaction among all interactions is AB. The significance of this interaction is explained by the fact that increasing the number of items in the pool decreases the effect of the range of variation of items in the pool. 

Session Video

%B IACAT 2017 Conference %I Niigata Seiryo University %C Niigata, Japan %8 08/2017 %G eng %U https://drive.google.com/file/d/1Bwe58kOQRgCSbB8x6OdZTDK4OIm3LQI3/view?usp=drive_web %0 Journal Article %J Applied Psychological Measurement %D 2017 %T Application of Binary Searching for Item Exposure Control in Cognitive Diagnostic Computerized Adaptive Testing %A Chanjin Zheng %A Chun Wang %X Cognitive diagnosis has emerged as a new generation of testing theory for educational assessment after the item response theory (IRT). One distinct feature of cognitive diagnostic models (CDMs) is that they assume the latent trait to be discrete instead of continuous as in IRT. From this perspective, cognitive diagnosis bears a close resemblance to searching problems in computer science and, similarly, item selection problem in cognitive diagnostic computerized adaptive testing (CD-CAT) can be considered as a dynamic searching problem. Previously, item selection algorithms in CD-CAT were developed from information indices in information science and attempted to achieve a balance among several objectives by assigning different weights. As a result, they suffered from low efficiency from a tug-of-war competition among multiple goals in item selection and, at the same time, put an undue responsibility of assigning the weights for these goals by trial and error on users. Based on the searching problem perspective on CD-CAT, this article adapts the binary searching algorithm, one of the most well-known searching algorithms in searching problems, to item selection in CD-CAT. The two new methods, the stratified dynamic binary searching (SDBS) algorithm for fixed-length CD-CAT and the dynamic binary searching (DBS) algorithm for variable-length CD-CAT, can achieve multiple goals without any of the aforementioned issues. The simulation studies indicate their performances are comparable or superior to the previous methods. %B Applied Psychological Measurement %V 41 %P 561-576 %U https://doi.org/10.1177/0146621617707509 %R 10.1177/0146621617707509 %0 Journal Article %J Educational and Psychological Measurement %D 2017 %T ATS-PD: An Adaptive Testing System for Psychological Disorders %A Ivan Donadello %A Andrea Spoto %A Francesco Sambo %A Silvana Badaloni %A Umberto Granziol %A Giulio Vidotto %X The clinical assessment of mental disorders can be a time-consuming and error-prone procedure, consisting of a sequence of diagnostic hypothesis formulation and testing aimed at restricting the set of plausible diagnoses for the patient. In this article, we propose a novel computerized system for the adaptive testing of psychological disorders. The proposed system combines a mathematical representation of psychological disorders, known as the “formal psychological assessment,” with an algorithm designed for the adaptive assessment of an individual’s knowledge. The assessment algorithm is extended and adapted to the new application domain. Testing the system on a real sample of 4,324 healthy individuals, screened for obsessive-compulsive disorder, we demonstrate the system’s ability to support clinical testing, both by identifying the correct critical areas for each individual and by reducing the number of posed questions with respect to a standard written questionnaire. %B Educational and Psychological Measurement %V 77 %P 792-815 %U https://doi.org/10.1177/0013164416652188 %R 10.1177/0013164416652188 %0 Journal Article %J Journal of Educational Measurement %D 2015 %T Assessing Individual-Level Impact of Interruptions During Online Testing %A Sinharay, Sandip %A Wan, Ping %A Choi, Seung W. %A Kim, Dong-In %X With an increase in the number of online tests, the number of interruptions during testing due to unexpected technical issues seems to be on the rise. For example, interruptions occurred during several recent state tests. When interruptions occur, it is important to determine the extent of their impact on the examinees' scores. Researchers such as Hill and Sinharay et al. examined the impact of interruptions at an aggregate level. However, there is a lack of research on the assessment of impact of interruptions at an individual level. We attempt to fill that void. We suggest four methodological approaches, primarily based on statistical hypothesis testing, linear regression, and item response theory, which can provide evidence on the individual-level impact of interruptions. We perform a realistic simulation study to compare the Type I error rate and power of the suggested approaches. We then apply the approaches to data from the 2013 Indiana Statewide Testing for Educational Progress-Plus (ISTEP+) test that experienced interruptions. %B Journal of Educational Measurement %V 52 %P 80–105 %U http://dx.doi.org/10.1111/jedm.12064 %R 10.1111/jedm.12064 %0 Journal Article %J Educational and Psychological Measurement %D 2015 %T a-Stratified Computerized Adaptive Testing in the Presence of Calibration Error %A Cheng, Ying %A Patton, Jeffrey M. %A Shao, Can %X a-Stratified computerized adaptive testing with b-blocking (AST), as an alternative to the widely used maximum Fisher information (MFI) item selection method, can effectively balance item pool usage while providing accurate latent trait estimates in computerized adaptive testing (CAT). However, previous comparisons of these methods have treated item parameter estimates as if they are the true population parameter values. Consequently, capitalization on chance may occur. In this article, we examined the performance of the AST method under more realistic conditions where item parameter estimates instead of true parameter values are used in the CAT. Its performance was compared against that of the MFI method when the latter is used in conjunction with Sympson–Hetter or randomesque exposure control. Results indicate that the MFI method, even when combined with exposure control, is susceptible to capitalization on chance. This is particularly true when the calibration sample size is small. On the other hand, AST is more robust to capitalization on chance. Consistent with previous investigations using true item parameter values, AST yields much more balanced item pool usage, with a small loss in the precision of latent trait estimates. The loss is negligible when the test is as long as 40 items. %B Educational and Psychological Measurement %V 75 %P 260-283 %U http://epm.sagepub.com/content/75/2/260.abstract %R 10.1177/0013164414530719 %0 Journal Article %J International Journal of Testing %D 2013 %T The Applicability of Multidimensional Computerized Adaptive Testing for Cognitive Ability Measurement in Organizational Assessment %A Makransky, Guido %A Glas, Cees A. W. %B International Journal of Testing %V 13 %P 123-139 %U http://www.tandfonline.com/doi/abs/10.1080/15305058.2012.672352 %R 10.1080/15305058.2012.672352 %0 Journal Article %J International Journal of Testing %D 2013 %T The applicability of multidimensional computerized adaptive testing to cognitive ability measurement in organizational assessment %A Makransky, G. %A Glas, C. A. W. %B International Journal of Testing %V 13 %P 123-139 %G English %N 2 %R 10.1080/15305058.2012.672352 %0 Journal Article %J Applied Psychological Measurement %D 2013 %T The Application of the Monte Carlo Approach to Cognitive Diagnostic Computerized Adaptive Testing With Content Constraints %A Mao, Xiuzhen %A Xin, Tao %X

The Monte Carlo approach which has previously been implemented in traditional computerized adaptive testing (CAT) is applied here to cognitive diagnostic CAT to test the ability of this approach to address multiple content constraints. The performance of the Monte Carlo approach is compared with the performance of the modified maximum global discrimination index (MMGDI) method on simulations in which the only content constraint is on the number of items that measure each attribute. The results of the two simulation experiments show that (a) the Monte Carlo method fulfills all the test requirements and produces satisfactory measurement precision and item exposure results and (b) the Monte Carlo method outperforms the MMGDI method when the Monte Carlo method applies either the posterior-weighted Kullback–Leibler algorithm or the hybrid Kullback–Leibler information as the item selection index. Overall, the recovery rate of the knowledge states, the distribution of the item exposure, and the utilization rate of the item bank are improved when the Monte Carlo method is used.

%B Applied Psychological Measurement %V 37 %P 482-496 %U http://apm.sagepub.com/content/37/6/482.abstract %R 10.1177/0146621613486015 %0 Book Section %B Testtheorie und Fragebogenkonstruktion %D 2012 %T Adaptives Testen [Adaptive testing]. %A Frey, A. %B Testtheorie und Fragebogenkonstruktion %I Heidelberg: Springer %C Berlin %G eng %0 Conference Paper %B Annual Conference of the International Association for Computerized Adaptive Testing %D 2011 %T Adaptive Item Calibration and Norming: Unique Considerations of a Global Deployment %A Alexander Schwall %A Evan Sinar %K CAT %K common item equating %K Figural Reasoning Test %K item calibration %K norming %B Annual Conference of the International Association for Computerized Adaptive Testing %8 10/2011 %G eng %0 Journal Article %J Psychiatry Research %D 2011 %T Applying computerized adaptive testing to the CES-D scale: A simulation study %A Smits, N. %A Cuijpers, P. %A van Straten, A. %X In this paper we studied the appropriateness of developing an adaptive version of the Center of Epidemiological Studies-Depression (CES-D, Radloff, 1977) scale. Computerized Adaptive Testing (CAT) involves the computerized administration of a test in which each item is dynamically selected from a pool of items until a pre-specified measurement precision is reached. Two types of analyses were performed using the CES-D responses of a large sample of adolescents (N=1392). First, it was shown that the items met the psychometric requirements needed for CAT. Second, CATs were simulated by using the existing item responses as if they had been collected adaptively. CATs selecting only a small number of items gave results which, in terms of depression measurement and criterion validity, were only marginally different from the results of full CES-D assessment. It was concluded that CAT is a very fruitful way of improving the efficiency of the CES-D questionnaire. The discussion addresses the strengths and limitations of the application of CAT in mental health research. %B Psychiatry Research %7 2011/01/07 %8 Jan 3 %@ 0165-1781 (Print)0165-1781 (Linking) %G Eng %0 Journal Article %J Psychiatry Research %D 2011 %T Applying computerized adaptive testing to the CES-D scale: A simulation study %A Smits, N. %A Cuijpers, P. %A van Straten, A. %B Psychiatry Research %V 188 %P 147–155 %N 1 %0 Book Section %B Elements of Adaptive Testing %D 2010 %T Adaptive Mastery Testing Using a Multidimensional IRT Model %A Glas, C. A. W. %A Vos, H. J. %B Elements of Adaptive Testing %P 409-431 %G eng %& 21 %R 10.1007/978-0-387-85461-8 %0 Book Section %B Elements of Adaptive Testing %D 2010 %T Adaptive Tests for Measuring Anxiety and Depression %A Walter, O. B. %B Elements of Adaptive Testing %P 123-136 %G eng %& 6 %R 10.1007/978-0-387-85461-8 %0 Book Section %B Elements of Adaptive Testing %D 2010 %T Assembling an Inventory of Multistage Adaptive Testing Systems %A Breithaupt, K %A Ariel, A. %A Hare, D. R. %B Elements of Adaptive Testing %P 247-266 %G eng %& 13 %R 10.1007/978-0-387-85461-8 %0 Journal Article %J Journal of Applied Testing Technology %D 2010 %T An automatic online calibration design in adaptive testing %A Makransky, G. %A Glas, C. A. W. %B Journal of Applied Testing Technology %V 11 %G English %0 Book Section %D 2009 %T Adaptive computer-based tasks under an assessment engineering paradigm %A Luecht, RM %C D. J. Weiss (Ed.), Proceedings of the 2009 GMAC Conference on Computerized Adaptive Testing. %G eng %0 Book Section %D 2009 %T Adaptive item calibration: A process for estimating item parameters within a computerized adaptive test %A Kingsbury, G. G. %X The characteristics of an adaptive test change the characteristics of the field testing that is necessary to add items to an existing measurement scale. The process used to add field-test items to the adaptive test might lead to scale drift or disrupt the test by administering items of inappropriate difficulty. The current study makes use of the transitivity of examinee and item in item response theory to describe a process for adaptive item calibration. In this process an item is successively administered to examinees whose ability levels match the performance of a given field-test item. By treating the item as if it were taking an adaptive test, examinees can be selected who provide the most information about the item at its momentary difficulty level. This should provide a more efficient procedure for estimating item parameters. The process is described within the context of the one-parameter logistic IRT model. The process is then simulated to identify whether it can be more accurate and efficient than random presentation of field-test items to examinees. Results indicated that adaptive item calibration might provide a viable approach to item calibration within the context of an adaptive test. It might be most useful for expanding item pools in settings with small sample sizes or needs for large numbers of items. %C D. J. Weiss (Ed.), Proceedings of the 2009 GMAC Conference on Computerized Adaptive Testing. %G eng %0 Journal Article %J Computers and Education %D 2009 %T An adaptive testing system for supporting versatile educational assessment %A Huang, Y-M. %A Lin, Y-T. %A Cheng, S-C. %K Architectures for educational technology system %K Distance education and telelearning %X With the rapid growth of computer and mobile technology, it is a challenge to integrate computer based test (CBT) with mobile learning (m-learning) especially for formative assessment and self-assessment. In terms of self-assessment, computer adaptive test (CAT) is a proper way to enable students to evaluate themselves. In CAT, students are assessed through a process that uses item response theory (IRT), a well-founded psychometric theory. Furthermore, a large item bank is indispensable to a test, but when a CAT system has a large item bank, the test item selection of IRT becomes more tedious. Besides the large item bank, item exposure mechanism is also essential to a testing system. However, IRT all lack the above-mentioned points. These reasons have motivated the authors to carry out this study. This paper describes a design issue aimed at the development and implementation of an adaptive testing system. The system can support several assessment functions and different devices. Moreover, the researchers apply a novel approach, particle swarm optimization (PSO) to alleviate the computational complexity and resolve the problem of item exposure. Throughout the development of the system, a formative evaluation was embedded into an integral part of the design methodology that was used for improving the system. After the system was formally released onto the web, some questionnaires and experiments were conducted to evaluate the usability, precision, and efficiency of the system. The results of these evaluations indicated that the system provides an adaptive testing for different devices and supports versatile assessment functions. Moreover, the system can estimate students' ability reliably and validly and conduct an adaptive test efficiently. Furthermore, the computational complexity of the system was alleviated by the PSO approach. By the approach, the test item selection procedure becomes efficient and the average best fitness values are very close to the optimal solutions. %B Computers and Education %V 52 %P 53-67 %@ 0360-1315 %G eng %0 Book Section %D 2009 %T Adequacy of an item pool measuring proficiency in English language to implement a CAT procedure %A Karino, C. A. %A Costa, D. R. %A Laros, J. A. %C D. J. Weiss (Ed.), Proceedings of the 2009 GMAC Conference on Computerized Adaptive Testing. %G eng %0 Book Section %D 2009 %T Applications of CAT in admissions to higher education in Israel: Twenty-two years of experience %A Gafni, N. %A Cohen, Y. %A Roded, K %A Baumer, M %A Moshinsky, A. %C D. J. Weiss (Ed.), Proceedings of the 2009 GMAC Conference on Computerized Adaptive Testing. %G eng %0 Book Section %D 2009 %T An approach to implementing adaptive testing using item response theory both offline and online %A Padaki, M. %A Natarajan, V. %C D. J. Weiss (Ed.), Proceedings of the 2009 GMAC Conference on Computerized Adaptive Testing. %G eng %0 Book Section %D 2009 %T Assessing the equivalence of Internet-based vs. paper-and-pencil psychometric tests. %A Baumer, M %A Roded, K %A Gafni, N. %C D. J. Weiss (Ed.), Proceedings of the 2009 GMAC Conference on Computerized Adaptive Testing. %G eng %0 Book Section %D 2009 %T An automatic online calibration design in adaptive testing %A Makransky, G. %C D. J. Weiss (Ed.), Proceedings of the 2009 GMAC Conference on Computerized Adaptive Testing. %G eng %0 Journal Article %J Zeitschrift für Psychologie / Journal of Psychology %D 2008 %T Adaptive measurement of individual change %A Kim-Kang, G. %A Weiss, D. J. %B Zeitschrift für Psychologie / Journal of Psychology %V 216 %P 49-58 %G eng %R 10.1027/0044-3409.216.1.49 %0 Journal Article %J Zeitschrift für Psychologie / Journal of Psychology %D 2008 %T Adaptive models of psychological testing %A van der Linden, W. J. %B Zeitschrift für Psychologie / Journal of Psychology %V 216(1) %P 3–11 %G eng %0 Journal Article %J Zeitschrift für Psychologie / Journal of Psychology %D 2008 %T Adaptive Models of Psychological Testing %A van der Linden, W. J. %B Zeitschrift für Psychologie / Journal of Psychology %V 216 %P 1-2 %N 1 %R 10.1027/0044-3409.216.1.49 %0 Journal Article %J American Journal of Physical Medicine and Rehabilitation %D 2008 %T Adaptive short forms for outpatient rehabilitation outcome assessment %A Jette, A. M. %A Haley, S. M. %A Ni, P. %A Moed, R. %K *Activities of Daily Living %K *Ambulatory Care Facilities %K *Mobility Limitation %K *Treatment Outcome %K Disabled Persons/psychology/*rehabilitation %K Female %K Humans %K Male %K Middle Aged %K Questionnaires %K Rehabilitation Centers %X OBJECTIVE: To develop outpatient Adaptive Short Forms for the Activity Measure for Post-Acute Care item bank for use in outpatient therapy settings. DESIGN: A convenience sample of 11,809 adults with spine, lower limb, upper limb, and miscellaneous orthopedic impairments who received outpatient rehabilitation in 1 of 127 outpatient rehabilitation clinics in the United States. We identified optimal items for use in developing outpatient Adaptive Short Forms based on the Basic Mobility and Daily Activities domains of the Activity Measure for Post-Acute Care item bank. Patient scores were derived from the Activity Measure for Post-Acute Care computerized adaptive testing program. Items were selected for inclusion on the Adaptive Short Forms based on functional content, range of item coverage, measurement precision, item exposure rate, and data collection burden. RESULTS: Two outpatient Adaptive Short Forms were developed: (1) an 18-item Basic Mobility Adaptive Short Form and (2) a 15-item Daily Activities Adaptive Short Form, derived from the same item bank used to develop the Activity Measure for Post-Acute Care computerized adaptive testing program. Both Adaptive Short Forms achieved acceptable psychometric properties. CONCLUSIONS: In outpatient postacute care settings where computerized adaptive testing outcome applications are currently not feasible, item response theory-derived Adaptive Short Forms provide the efficient capability to monitor patients' functional outcomes. The development of Adaptive Short Form functional outcome instruments linked by a common, calibrated item bank has the potential to create a bridge to outcome monitoring across postacute care settings and can facilitate the eventual transformation from Adaptive Short Forms to computerized adaptive testing applications easier and more acceptable to the rehabilitation community. %B American Journal of Physical Medicine and Rehabilitation %7 2008/09/23 %V 87 %P 842-52 %8 Oct %@ 1537-7385 (Electronic) %G eng %M 18806511 %0 Journal Article %J Psychiatric Services %D 2008 %T Are we ready for computerized adaptive testing? %A Unick, G. J. %A Shumway, M. %A Hargreaves, W. %K *Attitude of Health Personnel %K *Diagnosis, Computer-Assisted/instrumentation %K Humans %K Mental Disorders/*diagnosis %K Software %B Psychiatric Services %7 2008/04/02 %V 59 %P 369 %8 Apr %@ 1075-2730 (Print)1075-2730 (Linking) %G eng %M 18378833 %0 Journal Article %J Archives of Physical Medicine and Rehabilitation %D 2008 %T Assessing self-care and social function using a computer adaptive testing version of the pediatric evaluation of disability inventory %A Coster, W. J. %A Haley, S. M. %A Ni, P. %A Dumas, H. M. %A Fragala-Pinkham, M. A. %K *Disability Evaluation %K *Social Adjustment %K Activities of Daily Living %K Adolescent %K Age Factors %K Child %K Child, Preschool %K Computer Simulation %K Cross-Over Studies %K Disabled Children/*rehabilitation %K Female %K Follow-Up Studies %K Humans %K Infant %K Male %K Outcome Assessment (Health Care) %K Reference Values %K Reproducibility of Results %K Retrospective Studies %K Risk Factors %K Self Care/*standards/trends %K Sex Factors %K Sickness Impact Profile %X OBJECTIVE: To examine score agreement, validity, precision, and response burden of a prototype computer adaptive testing (CAT) version of the self-care and social function scales of the Pediatric Evaluation of Disability Inventory compared with the full-length version of these scales. DESIGN: Computer simulation analysis of cross-sectional and longitudinal retrospective data; cross-sectional prospective study. SETTING: Pediatric rehabilitation hospital, including inpatient acute rehabilitation, day school program, outpatient clinics; community-based day care, preschool, and children's homes. PARTICIPANTS: Children with disabilities (n=469) and 412 children with no disabilities (analytic sample); 38 children with disabilities and 35 children without disabilities (cross-validation sample). INTERVENTIONS: Not applicable. MAIN OUTCOME MEASURES: Summary scores from prototype CAT applications of each scale using 15-, 10-, and 5-item stopping rules; scores from the full-length self-care and social function scales; time (in seconds) to complete assessments and respondent ratings of burden. RESULTS: Scores from both computer simulations and field administration of the prototype CATs were highly consistent with scores from full-length administration (r range, .94-.99). Using computer simulation of retrospective data, discriminant validity, and sensitivity to change of the CATs closely approximated that of the full-length scales, especially when the 15- and 10-item stopping rules were applied. In the cross-validation study the time to administer both CATs was 4 minutes, compared with over 16 minutes to complete the full-length scales. CONCLUSIONS: Self-care and social function score estimates from CAT administration are highly comparable with those obtained from full-length scale administration, with small losses in validity and precision and substantial decreases in administration time. %B Archives of Physical Medicine and Rehabilitation %7 2008/04/01 %V 89 %P 622-629 %8 Apr %@ 1532-821X (Electronic)0003-9993 (Linking) %G eng %M 18373991 %2 2666276 %0 Conference Paper %B Software Engineering, Artificial Intelligence, Networking, and Parallel/Distributed Computing, 2008. SNPD'08. Ninth ACIS International Conference on %D 2008 %T An Automated Decision System for Computer Adaptive Testing Using Genetic Algorithms %A Phankokkruad, M. %A Woraratpanya, K. %X This paper proposes an approach to solve the triangle decision tree problem for computer adaptive testing (CAT) using genetic algorithms (GAs). In this approach, item response theory (IRT) parameters composed of discrimination, difficulty, and guess are firstly obtained and stored in an item bank. Then a fitness function, which is based on IRT parameters, of GAs for obtaining an optimal solution is set up. Finally, the GAs is applied to the parameters of the item bank so that an optimal decision tree is generated. Based on a six-level triangle-decision tree for examination items, the experimental results show that the optimal decision tree can be generated correctly when compared with the standard patterns. %B Software Engineering, Artificial Intelligence, Networking, and Parallel/Distributed Computing, 2008. SNPD'08. Ninth ACIS International Conference on %I IEEE %P 655–660 %0 Book Section %D 2007 %T Adaptive estimators of trait level in adaptive testing: Some proposals %A Raîche, G. %A Blais, J. G. %A Magis, D. %C D. J. Weiss (Ed.), Proceedings of the 2007 GMAC Conference on Computerized Adaptive Testing. %G eng %0 Book Section %D 2007 %T Adaptive testing with the multi-unidimensional pairwise preference model %A Stark, S. %A Chernyshenko, O. S. %C D. J. Weiss (Ed.), Proceedings of the 2007 GMAC Conference on Computerized Adaptive Testing. %G eng %0 Journal Article %J Quality of Life Research %D 2007 %T Applying item response theory and computer adaptive testing: The challenges for health outcomes assessment %A Fayers, P. M. %X OBJECTIVES: We review the papers presented at the NCI/DIA conference, to identify areas of controversy and uncertainty, and to highlight those aspects of item response theory (IRT) and computer adaptive testing (CAT) that require theoretical or empirical research in order to justify their application to patient reported outcomes (PROs). BACKGROUND: IRT and CAT offer exciting potential for the development of a new generation of PRO instruments. However, most of the research into these techniques has been in non-healthcare settings, notably in education. Educational tests are very different from PRO instruments, and consequently problematic issues arise when adapting IRT and CAT to healthcare research. RESULTS: Clinical scales differ appreciably from educational tests, and symptoms have characteristics distinctly different from examination questions. This affects the transferring of IRT technology. Particular areas of concern when applying IRT to PROs include inadequate software, difficulties in selecting models and communicating results, insufficient testing of local independence and other assumptions, and a need of guidelines for estimating sample size requirements. Similar concerns apply to differential item functioning (DIF), which is an important application of IRT. Multidimensional IRT is likely to be advantageous only for closely related PRO dimensions. CONCLUSIONS: Although IRT and CAT provide appreciable potential benefits, there is a need for circumspection. Not all PRO scales are necessarily appropriate targets for this methodology. Traditional psychometric methods, and especially qualitative methods, continue to have an important role alongside IRT. Research should be funded to address the specific concerns that have been identified. %B Quality of Life Research %7 2007/04/10 %V 16 %P 187-194 %@ 0962-9343 (Print) %G eng %M 17417722 %0 Journal Article %J Educational and Psychological Measurement %D 2007 %T Automated Simultaneous Assembly of Multistage Testlets for a High-Stakes Licensing Examination %A Breithaupt, Krista %A Hare, Donovan R. %X

Many challenges exist for high-stakes testing programs offering continuous computerized administration. The automated assembly of test questions to exactly meet content and other requirements, provide uniformity, and control item exposure can be modeled and solved by mixed-integer programming (MIP) methods. A case study of the computerized licensing examination of the American Institute of Certified Public Accountants is offered as one application of MIP techniques for test assembly. The solution illustrates assembly for a computer-adaptive multistage testing design. However, the general form of the constraint-based solution can be modified to generate optimal test designs for paper-based or computerized administrations, regardless of the specific psychometric model. An extension of this methodology allows for long-term planning for the production and use of test content on the basis of an exact psychometric test designs and administration schedules.

%B Educational and Psychological Measurement %V 67 %P 5-20 %U http://epm.sagepub.com/content/67/1/5.abstract %R 10.1177/0013164406288162 %0 Journal Article %J International Journal of Testing %D 2007 %T A “Rearrangement Procedure” For Scoring Adaptive Tests with Review Options %A Papanastasiou, Elena C. %A Reckase, Mark D. %B International Journal of Testing %V 7 %P 387-407 %U http://www.tandfonline.com/doi/abs/10.1080/15305050701632262 %R 10.1080/15305050701632262 %0 Journal Article %J Psychology Science %D 2006 %T Adaptive success control in computerized adaptive testing %A Häusler, Joachim %K adaptive success control %K computerized adaptive testing %K Psychometrics %X In computerized adaptive testing (CAT) procedures within the framework of probabilistic test theory the difficulty of an item is adjusted to the ability of the respondent, with the aim of maximizing the amount of information generated per item, thereby also increasing test economy and test reasonableness. However, earlier research indicates that respondents might feel over-challenged by a constant success probability of p = 0.5 and therefore cannot come to a sufficiently high answer certainty within a reasonable timeframe. Consequently response time per item increases, which -- depending on the test material -- can outweigh the benefit of administering optimally informative items. Instead of a benefit, the result of using CAT procedures could be a loss of test economy. Based on this problem, an adaptive success control algorithm was designed and tested, adapting the success probability to the working style of the respondent. Persons who need higher answer certainty in order to come to a decision are detected and receive a higher success probability, in order to minimize the test duration (not the number of items as in classical CAT). The method is validated on the re-analysis of data from the Adaptive Matrices Test (AMT, Hornke, Etzel & Rettig, 1999) and by the comparison between an AMT version using classical CAT and an experimental version using Adaptive Success Control. The results are discussed in the light of psychometric and psychological aspects of test quality. (PsycINFO Database Record (c) 2007 APA, all rights reserved) %B Psychology Science %I Pabst Science Publishers: Germany %V 48 %P 436-450 %@ 0033-3018 (Print) %G eng %M 2007-03313-004 %0 Journal Article %J Applied Measurement in Education %D 2006 %T Applying Bayesian item selection approaches to adaptive tests using polytomous items %A Penfield, R. D. %K adaptive tests %K Bayesian item selection %K computer adaptive testing %K maximum expected information %K polytomous items %K posterior weighted information %X This study applied the maximum expected information (MEI) and the maximum posterior- weighted information (MPI) approaches of computer adaptive testing item selection to the case of a test using polytomous items following the partial credit model. The MEI and MPI approaches are described. A simulation study compared the efficiency of ability estimation using the MEI and MPI approaches to the traditional maximal item information (MII) approach. The results of the simulation study indicated that the MEI and MPI approaches led to a superior efficiency of ability estimation compared with the MII approach. The superiority of the MEI and MPI approaches over the MII approach was greatest when the bank contained items having a relatively peaked information function. (PsycINFO Database Record (c) 2007 APA, all rights reserved) %B Applied Measurement in Education %I Lawrence Erlbaum: US %V 19 %P 1-20 %@ 0895-7347 (Print); 1532-4818 (Electronic) %G eng %M 2006-00588-001 %0 Journal Article %J Journal of Educational and Behavioral Statistics %D 2006 %T Assembling a computerized adaptive testing item pool as a set of linear tests %A van der Linden, W. J. %A Ariel, A. %A Veldkamp, B. P. %K Algorithms %K computerized adaptive testing %K item pool %K linear tests %K mathematical models %K statistics %K Test Construction %K Test Items %X Test-item writing efforts typically results in item pools with an undesirable correlational structure between the content attributes of the items and their statistical information. If such pools are used in computerized adaptive testing (CAT), the algorithm may be forced to select items with less than optimal information, that violate the content constraints, and/or have unfavorable exposure rates. Although at first sight somewhat counterintuitive, it is shown that if the CAT pool is assembled as a set of linear test forms, undesirable correlations can be broken down effectively. It is proposed to assemble such pools using a mixed integer programming model with constraints that guarantee that each test meets all content specifications and an objective function that requires them to have maximal information at a well-chosen set of ability values. An empirical example with a previous master pool from the Law School Admission Test (LSAT) yielded a CAT with nearly uniform bias and mean-squared error functions for the ability estimator and item-exposure rates that satisfied the target for all items in the pool. %B Journal of Educational and Behavioral Statistics %I Sage Publications: US %V 31 %P 81-99 %@ 1076-9986 (Print) %G eng %M 2007-08137-004 %0 Journal Article %J Applied Psychological Measurement %D 2006 %T Assessing CAT Test Security Severity %A Yi, Q., Zhang, J. %A Chang, Hua-Hua %B Applied Psychological Measurement %V 30(1) %P 62–63 %G eng %0 Book Section %D 2005 %T Adaptive orientation methods in computer adaptive testing %A Economides, A. A. %C Proceedings E-Learn 2005 World Conference on E-Learning in Corporate, Government, Healthcare, and Higher Education, pp. 1290-1295, Vancouver, Canada, AACE, October 2005. %G eng %0 Book %D 2005 %T Adaptive selection of personality items to inform a neural network predicting job performance %A Thissen-Roe, A. %C Unpublished doctoral dissertation, University of Washington %G eng %0 Book Section %B Outcomes assessment in cancer %D 2005 %T Applications of item response theory to improve health outcomes assessment: Developing item banks, linking instruments, and computer-adaptive testing %A Hambleton, R. K. %E C. C. Gotay %E C. Snyder %K Computer Assisted Testing %K Health %K Item Response Theory %K Measurement %K Test Construction %K Treatment Outcomes %X (From the chapter) The current chapter builds on Reise's introduction to the basic concepts, assumptions, popular models, and important features of IRT and discusses the applications of item response theory (IRT) modeling to health outcomes assessment. In particular, we highlight the critical role of IRT modeling in: developing an instrument to match a study's population; linking two or more instruments measuring similar constructs on a common metric; and creating item banks that provide the foundation for tailored short-form instruments or for computerized adaptive assessments. (PsycINFO Database Record (c) 2005 APA ) %B Outcomes assessment in cancer %I Cambridge University Press %C Cambridge, UK %P 445-464 %G eng %0 Journal Article %J Archives of Physical Medicine and Rehabilitation %D 2005 %T Assessing mobility in children using a computer adaptive testing version of the pediatric evaluation of disability inventory %A Haley, S. M. %A Raczek, A. E. %A Coster, W. J. %A Dumas, H. M. %A Fragala-Pinkham, M. A. %K *Computer Simulation %K *Disability Evaluation %K Adolescent %K Child %K Child, Preschool %K Cross-Sectional Studies %K Disabled Children/*rehabilitation %K Female %K Humans %K Infant %K Male %K Outcome Assessment (Health Care)/*methods %K Rehabilitation Centers %K Rehabilitation/*standards %K Sensitivity and Specificity %X OBJECTIVE: To assess score agreement, validity, precision, and response burden of a prototype computerized adaptive testing (CAT) version of the Mobility Functional Skills Scale (Mob-CAT) of the Pediatric Evaluation of Disability Inventory (PEDI) as compared with the full 59-item version (Mob-59). DESIGN: Computer simulation analysis of cross-sectional and longitudinal retrospective data; and cross-sectional prospective study. SETTING: Pediatric rehabilitation hospital, including inpatient acute rehabilitation, day school program, outpatient clinics, community-based day care, preschool, and children's homes. PARTICIPANTS: Four hundred sixty-nine children with disabilities and 412 children with no disabilities (analytic sample); 41 children without disabilities and 39 with disabilities (cross-validation sample). INTERVENTIONS: Not applicable. MAIN OUTCOME MEASURES: Summary scores from a prototype Mob-CAT application and versions using 15-, 10-, and 5-item stopping rules; scores from the Mob-59; and number of items and time (in seconds) to administer assessments. RESULTS: Mob-CAT scores from both computer simulations (intraclass correlation coefficient [ICC] range, .94-.99) and field administrations (ICC=.98) were in high agreement with scores from the Mob-59. Using computer simulations of retrospective data, discriminant validity, and sensitivity to change of the Mob-CAT closely approximated that of the Mob-59, especially when using the 15- and 10-item stopping rule versions of the Mob-CAT. The Mob-CAT used no more than 15% of the items for any single administration, and required 20% of the time needed to administer the Mob-59. CONCLUSIONS: Comparable score estimates for the PEDI mobility scale can be obtained from CAT administrations, with losses in validity and precision for shorter forms, but with a considerable reduction in administration time. %B Archives of Physical Medicine and Rehabilitation %7 2005/05/17 %V 86 %P 932-9 %8 May %@ 0003-9993 (Print) %G eng %M 15895339 %0 Journal Article %J Archives of Physical Medicine and Rehabilitation %D 2005 %T Assessing Mobility in Children Using a Computer Adaptive Testing Version of the Pediatric Evaluation of Disability Inventory %A Haley, S. %A Raczek, A. %A Coster, W. %A Dumas, H. %A Fragalapinkham, M. %B Archives of Physical Medicine and Rehabilitation %V 86 %P 932-939 %@ 00039993 %G eng %0 Journal Article %J Educational Technology & Society %D 2005 %T An Authoring Environment for Adaptive Testing %A Guzmán, E %A Conejo, R %A García-Hervás, E %K Adaptability %K Adaptive Testing %K Authoring environment %K Item Response Theory %X

SIETTE is a web-based adaptive testing system. It implements Computerized Adaptive Tests. These tests are tailor-made, theory-based tests, where questions shown to students, finalization of the test, and student knowledge estimation is accomplished adaptively. To construct these tests, SIETTE has an authoring environment comprising a suite of tools that helps teachers create questions and tests properly, and analyze students’ performance after taking a test. In this paper, we present this authoring environment in the
framework of adaptive testing. As will be shown, this set of visual tools, that contain some adaptable eatures, can be useful for teachers lacking skills in this kind of testing. Additionally, other systems that implement adaptive testing will be studied. 

%B Educational Technology & Society %V 8 %P 66-76 %G eng %N 3 %0 Journal Article %J International Journal of Testing %D 2005 %T Automated Simultaneous Assembly for Multistage Testing %A Breithaupt, Krista %A Ariel, Adelaide %A Veldkamp, Bernard P. %B International Journal of Testing %V 5 %P 319-330 %U http://www.tandfonline.com/doi/abs/10.1207/s15327574ijt0503_8 %R 10.1207/s15327574ijt0503_8 %0 Book Section %D 2004 %T The ABCs of Computerized Adaptive Testing %A Gershon, R. C. %C T. M. Wood and W. Zhi (Eds.), Measurement issues and practice in physical activity. Champaign, IL: Human kinetics. %G eng %0 Conference Paper %B Paper presented at the annual meeting of the National Council on Measurement in Education %D 2004 %T Achieving accuracy of retest calibration for a national CAT placement examination with a restricted test length %A Wang, X. B. %A Wiley, A. %B Paper presented at the annual meeting of the National Council on Measurement in Education %C San Diego CA %G eng %0 Journal Article %J Medical Care %D 2004 %T Activity outcome measurement for postacute care %A Haley, S. M. %A Coster, W. J. %A Andres, P. L. %A Ludlow, L. H. %A Ni, P. %A Bond, T. L. %A Sinclair, S. J. %A Jette, A. M. %K *Self Efficacy %K *Sickness Impact Profile %K Activities of Daily Living/*classification/psychology %K Adult %K Aftercare/*standards/statistics & numerical data %K Aged %K Boston %K Cognition/physiology %K Disability Evaluation %K Factor Analysis, Statistical %K Female %K Human %K Male %K Middle Aged %K Movement/physiology %K Outcome Assessment (Health Care)/*methods/statistics & numerical data %K Psychometrics %K Questionnaires/standards %K Rehabilitation/*standards/statistics & numerical data %K Reproducibility of Results %K Sensitivity and Specificity %K Support, U.S. Gov't, Non-P.H.S. %K Support, U.S. Gov't, P.H.S. %X BACKGROUND: Efforts to evaluate the effectiveness of a broad range of postacute care services have been hindered by the lack of conceptually sound and comprehensive measures of outcomes. It is critical to determine a common underlying structure before employing current methods of item equating across outcome instruments for future item banking and computer-adaptive testing applications. OBJECTIVE: To investigate the factor structure, reliability, and scale properties of items underlying the Activity domains of the International Classification of Functioning, Disability and Health (ICF) for use in postacute care outcome measurement. METHODS: We developed a 41-item Activity Measure for Postacute Care (AM-PAC) that assessed an individual's execution of discrete daily tasks in his or her own environment across major content domains as defined by the ICF. We evaluated the reliability and discriminant validity of the prototype AM-PAC in 477 individuals in active rehabilitation programs across 4 rehabilitation settings using factor analyses, tests of item scaling, internal consistency reliability analyses, Rasch item response theory modeling, residual component analysis, and modified parallel analysis. RESULTS: Results from an initial exploratory factor analysis produced 3 distinct, interpretable factors that accounted for 72% of the variance: Applied Cognition (44%), Personal Care & Instrumental Activities (19%), and Physical & Movement Activities (9%); these 3 activity factors were verified by a confirmatory factor analysis. Scaling assumptions were met for each factor in the total sample and across diagnostic groups. Internal consistency reliability was high for the total sample (Cronbach alpha = 0.92 to 0.94), and for specific diagnostic groups (Cronbach alpha = 0.90 to 0.95). Rasch scaling, residual factor, differential item functioning, and modified parallel analyses supported the unidimensionality and goodness of fit of each unique activity domain. CONCLUSIONS: This 3-factor model of the AM-PAC can form the conceptual basis for common-item equating and computer-adaptive applications, leading to a comprehensive system of outcome instruments for postacute care settings. %B Medical Care %V 42 %P I49-161 %G eng %M 14707755 %0 Book Section %B Evidence-based educational methods %D 2004 %T Adaptive computerized educational systems: A case study %A Ray, R. D. %E R. W. Malott %K Artificial %K Computer Assisted Instruction %K Computer Software %K Higher Education %K Individualized %K Instruction %K Intelligence %K Internet %K Undergraduate Education %X (Created by APA) Adaptive instruction describes adjustments typical of one-on-one tutoring as discussed in the college tutorial scenario. So computerized adaptive instruction refers to the use of computer software--almost always incorporating artificially intelligent services--which has been designed to adjust both the presentation of information and the form of questioning to meet the current needs of an individual learner. This chapter describes a system for Internet-delivered adaptive instruction. The author attempts to demonstrate a sharp difference between the teaching that takes place outside of the classroom in universities and the kind that is at least afforded, if not taken advantage of by many, students in a more personalized educational setting such as those in the small liberal arts colleges. The author describes a computer-based technology that allows that gap to be bridged with the advantage of at least having more highly prepared learners sitting in college classrooms. A limited range of emerging research that supports that proposition is cited. (PsycINFO Database Record (c) 2005 APA ) %B Evidence-based educational methods %S Educational Psychology Series %I Elsevier Academic Press %C San Diego, CA. USA %P 143-169 %G eng %& 10 %0 Journal Article %J WSEAS Transactions on Communications %D 2004 %T Adaptive exploration of user knowledge in computer based testing %A Lamboudis, D. %A Economides, A. A. %B WSEAS Transactions on Communications %V 3 (1) %P 322-327 %G eng %0 Journal Article %J Journal of Educational and Behavioral Statistics %D 2004 %T Adaptive Testing With Regression Trees in the Presence of Multidimensionality %A Yan, Duanli %A Lewis, Charles %A Stocking, Martha %X

It is unrealistic to suppose that standard item response theory (IRT) models will be appropriate for all the new and currently considered computer-based tests. In addition to developing new models, we also need to give attention to the possibility of constructing and analyzing new tests without the aid of strong models. Computerized adaptive testing currently relies heavily on IRT. Alternative, empirically based, nonparametric adaptive testing algorithms exist, but their properties are little known. This article introduces a nonparametric, tree-based algorithm for adaptive testing and shows that it may be superior to conventional, IRT-based adaptive testing in cases where the IRT assumptions are not satisfied. In particular, it shows that the tree-based approach clearly outperformed (one-dimensional) IRT when the pool was strongly two-dimensional.

%B Journal of Educational and Behavioral Statistics %V 29 %P 293-316 %U http://jeb.sagepub.com/cgi/content/abstract/29/3/293 %R 10.3102/10769986029003293 %0 Generic %D 2004 %T The AMC Linear Disability Score project in a population requiring residential care: psychometric properties %A Holman, R. %A Lindeboom, R. %A Vermeulen, M. %A de Haan, R. J. %K *Disability Evaluation %K *Health Status Indicators %K Activities of Daily Living/*classification %K Adult %K Aged %K Aged, 80 and over %K Data Collection/methods %K Female %K Humans %K Logistic Models %K Male %K Middle Aged %K Netherlands %K Pilot Projects %K Probability %K Psychometrics/*instrumentation %K Questionnaires/standards %K Residential Facilities/*utilization %K Severity of Illness Index %X BACKGROUND: Currently there is a lot of interest in the flexible framework offered by item banks for measuring patient relevant outcomes, including functional status. However, there are few item banks, which have been developed to quantify functional status, as expressed by the ability to perform activities of daily life. METHOD: This paper examines the psychometric properties of the AMC Linear Disability Score (ALDS) project item bank using an item response theory model and full information factor analysis. Data were collected from 555 respondents on a total of 160 items. RESULTS: Following the analysis, 79 items remained in the item bank. The remaining 81 items were excluded because of: difficulties in presentation (1 item); low levels of variation in response pattern (28 items); significant differences in measurement characteristics for males and females or for respondents under or over 85 years old (26 items); or lack of model fit to the data at item level (26 items). CONCLUSIONS: It is conceivable that the item bank will have different measurement characteristics for other patient or demographic populations. However, these results indicate that the ALDS item bank has sound psychometric properties for respondents in residential care settings and could form a stable base for measuring functional status in a range of situations, including the implementation of computerised adaptive testing of functional status. %B Health and Quality of Life Outcomes %7 2004/08/05 %V 2 %P 42 %8 Aug 3 %@ 1477-7525 (Electronic)1477-7525 (Linking) %G eng %M 15291958 %2 514531 %0 Book %D 2004 %T The application of cognitive diagnosis and computerized adaptive testing to a large-scale assessment %A McGlohen, MK %C Unpublished doctoral dissertation, University of Texas at Austin %G eng %0 Journal Article %J European Journal of Psychological Assessment %D 2004 %T Assisted self-adapted testing: A comparative study %A Hontangas, P. %A Olea, J. %A Ponsoda, V. %A Revuelta, J. %A Wise, S. L. %K Adaptive Testing %K Anxiety %K Computer Assisted Testing %K Psychometrics %K Test %X A new type of self-adapted test (S-AT), called Assisted Self-Adapted Test (AS-AT), is presented. It differs from an ordinary S-AT in that prior to selecting the difficulty category, the computer advises examinees on their best difficulty category choice, based on their previous performance. Three tests (computerized adaptive test, AS-AT, and S-AT) were compared regarding both their psychometric (precision and efficiency) and psychological (anxiety) characteristics. Tests were applied in an actual assessment situation, in which test scores determined 20% of term grades. A sample of 173 high school students participated. Neither differences in posttest anxiety nor ability were obtained. Concerning precision, AS-AT was as precise as CAT, and both revealed more precision than S-AT. It was concluded that AS-AT acted as a CAT concerning precision. Some hints, but not conclusive support, of the psychological similarity between AS-AT and S-AT was also found. (PsycINFO Database Record (c) 2005 APA ) (journal abstract) %B European Journal of Psychological Assessment %V 20 %P 2-9 %G eng %0 Conference Paper %B Paper presented at the annual meeting of the National Council on Measurement in Education %D 2004 %T Automated Simultaneous Assembly of Multi-Stage Testing for the Uniform CPA Examination %A Breithaupt, K %A Ariel, A. %A Veldkamp, B. %B Paper presented at the annual meeting of the National Council on Measurement in Education %C San Diego CA %G eng %0 Conference Paper %B Paper presented at the annual meeting of the American Educational Research Association %D 2003 %T Accuracy of reading and mathematics ability estimates under the shadow-test constraint MCAT %A Li, Y. H. %A Schafer, W. D. %B Paper presented at the annual meeting of the American Educational Research Association %C Chicago IL %G eng %0 Generic %D 2003 %T An adaptation of stochastic curtailment to truncate Wald’s SPRT in computerized adaptive testing %A Finkelman, M. %X

Computerized adaptive testing (CAT) has been shown to increase eÆciency in educational measurement. One common application of CAT is to classify students as either pro cient or not proficient in ability. A truncated form of Wald's sequential probability ratio test (SPRT), in which examination is halted after a prespeci ed number of questions, has been proposed to provide a diagnosis of prociency. This article studies the further truncation provided by stochastic curtailment, where an exam is stopped early if completion of the remaining questions would be unlikely to alter the classi cation of the examinee. In a simulation study presented, the increased truncation is shown to offer substantial improvement in test length with only a slight decrease in accuracy.

%I National Center for Research on Evaluation, Standards, and Student Testing %C Los Angeles %P 1-26 %8 Steptember, 2003 %G eng %0 Book Section %D 2003 %T Adaptive exploration of assessment results under uncertainty %A Lamboudis, D. %A Economides, A. A. %A Papastergiou, A. %C Proceedings 3rd IEEE ternational Conference on Advanced Learning Technologies, ICALT '03, 460-461, 2003. %G eng %0 Conference Paper %B Paper presented at the Annual meeting of the National Council on Measurement in Education %D 2003 %T An adaptive exposure control algorithm for computerized adaptive testing using a sharing item response theory model %A Segall, D. O. %B Paper presented at the Annual meeting of the National Council on Measurement in Education %C Chicago IL %G eng %0 Journal Article %J Applied Psychological Measurement %D 2003 %T Alpha-stratified adaptive testing with large numbers of content constraints %A van der Linden, W. J. %A Chang, Hua-Hua %B Applied Psychological Measurement %V 27 %P 107-120 %G eng %0 Conference Paper %B Paper presented at the annual meeting of the National Council on Measurement in Education %D 2003 %T The assembly of multiple form structures %A Armstrong, R. D. %A Little, J. %B Paper presented at the annual meeting of the National Council on Measurement in Education %C Chicago IL %G eng %0 Generic %D 2003 %T The assembly of multiple stage adaptive tests with discrete items %A Armstrong, R. D. %A Edmonds, J.J. %C Newtown, PA: Law School Admission Council Report %G eng %0 Conference Paper %B Paper presented at the annual meeting of the National Council on Measurement in Education %D 2003 %T Assessing CAT security breaches by the item pooling index %A Chang, Hua-Hua %A Zhang, J. %B Paper presented at the annual meeting of the National Council on Measurement in Education %C Chicago IL %G eng %0 Book Section %B Reusing online resources: A sustanable approach to e-learning %D 2003 %T Assessing question banks %A Bull, J. %A Dalziel, J. %A Vreeland, T. %K Computer Assisted Testing %K Curriculum Based Assessment %K Education %K Technology computerized adaptive testing %X In Chapter 14, Joanna Bull and James Daziel provide a comprehensive treatment of the issues surrounding the use of Question Banks and Computer Assisted Assessment, and provide a number of excellent examples of implementations. In their review of the technologies employed in Computer Assisted Assessment the authors include Computer Adaptive Testing and data generation. The authors reveal significant issues involving the impact of Intellectual Property rights and computer assisted assessment and make important suggestions for strategies to overcome these obstacles. (PsycINFO Database Record (c) 2005 APA )http://www-jime.open.ac.uk/2003/1/ (journal abstract) %B Reusing online resources: A sustanable approach to e-learning %I Kogan Page Ltd. %C London, UK %P 171-230 %G eng %0 Conference Paper %B Paper presented at the annual meeting of the American Educational Research Association %D 2003 %T Assessing the efficiency of item selection in computerized adaptive testing %A Weissman, A. %B Paper presented at the annual meeting of the American Educational Research Association %C Chicago IL %G eng %0 Journal Article %J British Journal of Mathematical and Statistical Psychology %D 2003 %T a-Stratified multistage CAT design with content-blocking %A Yi, Q. %A Chang, H.-H. %B British Journal of Mathematical and Statistical Psychology %V 56 %P 359–378 %0 Conference Paper %B Paper presented at the annual meeting of the American Educational Research Association %D 2002 %T Accuracy of the ability estimate and the item exposure rate under multidimensional adaptive testing with item constraints %A Li, Y. H. %A Yu, N. Y. %B Paper presented at the annual meeting of the American Educational Research Association %C New Orleans LA %G eng %0 Generic %D 2002 %T Adaptive testing without IRT in the presence of multidimensionality (Research Report 02-09) %A Yan, D. %A Lewis, C. %A Stocking, M. %C Princeton NJ: Educational Testing Service %G eng %0 Journal Article %J Seminars in Oncology %D 2002 %T Advances in quality of life measurements in oncology patients %A Cella, D. %A Chang, C-H. %A Lai, J. S. %A Webster, K. %K *Quality of Life %K *Sickness Impact Profile %K Cross-Cultural Comparison %K Culture %K Humans %K Language %K Neoplasms/*physiopathology %K Questionnaires %X Accurate assessment of the quality of life (QOL) of patients can provide important clinical information to physicians, especially in the area of oncology. Changes in QOL are important indicators of the impact of a new cytotoxic therapy, can affect a patient's willingness to continue treatment, and may aid in defining response in the absence of quantifiable endpoints such as tumor regression. Because QOL is becoming an increasingly important aspect in the management of patients with malignant disease, it is vital that the instruments used to measure QOL are reliable and accurate. Assessment of QOL involves a multidimensional approach that includes physical, functional, social, and emotional well-being, and the most comprehensive instruments measure at least three of these domains. Instruments to measure QOL can be generic (eg, the Nottingham Health Profile), targeted toward specific illnesses (eg, Functional Assessment of Cancer Therapy - Lung), or be a combination of generic and targeted. Two of the most widely used examples of the combination, or hybrid, instruments are the European Organization for Research and Treatment of Cancer Quality of Life Questionnaire Core 30 Items and the Functional Assessment of Chronic Illness Therapy. A consequence of the increasing international collaboration in clinical trials has been the growing necessity for instruments that are valid across languages and cultures. To assure the continuing reliability and validity of QOL instruments in this regard, item response theory can be applied. Techniques such as item response theory may be used in the future to construct QOL item banks containing large sets of validated questions that represent various levels of QOL domains. As QOL becomes increasingly important in understanding and approaching the overall management of cancer patients, the tools available to clinicians and researchers to assess QOL will continue to evolve. While the instruments currently available provide reliable and valid measurement, further improvements in precision and application are anticipated. %B Seminars in Oncology %V 29 %P 60-8 %8 Jun %G eng %M 12082656 %0 Journal Article %J Educational Media International %D 2002 %T Applicable adaptive testing models for school teachers %A Chang-Hwa, W. A. %A Chuang, C-L. %X The purpose of this study was to investigate the attitudinal effects on SPRT adaptive testing environment for junior high school students. Subjects were 39 eighth graders from a selected junior high school. Major instruments for the study were the Junior High School Natural Sciences Adaptive Testing System driven by the SPRT algorithm, and a self-developed attitudinal questionnaire, factors examined include: test anxiety, examinee preference, adaptability of the test, and acceptance of the test result. The major findings were that overall, junior high school students" attitudes towards computerized adaptive tests were positive, no significant correlations existed between test attitude and the test length. The results indicated that junior high school students generally have positive attitudes towards adaptive testing.Modèles de tests d"adaptation à l"usage des enseignants. L"objectif de cette étude était d"enquêter sur les effets causés par une passation de tests d"adaptation ( selon l"algorithme "Sequential Probability Radio Test " (SPRT) ) dans une classe de trente-neuf élèves de huitième année du secondaire inférieur. Les principaux instruments utilisés ont été ceux du système de tests d"adaptation (avec le SPRT) et destiné aux classes de sciences naturelles du degré secondaire inférieur. Un questionnaire d"attitude, développé par nos soins, a également été utilisé pour examiner les facteurs suivants: test d"anxiété, préférence des candidats, adaptabilité du test et acceptation des résultats. Les principales conclusions ont été que, dans l"ensemble, l"attitude des élèves du secondaire inférieur face aux tests d"adaptation informatisés a été positive, aucune corrélation significative existant entre cette attitude et la longueur des tests. Les résultats démontrent aussi que les élèves du secondaire ont une attitude généralement positive envers les tests d"adaptation.Test Modelle zur Anwendung durch Klassenlehrer Zweck dieser Untersuchung war, die Auswirkungen über die Einstellung von Jun. High School Schülern im Zusammenhang mit dem SPRT Testumfeld zu untersuchen. 39 Achtklässler einer Jun. High School nahmen an dem Test teil. Die Untersuchung stützte sich hauptsächlich auf das Jun. High School Natural. Sciences Adaptive Testing System, das auf dem SPRT Rechnungsverfahren basiert sowie einem selbst erstellten Fragebogen mit folgenden Faktoren: Testängste, Präferenzen der Testperson, Geeignetheit des Tests, Anerkennung des Testergebnisses. Es stellte sich heraus, dass die Einstellung der Studenten zu den Computer adaptierten Tests im allgemeinen positiv waren; es ergaben sich keine bedeutsamen Wechselwirkungen zwischen persönlicher Testeinstellung und Testlänge. Die Ergebnisse belegen, dass Jun. High School Schüler im allgemeinen eine positive Haltung zu adaptierten Tests haben. %B Educational Media International %V 39 %P 55-59 %G eng %M EJ654148 %0 Journal Article %J Applied Psychological Measurement %D 2002 %T Application of an empirical Bayes enhancement of Mantel-Haenszel differential item functioning analysis to a computerized adaptive test %A Zwick, R. %A Thayer, D. T. %B Applied Psychological Measurement %V 26 %P 57-76 %G eng %0 Book %D 2002 %T Assessing the efficiency of item selection in computerized adaptive testing %A Weissman, A. %C Unpublished doctoral dissertation, University of Pittsburgh. %G eng %0 Journal Article %J Drug and Alcohol Dependence %D 2002 %T Assessing tobacco beliefs among youth using item response theory models %A Panter, A. T. %A Reeve, B. B. %K *Attitude to Health %K *Culture %K *Health Behavior %K *Questionnaires %K Adolescent %K Adult %K Child %K Female %K Humans %K Male %K Models, Statistical %K Smoking/*epidemiology %X Successful intervention research programs to prevent adolescent smoking require well-chosen, psychometrically sound instruments for assessing smoking prevalence and attitudes. Twelve thousand eight hundred and ten adolescents were surveyed about their smoking beliefs as part of the Teenage Attitudes and Practices Survey project, a prospective cohort study of predictors of smoking initiation among US adolescents. Item response theory (IRT) methods are used to frame a discussion of questions that a researcher might ask when selecting an optimal item set. IRT methods are especially useful for choosing items during instrument development, trait scoring, evaluating item functioning across groups, and creating optimal item subsets for use in specialized applications such as computerized adaptive testing. Data analytic steps for IRT modeling are reviewed for evaluating item quality and differential item functioning across subgroups of gender, age, and smoking status. Implications and challenges in the use of these methods for tobacco onset research and for assessing the developmental trajectories of smoking among youth are discussed. %B Drug and Alcohol Dependence %V 68 %P S21-S39 %8 Nov %G eng %M 12324173 %0 Conference Paper %B Paper presented at the annual meeting of the American Educational Research Association %D 2001 %T Adaptation of a-stratified method in variable length computerized adaptive testing %A Wen, J.-B. %A Chang, Hua-Hua %A Hau, K.-T.  %B Paper presented at the annual meeting of the American Educational Research Association %C Seattle WA %G eng %0 Generic %D 2001 %T Application of data mining to response data in a computerized adaptive test %A Mendez, F. A. %C Paper presented at the Annual Meeting of the National Council on Measurement in Education, Seattle WA %G eng %0 Conference Paper %B Paper presented at the annual meeting of the National Council on Measurement in Education %D 2001 %T Application of score information for CAT pool development and its connection with "likelihood test information %A Krass, I. A. %B Paper presented at the annual meeting of the National Council on Measurement in Education %C Seattle WA %G eng %0 Journal Article %J Teacher Development %D 2001 %T Assessment in the twenty-first century: A role of computerised adaptive testing in national curriculum subjects %A Cowan, P. %A Morrison, H. %K computerized adaptive testing %X With the investment of large sums of money in new technologies forschools and education authorities and the subsequent training of teachers to integrate Information and Communications Technology (ICT) into their teaching strategies, it is remarkable that the old outdated models of assessment still remain. This article highlights the current problems associated with pen-and paper-testing and offers suggestions for an innovative and new approach to assessment for the twenty-first century. Based on the principle of the 'wise examiner' a computerised adaptive testing system which measures pupils' ability against the levels of the United Kingdom National Curriculum has been developed for use in mathematics. Using constructed response items, pupils are administered a test tailored to their ability with a reliability index of 0.99. Since the software administers maximally informative questions matched to each pupil's current ability estimate, no two pupils will receive the same set of items in the same order therefore removing opportunities for plagarism and teaching to the test. All marking is automated and a journal recording the outcome of the test and highlighting the areas of difficulty for each pupil is available for printing by the teacher. The current prototype of the system can be used on a school's network however the authors envisage a day when Examination Boards or the Qualifications and Assessment Authority (QCA) will administer Government tests from a central server to all United Kingdom schools or testing centres. Results will be issued at the time of testing and opportunities for resits will become more widespr %B Teacher Development %V 5 %P 241-57 %G eng %M EJ644183 %0 Conference Paper %B Paper presented at the Annual Meeting of the Psychometric Society %D 2001 %T a-stratified CAT design with content-blocking %A Yi, Q. %A Chang, Hua-Hua %B Paper presented at the Annual Meeting of the Psychometric Society %C King of Prussia, PA %G eng %0 Conference Paper %B Paper presented at the annual meeting of the American Educational Research Association %D 2001 %T a-stratified computerized adaptive testing with unequal item exposure across strata %A Deng, H. %A Chang, Hua-Hua %B Paper presented at the annual meeting of the American Educational Research Association %C Seattle WA %G eng %0 Journal Article %J Applied Psychological Measurement %D 2001 %T a-stratified multistage computerized adaptive testing with b blocking %A Chang, Hua-Hua %A Qian, J. %A Yang, Z. %K computerized adaptive testing %X Proposed a refinement, based on the stratification of items developed by D. Weiss (1973), of the computerized adaptive testing item selection procedure of H. Chang and Z. Ying (1999). Simulation studies using an item bank from the Graduate Record Examination show the benefits of the new procedure. (SLD) %B Applied Psychological Measurement %V 25 %P 333-41 %G eng %M EJ644200 %0 Journal Article %J Applied Psychological Measurement %D 2001 %T a-Stratified multistage computerized adaptive testing with b blocking %A Chang, Hua-Hua %A Qian, J. %A Ying, Z. %X Chang & Ying’s (1999) computerized adaptive testing item-selection procedure stratifies the item bank according to a parameter values and requires b parameter values to be evenly distributed across all strata. Thus, a and b parameter values must be incorporated into how strata are formed. A refinement is proposed, based on Weiss’ (1973) stratification of items according to b values. Simulation studies using a retired item bank of a Graduate Record Examination test indicate that the new approach improved control of item exposure rates and reduced mean squared errors. %B Applied Psychological Measurement %V 25 %P 333-341 %@ 0146-6216 %G eng %0 Generic %D 2000 %T Adaptive mastery testing using a multidimensional IRT model and Bayesian sequential decision theory (Research Report 00-06) %A Glas, C. A. W. %A Vos, H. J. %C Enschede, The Netherlands: University of Twente, Faculty of Educational Science and Technology, Department of Measurement and Data Analysis %G eng %0 Journal Article %J Psicothema %D 2000 %T Algoritmo mixto mínima entropía-máxima información para la selección de ítems en un test adaptativo informatizado %A Dorronsoro, J. R. %A Santa-Cruz, C. %A Rubio Franco, V. J. %A Aguado García, D. %K computerized adaptive testing %X El objetivo del estudio que presentamos es comparar la eficacia como estrat egia de selección de ítems de tres algo ritmos dife rentes: a) basado en máxima info rmación; b) basado en mínima entropía; y c) mixto mínima entropía en los ítems iniciales y máxima info rmación en el resto; bajo la hipótesis de que el algo ritmo mixto, puede dotar al TAI de mayor eficacia. Las simulaciones de procesos TAI se re a l i z a ron sobre un banco de 28 ítems de respuesta graduada calibrado según el modelo de Samejima, tomando como respuesta al TAI la respuesta ori ginal de los sujetos que fueron utilizados para la c a l i b ración. Los resultados iniciales mu e s t ran cómo el cri t e rio mixto es más eficaz que cualquiera de los otros dos tomados indep e n d i e n t e m e n t e. Dicha eficacia se maximiza cuando el algo ritmo de mínima entropía se re s t ri n ge a la selección de los pri m e ros ítems del TAI, ya que con las respuestas a estos pri m e ros ítems la estimación de q comienza a ser re l evante y el algo ritmo de máxima informaciónse optimiza.Item selection algo rithms in computeri zed adap t ive testing. The aim of this paper is to compare the efficacy of three different item selection algo rithms in computeri zed adap t ive testing (CAT). These algorithms are based as follows: the first one is based on Item Info rm ation, the second one on Entropy, and the last algo rithm is a mixture of the two previous ones. The CAT process was simulated using an emotional adjustment item bank. This item bank contains 28 graded items in six categories , calibrated using Samejima (1969) Graded Response Model. The initial results show that the mixed criterium algorithm performs better than the other ones. %B Psicothema %V 12 %P 12-14 %G eng %0 Conference Paper %B Paper presented at the annual meeting of the National Council on Measurement in Education %D 2000 %T Applying specific information item selection to a passage-based test %A Thompson, T.D. %A Davey, T. %B Paper presented at the annual meeting of the National Council on Measurement in Education %C New Orleans, LA, April %G eng %0 Conference Paper %B Paper presented at the annual meeting of the American Educational Research Association %D 2000 %T Assembling parallel item pools for computerized adaptive testing %A Wang, T. %A Fan, M. Yi, Q. %A Ban, J. C. %A Zhu, D. %B Paper presented at the annual meeting of the American Educational Research Association %C New Orleans %G eng %0 Generic %D 1999 %T Adaptive testing with equated number-correct scoring (Research Report 99-02) %A van der Linden, W. J. %C Enschede, The Netherlands: University of Twente, Faculty of Educational Science and Technology, Department of Measurement and Data Analysis %G eng %0 Conference Paper %B Paper presented at the annual meeting of the American Educational Research Association %D 1999 %T Adjusting computer adaptive test starting points to conserve item pool %A Zhu, D. %A Fan. M. %B Paper presented at the annual meeting of the American Educational Research Association %C Montreal, Canada %G eng %0 Conference Paper %B Paper presented at the annual meeting of the American Educational Research Association %D 1999 %T Adjusting "scores" from a CAT following successful item challenges %A Wang, T. %A Yi, Q. %A Ban, J. C. %A Harris, D. J. %A Hanson, B. A. %B Paper presented at the annual meeting of the American Educational Research Association %C Montreal, Canada %G eng %0 Conference Paper %B Paper presented at the annual meeting of the National Council on Measurement in Education< Montreal %D 1999 %T Alternative item selection strategies for improving test security and pool usage in computerized adaptive testing %A Robin, F. %B Paper presented at the annual meeting of the National Council on Measurement in Education< Montreal %C Canada %G eng %0 Journal Article %J Dissertation Abstracts International: Section B: the Sciences & Engineering %D 1999 %T Alternative methods for the detection of item preknowledge in computerized adaptive testing %A McLeod, Lori Davis %K computerized adaptive testing %B Dissertation Abstracts International: Section B: the Sciences & Engineering %V 59 %P 3765 %G eng %0 Journal Article %J Applied Psychological Measurement %D 1999 %T a-stratified multistage computerized adaptive testing %A Chang, Hua-Hua %A Ying, Z. %K computerized adaptive testing %X For computerized adaptive tests (CAT) based on the three-parameter logistic mode it was found that administering items with low discrimination parameter (a) values early in the test and administering those with high a values later was advantageous; the skewness of item exposure distributions was reduced while efficiency was maintain in trait level estimation. Thus, a new multistage adaptive testing approach is proposed that factors a into the item selection process. In this approach, the items in the item bank are stratified into a number of levels based on their a values. The early stages of a test use items with lower as and later stages use items with higher as. At each stage, items are selected according to an optimization criterion from the corresponding level. Simulation studies were performed to compare a-stratified CATs with CATs based on the Sympson-Hetter method for controlling item exposure. Results indicated that this new strategy led to tests that were well-balanced, with respect to item exposure, and efficient. The a-stratified CATs achieved a lower average exposure rate than CATs based on Bayesian or information-based item selection and the Sympson-Hetter method. (PsycINFO Database Record (c) 2003 APA, all rights reserved). %B Applied Psychological Measurement %V 23 %P 211-222 %G eng %0 Journal Article %J Applied Psychological Measurement %D 1999 %T a-stratified multistage computerized adaptive testing %A Chang, Hua-Hua %A Ying, Z. %B Applied Psychological Measurement %V 23 %P 211-222 %G eng %0 Journal Article %J Applied Psychological Measurement %D 1999 %T a-stratified multistage computerized adaptive testing %A Chang, Hua-Hua %A Ying, Z. %B Applied Psychological Measurement %V 23 %P 211–222 %G eng %0 Conference Paper %B Paper presented at the annual meeting of the National Council on Measurement in Education %D 1999 %T Automated flawed item detection and graphical item used in on-line calibration of CAT-ASVAB. %A Krass, I. A. %A Thomasson, G. L. %B Paper presented at the annual meeting of the National Council on Measurement in Education %C Montreal, Canada %G eng %0 Generic %D 1998 %T Adaptive mastery testing using the Rasch model and Bayesian sequential decision theory (Research Report 98-15) %A Glas, C. A. W. %A Vos, H. J. %C Enschede, The Netherlands: University of Twente, Faculty of Educational Science and Technology, Department of Measurement and Data Analysis %G eng %0 Conference Paper %B Paper presented at the annual meeting of the National Council on Measurement in Education %D 1998 %T Adaptive testing without IRT %A Yan, D. %A Lewis, C. %A Stocking, M. %B Paper presented at the annual meeting of the National Council on Measurement in Education %C San Diego CA %G eng %0 Conference Paper %B Paper presented at an Educational Testing Service-sponsored colloquium entitled Computer-based testing: Building the foundations for future assessments %D 1998 %T Alternatives for scoring computerized adaptive tests %A Dodd, B. G. %A Fitzpatrick, S. J. %B Paper presented at an Educational Testing Service-sponsored colloquium entitled Computer-based testing: Building the foundations for future assessments %C Philadelphia PA %G eng %0 Book Section %B Computer-based testing %D 1998 %T Alternatives for scoring computerized adaptive tests %A Dodd, B. G. %A Fitzpatrick, S. J. %E J. J. Fremer %E W. C. Ward %B Computer-based testing %I Lawrence Erlbaum Associates, Inc. %C Mahwah, N.J., USA %G eng %0 Conference Paper %B Paper presented at the annual meeting of the Society for Industrial and Organization Psychology %D 1998 %T Application of an IRT ideal point model to computer adaptive assessment of job performance %A Stark, S. %A F Drasgow %B Paper presented at the annual meeting of the Society for Industrial and Organization Psychology %C Dallas TX %G eng %0 Conference Paper %B Paper presented at the annual meeting of the National Council on Measurement in Education %D 1998 %T Application of direct optimization for on-line calibration in computerized adaptive testing %A Krass, I. A. %B Paper presented at the annual meeting of the National Council on Measurement in Education %C San Diego CA %G eng %0 Book %D 1998 %T Applications of network flows to computerized adaptive testing %A Cordova, M. J. %C Dissertation, Rutgers Center for Operations Research (RUTCOR), Rutgers University, New Brunswick NJ %G eng %0 Journal Article %J Dissertation Abstracts International: Section B: the Sciences & Engineering %D 1998 %T Applications of network flows to computerized adaptive testing %A Claudio, M. J. C. %K computerized adaptive testing %X Recently, the concept of Computerized Adaptive Testing (CAT) has been receiving ever growing attention from the academic community. This is so because of both practical and theoretical considerations. Its practical importance lies in the advantages of CAT over the traditional (perhaps outdated) paper-and-pencil test in terms of time, accuracy, and money. The theoretical interest is sparked by its natural relationship to Item Response Theory (IRT). This dissertation offers a mathematical programming approach which creates a model that generates a CAT that takes care of many questions concerning the test, such as feasibility, accuracy and time of testing, as well as item pool security. The CAT generated is designed to obtain the most information about a single test taker. Several methods for eatimating the examinee's ability, based on the (dichotomous) responses to the items in the test, are also offered here. (PsycINFO Database Record (c) 2003 APA, all rights reserved). %B Dissertation Abstracts International: Section B: the Sciences & Engineering %V 59 %P 0855 %G eng %0 Conference Paper %B Paper presented at the annual meeting of the National Council on Measurement in Education %D 1997 %T The accuracy of examinee judgments of relative item difficulty: Implication for computerized adaptive testing %A Wise, S. L. %A Freeman, S.A. %A Finney, S. J. %A Enders, C. K. %A Severance, D.D. %B Paper presented at the annual meeting of the National Council on Measurement in Education %C Chicago %G eng %0 Journal Article %J Personnel Psychology %D 1997 %T Adapting to adaptive testing %A Overton, R. C. %A Harms, H. J. %A Taylor, L. R. %A Zickar, M.. J. %B Personnel Psychology %V 50 %P 171-185 %G eng %0 Conference Paper %B Paper presented at the annual meeting of the National Council on Measurement in Education %D 1997 %T Administering and scoring the computerized adaptive testing %A A Zara %B Paper presented at the annual meeting of the National Council on Measurement in Education %C Chicago IL %G eng %0 Conference Paper %B Paper presented at the annual meeting of the National Council on Measurement in Education %D 1997 %T Alternate methods of scoring computer-based adaptive tests %A Green, B. F. %B Paper presented at the annual meeting of the National Council on Measurement in Education %C Chicago IL %G eng %0 Journal Article %J Journal of Educational and Behavioral Statistics %D 1997 %T An alternative method for scoring adaptive tests %A Stocking, M. L. %B Journal of Educational and Behavioral Statistics %V 21 %P 365-389 %G eng %0 Generic %D 1997 %T Applications of Bayesian decision theory to sequential mastery testing (Research Report 97-06) %A Vos, H. J. %C Twente, The Netherlands: Department of Educational Measurement and Data Analysis %G eng %0 Conference Paper %B Paper presented at the annual meeting of the National Council on Measurement in Education %D 1997 %T Applications of multidimensional adaptive testing %A Segall, D. O. %B Paper presented at the annual meeting of the National Council on Measurement in Education %C Montreal, Canada %G eng %0 Conference Paper %B Paper presented at the annual meeting of the National Council on Measurement in Education %D 1997 %T Assessing speededness in variable-length computer-adaptive tests %A Bontempo, B %A Julian, E. R %A Gorham, J. L. %B Paper presented at the annual meeting of the National Council on Measurement in Education %C Chicago IL %G eng %0 Book Section %D 1996 %T Adaptive assessment and training using the neighbourhood of knowledge states %A Dowling, C. E. %A Hockemeyer, C. %A Ludwig, A .H. %C Frasson, C. and Gauthier, G. and Lesgold, A. (eds.) Intelligent Tutoring Systems, Third International Conference, ITS'96, Montral, Canada, June 1996 Proceedings. Lecture Notes in Computer Science 1086. Berlin Heidelberg: Springer-Verlag 578-587. %G eng %0 Book Section %D 1996 %T Adaptive assessment using granularity hierarchies and Bayesian nets %A Collins, J. A. %A Greer, J. E. %A Huang, S. X. %C Frasson, C. and Gauthier, G. and Lesgold, A. (Eds.) Intelligent Tutoring Systems, Third International Conference, ITS'96, Montréal, Canada, June 1996 Proceedings. Lecture Notes in Computer Science 1086. Berlin Heidelberg: Springer-Verlag 569-577. %G eng %0 Book %D 1996 %T Adaptive testing with granularity %A Collins, J. A. %C Masters thesis, University of Saskatchewan, Department of Computer Science %G eng %0 Journal Article %J Journal of Educational and Behavioral Statistics %D 1996 %T An alternative method for scoring adaptive tests %A Stocking, M. L. %B Journal of Educational and Behavioral Statistics %V 21 %P 365-389 %G eng %0 Journal Article %J Dissertation Abstracts International: Section B: the Sciences & Engineering %D 1995 %T Assessment of scaled score consistency in adaptive testing from a multidimensional item response theory perspective %A Fan, Miechu %K computerized adaptive testing %X The purpose of this study was twofold: (a) to examine whether the unidimensional adaptive testing estimates are comparable for different ability levels of examinees when the true examinee-item interaction is correctly modeled using a compensatory multidimensional item response theory (MIRT) model; and (b) to investigate the effects of adaptive testing estimation when the procedure of item selection of computerized adaptive testing (CAT) is controlled by either content-balancing or selecting the most informative item in a user specified direction at the current estimate of unidimensional ability. A series of Monte Carlo simulations were conducted in this study. Deviation from the reference composite angle was used as an index of the theta1,theta2-composite consistency across the different levels of unidimensional CAT estimates. In addition, the effect of the content-balancing item selection procedure and the fixed-direction item selection procedure were compared across the different ability levels. The characteristics of item selection, test information and the relationship between unidimensional and multidimensional models were also investigated. In addition to employing statistical analysis to examine the robustness of the CAT procedure violations of unidimensionality, this research also included graphical analyses to present the results. The results were summarized as follows: (a) the reference angles for the no-control-item-selection method were disparate across the unidimensional ability groups; (b) the unidimensional CAT estimates from the content-balancing item selection method did not offer much improvement; (c) the fixed-direction-item selection method did provide greater consistency for the unidimensional CAT estimates across the different levels of ability; (d) and, increasing the CAT test length did not provide greater score scale consistency. Based on the results of this study, the following conclusions were drawn: (a) without any controlling (PsycINFO Database Record (c) 2003 APA, all rights reserved). %B Dissertation Abstracts International: Section B: the Sciences & Engineering %V 55 %P 5598 %G eng %0 Journal Article %J Educational and Psychological Measurement %D 1994 %T ADTEST: A computer-adaptive tests based on the maximum information principle %A Ponsoda, V. %A Olea, J., %A Revuelta, J. %B Educational and Psychological Measurement %V 54 %P 680-686 %G eng %0 Journal Article %J Applied Psychological Measurement %D 1993 %T The application of an automated item selection method to real data %A Stocking, M. L. %A Swanson, L. %A Pearlman, M. %B Applied Psychological Measurement %V 17 %P 167-176 %G eng %0 Journal Article %J Dissertation Abstracts International %D 1993 %T An application of Computerized Adaptive Testing to the Test of English as a Foreign Language %A Moon, O. %K computerized adaptive testing %B Dissertation Abstracts International %V 53 %P 4257-4258 %G eng %0 Journal Article %J Educational Measurement: Issues and Practice %D 1993 %T Assessing the utility of item response models: computerized adaptive testing %A Kingsbury, G. G. %A Houser, R.L. %K computerized adaptive testing %B Educational Measurement: Issues and Practice %V 12 %P 21-27 %G eng %0 Conference Paper %B Paper presented at the annual meeting of the American Educational Research Association %D 1992 %T Ability measure equivalence of computer adaptive and paper and pencil tests: A research synthesis %A Bergstrom, B. %B Paper presented at the annual meeting of the American Educational Research Association %C San Francisco %0 Journal Article %J Applied Measurement in Education %D 1992 %T Altering the level of difficulty in computer adaptive testing %A Bergstrom, Betty A. %A Lunz, M. E. %A Gershon, R. C. %K computerized adaptive testing %X Examines the effect of altering test difficulty on examinee ability measures and test length in a computer adaptive test. The 225 Ss were randomly assigned to 3 test difficulty conditions and given a variable length computer adaptive test. Examinees in the hard, medium, and easy test condition took a test targeted at the 50%, 60%, or 70% probability of correct response. The results show that altering the probability of a correct response does not affect estimation of examinee ability and that taking an easier computer adaptive test only slightly increases the number of items necessary to reach specified levels of precision. (PsycINFO Database Record (c) 2002 APA, all rights reserved). %B Applied Measurement in Education %V 5 %P 137-149 %G eng %0 Journal Article %J Psychometrika %D 1992 %T The application of latent class models in adaptive testing %A Macready, G. B. %A Dayton, C. M. %B Psychometrika %V 57 %P 71-88 %G eng %0 Conference Paper %B ERIC Document No. TM022404 %D 1992 %T Assessing existing item bank depth for computer adaptive testing %A Bergstrom, Betty A. %A Stahl, J. A. %B ERIC Document No. TM022404 %G eng %0 Generic %D 1991 %T An analysis of CAT-ASVAB scores in the Marine Corps JPM data (CRM- 91-161) %A Divgi, D. R. %C Alexandria VA: Center for Naval Analysis %G eng %0 Conference Paper %B Paper presented at the annual meeting of the American Educational Research Association %D 1991 %T Applications of computer-adaptive testing in Maryland %A Baghi, H %A Gabrys, R. %A Ferrara, S. %B Paper presented at the annual meeting of the American Educational Research Association %C Chicago IL %G eng %0 Generic %D 1991 %T Automatic item selection (AIS) methods in the ETS testing environment (Research Memorandum 91-5) %A Stocking, M. L. %A Swanson, L. %A Pearlman, M. %C Princeton NJ: Educational Testing Service %G eng %0 Journal Article %J Educational Measurement: Issues and Practice %D 1990 %T Adapting adaptive testing: Using the MicroCAT Testing System in a local school district %A Kingsbury, G. G. %B Educational Measurement: Issues and Practice %V 29 (2) %P 3-6 %G eng %0 Generic %D 1990 %T An adaptive algebra test: A testlet-based, hierarchically structured test with validity-based scoring %A Wainer, H., %A Lewis, C. %A Kaplan, B, %A Braswell, J. %C ETS Technical Report 90-92 %G eng %0 Journal Article %J Journal of Marketing Research %D 1990 %T Adaptive designs for Likert-type data: An approach for implementing marketing research %A Singh, J. %A Howell, R. D. %A Rhoads, G. K. %B Journal of Marketing Research %V 27 %P 304-321 %G eng %0 Journal Article %J Measurement and Evaluation in Counseling and Development %D 1990 %T Applying computerized adaptive testing in schools %A Olson, J. B %B Measurement and Evaluation in Counseling and Development %V 23 %P 311-38 %G eng %0 Conference Paper %B A paper presented to the annual meeting of the National Council of Measurement in Education %D 1990 %T Assessing the utility of item response models: Computerized adaptive testing %A Kingsbury, G. G. %A Houser, R.L. %B A paper presented to the annual meeting of the National Council of Measurement in Education %C Boston MA %G eng %0 Journal Article %J Applied Psychological Measurement %D 1989 %T Adaptive and Conventional Versions of the DAT: The First Complete Test Battery Comparison %A Henly, S. J. %A Klebe, K. J. %A J. R. McBride %A Cudeck, R. %B Applied Psychological Measurement %V 13 %P 363-371 %G English %N 4 %0 Journal Article %J Applied Psychological Measurement %D 1989 %T Adaptive and conventional versions of the DAT: The first complete test battery comparison %A Henly, S. J. %A Klebe, K. J. %A J. R. McBride %A Cudeck, R. %B Applied Psychological Measurement %V 13 %P 363-371 %G eng %0 Journal Article %J Applied Psychological Measurement %D 1989 %T Adaptive Estimation When the Unidimensionality Assumption of IRT is Violated %A Folk, V.G. %A Green, B. F. %B Applied Psychological Measurement %V 13 %P 373-390 %G English %N 4 %0 Journal Article %J Applied Psychological Measurement %D 1989 %T Adaptive estimation when the unidimensionality assumption of IRT is violated %A Folk, V.G., %A Green, B. F. %B Applied Psychological Measurement %V 13 %P 373-389 %G eng %0 Journal Article %J Educational Measurement: Issues and Practice %D 1989 %T Adaptive testing: The evolution of a good idea %A Reckase, M. D. %K computerized adaptive testing %B Educational Measurement: Issues and Practice %V 8 %P 11-15 %@ 1745-3992 %G eng %0 Journal Article %J Dissertation Abstracts International %D 1989 %T Application of computerized adaptive testing to the University Entrance Exam of Taiwan, R.O.C %A Hung, P-H. %K computerized adaptive testing %B Dissertation Abstracts International %V 49 %P 3662 %G eng %0 Thesis %B Faculty of Behavioural and Social Sciences %D 1989 %T An applied study on computerized adaptive testing %A Schoonman, W. %K computerized adaptive testing %X (from the cover) The rapid development and falling prices of powerful personal computers, in combination with new test theories, will have a large impact on psychological testing. One of the new possibilities is computerized adaptive testing. During the test administration each item is chosen to be appropriate for the person being tested. The test becomes tailor-made, resolving some of the problems with classical paper-and-pencil tests. In this way individual differences can be measured with higher efficiency and reliability. Scores on other meaningful variables, such as response time, can be obtained easily using computers. /// In this book a study on computerized adaptive testing is described. The study took place at Dutch Railways in an applied setting and served practical goals. Topics discussed include the construction of computerized tests, the use of response time, the choice of algorithms and the implications of using a latent trait model. After running a number of simulations and calibrating the item banks, an experiment was carried out. In the experiment a pretest was administered to a sample of over 300 applicants, followed by an adaptive test. In addition, a survey concerning the attitudes of testees towards computerized testing formed part of the design. %B Faculty of Behavioural and Social Sciences %I University of Groingen %C Groningen, The Netherlands %P 185 %G eng %9 Dissertation %0 Book %D 1989 %T An applied study on computerized adaptive testing %A Schoonman, W. %C Amsterdam, The Netherlands: Swets and Zeitlinger %G eng %0 Conference Paper %B Paper presented at the annual meeting of the National Council on Measurement in Education %D 1989 %T Assessing the impact of using item parameter estimates obtained from paper-and-pencil testing for computerized adaptive testing %A Kingsbury, G. G. %A Houser, R.L. %B Paper presented at the annual meeting of the National Council on Measurement in Education %C San Francisco %G eng %0 Book %D 1988 %T Application of appropriateness measurement to a problem in computerized adaptive testing %A Candell, G. L. %C Unpublished doctoral dissertation, University of Illinois %G eng %0 Journal Article %J School Psychology Review %D 1988 %T Assessment of academic skills of learning disabled students with classroom microcomputers %A Watkins, M. W. %A Kush, J. C. %B School Psychology Review %V 17 %P 81-88 %G eng %0 Book %D 1987 %T An adaptive test of musical memory: An application of item response theory to the assessment of musical ability %A Vispoel, W. P. %C Doctoral dissertation, University of Illinois. Dissertation Abstracts International, 49, 79A. %G eng %0 Journal Article %J Applied Psychology: An International Review %D 1987 %T Adaptive testing %A Weiss, D. J. %A Vale, C. D. %B Applied Psychology: An International Review %V 36 %P 249-262 %G eng %0 Generic %D 1987 %T Adaptive testing, information, and the partial credit model %A Adams, R. J. %C Melbourne, Australia: University of Melbourne, Center for the Study of Higher Education %0 Journal Article %J Educational and Psychological Measurement %D 1986 %T An application of computer adaptive testing with communication handicapped examinees %A Garrison, W. M. %A Baumgarten, B. S. %K computerized adaptive testing %X This study was conducted to evaluate a computerized adaptive testing procedure for the measurement of mathematical skills of entry level deaf college students. The theoretical basis of the study was the Rasch model for person measurement. Sixty persons were tested using an Apple II Plus microcomputer. Ability estimates provided by the computerized procedure were compared for stability with those obtained six to eight weeks earlier from conventional (written) testing of the same subject matter. Students' attitudes toward their testing experiences also were measured. Substantial increases in measurement efficiency (by reducing test length) were realized through the adaptive testing procedure. Because the item pool used was not specifically designed for adaptive testing purposes, the psychometric quality of measurements resulting from the different testing methods was approximately equal. Attitudes toward computerized testing were favorable. %B Educational and Psychological Measurement %V 46 %P 23-35 %@ 0013-1644 %G eng %M 10.1177/0013164486461003 %0 Journal Article %J Dissertation Abstracts International %D 1985 %T Adaptive self-referenced testing as a procedure for the measurement of individual change due to instruction: A comparison of the reliabilities of change estimates obtained from conventional and adaptive testing procedures %A Kingsbury, G. G. %K computerized adaptive testing %B Dissertation Abstracts International %V 45 %P 3057 %G eng %0 Journal Article %J Journal of Consulting and Clinical Psychology %D 1985 %T Adaptive testing by computer %A Weiss, D. J. %B Journal of Consulting and Clinical Psychology %V 53 %P 774-789 %G eng %0 Journal Article %J Applied Psychological Measurement %D 1985 %T ALPHATAB: A lookup table for Bayesian computerized adaptive testing %A De Ayala, R. J., %A Koch, W. R. %B Applied Psychological Measurement %V 9 %P 326 %G eng %0 Generic %D 1985 %T Armed Services Vocational Aptitude Battery: Development of an adaptive item pool (AFHLR-TR-85-19; Technical Rep No 85-19) %A Prestwood, J. S. %A Vale, C. D. %A Massey, R. H. %A Welsh, J. R. %C Brooks Air Force Base TX: Air Force Human Resources Laboratory %G eng %0 Book %D 1984 %T Adaptive self-referenced testing as a procedure for the measurement of individual change in instruction: A comparison of the reliabilities of change estimates obtained from conventional and adaptive testing procedures %A Kingsbury, G. G. %C Unpublished doctoral dissertation, Univerity of Minnesota, Minneapolis %G eng %0 Generic %D 1984 %T Adaptive testing (Final Report Contract OPM-29-80) %A Trollip, S. R. %C Urbana-Champaign IL: University of Illinois, Aviation Research Laboratory %G eng %0 Generic %D 1984 %T Analysis of experimental CAT ASVAB data %A Allred, L. A %A Green, B. F. %C Baltimore MD: Johns Hopkins University, Department of Psychology %0 Generic %D 1984 %T Analysis of speeded test data from experimental CAT system %A Greaud, V. A., %A Green, B. F. %C Baltimore MD: Johns Hopkins University, Department of Psychology %G eng %0 Generic %D 1984 %T Application of adaptive testing to a fraction test (Research Report 84-3-NIE) %A Tatsuoka, K. K. %A Tatsuoka, M. M. %A Baillie, R. %C Urbana IL: Univerity of Illinois, Computer-Based Education Research Laboratory %G eng %0 Book Section %D 1983 %T Adaptive testing by computer %A Green, B. F. %C R. B. Ekstrom (ed.), Measurement, technology, and individuality in education. New directions for testing and measurement, Number 17. San Francisco: Jossey-Bass. %G eng %0 Generic %D 1983 %T Alternate forms reliability and concurrent validity of adaptive and conventional tests with military recruits %A Kiely, G. L. %A A Zara %A Weiss, D. J. %C Minneapolis MN: University of Minnesota, Department of Psychology, Computerized Adaptive Testing Laboratory %G eng %0 Journal Article %J Journal of Computer-Based Instruction %D 1983 %T An application of computerized adaptive testing in U. S. Army recruiting. %A Sands, W. A. %A Gade, P. A. %B Journal of Computer-Based Instruction %V 10 %P 87-89 %G eng %0 Journal Article %J Dissertation Abstracts International %D 1982 %T Ability measurement, test bias reduction, and psychological reactions to testing as a function of computer adaptive testing versus conventional testing %A Orban, J. A. %K computerized adaptive testing %B Dissertation Abstracts International %V 42 %P 4233 %G eng %0 Journal Article %J Applied Psychological Measurement %D 1982 %T Adaptive EAP estimation of ability in a microcomputer environment %A Bock, B. D., %A Mislevy, R. J. %B Applied Psychological Measurement %V 6 %P 431-444 %G eng %0 Generic %D 1982 %T An adaptive Private Pilot Certification Exam %A Trollip, S. R. %A Anderson, R. I. %C Aviation, Space, and Environmental Medicine %G eng %0 Conference Paper %B Paper presented at the annual meeting of the American Educational Research Association %D 1982 %T Assessing mathematics achievement with a tailored testing program %A Garrison, W. M. %A Baumgarten, B. S. %B Paper presented at the annual meeting of the American Educational Research Association %C New York %G eng %0 Journal Article %J International Journal of Man-Machine Studies %D 1982 %T Automated tailored testing using Raven’s Matrices and the Mill Hill vocabulary tests %A Watts, K., %A Baddeley, A. D. %A Williams, M. %B International Journal of Man-Machine Studies %V 17 %P 331-344 %G eng %0 Book %D 1981 %T Ability measurement, test bias reduction, and psychological reactions to testing as a function of computer adaptive testing versus conventional %A Orban, J. A. %C Unpublished doctoral dissertation, Virginia Polytechnic Institute and State University. Dissertational Abstracts International, 1982, 42,(10-B), 4233 %G eng %0 Generic %D 1981 %T Adaptive testing without a computer %A Friedman, D. %A Steinberg, A, %A Ree, M. J. %C Catalog of Selected Documents in Psychology, Nov 1981, 11, 74-75 (Ms. No. 2350). AFHRL Technical Report 80-66. %G eng %0 Book Section %D 1980 %T Adaptive verbal ability testing in a military setting %A J. R. McBride %C D. J. Weiss (Ed.), Proceedings of the 1979 Computerized Adaptive Testing Conference (pp. 4-15). Minneapolis MN: University of Minnesota, Department of Psychology, Psychometric Methods Program, Computerized Adaptive Testing Laboratory. %G eng %0 Generic %D 1980 %T An alternate-forms reliability and concurrent validity comparison of Bayesian adaptive and conventional ability tests (Research Report 80-5) %A Kingsbury, G. G. %A Weiss, D. J. %C Minneapolis, Department of Psychology, Psychometric Methods Program, Computerized Adaptive Testing Laboratory %G eng %0 Generic %D 1979 %T Adaptive mental testing: The state of the art (Technical Report 423) %A J. R. McBride %C Alexandria VA: U.S. Army Research Institute for the Behavioral and Social Sciences. %G eng %0 Generic %D 1979 %T An adaptive testing strategy for mastery decisions (Research Report 79-5) %A Kingsbury, G. G. %A Weiss, D. J. %C Minneapolis: University of Minnesota, Department of Psychology, Psychometric Methods Program %G eng %0 Generic %D 1979 %T Adaptive tests' usefulness for military personnel screening %A J. R. McBride %C In M. Wiskoff, Chair, Military Applications of Computerized Adaptive Testing. Symposium presented at the Annual Convention of the American Psychological Association, New York. %G eng %0 Conference Paper %B Presentation to the convention of the Western Psychological Association %D 1978 %T An adaptive test designed for paper-and-pencil testing %A J. R. McBride %B Presentation to the convention of the Western Psychological Association %C San Francisco, CA %G eng %0 Book Section %D 1978 %T Applications of latent trait theory to criterion-referenced testing %A J. R. McBride %C D.J. Weiss (Ed.), Proceedings of the 1977 Computerized Adaptive Testing Conference. Minneapolis, MN: University of Minnesota. %G eng %0 Book Section %D 1978 %T Applications of sequential testing procedures to performance testing %A Epstein, K. I. %A Knerr, C. S. %C D. J. Weiss (Ed.), Proceedings of the 1977 Computerized Adaptive Testing Conference. Minneapolis MN: University of Minnesota, Department of Psychology, Psychometric Methods Program. %G eng %0 Book Section %D 1977 %T Adaptive Branching in a Multi-Content Achievement Test %A Pennell, R. J. %A Harris, D. A. %C D. J. Weiss (Ed.), Proceedings of the 1977 Computerized Adaptive Testing Conference. Minneapolis MN: University of Minnesota, Department of Psychology, Psychometric Methods Program %0 Book Section %D 1977 %T An adaptive test of arithmetic reasoning %A J. R. McBride %C the Proceedings of the Nineteenth Military Testing Association conference, San Antonio, TX. %G eng %0 Book Section %D 1977 %T Adaptive testing and the problem of classification %A Vale, C. D. %C D. Weiss (Ed.), Applications of computerized adaptive testing (Research Report 77-1). Minneapolis MN: University of Minnesota, Department of Psychology, Psychometric Methods Program. %G eng %0 Book Section %D 1977 %T Adaptive Testing Applied to Hierarchically Structured Objectives-Based Programs %A Hambleton, R. K. %A Eignor, D. R. %C D. J. Weiss (Ed.), Proceedings of the 1977 Computerized Adaptive Testing Conference. Minneapolis MN: University of Minnesota, Department of Psychology, Psychometric Methods Program %0 Generic %D 1977 %T An adaptive testing strategy for achievement test batteries (Research Rep No 77-6) %A Brown, J. M %A Weiss, D. J. %C Minneapolis: University of Minnesota, Department of Psychology, Psychometric Methods Program %0 Journal Article %J Behavior Research Methods and Instrumentation %D 1977 %T Application of tailored testing to achievement measurement %A English, R. A. %A Reckase, M. D. %A Patience, W. M. %B Behavior Research Methods and Instrumentation %V 9 %P 158-161 %G eng %0 Book %D 1977 %T An application of the Rasch one-parameter logistic model to individual intelligence testing in a tailored testing environment %A Ireland, C. M. %C Dissertation Abstracts International, 37 (9-A), 5766 %G eng %0 Book Section %D 1977 %T Applications of adaptive testing in measuring achievement and performance %A Bejar, I. I. %C D. J. Weiss (Ed.), Applications of computerized adaptive testing (Research Report 77-1). Minneapolis MN: University of Minnesota, Department of Psychology, Psychometric Methods Program. %G eng %0 Generic %D 1977 %T Applications of computerized adaptive testing (Research Report 77-1) %A Weiss, D. J. %C Minneapolis MN: University of Minnesota, Department of Psychology, Psychometric Methods Program, Computerized Adaptive Testing Laboratory %G eng %0 Conference Proceedings %B 1977 Computerized Adaptive Testing Conference %D 1977 %T Applications of sequential testing procedures to performance testing %A Epstein, K. I. %A Knerr, C. S. %B 1977 Computerized Adaptive Testing Conference %I University of Minnesota %C Minneapolis, MN. USA %G eng %0 Generic %D 1976 %T Adaptive mental testing: The state of the art (Technical Report 423) %A J. R. McBride %C Washington DC: U.S. Army Research Institute for the Social and Behavioral Sciences. %G eng %0 Journal Article %J Programmed Learning and Educational Technology %D 1976 %T Adaptive testing: A Bayesian procedure for the efficient measurement of ability %A Wood, R. %B Programmed Learning and Educational Technology %V 13(2) %P 36 %G eng %0 Book Section %D 1976 %T Adaptive testing research at Minnesota: Overview, recent results, and future directions %A Weiss, D. J. %C C. L. Clark (Ed.), Proceedings of the First Conference on Computerized Adaptive Testing (pp. 24-35). Washington DC: United States Civil Service Commission. %G eng %0 Book Section %D 1976 %T Adaptive testing research at Minnesota: Some properties of a Bayesian sequential adaptive mental testing strategy %A J. R. McBride %C C. K. Clark (Ed.), Proceedings of the First Conference on Computerized Adaptive Testing (pp. 36-53). Washington DC: U.S. Government Printing Office. %G eng %0 Journal Article %J British Journal of Mathematical and Statistical Psychology %D 1974 %T An application of latent trait mental test theory %A Jensema, C J %B British Journal of Mathematical and Statistical Psychology %V 27 %P 29-48 %G eng %0 Conference Paper %B Paper presented at the annual meeting of the American Educational Research Association. %D 1974 %T An application of the Rasch simple logistic model to tailored testing %A Reckase, M. D. %B Paper presented at the annual meeting of the American Educational Research Association. %C St. Loius MO %G eng %0 Generic %D 1973 %T Ability measurement: Conventional or adaptive? (Research Report 73-1) %A Weiss, D. J. %A Betz, N. E. %C Minneapolis: University of Minnesota, Department of Psychology, Psychometric Methods Program %G eng %0 Generic %D 1972 %T An application of latent trait mental test theory to the Washington Pre-College Testing Battery %A Jensema, C J %C Unpublished doctoral dissertation, University of Washington %G eng %0 Generic %D 1971 %T The application of item generators for individualizing mathematics testing and instruction (Report 1971/14) %A Ferguson, R. L. %A Hsu, T. %C Pittsburgh PA: University of Pittsburgh Learning Research and Development Center %G eng %0 Journal Article %J Proceedings of the Annual Convention of the American Psychological Association %D 1970 %T Adaptive testing of cognitive skills %A Wargo, M. J. %B Proceedings of the Annual Convention of the American Psychological Association %V 5 (part 1) %P 167-168 %G eng %0 Journal Article %J American Psychologist %D 1969 %T Automation of psychological testing %A Elwood, D. L. %B American Psychologist %V 24 %P 287-289 %G eng %0 Journal Article %J Journal of Psychology %D 1965 %T Adaptive testing in an older population %A Greenwood, D. I. %A Taylor, C. %B Journal of Psychology %V 60 %P 193-198 %G eng %0 Book %D 1961 %T An analysis of the application of utility theory to the development of two-stage testing models %A Rosenbach, J. H. %C Unpublished doctoral dissertation, University of Buffalo %G eng %0 Journal Article %J Journal of the American Statistical Association %D 1946 %T An application of sequential sampling to testing students %A Cowden, D. J. %B Journal of the American Statistical Association %V 41 %P 547-556 %G eng