TY - JOUR T1 - Mutual Information Item Selection in Adaptive Classification Testing JF - Educational and Psychological Measurement Y1 - 2007 A1 - Weissman, Alexander AB -

A general approach for item selection in adaptive multiple-category classification tests is provided. The approach uses mutual information (MI), a special case of the Kullback-Leibler distance, or relative entropy. MI works efficiently with the sequential probability ratio test and alleviates the difficulties encountered with using other local- and global-information measures in the multiple-category classification setting. Results from simulation studies using three item selection methods, Fisher information (FI), posterior-weighted FI (FIP), and MI, are provided for an adaptive four-category classification test. Both across and within the four classification categories, it is shown that in general, MI item selection classifies the highest proportion of examinees correctly and yields the shortest test lengths. The next best performance is observed for FIP item selection, followed by FI.

VL - 67 UR - http://epm.sagepub.com/content/67/1/41.abstract ER -