Abstract
In this paper, we propose a generalized expectation model selection (GEMS) algorithm for latent variable selection in multidimensional item response theory models which are commonly used for identifying the relationships between the latent traits and test items. Under some mild assumptions, we prove the numerical convergence of GEMS for model selection by minimizing the generalized information criteria of observed data in the presence of missing data. For latent variable selection in the multidimensional two-parameter logistic (M2PL) models, we present an efficient implementation of GEMS to minimize the Bayesian information criterion. To ensure parameter identifiability, the variances of all latent traits are assumed to be unity and each latent trait is required to have an item exclusively associated with it. The convergence of GEMS for the M2PL models is verified. Simulation studies show that GEMS is computationally more efficient than the expectation model selection (EMS) algorithm and the expectation maximization based L1-penalized method (EML1), and it yields better correct rate of latent variable selection and mean squared error of parameter estimates than the EMS and EML1. The GEMS algorithm is illustrated by analyzing a real dataset related to the Eysenck Personality Questionnaire.
| Original language | English |
|---|---|
| Article number | 49 |
| Pages (from-to) | 1-15 |
| Number of pages | 16 |
| Journal | Statistics and Computing |
| Volume | 34 |
| Issue number | 1 |
| Early online date | 25 Nov 2024 |
| DOIs | |
| Publication status | E-pub ahead of print - 25 Nov 2024 |
Keywords
- Bayesian information criterion
- Exploratory item factor analysis
- Generalized expectation model selection algorithm
- Multidimensional item response theory model