An NCME Instructional Module on Item‐Fit Statistics for Item Response Theory Models
Allison J. Ames, University of North Carolina at Greensboro, Educational Research Methodology, Department of Educational Research Methodology ERM Department, School of Education Building, Room #254, Greensboro, NC 27402; ajames@uncg.edu. Randall D. Penfield, University of North Carolina at Greensboro, Educational Research Methodology, 1300 Spring Garden St., Greensboro, NC 27412; rdpenfie@uncg.edu
Abstract
Drawing valid inferences from item response theory (IRT) models is contingent upon a good fit of the data to the model. Violations of model‐data fit have numerous consequences, limiting the usefulness and applicability of the model. This instructional module provides an overview of methods used for evaluating the fit of IRT models. Upon completing this module, the reader will have an understanding of traditional and Bayesian approaches for evaluating model‐data fit of IRT models, the relative advantages of each approach, and the software available to implement each method.
Citing Literature
Number of times cited according to CrossRef: 17
- Federica Conte, Giulio Costantini, Luca Rinaldi, Tiziano Gerosa, Luisa Girelli, Intellect is not that expensive: differential association of cultural and socio-economic factors with crystallized intelligence in a sample of Italian adolescents, Intelligence, 10.1016/j.intell.2020.101466, 81, (101466), (2020).
- Derek Sauder, Christine DeMars, Applying a Multiple Comparison Control to IRT Item-fit Testing, Applied Measurement in Education, 10.1080/08957347.2020.1789138, (1-16), (2020).
- Tim Kaiser, Luis Oberrauch, Günther Seeber, Measuring economic competence of secondary school students in Germany, The Journal of Economic Education, 10.1080/00220485.2020.1804504, (1-16), (2020).
- Hwanggyu Lim, Craig S. Wells, irtplay : An R Package for Online Item Calibration, Scoring, Evaluation of Model Fit, and Useful Functions for Unidimensional IRT , Applied Psychological Measurement, 10.1177/0146621620921247, (014662162092124), (2020).
- William R. Dardick, Brandi A. Weiss, An Investigation of Chi-Square and Entropy Based Methods of Item-Fit Using Item Level Contamination in Item Response Theory, Journal of Modern Applied Statistical Methods, 10.22237/jmasm/1604190480, 18, 2, (2-43), (2020).
- Juan Xia, Zheng Tang, Peng Wu, Jiwei Wang, Jinming Yu, Use of item response theory to develop a shortened version of the EORTC QLQ-BR23 scales, Scientific Reports, 10.1038/s41598-018-37965-x, 9, 1, (2019).
- Yoon Soo Park, Item Response Theory, Assessment in Health Professions Education, 10.4324/9781315166902, (287-297), (2019).
- Hyeri Hong, Gregory E. Hamot, Differential effects of state testing policies and school characteristics on social studies educators’ gate-keeping autonomy: A multilevel model, Theory & Research in Social Education, 10.1080/00933104.2019.1655508, (1-27), (2019).
- Carmen Köhler, Alexander Robitzsch, Johannes Hartig, A Bias-Corrected RMSD Item Fit Statistic: An Evaluation and Comparison to Alternatives, Journal of Educational and Behavioral Statistics, 10.3102/1076998619890566, (107699861989056), (2019).
- Carola Salvi, Giulio Costantini, Adriana Pace, Massimiliano Palmiero, Validation of the Italian Remote Associate Test, The Journal of Creative Behavior, 10.1002/jocb.345, 54, 1, (62-74), (2018).
- Jason D. Rights, Sonya K. Sterba, Sun-Joo Cho, Kristopher J. Preacher, Addressing model uncertainty in item response theory person scores through model averaging, Behaviormetrika, 10.1007/s41237-018-0052-1, 45, 2, (495-503), (2018).
- Juliane Rutsch, Tobias Dörfler, Vignettentest zur Erfassung des fachdidaktischen Wissens im Leseunterricht bei angehenden Lehrkräften, Diagnostica, 10.1026/0012-1924/a000188, 64, 1, (2-13), (2018).
- Yuan-Ling Liaw, Yuanyue Wu, David Rutkowski, Leslie Rutkowski, Evaluating PISA scales across Chinese economies, Asia Pacific Journal of Education, 10.1080/02188791.2018.1491388, (1-20), (2018).
- William B. Walstad, Ken Rebeck, The Test of Financial Literacy : Development and measurement characteristics , The Journal of Economic Education, 10.1080/00220485.2017.1285739, 48, 2, (113-122), (2017).
- Carmen Köhler, Johannes Hartig, Practical Significance of Item Misfit in Educational Assessments, Applied Psychological Measurement, 10.1177/0146621617692978, 41, 5, (388-400), (2017).
- Seçil ÖMÜR SÜNBÜL, Semih AŞİRET, Madde Tepki Kuramı’na Dayalı Madde-Uyum İndekslerinin I.Tip Hata ve Güç Oranlarının İncelenmesi, Eğitimde ve Psikolojide Ölçme ve Değerlendirme Dergisi, 10.21031/epod.301529, (79-79), (2017).
- Hyunsoo Seol, Using the Bootstrap Method to Evaluate the Critical Range of Misfit for Polytomous Rasch Fit Statistics, Psychological Reports, 10.1177/0033294116649434, 118, 3, (937-956), (2016).




