Volume 34, Issue 3
ITEMS Module

An NCME Instructional Module on Item‐Fit Statistics for Item Response Theory Models

Allison J. Ames

University of North Carolina at Greensboro

Search for more papers by this author
Randall D. Penfield

University of North Carolina at Greensboro

Search for more papers by this author
First published: 30 April 2015
Citations: 17

Allison J. Ames, University of North Carolina at Greensboro, Educational Research Methodology, Department of Educational Research Methodology ERM Department, School of Education Building, Room #254, Greensboro, NC 27402; ajames@uncg.edu. Randall D. Penfield, University of North Carolina at Greensboro, Educational Research Methodology, 1300 Spring Garden St., Greensboro, NC 27412; rdpenfie@uncg.edu

Abstract

Drawing valid inferences from item response theory (IRT) models is contingent upon a good fit of the data to the model. Violations of model‐data fit have numerous consequences, limiting the usefulness and applicability of the model. This instructional module provides an overview of methods used for evaluating the fit of IRT models. Upon completing this module, the reader will have an understanding of traditional and Bayesian approaches for evaluating model‐data fit of IRT models, the relative advantages of each approach, and the software available to implement each method.

Number of times cited according to CrossRef: 17

  • Intellect is not that expensive: differential association of cultural and socio-economic factors with crystallized intelligence in a sample of Italian adolescents, Intelligence, 10.1016/j.intell.2020.101466, 81, (101466), (2020).
  • Applying a Multiple Comparison Control to IRT Item-fit Testing, Applied Measurement in Education, 10.1080/08957347.2020.1789138, (1-16), (2020).
  • Measuring economic competence of secondary school students in Germany, The Journal of Economic Education, 10.1080/00220485.2020.1804504, (1-16), (2020).
  • irtplay : An R Package for Online Item Calibration, Scoring, Evaluation of Model Fit, and Useful Functions for Unidimensional IRT , Applied Psychological Measurement, 10.1177/0146621620921247, (014662162092124), (2020).
  • An Investigation of Chi-Square and Entropy Based Methods of Item-Fit Using Item Level Contamination in Item Response Theory, Journal of Modern Applied Statistical Methods, 10.22237/jmasm/1604190480, 18, 2, (2-43), (2020).
  • Use of item response theory to develop a shortened version of the EORTC QLQ-BR23 scales, Scientific Reports, 10.1038/s41598-018-37965-x, 9, 1, (2019).
  • Item Response Theory, Assessment in Health Professions Education, 10.4324/9781315166902, (287-297), (2019).
  • Differential effects of state testing policies and school characteristics on social studies educators’ gate-keeping autonomy: A multilevel model, Theory & Research in Social Education, 10.1080/00933104.2019.1655508, (1-27), (2019).
  • A Bias-Corrected RMSD Item Fit Statistic: An Evaluation and Comparison to Alternatives, Journal of Educational and Behavioral Statistics, 10.3102/1076998619890566, (107699861989056), (2019).
  • Validation of the Italian Remote Associate Test, The Journal of Creative Behavior, 10.1002/jocb.345, 54, 1, (62-74), (2018).
  • Addressing model uncertainty in item response theory person scores through model averaging, Behaviormetrika, 10.1007/s41237-018-0052-1, 45, 2, (495-503), (2018).
  • Vignettentest zur Erfassung des fachdidaktischen Wissens im Leseunterricht bei angehenden Lehrkräften, Diagnostica, 10.1026/0012-1924/a000188, 64, 1, (2-13), (2018).
  • Evaluating PISA scales across Chinese economies, Asia Pacific Journal of Education, 10.1080/02188791.2018.1491388, (1-20), (2018).
  • The Test of Financial Literacy : Development and measurement characteristics , The Journal of Economic Education, 10.1080/00220485.2017.1285739, 48, 2, (113-122), (2017).
  • Practical Significance of Item Misfit in Educational Assessments, Applied Psychological Measurement, 10.1177/0146621617692978, 41, 5, (388-400), (2017).
  • Madde Tepki Kuramı’na Dayalı Madde-Uyum İndekslerinin I.Tip Hata ve Güç Oranlarının İncelenmesi, Eğitimde ve Psikolojide Ölçme ve Değerlendirme Dergisi, 10.21031/epod.301529, (79-79), (2017).
  • Using the Bootstrap Method to Evaluate the Critical Range of Misfit for Polytomous Rasch Fit Statistics, Psychological Reports, 10.1177/0033294116649434, 118, 3, (937-956), (2016).

The full text of this article hosted at iucr.org is unavailable due to technical difficulties.