Published Online: 21 July 2017
AIP Conference Proceedings 1863, 550002 (2017); https://doi.org/10.1063/1.4992683
more...View Affiliations
The Item Response Theory (IRT) has become one of the most popular scoring frameworks for measurement data, frequently used in computerized adaptive testing, cognitively diagnostic assessment and test equating. According to Andrade et al. (2000), IRT can be defined as a set of mathematical models (Item Response Models – IRM) constructed to represent the probability of an individual giving the right answer to an item of a particular test. The number of Item Responsible Models available to measurement analysis has increased considerably in the last fifteen years due to increasing computer power and due to a demand for accuracy and more meaningful inferences grounded in complex data. The developments in modeling with Item Response Theory were related with developments in estimation theory, most remarkably Bayesian estimation with Markov chain Monte Carlo algorithms (Patz & Junker, 1999). The popularity of Item Response Theory has also implied numerous overviews in books and journals, and many connections between IRT and other statistical estimation procedures, such as factor analysis and structural equation modeling, have been made repeatedly (Van der Lindem & Hambleton, 1997). As stated before the Item Response Theory covers a variety of measurement models, ranging from basic one-dimensional models for dichotomously and polytomously scored items and their multidimensional analogues to models that incorporate information about cognitive sub-processes which influence the overall item response process. The aim of this work is to introduce the main concepts associated with one-dimensional models of Item Response Theory, to specify the logistic models with one, two and three parameters, to discuss some properties of these models and to present the main estimation procedures.
  1. 1. Andrade, D.F., Tavares, H.R., & Valle, R.C. Teoria da resposta ao Item: Conceitos e Aplicações. SINAPE. Associação Brasileira de Estatística, 2000. Google Scholar
  2. 2. Andrich, D. A rating formulation for ordered response categories. Psychometrika,43, pp. 561–573,1978. Google Scholar
  3. 3. Baker, F.B. The Basics of Item Response Theory (2nd edition). ERIC Clearinghouse on Assessment and Evaluation, USA, 2001. Google Scholar
  4. 4. Birnbaum, A. Some Latent Trait Models and Their Use in Inferring an Examinee’s Ability. In F.M. Lord & M.R. Novick. Statistical Theories of Mental Test Scores. Reading, MA: Addison-Wesley, 1968. Google Scholar
  5. 5. Bock, R. D. Estimating item parameters and latent ability when responses are scored in two or more nominal categories. Psychometrika, 37, pp. 29–51, 1972. https://doi.org/10.1007/BF02291411, Google ScholarCrossref
  6. 6. Bock, R. D. & Zimowski, M. F. Multiple group IRT. In Van der Linden, W.J. & Hambleton, R.K. (Eds.) Handbook of Modern Item Response Theory. New-York: Springer-Verlag, pp. 433–488,1997. Google ScholarCrossref
  7. 7. Edelen, M.O. & Reeve, B.B. Applying item response theory (IRT) modeling to questionnaire development, evaluation, and refinement Qual Life Res 16: 5, 2007. https://doi.org/10.1007/s11136-007-9198-0, Google ScholarCrossref
  8. 8. Hambleton, R.K., Swaminathan, H., & Rogers, H.J. Fundamentals of Item Response Theory. Sage Publications Inc, 1991. Google Scholar
  9. 9. Hays, R.D., Morales, L.S., Reise, S.P. Response Theory and Health Outcomes Measurement in the 21st Century. Med Care. Sep; 38(9 Suppl): II28–II42, 2000. Google Scholar
  10. 10. Lord, F. M. A theory of test scores (No. 7). Psychometric Monograph, 1952. Google Scholar
  11. 11. Masters, G. N. A Rasch model for partial credit scoring. Psychometrika, 47, pp. 149–174, 1982. https://doi.org/10.1007/BF02296272, Google ScholarCrossref
  12. 12. Muraki, E. A generalized partial credit model: Application of EM algorithm. Applied Psychological Measurement, 16, pp. 159–176, 1992. https://doi.org/10.1177/014662169201600206, Google ScholarCrossref
  13. 13. Patz, R.J. & Jumker, B.W. A straightforward approach to Markov chain Monte Carlo methods for item response models. Journal of Educational and Behavioral Statistics, 24, pp. 146–178, 1999. https://doi.org/10.2307/1165199, Google ScholarCrossref
  14. 14. Rasch, G. Probabilistic Models for Some Intelligence and Attainment Tests. Copenhagen: Danish Institute for Educational Research, 1960. Google Scholar
  15. 15. Reise, SP and Revicki, D.A. Handbook of Item Response Theory Modeling. Applications to Typical Performance Assessment. Taylor & Francis, 2015. Google Scholar
  16. 16. Reise SP and Waller NG. Item response theory and clinical measurement. Annu Rev Clin Psychol. 5:27–48, 2009. https://doi.org/10.1146/annurev.clinpsy.032408.153553, Google ScholarCrossref
  17. 17. Rupp, A.A. Item Response Modelling with BILOG_MG and MULTILOG for Windows. International Journal of Testing, 3(4), pp. 365–384, 2003. https://doi.org/10.1207/S15327574IJT0304_5, Google ScholarCrossref
  18. 18. Samejima, F. A. Estimation of latent ability using a response pattern of graded scores. Psychometric Monograph, 17, 1969. Google Scholar
  19. 19. Van der Linden, W.J. & Hambleton, R.K. Handbook of Modern Item Response Theory. New-York: Springer-Verlag, 1997. Google ScholarCrossref
  20. 20. Wright, B. D. Sample-free test calibration and person measurement. Proceedings of the 1967 Invitational Conference on Testing Problems. Princeton, N. J.: Educational Testing Service, 1968. Google Scholar
  21. 21. Zimowski, M. F., Muraki, E., Mislevy, R. J. & Bock, R.D. BILOG-MG: Multiple-Group IRT Analysis and Test Maintenance for BinaryItems. Chicago: Scientific Software, Inc., 1996. Google Scholar
  22. Published by AIP Publishing.