Tensor Flow, etc.
Concept 1, etc.
psychpackage documentation for more information about these options.
Many educators design multiple-choice question examination. How do we know that these tests are valid and reliable? How can we improve upon the test by way of modifying, revising and deleting items based on student responses?
In a paper in the highly regarded Journal of Engineering Education, Jorion, et al (2016) developed "an analytical framework for evaluating the validity of concept inventory claims". We believe that we can use this framework to help educators design their multiple-choice tests as well, especially, if they are designed as the final mastery examination in a course. An open source software to analyze a multiple-choice question examination would be encouraging to educators who have minimal programming experience and promising to contributors who would enhance the program.
Garrick Aden-Buie is a doctoral candidate in Industrial and Management Systems Engineering at the University of South Florida. He is an avid R enthusiast and programmer. His research focus is on collecting, storing, processing, visualizing and learning from passive sensors networks in smart homes. He is also passionate about bringing together education, data science and interactive R tools to improve education outcomes in higher education.
Autar Kaw is a professor of mechanical engineering and Jerome Krivanek Distinguished Teacher at the University of South Florida. He is a recipient of the 2012 U.S. Professor of the Year Award from the Council for Advancement and Support of Education (CASE) and Carnegie Foundation for Advancement of Teaching. Professor Kaw's related main scholarly interests are in engineering education research, open courseware development, and the state and future of higher education. His education research has been funded by National Science Foundation since 2002.
Baker, F. B. (2001). The basics of item response theory (2nd ed.). ERIC Clearinghouse on Assessment; Evaluation. Retrieved from http://echo.edres.org:8080/irt/baker/
Bond, T. G., & Fox, C. M. (2007). Applying the rasch model: Fundamental measurement in the human sciences (1st ed.). Mahwah, N.J.: Lawrence Erlbaum Associates Publishers.
DiBello, L. V., Henson, R. A., & Stout, W. F. (2015). A family of generalized diagnostic classification models for multiple choice option-based scoring. Applied Psychological Measurement, 39(1), 62–79. https://doi.org/10.1177/0146621614561315
Haertel, E. H., & Lorie, W. A. (2004). Validating standards-based test score interpretations. Measurement: Interdisciplinary Research and Perspectives, 2(2), 61–103. https://doi.org/10.1207/s15366359mea0202_1
Jorion, N., Gane, B. D., James, K., Schroeder, L., DiBello, L. V., & Pellegrino, J. W. (2015). An analytic framework for evaluating the validity of concept inventory claims. Journal of Engineering Education, 104(4), 454–496. https://doi.org/10.1002/jee.20104
Revelle, W. (2017). Northwestern University; http://personality-project.org/r/book/.
Sleeper, R. (2011). Keep, toss or revise? Tips for post-exam item analysis. http://www.ttuhsc.edu/sop/administration/enhancement/documents/Sleeper_Handout.ppt (URL no longer valid).
Revelle, W. (2016). Psych: Procedures for psychological, psychometric, and personality research. Evanston, Illinois: Northwestern University; https://CRAN.R-project.org/package=psych. Retrieved from https://CRAN.R-project.org/package=psych
Rizopoulos, D. (2006). Ltm: An r package for latent variable modelling and item response theory analyses. Journal of Statistical Software, 17(5), 1–25. Retrieved from http://www.jstatsoft.org/v17/i05/
Fletcher, T. D. (2010). Psychometric: Applied psychometric theory. Retrieved from https://CRAN.R-project.org/package=psychometric
MCTestAnalysis was built in R using the following packages: