Assessing the psychometric properties of AI-generated multiple-choice exams in a psychology subject
More Detail
1 Faculty of Psychology Program, Social Sciences Department, College of Arts and Sciences, Southern Luzon State University, Lucban, Quezon, Philippines
* Corresponding Author
Journal of Pedagogical Sociology and Psychology, 2025 - Volume 7 Issue 3, pp. 18-34
https://doi.org/10.33902/jpsp.202536891
Article Type: Research Article
Published Online: 13 Sep 2025
Views: 7 | Downloads: 2
This study assessed the psychometric properties of AI-generated multiple-choice questions in undergraduate psychology education, specifically focusing on an Experimental Psychology course. Using a mixed-methods approach, we evaluated 80 multiple-choice questions created by ChatGPT-4 through expert content validation, administration to undergraduate psychology students, and comprehensive psychometric analysis. Results indicated that AI-generated items demonstrated reasonable content validity with strongest ratings for clarity and weakest for distractor quality. The assessment showed acceptable internal consistency and moderate test-retest reliability. Item difficulty analysis revealed a slight tendency toward easier items, with 58.75% of questions in the medium difficulty range. Discrimination indices were generally good, though 11.25% of items showed poor discrimination. Cognitive level analysis identified a significant imbalance in the distribution of items across Bloom's taxonomy, with 73.75% targeting lower-order thinking skills (remembering, understanding) and only 8.75% addressing higher-order skills (analyzing, evaluating). These findings suggest that while AI can generate psychometrically sound multiple-choice questions for psychology assessment, significant limitations exist in assessing higher-order cognitive skills. The results support a hybrid approach that leverages AI for generating fundamental knowledge assessment items while preserving human expertise for higher cognitive domain evaluation. This study contributes to the emerging understanding of AI applications in psychology education by providing domain-specific insights into the capabilities and limitations of AI-generated assessment tools.
In-text citation: (Baudin, 2025)
Reference: Baudin, J. S. P. (2025). Assessing the psychometric properties of AI-generated multiple-choice exams in a psychology subject.
Journal of Pedagogical Sociology and Psychology, 7(3), 18-34.
https://doi.org/10.33902/jpsp.202536891
In-text citation: (1), (2), (3), etc.
Reference: Baudin JSP. Assessing the psychometric properties of AI-generated multiple-choice exams in a psychology subject.
Journal of Pedagogical Sociology and Psychology. 2025;7(3), 18-34.
https://doi.org/10.33902/jpsp.202536891
In-text citation: (1), (2), (3), etc.
Reference: Baudin JSP. Assessing the psychometric properties of AI-generated multiple-choice exams in a psychology subject. Journal of Pedagogical Sociology and Psychology. 2025;7(3):18-34.
https://doi.org/10.33902/jpsp.202536891
In-text citation: (Baudin, 2025)
Reference: Baudin, Jomar Saif P.. "Assessing the psychometric properties of AI-generated multiple-choice exams in a psychology subject".
Journal of Pedagogical Sociology and Psychology 2025 7 no. 3 (2025): 18-34.
https://doi.org/10.33902/jpsp.202536891
In-text citation: (Baudin, 2025)
Reference: Baudin, J. S. P. (2025). Assessing the psychometric properties of AI-generated multiple-choice exams in a psychology subject.
Journal of Pedagogical Sociology and Psychology, 7(3), pp. 18-34.
https://doi.org/10.33902/jpsp.202536891
In-text citation: (Baudin, 2025)
Reference: Baudin, Jomar Saif P. "Assessing the psychometric properties of AI-generated multiple-choice exams in a psychology subject".
Journal of Pedagogical Sociology and Psychology, vol. 7, no. 3, 2025, pp. 18-34.
https://doi.org/10.33902/jpsp.202536891
This is an open access article distributed under the
Creative Commons Attribution License which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.