Item Analysis of an English Summative Test: A Classical Test Theory Approach in Indonesian Junior High School Context

Authors

  • Rahmawaty Ahmad Universitas Negeri Gorontalo, Indonesia
  • Rasuna Rasid Talib Universitas Negeri Gorontalo, Indonesia
  • Magvirah El Walidayni Kau Universitas Negeri Gorontalo, Indonesia

DOI:

https://doi.org/10.24256/ideas.v14i1.10243

Keywords:

classical test theory, discriminating power, item analysis, summative test, validity

Abstract

Assessment plays a crucial role in evaluating students’ learning outcomes; however, the quality of summative test instruments often remains questionable due to the limited use of systematic evaluation procedures. This study aims to examine the quality of English summative test items administered to eighth-grade students at SMP Negeri 1 Suwawa Timur using Classical Test Theory (CTT). A descriptive quantitative design was employed, analyzing 40 multiple-choice items completed by 56 students using Anates V4 software. The analysis focused on item validity, difficulty level, discriminating power, and distractor effectiveness. The results revealed that only 55% of the items were valid, while 45% were invalid. Most items (62.5%) were of moderate difficulty, yet no easy items were identified, indicating an imbalance. Furthermore, a substantial proportion of items demonstrated weak discriminating power, and several distractors were ineffective. These findings suggest that although some test items meet acceptable standards, a considerable number require revision. This study highlights the importance of systematic item analysis in improving the validity, reliability, and overall quality of assessment instruments.

References

Arifin, Z. (2019). Evaluasi pembelajaran. Remaja Rosdakarya.

Arikunto, S. (2018). Dasar-dasar evaluasi pendidikan (3rd ed.). Bumi Aksara.

Atmowardoyo, H. (2018). Research methods in TEFL studies: Descriptive research, case study, error analysis, and R&D. Journal of Language Teaching and Research, 9(1), 197–204.

Bani, M., & Masruddin, M. (2021). Development of Android-based harmonic oscillation pocket book for senior high school students. JOTSE: Journal of Technology and Science Education, 11(1), 93-103.

Creswell, J. W., & Creswell, J. D. (2019). Research design: Qualitative, quantitative, and mixed methods approach (5th ed.). Sage Publications.

Fiska, N., et al. (2021). Analysis of item validity in educational assessment. Journal of Educational Measurement, 12(2), 45–56.

Hartati, S., & Yogi, A. (2019). Item analysis in educational evaluation. Indonesian Journal of Education, 8(1), 67–75.

Himawan, R., et al. (2024). Distractor effectiveness in multiple-choice tests. Journal of Educational Assessment, 15(1), 23–35.

Mistiani, R. (2020). Classical test theory in educational measurement. Educational Research Journal, 10(2), 89–98.

Muluki, A. (2020). Constructing effective distractors in multiple-choice tests. Language Testing Journal, 7(1), 55–63.

Novriyanti, D., & Arthur, R. (2024). Discriminating power in educational testing. Assessment in Education Journal, 14(2), 101–115.

Puspitaningsih, R., et al. (2019). Validity in educational assessment. Journal of Educational Studies, 11(3), 145–156.

Ratnawulan, E. (2014). Evaluasi pembelajaran. Pustaka Setia.

Rustan, E. (2025). Developing a Web-Based Mobile Game to Enhance Students’ Motivation in learning English Vocabulary. IDEAS: Journal on English Language Teaching and Learning, Linguistics and Literature, 13(1), 46-73.

Sugiyono. (2013). Metode penelitian pendidikan. Alfabeta.

Suek, A. (2021). The importance of item analysis in improving test quality. Journal of Educational Evaluation, 9(2), 88–96.

Downloads

Published

2026-05-06

Citation Check