Mapping the Landscape of Artificial Intelligence (AI)-Powered Assessment: A Bibliometric Analysis of Scopus and Web of Science (WoS)

Authors

  • Nurul Ashikin Izhar School of Education Studies, Universiti Sains Malaysia, 11700 Gelugor, Pulau Pinang, Malaysia
  • Yahya Al-Dheleai Digital Buddy International B.V. Parallelweg 30, Den Bosch, The Netherlands

DOI:

https://doi.org/10.37134/ejoss.vol12.Sp.13.2026

Keywords:

artificial intelligence, AI-powered assessment, assessment , bibliometric analysis , BLR

Abstract

The surge of interest in Artificial Intelligence (AI) in higher education has led to rapid growth in research regarding its potential applications in assessment. This study analyzes publication trends, document types, and citation patterns related to AI in educational assessment, alongside the emergence of AI-powered literature search platforms. Data from Scopus and Web of Science (WoS) databases (2020–2024) were retrieved for analysis. A total of 42 publications were analyzed using VOSviewer for keyword mapping and cluster identification, while Harzing’s Publish or Perish was utilized for citation metrics. The results show a consistent increase in publications related to AI-based assessment, with articles being the primary format. Keyword analysis revealed dominant clusters centered on student perceptions and automated grading systems. This study provides an updated bibliometric landscape that guides researchers in identifying research gaps and emerging directions in AI assessment, while highlighting how AI-powered search tools can enhance systematic literature mapping.

Downloads

Download data is not yet available.

References

Bachman, L. F., & Palmer, A. S. (2010). Language assessment in practice: Developing language assessments and justifying their use in the real world. Oxford University Press.

Butler, Y. G. (2018). English language education among young learners in East Asia: A review of current research. Language Teaching, 51(1), 1–34. https://doi.org/10.1017/S0261444817000232

Council of Europe. (2020). Common European Framework of Reference for Languages: Learning, teaching, assessment – Companion volume. Council of Europe Publishing. https://rm.coe.int/cefr-companion-volume-with-new-descriptors-2020

DeVellis, R. F. (2017). Scale development: Theory and applications (4th ed.). Sage Publications.

Hashim, H., & Yunus, M. M. (2020). English language learning challenges in rural schools and strategies to improve proficiency. Creative Education, 11(6), 1045–1055. https://doi.org/10.4236/ce.2020.116078

Hashim, H., & Yunus, M. M. (2020). ESL teachers’ perceptions on the use of speaking assessment in Malaysian primary schools. International Journal of Learning, Teaching and Educational Research, 19(6), 1–15. https://doi.org/10.26803/ijlter.19.6.1

Leong, L. M., & Ahmadi, S. M. (2017). An analysis of factors influencing learners’ English speaking skill. International Journal of Research in English Education, 2(1), 34–41. https://doi.org/10.18869/acadpub.ijree.2.1.34

McHugh, M. L. (2012). Interrater reliability: The kappa statistic. Biochemia Medica, 22(3), 276–282. https://doi.org/10.11613/BM.2012.031

Mohammed, S. A., Rahman, M. S., & Yunus, M. M. (2021). Teachers’ readiness in implementing CEFR-aligned English curriculum in rural Malaysia. Asian Journal of Education and Social Studies, 17(4), 21–34. https://doi.org/10.9734/AJESS/2021/v17i430422

Mohammed, A. A., Yunus, M. M., & Hashim, H. (2021). Challenges in assessing speaking skills among Malaysian ESL learners. Asian Journal of University Education, 17(3), 1–11.

https://doi.org/10.24191/ajue.v17i3.14575

Mustapha, S. M., & Yahaya, A. (2013). Low English proficiency among Malaysian learners: Causes and strategies. English Language Teaching, 6(12), 155–163. https://doi.org/10.5539/elt.v6n12p155

Noor, N. M., Mohamed, A. R., & Rahman, N. A. (2020). Achieving CEFR benchmarks in Malaysian ESL primary schools: Challenges and implications. Asian EFL Journal, 24(3), 75–98.

Noor, N. M., Aman, I., Mustaffa, R., & Seong, T. K. (2020). Language proficiency and learning challenges among rural ESL learners in Malaysia. Issues in Educational Research, 30(2), 533–551.

North, B. (2014). Putting the Common European Framework of Reference to good use. Language Teaching, 47(2), 228–249. https://doi.org/10.1017/S0261444812000205

Polit, D. F., & Beck, C. T. (2006). The content validity index: Are you sure you know what’s being reported? Critique and recommendations. Research in Nursing & Health, 29(5), 489–497. https://doi.org/10.1002/nur.20147

Shang, Y. (2024). A meta-analysis of the reliability of second language proficiency assessments. Language Assessment Quarterly, 21(2), 145–169. https://doi.org/10.1080/15434303.2024.1187649

Stemler, S. E. (2004). A comparison of consensus, consistency, and measurement approaches to estimating interrater reliability. Practical Assessment, Research & Evaluation, 9(4), 1–19. https://doi.org/10.7275/96jp-xz07

Sureeyatanapas, P. (2024). Analysing marking reliability in English speaking proficiency tests: A marking consistency approach. Language Testing in Asia, 14(1), 65–79. https://doi.org/10.1186/s40468-023-00271-z

Tavakol, M., & Dennick, R. (2011). Making sense of Cronbach’s alpha. International Journal of Medical Education, 2, 53–55. https://doi.org/10.5116/ijme.4dfb.8dfd

Downloads

Published

2026-01-20

How to Cite

Izhar, N. A., & Al-Dheleai, Y. (2026). Mapping the Landscape of Artificial Intelligence (AI)-Powered Assessment: A Bibliometric Analysis of Scopus and Web of Science (WoS). EDUCATUM Journal of Social Sciences, 12(Special Issue), 107-113. https://doi.org/10.37134/ejoss.vol12.Sp.13.2026