310
v1v2 (latest)

Enhancing Arabic Automated Essay Scoring with Synthetic Data and Error Injection

Workshop on Innovative Use of NLP for Building Educational Applications (UNBEA), 2025
Main:8 Pages
6 Figures
Bibliography:2 Pages
6 Tables
Appendix:1 Pages
Abstract

Automated Essay Scoring (AES) plays a crucial role in assessing language learners' writing quality, reducing grading workload, and providing real-time feedback. The lack of annotated essay datasets inhibits the development of Arabic AES systems. This paper leverages Large Language Models (LLMs) and Transformer models to generate synthetic Arabic essays for AES. We prompt an LLM to generate essays across the Common European Framework of Reference (CEFR) proficiency levels and introduce and compare two approaches to error injection. We create a dataset of 3,040 annotated essays with errors injected using our two methods. Additionally, we develop a BERT-based Arabic AES system calibrated to CEFR levels. Our experimental results demonstrate the effectiveness of our synthetic dataset in improving Arabic AES performance. We make our code and data publicly available.

View on arXiv
Comments on this paper