Artificial Intelligence in Assessment for Learning: Comparing Automated and Teacher Feedback in Primary and Lower Secondary Education
Beatrice Doria, University of Padua (Italy)
Giorgia Slaviero, University of Padua (Italy)
Valentina Grion, Online University Pegaso (Italy)
Abstract
The growing diffusion of generative artificial intelligence (AI) in school contexts has intensified pedagogical reflection on formative assessment and feedback processes, particularly within assessment for learning frameworks [1,2,3]. Recent studies highlight the potential of AI to enhance the timeliness, consistency and sustainability of feedback practices in educational settings [4,5], yet empirical evidence in compulsory education remains limited, especially in contexts characterised by teacher supervision and criterion-referenced assessment. This study investigates the effectiveness, clarity and pedagogical sustainability of AI-generated feedback compared with teacher-provided feedback in primary and lower secondary education. The research was conducted during the 2024/2025 school year in two comprehensive schools in Northern Italy and involved 100 students across seven classes and five teachers from different subject areas. Adopting a mixed-method empirical–comparative design, the study integrated quantitative analyses of learning outcomes with survey data on students’ perceptions. Written curricular tasks were assessed using shared analytic rubrics, and feedback was generated either by teachers or by a generative AI system (GPT-4o), guided by carefully designed prompts aligned with the same assessment criteria [6]. After receiving feedback, students revised their work, enabling pre–post comparisons of performance. Results show statistically significant improvements in both groups, with higher average learning gains associated with AI-generated feedback in five out of seven classes. High levels of agreement between human and automated assessments were observed (ICC > .92), indicating strong consistency and reliability. In line with research on formative feedback and feedback literacy [7], students reported high clarity and usefulness of feedback regardless of its source and demonstrated strong engagement in revision processes. Overall, the findings suggest that AI-generated feedback, when embedded within a pedagogically intentional and teacher-supervised framework, can effectively support assessment for learning by enhancing timeliness, coherence and inclusivity, while complementing rather than replacing teachers’ professional judgment.
|
Keywords |
Artificial intelligence, assessment for learning, formative feedback, primary education, student assessment |
|
REFERENCES |
[1] Sadler, D. R. (1989). Formative assessment and the design of instructional systems. Instructional Science, 18(2), 119–144. [2] Black, P., & Wiliam, D. (2009). Developing the theory of formative assessment. Educational Assessment, Evaluation and Accountability, 21(1), 5–31. [3] Hattie, J., & Timperley, H. (2007). The Power of Feedback. Review of Educational Research, 77(1), 81–112. [4] Adiguzel, T., Kaya, M. H., & Cansu, F. K. (2023). Revolutionizing education with AI: Exploring the transformative potential of ChatGPT. Contemporary Educational Technology, 15(3), ep429. https://doi.org/10.30935/cedtech/13152 [5] Baidoo-Anu, D., & Owusu Ansah, L. (2023). Education in the era of generative artificial intelligence (AI): Opportunities, challenges, and implications. Journal of Artificial Intelligence, 7(1), Article 1337500. [6] Nicol, D., & Macfarlane‐Dick, D. (2006). Formative assessment and self‐regulated learning: A model and seven principles of good feedback practice. Studies in Higher Education, 31(2), 199–218. [7] Carless, D., & Boud, D. (2018). The development of student feedback literacy: enabling uptake of feedback. Assessment & Evaluation in Higher Education, 43(8), 1315–1325. |
The Future of Education




























