SPECIAL ISSUE: ARTIFICIAL INTELLIGENCE AND MACHINE LEARNING TECHNOLOGIES
Evaluating the effectiveness of large language models in identifying communicatively significant errors in written works of students learning russian as a foreign language
Abstract:
This article examines the capability of contemporary large language models (LLMs), such as GPT-5 and DeepSeek-R1, to identify and classify communicative errors in written works of students learning Russian as a foreign language (RFL). While existing tools primarily focus on formal errors, this study emphasizes the communicative aspect, evaluating the extent to which an error disrupts comprehension (communicatively significant errors) or merely affects linguistic norms (communicatively insignificant errors). To this end, a corpus of written works by B2-level students (TORFL-2) was created and annotated by experts, and a multi-stage pipeline for testing LLMs was developed, incorporating structured prompting and heuristic voting methods to enhance result reliability. The experiment revealed that while models can localize errors with certain accuracy, they experience considerable difficulties in their proper communicative classification. The models tend to systematically underestimate the degree of error impact on comprehension, confuse error types, and encounter challenges in identifying multiple errors within a single fragment. The study demonstrates both the potential and current limitations of LLMs as tools for automated communicatively-oriented feedback in educational technologies.
Keywords:LLM, natural language processing, RFL, TORFL, communicatively significant errors, communicatively insignificant errors, automated text assessment, error classification.