Item

Intent Matters: Enhancing AI Tutoring with Fine-Grained Pedagogical Intent Annotation

Petukhova, Kseniia
Kochmar, Ekaterina
Supervisor
Department
Natural Language Processing
Embargo End Date
Type
Conference proceeding
Date
2025
License
Language
English
Collections
Research Projects
Organizational Units
Journal Issue
Abstract
Large language models (LLMs) hold great promise for educational applications, particularly in intelligent tutoring systems. However, effective tutoring requires alignment with pedagogical strategies – something current LLMs lack without task-specific adaptation. In this work, we explore whether fine-grained annotation of teacher intents can improve the quality of LLM-generated tutoring responses. We focus on MathDial, a dialog dataset for math instruction, and apply an automated annotation framework to re-annotate a portion of the dataset using a detailed taxonomy of eleven pedagogical intents. We then fine-tune an LLM using these new annotations and compare its performance to models trained on the original four-category taxonomy. Both automatic and qualitative evaluations show that the fine-grained model produces more pedagogically aligned and effective responses. Our findings highlight the value of intent specificity for controlled text generation in educational settings, and we release our annotated data and code to facilitate further research.
Citation
K. Petukhova and E. Kochmar, “Intent Matters: Enhancing AI Tutoring with Fine-Grained Pedagogical Intent Annotation,” pp. 860–872, Aug. 2025, doi: 10.18653/V1/2025.BEA-1.63
Source
Proceedings of the 20th Workshop on Innovative Use of NLP for Building Educational Applications
Conference
20th Workshop on Innovative Use of NLP for Building Educational Applications
Keywords
Subjects
Source
20th Workshop on Innovative Use of NLP for Building Educational Applications
Publisher
Association for Computational Linguistics
Full-text link