ERIC Number: EJ1405358
Record Type: Journal
Publication Date: 2024
Pages: 12
Abstractor: As Provided
ISBN: N/A
ISSN: N/A
EISSN: EISSN-1939-1382
Available Date: N/A
Enhancing Educational Dialogue Act Classification with Discourse Context and Sample Informativeness
IEEE Transactions on Learning Technologies, v17 p258-269 2024
Automating the classification of instructional strategies from a large-scale online tutorial dialogue corpus is indispensable to the design of dialogue-based intelligent tutoring systems. Despite many existing studies employing supervised machine learning (ML) models to automate the classification process, they concluded that building a well-performed ML model is nontrivial since the sample size is commonly limited for model training. Based on this fact, we posited that the model performance can be further optimized by the design of input features and the selection of informative instances for model training. By reviewing the existing dialogue-related research, we found that contextual information (i.e., the content of preceding utterances) was a potential feature in many text classification tasks but underexplored in the classification of an instructional strategy. In addition, training a well-performed ML model (e.g., deep neural network) to recognize instructional strategies usually requires a large amount of manually annotated data, which are labor intensive. To alleviate the demand of manual annotation, researchers proposed the use of statistical active learning methods to select informative instances, but this method was rarely used for recognizing the instructional strategy in online tutoring dialogue. Therefore, our study aimed to investigate the improvement of automating the classification of dialogue acts--a popular approach to the detection of instructional strategies--from two perspectives. First, we explored whether and to what extent the incorporation of contextual information can boost a model's prediction performance. Then, we investigated the extent to which the recent active learning methods can alleviate the labor-intensive issues in training the ML model for recognizing the instructional strategies. Our study showed that: 1) the ML models trained on the features that included the contextual information achieved better performance than that of the models excluding it; 2) the effectiveness of the contextual information decayed after the ML model achieved an optimal performance; and 3) compared with the random baseline, active learning methods can select informative samples from the training dataset to train ML models, which can alleviate the labor-intensive issues.
Descriptors: Classification, Dialogs (Language), Teaching Methods, Computer Assisted Instruction, Models, Artificial Intelligence, Active Learning, Context Effect, Prediction, Comparative Analysis, Computational Linguistics, Natural Language Processing
Institute of Electrical and Electronics Engineers, Inc. 445 Hoes Lane, Piscataway, NJ 08854. Tel: 732-981-0060; Web site: http://ieeexplore.ieee.org/xpl/RecentIssue.jsp?punumber=4620076
Publication Type: Journal Articles; Reports - Research
Education Level: N/A
Audience: N/A
Language: English
Sponsor: N/A
Authoring Institution: N/A
Grant or Contract Numbers: N/A
Author Affiliations: N/A