Enhancing the Automatic Classification of Metadiscourse in Low-Proficiency Learners' Spoken and Written English Texts Using XLNet

Open Access
Authors
Publication date 2025
Host editors
  • Michael Strube
  • Chloe Braud
  • Christian Hardmeier
  • Junyi Jessy Li
  • Sharid Loaiciga
  • Amir Zeldes
  • Chuyuan Li
Book title The 6th Workshop on Computational Approaches to Discourse (CODI 2025) : proceedings of the workshop
Book subtitle CODI 2025 : November 9, 2025
ISBN (electronic)
  • 9798891763432
Event 6th Workshop on computational approaches to discourse, context, and document-level inferences
Pages (from-to) 27-41
Publisher Kerrville, TX: Association for Computational Linguistics
Organisations
  • Interfacultary Research - Institute for Logic, Language and Computation (ILLC)
  • Faculty of Humanities (FGw) - Amsterdam Institute for Humanities Research (AIHR) - Amsterdam Center for Language and Communication (ACLC)
  • Faculty of Humanities (FGw) - Amsterdam Institute for Humanities Research (AIHR)
Abstract
This study aims to enhance the automatic identification and classification of metadiscourse markers in English texts, evaluating various large language models for the purpose. Metadiscourse is a commonly used rhetorical strategy in both written and spoken language to guide addressees through discourse. Due to its linguistic complexity and dependency on the context, automated metadiscourse classification is challenging. With a hypothesis that LLMs may handle complicated tasks more effectively than supervised machine learning approaches, we tune and evaluate seven encoder language models on the task using a dataset totalling 575,541 tokens and annotated with 24 labels. The results show a clear improvement over supervised machine learning approaches as well as an untuned Llama3.3-70B-Instruct baseline, with XLNet-large achieving an accuracy and F1-score of 0.91 and 0.93, respectively. However, four less frequent categories record F-scores below 0.5, highlighting the need for more balanced data representation.
Document type Conference contribution
Language English
Published at https://doi.org/10.18653/v1/2025.codi-1.3
Downloads
2025.codi-1.3 (Final published version)
Permalink to this page
Back