Hybrid Models for Sentence Readability Assessment

Research output: Chapters, Conference Papers, Creative and Literary WorksRGC 32 - Refereed conference paper (with host publication)peer-review

3 Scopus Citations
View graph of relations

Detail(s)

Original languageEnglish
Title of host publicationProceedings of the 18th Workshop on Innovative Use of NLP for Building Educational Applications (BEA 2023)
PublisherAssociation for Computational Linguistics
Pages448-454
Publication statusPublished - 13 Jul 2023

Publication series

NameProceedings of the Annual Meeting of the Association for Computational Linguistics
ISSN (Print)0736-587X

Conference

Title18th Workshop on Innovative Use of NLP for Building Educational Applications (BEA 2023)
LocationHybrid
PlaceCanada
CityToronto
Period13 July 2023

Abstract

Automatic readability assessment (ARA) predicts how difficult it is for the reader to understand a text. While ARA has traditionally been performed at the passage level, there has been increasing interest in ARA at the sentence level, given its applications in downstream tasks such as text simplification and language exercise generation. Recent research has suggested the effectiveness of hybrid approaches for ARA, but they have yet to be applied on the sentence level. We present the first study that compares neural and hybrid models for sentence-level ARA. We conducted experiments on graded sentences from the Wall Street Journal (WSJ) and a dataset derived from the OneStopEnglish corpus. Experimental results show that both neural and hybrid models outperform traditional classifiers trained on linguistic features. Hybrid models obtained the best accuracy on both datasets, surpassing the previous best result reported on the WSJ dataset by almost 13% absolute.

© 2023 Association for Computational Linguistics

Citation Format(s)

Hybrid Models for Sentence Readability Assessment. / Liu, Fengkai; Lee, John S. Y.
Proceedings of the 18th Workshop on Innovative Use of NLP for Building Educational Applications (BEA 2023). Association for Computational Linguistics, 2023. p. 448-454 (Proceedings of the Annual Meeting of the Association for Computational Linguistics).

Research output: Chapters, Conference Papers, Creative and Literary WorksRGC 32 - Refereed conference paper (with host publication)peer-review