Projects per year
Abstract
Automatic readability assessment (ARA) aims to determine the cognitive load of a reader to comprehend a given text. ARA research has been mostly conducted at the text level, with numerous studies showing strong performance of neural and hybrid models. ARA at the sentence level, however, has received less attention, even though many applications in natural language processing (NLP) require assessment of the difficulty of individual sentences. This article compares the performance of neural models, hybrid models and large language models (LLMs) for sentence-level ARA, making three main contributions. First, we construct the first Chinese sentence-level ARA datasets, with nearly 70K sentences, to facilitate evaluation on Chinese. Second, we present the first experimental results on applying LLMs to sentence-level ARA. Finally, while previous work focused mostly on English data, we show that hybrid models outperform traditional classifiers, neural models, and LLMs in both English and Chinese data. The best hybrid model obtained state-of-the-art results on the Wall Street Journal dataset, surpassing the previous best result by almost 15% absolute. It also achieved competitive results on the CEFR-SP dataset. In detailed analyses, we identify the linguistic features that most significantly contributed to the performance of the hybrid model. We show that 10 linguistic features are correlated to readability across all datasets, and models trained on this reduced feature set achieve performance that rivals the full set. These results not only yield new insights into hybrid models for sentence-level ARA, but also set new benchmarks for future research in both English and Chinese.
© The Author(s), under exclusive licence to Springer Nature B.V. 2025
© The Author(s), under exclusive licence to Springer Nature B.V. 2025
Original language | English |
---|---|
Number of pages | 32 |
Journal | Language Resources and Evaluation |
DOIs | |
Publication status | Online published - 9 Feb 2025 |
Funding
This work was partly supported by the Language Fund from the Standing Committee on Language Education and Research (project EDB(LE)/P&R/EL/203/14); the General Research Fund (Project 11207320); and the CityU Strategic Research Grant (project 7005803).
Research Keywords
- Sentence readability assessment
- Automatic readability assessment
- Hybrid models
- Large language models
Fingerprint
Dive into the research topics of 'Automatic readability assessment for sentences: neural, hybrid and large language models'. Together they form a unique fingerprint.Projects
- 1 Finished
-
GRF: Semantic Modeling for Sentence-level Readability Assessment
LEE, J. S. Y. (Principal Investigator / Project Coordinator), LIU, M. (Co-Investigator) & Sun, W. (Co-Investigator)
1/01/21 → 17/06/24
Project: Research