Probing Natural Language Inference Models through Semantic Fragments

Kyle Richardson, Hai Hu, Lawrence S. Moss, Ashish Sabharwal

Research output: Chapters, Conference Papers, Creative and Literary WorksRGC 32 - Refereed conference paper (with host publication)peer-review

96 Citations (Scopus)

Abstract

Do state-of-the-art models for language understanding already have, or can they easily learn, abilities such as boolean coordination, quantification, conditionals, comparatives, and monotonicity reasoning (i.e., reasoning about word substitutions in sentential contexts)? While such phenomena are involved in natural language inference (NLI) and go beyond basic linguistic understanding, it is unclear the extent to which they are captured in existing NLI benchmarks and effectively learned by models. To investigate this, we propose the use of semantic fragments-systematically generated datasets that each target a different semantic phenomenon-for probing, and efficiently improving, such capabilities of linguistic models. This approach to creating challenge datasets allows direct control over the semantic diversity and complexity of the targeted linguistic phenomena, and results in a more precise characterization of a model's linguistic behavior. Our experiments, using a library of 8 such semantic fragments, reveal two remarkable findings: (a) State-of-the-art models, including BERT, that are pre-trained on existing NLI benchmark datasets perform poorly on these new fragments, even though the phenomena probed here are central to the NLI task; (b) On the other hand, with only a few minutes of additional fine-tuning-with a carefully selected learning rate and a novel variation of “inoculation”-a BERT-based model can master all of these logic and monotonicity fragments while retaining its performance on established NLI benchmarks. Copyright © 2020, Association for the Advancement of Artificial Intelligence (www.aaai.org). All rights reserved.
Original languageEnglish
Title of host publicationAAAI-20 / IAAI-20 / EAAI- 20 Proceedings
Subtitle of host publicationThe Thirty-Fourth AAAI Conference on Artificial Intelligence, The Thirty-Second Conference on Innovative Applications of Artificial Intelligence, The Tenth Symposium on Educational Advances in Artificial Intelligence
PublisherAAAI Press
Pages8713-8721
Number of pages9
ISBN (Electronic)9781713827368
ISBN (Print)9781577358350
DOIs
Publication statusPublished - 2020
Externally publishedYes
Event34th AAAI Conference on Artificial Intelligence, AAAI 2020 - New York, United States
Duration: 7 Feb 202012 Feb 2020

Publication series

NameProceedings of the AAAI Conference on Artificial Intelligence
Number5
Volume34
ISSN (Print)2159-5399
ISSN (Electronic)2374-3468

Conference

Conference34th AAAI Conference on Artificial Intelligence, AAAI 2020
PlaceUnited States
CityNew York
Period7/02/2012/02/20

Funding

Part of this work is supported by grant #586136 from the Simons Foundation. Hai Hu is supported by the China Scholarship Council.

Fingerprint

Dive into the research topics of 'Probing Natural Language Inference Models through Semantic Fragments'. Together they form a unique fingerprint.

Cite this