Do you have the right scissors? Tailoring pre-trained language models via Monte-Carlo methods

Ning Miao, Yuxuan Song, Hao Zhou, Lei Li

Research output: Chapters, Conference Papers, Creative and Literary WorksRGC 32 - Refereed conference paper (with host publication)peer-review

4 Citations (Scopus)

Abstract

It has been a common approach to pre-train a language model on a large corpus and fine-tune it on task-specific data. In practice, we observe that fine-tuning a pre-trained model on a small dataset may lead to over- and/or under-estimation problem. In this paper, we propose MC-Tailor, a novel method to alleviate the above issue in text generation tasks by truncating and transferring the probability mass from over-estimated regions to underestimated ones. Experiments on a variety of text generation datasets show that MC-Tailor consistently and significantly outperforms the fine-tuning approach. Our code is available at https://github.com/NingMiao/MC-tailor. © 2020 Association for Computational Linguistics.
Original languageEnglish
Title of host publicationACL 2020 - The 58th Annual Meeting of the Association for Computational Linguistics, Proceedings of the Conference
EditorsDan Jurafsky, Joyce Chai, Natalie Schluter, Joel Tetreault
Place of PublicationStroudsburg, PA
PublisherAssociation for Computational Linguistics
Pages3436-3441
ISBN (Print)9781952148255
DOIs
Publication statusPublished - Jul 2020
Externally publishedYes
Event58th Annual Meeting of the Association for Computational Linguistics (ACL 2020) - Virtual, United States
Duration: 5 Jul 202010 Jul 2020
https://acl2020.org/

Publication series

NameProceedings of the Annual Meeting of the Association for Computational Linguistics
ISSN (Print)0736-587X

Conference

Conference58th Annual Meeting of the Association for Computational Linguistics (ACL 2020)
Abbreviated titleACL2020
Country/TerritoryUnited States
Period5/07/2010/07/20
Internet address

Fingerprint

Dive into the research topics of 'Do you have the right scissors? Tailoring pre-trained language models via Monte-Carlo methods'. Together they form a unique fingerprint.

Cite this