To adapt or to fine-tune: a case study on abstractive summarization

  • Zheng Zhao*
  • , Pinzhen Chen
  • *Corresponding author for this work

Research output: Chapter in Book/Report/Conference proceedingConference contribution

1 Citation (Scopus)

Abstract

Recent advances in the field of abstractive summarization leverage pre-trained language models rather than train a model from scratch. However, such models are sluggish to train and accompanied by a massive overhead. Researchers have proposed a few lightweight alternatives such as smaller adapters to mitigate the drawbacks. Nonetheless, it remains uncertain whether using adapters benefits the task of summarization, in terms of improved efficiency without an unpleasant sacrifice in performance. In this work, we carry out multifaceted investigations on fine-tuning and adapters for summarization tasks with varying complexity: language, domain, and task transfer. In our experiments, fine-tuning a pre-trained language model generally attains a better performance than using adapters; the performance gap positively correlates with the amount of training data used. Notably, adapters exceed fine-tuning under extremely low-resource conditions. We further provide insights on multilinguality, model convergence, and robustness, hoping to shed light on the pragmatic choice of fine-tuning or adapters in abstractive summarization.

Original languageEnglish
Title of host publicationChinese Computational Linguistics: 21st China National Conference (CCL 2022): Proceedings
EditorsMaosong Sun, Yang Liu, Wanxiang Che, Yang Feng, Xipeng Qiu, Gaoqi Rao, Yubo Chen
Place of PublicationNanchang, China
PublisherSpringer Cham
Pages133–146
Number of pages14
ISBN (Electronic)9783031183157
ISBN (Print)9783031183140
DOIs
Publication statusPublished - 04 Oct 2022
Externally publishedYes

Publication series

NameLecture Notes in Computer Science
Volume13603
ISSN (Print)0302-9743
ISSN (Electronic)1611-3349

Keywords

  • summarization
  • pre-trained language models
  • transfer learning

Fingerprint

Dive into the research topics of 'To adapt or to fine-tune: a case study on abstractive summarization'. Together they form a unique fingerprint.

Cite this