首页    期刊浏览 2024年12月12日 星期四
登录注册

文章基本信息

  • 标题:GlobalizingBERT-based Transformer Architectures for Long Document Summarization
  • 本地全文:下载
  • 作者:Quentin Grail ; Julien Perez ; Eric Gaussier
  • 期刊名称:Conference on European Chapter of the Association for Computational Linguistics (EACL)
  • 出版年度:2021
  • 卷号:2021
  • 页码:1792-1810
  • DOI:10.18653/v1/2021.eacl-main.154
  • 语种:English
  • 出版社:ACL Anthology
  • 摘要:Fine-tuning a large language model on downstream tasks has become a commonly adopted process in the Natural Language Processing (NLP) (CITATION). However, such a process, when associated with the current transformer-based (CITATION) architectures, shows several limitations when the target task requires to reason with long documents. In this work, we introduce a novel hierarchical propagation layer that spreads information between multiple transformer windows. We adopt a hierarchical approach where the input is divided in multiple blocks independently processed by the scaled dot-attentions and combined between the successive layers. We validate the effectiveness of our approach on three extractive summarization corpora of long scientific papers and news articles. We compare our approach to standard and pre-trained language-model-based summarizers and report state-of-the-art results for long document summarization and comparable results for smaller document summarization.
国家哲学社会科学文献中心版权所有