首页    期刊浏览 2025年03月03日 星期一
登录注册

文章基本信息

  • 标题:Joint Energy-based Model Training for Better Calibrated Natural Language Understanding Models
  • 本地全文:下载
  • 作者:Tianxing He ; Bryan McCann ; Caiming Xiong
  • 期刊名称:Conference on European Chapter of the Association for Computational Linguistics (EACL)
  • 出版年度:2021
  • 卷号:2021
  • 页码:1754-1761
  • DOI:10.18653/v1/2021.eacl-main.151
  • 语种:English
  • 出版社:ACL Anthology
  • 摘要:In this work, we explore joint energy-based model (EBM) training during the finetuning of pretrained text encoders (e.g., Roberta) for natural language understanding (NLU) tasks. Our experiments show that EBM training can help the model reach a better calibration that is competitive to strong baselines, with little or no loss in accuracy. We discuss three variants of energy functions (namely scalar, hidden, and sharp-hidden) that can be defined on top of a text encoder, and compare them in experiments. Due to the discreteness of text data, we adopt noise contrastive estimation (NCE) to train the energy-based model. To make NCE training more effective, we train an auto-regressive noise model with the masked language model (MLM) objective.
国家哲学社会科学文献中心版权所有