Timezone: »
Text generation is of great importance to many natural language processing applications. However, maximization-based decoding methods (e.g., beam search) of neural language models often lead to degenerate solutions---the generated text is unnatural and contains undesirable repetitions. Existing approaches introduce stochasticity via sampling or modify training objectives to decrease the probabilities of certain tokens (e.g., unlikelihood training). However, they often lead to solutions that lack coherence. In this work, we show that an underlying reason for model degeneration is the anisotropic distribution of token representations. We present a contrastive solution: (i) SimCTG, a contrastive training objective to calibrate the model's representation space, and (ii) a decoding method---contrastive search---to encourage diversity while maintaining coherence in the generated text. Extensive experiments and analyses on three benchmarks from two languages demonstrate that our proposed approach outperforms state-of-the-art text generation methods as evaluated by both human and automatic metrics.
Author Information
Yixuan Su (University of Cambridge)
Tian Lan (Tencent Technology (Shenzhen) Co., Ltd.)
Yan Wang (Tencent AI Lab)
Yan Wang is a senior researcher of Natural Language Processing Center, Tencent AI Lab. His research interests include dialogue systems, text generation, and question answering. He has published over 30 research papers in leading conferences and journals, such as ACL, EMNLP, NAACL, AAAI, and TASLP. He received an outstanding paper award in ACL 2021 for one of his works on retrieval-augmented text generation. He served in the program committee of some conferences including ACL, EMNLP, NAACL, AAAI, etc, and was selected as a session chair in ACL 2021 and senior program committee member in AAAI 2022.
Dani Yogatama (Google DeepMind)
Lingpeng Kong (Department of Computer Science, The University of Hong Kong)
Nigel Collier (University of Cambridge)
Related Events (a corresponding poster, oral, or spotlight)
-
2022 Spotlight: A Contrastive Framework for Neural Text Generation »
Tue. Dec 6th 05:00 -- 07:00 PM Room
More from the Same Authors
-
2021 : LiRo: Benchmark and leaderboard for Romanian language tasks »
Stefan Dumitrescu · Petru Rebeja · Beata Lorincz · Mihaela Gaman · Andrei Avram · Mihai Ilie · Andrei Pruteanu · Adriana Stan · Lorena Rosia · Cristina Iacobescu · Luciana Morogan · George Dima · Gabriel Marchidan · Traian Rebedea · Madalina Chitez · Dani Yogatama · Sebastian Ruder · Radu Tudor Ionescu · Razvan Pascanu · Viorica Patraucean -
2021 Spotlight: Mind the Gap: Assessing Temporal Generalization in Neural Language Models »
Angeliki Lazaridou · Adhi Kuncoro · Elena Gribovskaya · Devang Agrawal · Adam Liska · Tayfun Terzi · Mai Gimenez · Cyprien de Masson d'Autume · Tomas Kocisky · Sebastian Ruder · Dani Yogatama · Kris Cao · Susannah Barlow · Phil Blunsom -
2022 Poster: Measuring and Reducing Model Update Regression in Structured Prediction for NLP »
Deng Cai · Elman Mansimov · Yi-An Lai · Yixuan Su · Lei Shu · Yi Zhang -
2022 Poster: CoNT: Contrastive Neural Text Generation »
Chenxin An · Jiangtao Feng · Kai Lv · Lingpeng Kong · Xipeng Qiu · Xuanjing Huang -
2022 Spotlight: CoNT: Contrastive Neural Text Generation »
Chenxin An · Jiangtao Feng · Kai Lv · Lingpeng Kong · Xipeng Qiu · Xuanjing Huang -
2021 Poster: End-to-End Training of Multi-Document Reader and Retriever for Open-Domain Question Answering »
Devendra Singh · Siva Reddy · Will Hamilton · Chris Dyer · Dani Yogatama -
2021 Poster: Mind the Gap: Assessing Temporal Generalization in Neural Language Models »
Angeliki Lazaridou · Adhi Kuncoro · Elena Gribovskaya · Devang Agrawal · Adam Liska · Tayfun Terzi · Mai Gimenez · Cyprien de Masson d'Autume · Tomas Kocisky · Sebastian Ruder · Dani Yogatama · Kris Cao · Susannah Barlow · Phil Blunsom