Back to Search Start Over

A Learning-Exploring Method to Generate Diverse Paraphrases with Multi-Objective Deep Reinforcement Learning

Authors :
Mingtong Liu
Changjian Hu
Yao Meng
Erguang Yang
Yujie Zhang
Deyi Xiong
Jinan Xu
Yufeng Chen
Source :
COLING
Publication Year :
2020
Publisher :
International Committee on Computational Linguistics, 2020.

Abstract

Paraphrase generation (PG) is of great importance to many downstream tasks in natural language processing. Diversity is an essential nature to PG for enhancing generalization capability and robustness of downstream applications. Recently, neural sequence-to-sequence (Seq2Seq) models have shown promising results in PG. However, traditional model training for PG focuses on optimizing model prediction against single reference and employs cross-entropy loss, which objective is unable to encourage model to generate diverse paraphrases. In this work, we present a novel approach with multi-objective learning to PG. We propose a learning-exploring method to generate sentences as learning objectives from the learned data distribution, and employ reinforcement learning to combine these new learning objectives for model training. We first design a sample-based algorithm to explore diverse sentences. Then we introduce several reward functions to evaluate the sampled sentences as learning signals in terms of expressive diversity and semantic fidelity, aiming to generate diverse and high-quality paraphrases. To effectively optimize model performance satisfying different evaluating aspects, we use a GradNorm-based algorithm that automatically balances these training objectives. Experiments and analyses on Quora and Twitter datasets demonstrate that our proposed method not only gains a significant increase in diversity but also improves generation quality over several state-of-the-art baselines.

Details

Database :
OpenAIRE
Journal :
Proceedings of the 28th International Conference on Computational Linguistics
Accession number :
edsair.doi...........e48a956430070e92207711c9f172278e
Full Text :
https://doi.org/10.18653/v1/2020.coling-main.209