Not Logged In

CGMH: Constrained Sentence Generation by Metropolis-Hastings Sampling

Full Text: 4659-Article Text-7698-1-10-20190707.pdf PDF

In real-world applications of natural language generation, there are often constraints on the target sentences in addition to fluency and naturalness requirements. Existing language generation techniques are usually based on recurrent neural networks (RNNs). However, it is non-trivial to impose constraints on RNNs while maintaining generation quality, since RNNs generate sentences sequentially (or with beam search) from the first word to the last. In this paper, we propose CGMH, a novel approach using Metropolis-Hastings sampling for constrained sentence generation. CGMH allows complicated constraints such as the occurrence of multiple keywords in the target sentences, which cannot be handled in traditional RNN-based approaches. Moreover, CGMH works in the inference stage, and does not require parallel corpora for training. We evaluate our method on a variety of tasks, including keywords-to-sentence generation, unsupervised sentence paraphrasing, and unsupervised sentence error correction. CGMH achieves high performance compared with previous supervised methods for sentence generation. Our code is released at https://github.com/NingMiao/CGMH

Citation

N. Miao, H. Zhou, L. Mou, R. Yan, L. Li. "CGMH: Constrained Sentence Generation by Metropolis-Hastings Sampling". National Conference on Artificial Intelligence (AAAI), pp 6834-6842, January 2019.

Keywords:  
Category: In Conference
Web Links: doi
  AAAI

BibTeX

@incollection{Miao+al:AAAI19,
  author = {Ning Miao and Hao Zhou and Lili Mou and Rui Yan and Lei Li},
  title = {CGMH: Constrained Sentence Generation by Metropolis-Hastings
    Sampling},
  Pages = {6834-6842},
  booktitle = {National Conference on Artificial Intelligence (AAAI)},
  year = 2019,
}

Last Updated: February 02, 2021
Submitted by Sabina P

University of Alberta Logo AICML Logo