T-CVAE: Transformer-Based Conditioned Variational Autoencoder for Story Completion

T-CVAE: Transformer-Based Conditioned Variational Autoencoder for Story Completion

Tianming Wang, Xiaojun Wan

Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence
Main track. Pages 5233-5239. https://doi.org/10.24963/ijcai.2019/727

Story completion is a very challenging task of generating the missing plot for an incomplete story, which requires not only understanding but also inference of the given contextual clues. In this paper, we present a novel conditional variational autoencoder based on Transformer for missing plot generation. Our model uses shared attention layers for encoder and decoder, which make the most of the contextual clues, and a latent variable for learning the distribution of coherent story plots. Through drawing samples from the learned distribution, diverse reasonable plots can be generated. Both automatic and manual evaluations show that our model generates better story plots than state-of-the-art models in terms of readability, diversity and coherence.
Keywords:
Natural Language Processing: Natural Language Generation
Natural Language Processing: Natural Language Processing