[1] Goodfellow I J, Pouget-Abadie J, Mirza M, et al. Generative adversarial nets[C]//Proceedings of the Annual Conference on Neural Information Processing Systems 2014, Montreal, Dec 8-13, 2014. Red Hook: Curran Associates, 2014: 2672-2680.
[2] Yu L T, Zhang W N, Wang J, et al. SeqGAN: sequence gene-rative adversarial nets with policy gradient[C]//Proceedings of the 31st AAAI Conference on Artificial Intelligence, San Francisco, Feb 4-9, 2017. Menlo Park: AAAI, 2017: 2852-2858.
[3] Sutskever I, Vinyals O, Le Q V, et al. Sequence to sequence learning with neural networks[C]//Proceedings of the Annual Conference on Neural Information Processing Systems 2014, Montreal, Dec 8-13, 2014. Red Hook: Curran Associates, 2014: 3104-3112.
[4] Bengio S, Vinyals O, Jaitly N, et al. Scheduled sampling for sequence prediction with recurrent neural networks[C]//Pro-ceedings of the Annual Conference on Neural Information Processing Systems 2015, Montreal, Dec 7-12, 2015. Red Hook: Curran Associates, 2015: 1171-1179.
[5] Denton E L, Chintala S, Szlam A. Deep generative image mo-dels using a Laplacian pyramid of adversarial networks[C]//Proceedings of the Annual Conference on Neural Information Processing Systems 2015, Montreal, Dec 7-12, 2015. Red Hook: Curran Associates, 2015: 1486-1494.
[6] Huszár F. How (not) to train your generative model: scheduled sampling, likelihood, adversary?[J]. arXiv:1511.05101, 2015.
[7] Lin K, Li D Q, He X D, et al. Adversarial ranking for lan-guage generation[C]//Proceedings of the Annual Conference on Neural Information Processing Systems 2017, Long Beach, Dec 4-9, 2017. Red Hook: Curran Associates, 2017: 3155-3165.
[8] Guo J X, Lu S D, Cai H, et al. Long text generation via adve-rsarial training with leaked information[C]//Proceedings of the 32nd AAAI Conference on Artificial Intelligence, the 30th Innovative Applications of Artificial Intelligence, and the 8th AAAI Symposium on Educational Advances in Artificial Intelligence, New Orleans, Feb 2-7, 2018. Menlo Park: AAAI, 2018: 5141-5148.
[9] Wang K, Wan X J. SentiGAN: generating sentimental texts via mixture adversarial networks[C]//Proceedings of the 27th International Joint Conference on Artificial Intelligence, Sto-ckholm, Jul 13-19, 2018: 4446-4452.
[10] Hochreiter S, Schmidhuber J. Long short-term memory[J]. Neural Computation, 1997, 9(8): 1735-1780.
[11] Cho K, Van Merrienboer B, GülçehreÇ, et al. Learning phrase representations using RNN encoder-decoder for statistical machine translation[C]//Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing, Doha, Oct 25-29, 2014. Stroudsburg: ACL, 2014: 1724-1734.
[12] Bahdanau D, Cho K, Bengio Y. Neural machine translation by jointly learning to align and translate[J]. arXiv:1409.0473, 2014.
[13] Bing L D, Li P J, Liao Y, et al. Abstractive multi-document summarization via phrase selection and merging[J]. arXiv:1506.01597, 2015.
[14] Hu B T, Lu Z D, Li H, et al. Convolutional neural network architectures for matching natural language sentences[C]//Proceedings of the Annual Conference on Neural Information Processing Systems 2014, Montreal, Dec 8-13, 2014. Red Hook: Curran Associates, 2014: 2042-2050.
[15] Srivastava N, Hinton G H, Krizhevsky A, et al. Dropout: a simple way to prevent neural networks from overfitting[J]. Journal of Machine Learning Research, 2014, 15(1): 1929-1958.
[16] Lin C Y. Rouge: a package for automatic evaluation of sum-maries[C]//Proceedings of the Text Summarization Branches Out, Barcelona, Jul 2014. Stroudsburg: ACL, 2004: 74-81.
[17] Stolcke A. SRILM-an extensible language modeling toolkit[C]//Proceedings of the 7th International Conference on Spoken Language Processing, Denver, Sep 16-20, 2002: 1-4. |