ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1503.05034
30
28

gengengenCNN: A Convolutional Architecture for Word Sequence Prediction

17 March 2015
Mingxuan Wang
Zhengdong Lu
Hang Li
Wenbin Jiang
Qun Liu
ArXivPDFHTML
Abstract

We propose a novel convolutional architecture, named gengengenCNN, for word sequence prediction. Different from previous work on neural network-based language modeling and generation (e.g., RNN or LSTM), we choose not to greedily summarize the history of words as a fixed length vector. Instead, we use a convolutional neural network to predict the next word with the history of words of variable length. Also different from the existing feedforward networks for language modeling, our model can effectively fuse the local correlation and global correlation in the word sequence, with a convolution-gating strategy specifically designed for the task. We argue that our model can give adequate representation of the history, and therefore can naturally exploit both the short and long range dependencies. Our model is fast, easy to train, and readily parallelized. Our extensive experiments on text generation and nnn-best re-ranking in machine translation show that gengengenCNN outperforms the state-of-the-arts with big margins.

View on arXiv
Comments on this paper