Skip to content

Universal seq2seq model for question generation, text summarization, machine translation etc. written in Python and Tensorflow 1.4 with Tensorflow's Beam Search API decoder

Notifications You must be signed in to change notification settings

bhavikapanara/seq2seq

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

9 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

seq2seq

Universal sequence-to-sequence model with attention and beam search for inference decoding. Should work for text summarization, neural machine translation, question generation etc. although might require different hyperparameters or data preprocessing.

In my case I've used it for question generation - I've trained the model on reversed SQuAD dataset with paragraphs as my input and questions as my targets. The SQuAD parsing script is also included in the depository.

The data and checkpoints are not included as they simply weight too much, but feel free to experiment around with the model and let me know your results! :)

Requirements

Workflow

Model uses LSTM cells, Bahdanau attention, Conceptnet Numberbatch for word vector and spaCy for data preprocessing. I believe every function is nicely commented, so I won't go too much into details here as the code speaks for itself.

For the data preprocessing I'm using spaCy as I've came to rely heavily on it when it comes to any NLP tasks - it's simply brilliant. In this case, spaCy is used to remove stopwords and punctuation from the dataset as well as exchange any entities in the text into their corresponding labels, such as LOC, PERSON, DATE etc, so the model learns dependencies between paragraph and question and does not overfit.

Sample

The model doesn't produce any staggering results for the question generation task with the dataset I've used, however, I can imaging with more examples the output would make more sense. Possibly training the model without changing the entities with their labels, or more experimenting with data preprocessing (lemmatization, keeping only important nouns/verbs etc) could also yield better retuls.

Original Text:
'Archbishop Albrecht of Mainz and Magdeburg did not reply to Luther's letter containing the 95 Theses. He had the theses checked for heresy and in December 1517 forwarded them to Rome. He needed the revenue from the indulgences to pay off a papal dispensation for his tenure of more than one bishopric. As Luther later noted, "the pope had a finger in the pie as well, because one half was to go to the building of St Peter's Church in Rome".'

Converted Text:
ORG PERSON reply ORG 's letter containing 95 theses theses checked heresy DATE forwarded GPE needed revenue indulgences pay papal dispensation tenure bishopric ORG later noted pope finger pie CARDINAL building ORG GPE

Generated Questions:
-- : what was the issue of ORG
-- : what was the issue of ORG in GPE
-- : what was ORG 's stance of the printed taken in GPE

About

Universal seq2seq model for question generation, text summarization, machine translation etc. written in Python and Tensorflow 1.4 with Tensorflow's Beam Search API decoder

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 100.0%