-
Notifications
You must be signed in to change notification settings - Fork 4
Recurrent Neural Networks
Recurrent neural networks are great at encapsulating sequence or time-series data, due to the connections formed between nodes create a directed graph along the sequence. A great resource that explains LSTM's better than I could here is found here. Check it out.
There has been a great study on the optimal parameters for LSTM networks in sequence classification tasks here. In section 7 there is a comparison of different features and additions and the impact they have on the network. I will summarize them here:
Word embeddings improved accuracy when used accross the board.
Adam and Nadam proved the best, followed by RMSProp
CRF instead of softmax proved to be better.
Variational Dropout performed significantly better than naive or no dropout.
Gradient clipping did not help at all, however gradient normalization with T=1 proved to significatly increase accuracy.
The BIO and IOBES tagging scheme performed consistently better than the IOB tagging scheme.
If the number of recurrent units is kept constant, two stacked BiLSTM-layers resulted in the best performance.
The optimal size for the mini-batch appears to depend on the task. For POS tagging and event recognition, a size of 1 was optimal,for chunking a size of 8 and for NER and Entity Recognition a size of 31.
Character-based representations were in a lot of tested configurations not that helpful and could not improve the performance of the network.
The number of recurrent units, as long as it is not far too large or far too small, has only a minor effect on the results. A value of about 100 for each LSTM-network appears to be a good rule of thumb for the tested tasks.
Theano as well as Tensorflow performed equally in terms of test performance.
- ACLSW 2019
- Our datasets
- Experiment Results
- Research Analysis
- Hypothesis
- Machine Learning
- Deep Learning
- Paper Section Drafts
- Word Embeddings
- References/Resources
- Correspondence with H. Aghakhani
- The Gotcha! Collection