Sentiment Analysis Based on a Deep Stochastic Network and Active Learning
DOI:
https://doi.org/10.26438/ijcse/v5i9.16Keywords:
Fasttext, Recurrent Neural Network, Gated Recurrent Unit, Active LearningAbstract
this paper proposes a novel approach for sentiment analysis. The growing importance of sentiment analysis commensurate with the use of social media such as reviews, forum discussions, blogs, micro blogs like Twitter, and other social networks. We require efficient and higher accuracy algorithms in sentiment polarity classification as well as sentiment strength detection. In comparison to pure vocabulary based system, deep learning algorithms show significantly higher performance. The goal of this research is to modify a Recurrent Neural Network (RNN) with Gated Recurrent Unit (GRU) by introducing stochastic depth in a hidden layer and comparing it with baseline Naïve Bayes, vanilla RNN and GRU-RNN models. To improve our results, we also incorporated Active Learning with Uncertainty Sampling approach. Movie review dataset from Rotten Tomatoes was used, the dataset includes 215,154 fine grained labelled phrases in addition to 11,855 full sentences. We performed pre-processing on the data and used an embedding matrix with pre-trained word vectors as features for training our model. These word vectors were generated using character level n-grams with fasttext on Wikipedia data.
References
Bojanowski, Piotr, et al, “Enriching word vectors with subword information”, arXiv preprint arXiv:1607.04606 (2016).
Huang, Gao, et al, “Deep networks with stochastic depth”, European Conference on Computer Vision. Springer International Publishing, 2016.
Socher, Richard, et al, “Recursive deep models for semantic compositionality over a sentiment treebank”, Proceedings of the 2013 conference on empirical methods in natural language processing. 2013.
Lewis, David D., and William A. Gale, “A sequential algorithm for training text classifiers”, Proceedings of the 17th annual international ACM SIGIR conference on Research and development in information retrieval. Springer-Verlag New York, Inc., 1994.
McCallum, Andrew, and Kamal Nigam, “A comparison of event models for naive bayes text classification”, AAAI-98 workshop on learning for text categorization. Vol. 752. 1998.
Mikolov, Tomas, et al, “Efficient estimation of word representations in vector space”, arXiv preprint arXiv, pp.1301.3781 (2013).
Mikolov, Tomas, et al, “Recurrent neural network based language model”, Interspeech. Vol. 2. 2010.
Settles, Burr, “Active learning literature survey”, University of Wisconsin, Madison, Vol.52, pp.55-66, 2010.
Downloads
Published
How to Cite
Issue
Section
License

This work is licensed under a Creative Commons Attribution 4.0 International License.
Authors contributing to this journal agree to publish their articles under the Creative Commons Attribution 4.0 International License, allowing third parties to share their work (copy, distribute, transmit) and to adapt it, under the condition that the authors are given credit and that in the event of reuse or distribution, the terms of this license are made clear.
