Sentiment Analysis Based on a Deep Stochastic Network and Active Learning

Authors

  • Jain T Dept. of CSE, Indian Institute of Technology (IIT), Delhi, India
  • Agarwal Dept. of CSE, Indian Institute of Technology (IIT), Delhi, India
  • Pancholia R Dept. of CSE, Birla Institute of Technology and Science, Pilani, India

DOI:

https://doi.org/10.26438/ijcse/v5i9.16

Keywords:

Fasttext, Recurrent Neural Network, Gated Recurrent Unit, Active Learning

Abstract

this paper proposes a novel approach for sentiment analysis. The growing importance of sentiment analysis commensurate with the use of social media such as reviews, forum discussions, blogs, micro blogs like Twitter, and other social networks. We require efficient and higher accuracy algorithms in sentiment polarity classification as well as sentiment strength detection. In comparison to pure vocabulary based system, deep learning algorithms show significantly higher performance. The goal of this research is to modify a Recurrent Neural Network (RNN) with Gated Recurrent Unit (GRU) by introducing stochastic depth in a hidden layer and comparing it with baseline Naïve Bayes, vanilla RNN and GRU-RNN models. To improve our results, we also incorporated Active Learning with Uncertainty Sampling approach. Movie review dataset from Rotten Tomatoes was used, the dataset includes 215,154 fine grained labelled phrases in addition to 11,855 full sentences. We performed pre-processing on the data and used an embedding matrix with pre-trained word vectors as features for training our model. These word vectors were generated using character level n-grams with fasttext on Wikipedia data.

References

Bojanowski, Piotr, et al, “Enriching word vectors with subword information”, arXiv preprint arXiv:1607.04606 (2016).

Huang, Gao, et al, “Deep networks with stochastic depth”, European Conference on Computer Vision. Springer International Publishing, 2016.

Socher, Richard, et al, “Recursive deep models for semantic compositionality over a sentiment treebank”, Proceedings of the 2013 conference on empirical methods in natural language processing. 2013.

Lewis, David D., and William A. Gale, “A sequential algorithm for training text classifiers”, Proceedings of the 17th annual international ACM SIGIR conference on Research and development in information retrieval. Springer-Verlag New York, Inc., 1994.

McCallum, Andrew, and Kamal Nigam, “A comparison of event models for naive bayes text classification”, AAAI-98 workshop on learning for text categorization. Vol. 752. 1998.

Mikolov, Tomas, et al, “Efficient estimation of word representations in vector space”, arXiv preprint arXiv, pp.1301.3781 (2013).

Mikolov, Tomas, et al, “Recurrent neural network based language model”, Interspeech. Vol. 2. 2010.

Settles, Burr, “Active learning literature survey”, University of Wisconsin, Madison, Vol.52, pp.55-66, 2010.

Downloads

Published

2025-11-12
CITATION
DOI: 10.26438/ijcse/v5i9.16
Published: 2025-11-12

How to Cite

[1]
T. Jain, K. Agarwal, and R. Pancholia, “Sentiment Analysis Based on a Deep Stochastic Network and Active Learning”, Int. J. Comp. Sci. Eng., vol. 5, no. 9, pp. 1–6, Nov. 2025.

Issue

Section

Research Article