• Open access
  • Published: 11 December 2017

Optimization of sentiment analysis using machine learning classifiers

  • Jaspreet Singh   ORCID: orcid.org/0000-0002-9018-1233 1 ,
  • Gurvinder Singh 1 &
  • Rajinder Singh 1  

Human-centric Computing and Information Sciences volume  7 , Article number:  32 ( 2017 ) Cite this article

34k Accesses

97 Citations

2 Altmetric

Metrics details

Words and phrases bespeak the perspectives of people about products, services, governments and events on social media. Extricating positive or negative polarities from social media text denominates task of sentiment analysis in the field of natural language processing. The exponential growth of demands for business organizations and governments, impel researchers to accomplish their research in sentiment analysis. This paper leverages four state-of-the-art machine learning classifiers viz. Naïve Bayes, J48, BFTree and OneR for optimization of sentiment analysis. The experiments are performed using three manually compiled datasets; two of them are captured from Amazon and one dataset is assembled from IMDB movie reviews. The efficacies of these four classification techniques are examined and compared. The Naïve Bayes found to be quite fast in learning whereas OneR seems more promising in generating the accuracy of 91.3% in precision, 97% in F-measure and 92.34% in correctly classified instances.

Introduction to sentiment analysis

The popularity of rapidly growing online social networks and electronic media based societies has influenced the young researchers to pursue their work on sentiment analysis. These days organizations quite keen assess their customers or public opinion about their products from social media text [ 1 ]. The online service providers are hooked on assessing social media data on blogs, online forums, comments, tweets and product reviews. This assessment is exploited for their decision making or amelioration of their services or quality of products. The applications of sentiment analysis encompass the areas like social event planning, election campaigning, healthcare monitoring, consumer products and awareness services. The immoderate use of internet by business organizations all around the globe has noticed that opinionated web text has molded our business plays and socio-economic systems. The computational power is fueled by burgeon of machine learning techniques. This work focused on four text classifiers utilized for sentiment analysis viz. Naïve Bayes, J48, BFTree and OneR algorithm. The “ Machine learning techniques for sentiment analysis ” section of this paper provides the intuition behind the task of sentiment classification by leveraging the modeling of aforementioned four classifiers. The architecture of proposed model using four sentiment classifiers is disposed in “ Proposed methodology for optimization of sentiment prediction using weka ” section. The related work with recent contributions of machine learning in the field of sentiment classification is described in “ Related work ” section. In “ Datasets taken ” section, the three manually annotated datasets are described along with their preprocessing. The experimental results and discussion of efficacies of classifiers are cataloged in “ Results and discussions ” section followed by the ending remarks along with a future direction in “ Conclusion ” section.

Levels of sentiment

Due to scarcity of opinion text available in digital form, very less research interest on computational linguistics in the last decade of twentieth century was witnessed [ 2 , 3 , 4 ]. The escalation of social media text on internet attracts young researchers to define the level of granularities of text. The web text is classified into three levels viz. document level, sentence level and word level. In [ 5 ], the fourth level granularity is defined by using deep convolution neural network. This fourth level is character level feature extraction approach used for extracting features of each character window from given word (Table  1 ).

Machine learning techniques for sentiment analysis

The social networking sites dispense their data conveniently and freely on the web. This availability of data entices the interest of young researchers to plunge them in the field of sentiment analysis. People express their emotions and perspectives on the social media discussion forums [ 6 ]. The business organizations employ researchers to investigate the unrevealed facts about their products and services. Spontaneous and automatic determination of sentiments from reviews is the main concern of multinational organizations [ 7 , 8 , 9 , 10 ]. The machine learning techniques have improved accuracy of sentiment analysis and expedite automatic evaluation of data these days. This work attempted to utilize four machine learning techniques for the task of sentiment analysis. The modeling of four techniques is briefly discussed below.

Naïve Bayes used for sentiment classification

The dichotomy of sentiment is generally decided by the mindset of an author of text whether he is positively or negatively oriented towards his saying [ 6 , 11 , 12 , 13 ]. Naïve Bayes classifier is a popular supervised classifier, furnishes a way to express positive, negative and neutral feelings in the web text. Naïve Bayes classifier utilizes conditional probability to classify words into their respective categories. The benefit of using Naïve Bayes on text classification is that it needs small dataset for training. The raw data from web undergoes preprocessing, removal of numeric, foreign words, html tags and special symbols yielding the set of words. The tagging of words with labels of positive, negative and neutral tags is manually performed by human experts. This preprocessing produces word-category pairs for training set. Consider a word ‘y’ from test set (unlabeled word set) and a window of n-words (x 1 , x 2 , …… x n ) from a document. The conditional probability of given data point ‘y’ to be in the category of n-words from training set is given by:

Consider an example of a movie review for movie “Exposed”. The experimentation with Naïve Bayes yields the following results.

J48 algorithm used for sentiment prediction

The hierarchical mechanism divides feature space into distinct regions followed by the categorization of sample into category labels. J48 is a decision tree based classifier used to generate rules for the prediction of target terms. It has an ability to deal with larger training datasets than other classifiers [ 14 ]. The word features for sentences of corpus taken from labeled arff file of training set are represented in the leaf nodes of decision tree. In the test set every time when a near feature qualifies the label condition of internal feature node, its level is lifted up in the same branch of decision tree. The assignment of labels to the word features of test set gradually generates different two branches of decision tree. J48 algorithm uses entropy function for testing the classification of terms from the test set.

where (Term) can be unigram, bigram and trigram. In this study we have considered unigrams and bigrams. The example in the Table  2 contains bigrams like “Horrible acting”, “Bad writing” and “Very misleading” are labeled with negative sentiment whereas the term “More enjoyable” reflects positive sentiment towards the movie. The decision tree of J48 algorithm for obtaining sentiment form text is represented in the Fig.  1 below.

J48’s Decision Tree for terms of Example in Table  2

BFTREE algorithm used for sentiment prediction

Another classification approach outperforms J48, C4.5 and CART by expanding only best node in the depth first order. BFTree algorithm excavates the training file for locating best supporting matches of positive and negative terms in the test file. BFTree algorithm keeps heuristic information gain to identify best node by probing all collected word features. The only difference between J48 and BFTree classifier is the computation order in which decision tree is built. The decision tree disparate feature terms of plain text taken from movie reviews and classify them at document level by tagging appropriate labels. BFTree extracts best node from labeled and trained binary tree nodes to reduce the error computed from information gain.

where S is word feature term of test set and A is the attribute of sampled term from training set. V(A) denotes set of all possible values of A. The binary tree stops growing when an attribute A captures single value or when value of information gain vanishes.

OneR algorithm used for sentiment prediction

OneR algorithm is a classification approach which restricts decision tree to level one thereby generating one rule. One rule makes prediction on word feature terms with minimal error rate due to repetitive assessment of word occurrences. The classification of most frequent terms of a particular sentence is made on the basis of class of featured terms from training set. The demonstration of OneR algorithm for sentiment prediction with smallest error of classification is given below:

Select a featured term from training set.

Train a model using step 3 and step 4.

For each prediction term.

For each value of that predictor.

Count frequency of each value of target term.

Find most frequent class.

Make a rule and assign that class to predictor.

Calculate total error of rules of each predictor.

Choose predictor with smallest error.

Proposed methodology for optimization of sentiment prediction using weka

The preprocessing of raw text from web is done in python 3.5 using NLTK and bs4 libraries. Each review in the first dataset is parsed with NLTK’s parser and title of the review is considered as a feature. We have obtained 15 features from first dataset and 42 features from each of second and third dataset. The CSV files generated from Python are converted to ARFF files for WEKA 3.8. Only two sentiment labels namely Pos for positive and Neg for negative are used for assigning sentences. The working methodology of proposed work for optimization of sentiment prediction is given below in Fig.  2 .

Proposed methodology

After loading files with ARFF loader, the class assigner picks up appropriate class labels from dataset and performs feature selection on the basis of frequently used headings and most frequent titles. The feature selector module is implemented using three feature selection methods namely Document Frequency (DF), Mutual Information (MI) and Information Gain (IG). The mathematical modeling of these feature selection methods requires some probability distributions and statistical notations described below:

P(w): Probability that a document ‘d’ contains term ‘w’.

P(c’): Probability that document ‘d’ does not belongs to category ‘c’.

P(w, c): Joint probability that document ‘d’ contains word term ‘w’ of category ‘c’.

P(c/w): Conditional probability that a document ‘d’ belongs to category ‘c’ under the condition that ‘d’ contains word term ‘w’.

Similarly other notations like P(w’), P(w/c), P(w/c’), P(c/w’) and P(c’/w) are taken and {c} is the set of categories.

N 1 : Number of documents that exhibit category ‘c’ and contain term ‘w’.

N 2 : Number of documents that do not belong to category ‘c’ but contains term ‘w’.

N 3 : Number of documents that belong to category ‘c’ and do not contain term ‘w’.

N 4 : Number of documents that neither belong to category ‘c’ nor contain term ‘w’.

N: Total number of document reviews.

DF method qualifies only those documents in which a higher frequency terms are considered.

The MI method measures features of text by computing similarity of word terms ‘w’ and category ‘c’.

The IG-construct measures similarity information for category by exploiting probabilities of absence or presence of terms in a document review.

The normalization module converts all letters into lowercase, removal of punctuation marks and special symbols, conversion of numbers into words, expansion of abbreviation and limiting the average length of twenty words in a sentence. Each sentence is delimited by a newline character. The Python’s NLTK and bs4 libraries are used for this purpose. Data splitter take the ratio of (80:20) of (Train: Test) subsets. We have used manual splitting of dataset at the time of retrieval of data from web. The four classifiers are trained with training subsets followed by performance evaluation. The evaluation metrics taken in the experiment are precision, recall, accuracy and F-measure.

Related work

Existing approaches of sentiment prediction and optimization widely includes SVM and Naïve Bayes classifiers. Hierarchical machine learning approaches yields moderate performance in classification tasks whereas SVM and Multinomial Naïve Bayes are proved better in terms of accuracy and optimization. Sentiment analysis using neural network architectures has appeared in very few works. The sentiment prediction methods using recursive neural networks and deep convolution neural networks are bit complex in capturing compositionality of words. Extracting character level features and embeddings of complex words is found hard in many neural network architectures whereas extracting sentence level or word level features such as morphological tags and stems are more effectively achieved in convolutional neural networks. A very few researchers have used J48, BFTree and OneR for the task of sentiment prediction. These three classifiers are utilized for other classification tasks like emotion recognition from text and twitter’s text categorizations. The summary of benchmarks related to machine learning techniques in terms of accuracy of classification is listed in the Table  2 . SVM and Naive Bayes are proving better in terms of benchmarks than other machine learning techniques (Table  3 ).

Datasets taken

Three Datasets are manually annotated from http://www.amazon.in . First dataset consists of product reviews of Woodland’s wallet are taken from 12th October 2016 to 25th October 2016 for training set containing 88 reviews and from 25th October 2016 to 30th October 2016 for testing set containing 12 randomly chosen product reviews with their sentiments prediction using four machine learning algorithms. Second dataset consists of 7465 Digital Camera reviews of Sony are taken from 01st October 2016 to 25th October 2016 for training set and 1000 reviews are from 25th October 2016 to 30th October 2016 for test dataset. Third dataset consists of movie reviews taken from http://www.imdb.com . It contains 2421 reviews for training set and 500 reviews for test set.

Results and discussions

The experiment is carried out by using freeware WEKA software tool for classification of sentiments in the text. Standard implementations of Naïve Bayes, J48, BFTree and OneR algorithms are exploited from WEKA version 3.8. The classification accuracy of first dataset shows 100% classification accuracy with Naïve Bayes in some of the epochs because of small size of dataset. The average of 29 epochs for all four classifiers on second and third datasets is presented in Table  4 below. Naïve Bayes shows faster learning among four classifiers whereas J48 found to be slower. OneR classifier is leading from other three classifiers in percentage of correctly classified instances. The accuracy of J48 algorithm is promising in true positive and false positive rates.

Results of classification accuracies for the test subsets with 42 and 15 attributes are recorded. The average accuracies of 29 runs on three datasets is presented in Table  5 below. All four classifiers improved in accuracies with the increase of features from 15 to 42. This shows the direct proportionality of multiple features with learning capability for machine learning algorithms.

This paper exploits four machine learning classifiers for sentiment analysis using three manually annotated datasets. The mean of 29 epochs of experimentation recorded in Table  4 shows that OneR is more precise in terms of percentage of correctly classified instances. On the other hand, Naïve Bayes exhibits faster learning rate and J48 reveals adequacy in the true positive and false positive rates. Table  5 reveals the truth that J48 and OneR are better for smaller dataset of woodland’s wallet reviews. The preprocessing of proposed methodology is limited to extract foreign words, emoticons and elongated words with their appropriate sentiments. The future work in the task of sentiment analysis has scope to improve preprocessing with word embeddings using deep neural networks and can also extend this study through convolution neural networks.

Parvathy G, Bindhu JS (2016) A probabilistic generative model for mining cybercriminal network from online social media: a review. Int J Comput Appl 134(14):1–4. doi: 10.5120/ijca2016908121

Google Scholar  

Cambria E, White B (2014) Jumping NLP curves: a review of natural language processing research. IEEE Comput Intell Mag 9(2):48–57. doi: 10.1109/mci.2014.2307227

Article   Google Scholar  

Pang B, Lee L, Vaithyanathan S (2002) Thumbs up? In: Proceedings of the ACL-02 conference on empirical methods in natural language processing—EMNLP ‘02. doi: 10.3115/1118693.1118704

Poria S, Cambria E, Gelbukh A, Bisio F, Hussain A (2015) Sentiment data flow analysis by means of dynamic linguistic patterns. IEEE Comput Intell Mag 10(4):26–36. doi: 10.1109/mci.2015.2471215

Nogueira dos Santos C, Gatti M (2014) Deep convolution neural networks for sentiment analysis of short texts. In: Proceedings of COLING 2014, the 25th international conference on computational linguistics. p 69–78

Kiritchenko S, Mohammad S, Salameh M (2016) SemEval-2016 task 7: determining sentiment intensity of English and Arabic phrases. In: Proceedings of the 10th international workshop on semantic evaluation (SemEval-2016). doi:10.18653/v1/s16-1004

Cernian A, Sgarciu V, Martin B (2015) Sentiment analysis from product reviews using SentiWordNet as lexical resource. In: 2015 7th international conference on electronics, computers and artificial intelligence (ECAI). doi: 10.1109/ecai.2015.7301224

Hammer HL, Solberg PE, Øvrelid L (2014) Sentiment classification of online political discussions: a comparison of a word-based and dependency-based method. In: Proceedings of the 5th workshop on computational approaches to subjectivity, sentiment and social media analysis. doi: 10.3115/v1/w14-2616

Zadeh L (2006) Toward human-level machine intelligence. In: 2006 18th IEEE international conference on tools with artificial intelligence (ICTAI’06). doi: 10.1109/ictai.2006.114

Joachims T (2002) Text classification. Learning to classify text using support vector machines. p 7-33. doi: 10.1007/978-1-4615-0907-3_2

Wanton TM, Porrata AP, Guijarro AM, Balahur A (2010) Opinion polarity detection—using word sense disambiguation to determine the polarity of opinions. In: Proceedings of the 2nd international conference on agents and artificial intelligence. doi: 10.5220/0002703504830486

Xia Y, Cambria E, Hussain A, Zhao H (2014) Word polarity disambiguation using bayesian model and opinion-level features. Cogn Comput 7(3):369–380. doi: 10.1007/s12559-014-9298-4

Dey L, Chakraborty S, Biswas A, Bose B, Tiwari S (2016) Sentiment analysis of review datasets using Naïve Bayes’ and K-NN classifier. Int J Inform Eng Electron Bus 8(4):54–62. doi: 10.5815/ijieeb.2016.04.07

Nie CY, Wang J, He F, Sato R (2015) Application of J48 decision tree classifier in emotion recognition based on chaos characteristics. In: Proceedings of the 2015 international conference on automation, mechanical control and computational engineering. doi: 10.2991/amcce-15.2015.330

Tan S, Zhang J (2008) An empirical study of sentiment analysis for Chinese documents. Expert Syst Appl 34(4):2622–2629. doi: 10.1016/j.eswa.2007.05.028

Mohammad SM, Zhu X, Kiritchenko S, Martin J (2015) Sentiment, emotion, purpose, and style in electoral tweets. Inf Process Manage 51(4):480–499. doi: 10.1016/j.ipm.2014.09.003

Kiritchenko S, Mohammad SM (2016) sentiment composition of words with opposing polarities. In: Proceedings of the 2016 conference of the north american chapter of the association for computational linguistics: human language technologies. doi:10.18653/v1/n16-1128

Dashtipour K, Poria S, Hussain A, Cambria E, Hawalah AY, Gelbukh A, Zhou Q (2016) Multilingual sentiment analysis: state of the art and independent comparison of techniques. Cogn Comput 8(4):757–771. doi: 10.1007/s12559-016-9415-7

Sobhani P, Mohammad S, Kiritchenko S (2016) Detecting stance in tweets and analyzing its interaction with sentiment. In: Proceedings of the 5th joint conference on lexical and computational semantics. doi:10.18653/v1/s16-2021

Poria S, Cambria E, Winterstein G, Huang G (2014) Sentic patterns: dependency-based rules for concept-level sentiment analysis. Knowl Based Syst 69:45–63. doi: 10.1016/j.knosys.2014.05.005

Socher R (2016) deep learning for sentiment analysis—invited talk. In: Proceedings of the 7th workshop on computational approaches to subjectivity, sentiment and social media analysis. doi:10.18653/v1/w16-0408

Turney PD, Mohammad SM (2014) Experiments with three approaches to recognizing lexical entailment. Nat Lang Eng 21(03):437–476. doi: 10.1017/s1351324913000387

Mohammad S, Kiritchenko S, Sobhani P, Zhu X, Cherry C (2016) SemEval-2016 task 6: detecting stance in tweets. In: Proceedings of the 10th international workshop on semantic evaluation (SemEval-2016). doi:10.18653/v1/s16-1003

Poria S, Chaturvedi I, Cambria E, Hussain A (2016) Convolutional MKL based multimodal emotion recognition and sentiment analysis. In: 2016 IEEE 16th international conference on data mining (ICDM). doi: 10.1109/icdm.2016.0055

Wang J, Yu L, Lai KR, Zhang X (2016) Dimensional sentiment analysis using a regional CNN-LSTM model. In: Proceedings of the 54th annual meeting of the association for computational linguistics, vol 2: short papers. doi:10.18653/v1/p16-2037

Download references

Authors’ contributions

JS made substantial contributions to conception and design, or acquisition of data, or analysis and interpretation of data. GS helped in revision and has given final approval of the version to be published. RS agreed to be accountable for all aspects of the work in ensuring that questions related to the accuracy or integrity of any part of the work are appropriately investigated and resolved. All authors read and approved the final manuscript.

Acknowledgements

This research was supported by Department of Computer Science, Guru Nanak Dev University, Amritsar. I thank Dr. Gurvinder Singh and Dr. Rajinder Singh for their participation in experimental work and their assistance to improve the manuscript.

Competing interests

This research work has non-financial Academic and intellectual competing interests.

Availability of data and materials

I submit that I can make the experimental data and materials available after the completion of my thesis.

Consent for publication

We hereby grant and assign all rights to Human-centric Computing and Information Sciences for publication.

Ethics approval and consent to participate

The article submitted is an original work and has neither been published in any other peer-reviewed journal nor under consideration for publication by any other journal.

Funding information

No funding was received from any funder.

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Author information

Authors and affiliations.

Department of Computer Science, Guru Nanak Dev University, Amritsar, India

Jaspreet Singh, Gurvinder Singh & Rajinder Singh

You can also search for this author in PubMed   Google Scholar

Corresponding author

Correspondence to Jaspreet Singh .

Rights and permissions

Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License ( http://creativecommons.org/licenses/by/4.0/ ), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made.

Reprints and permissions

About this article

Cite this article.

Singh, J., Singh, G. & Singh, R. Optimization of sentiment analysis using machine learning classifiers. Hum. Cent. Comput. Inf. Sci. 7 , 32 (2017). https://doi.org/10.1186/s13673-017-0116-3

Download citation

Received : 24 December 2016

Accepted : 04 September 2017

Published : 11 December 2017

DOI : https://doi.org/10.1186/s13673-017-0116-3

Share this article

Anyone you share the following link with will be able to read this content:

Sorry, a shareable link is not currently available for this article.

Provided by the Springer Nature SharedIt content-sharing initiative

  • Sentiment analysis
  • Social media text
  • Movie reviews
  • Product reviews
  • Machine learning classifiers

research paper on sentiment analysis using machine learning

Design Amharic Text Sentiment Analysis Model Using Machine Learning Techniques. In Case of Restaurant Reviews

Ieee account.

  • Change Username/Password
  • Update Address

Purchase Details

  • Payment Options
  • Order History
  • View Purchased Documents

Profile Information

  • Communications Preferences
  • Profession and Education
  • Technical Interests
  • US & Canada: +1 800 678 4333
  • Worldwide: +1 732 981 0060
  • Contact & Support
  • About IEEE Xplore
  • Accessibility
  • Terms of Use
  • Nondiscrimination Policy
  • Privacy & Opting Out of Cookies

A not-for-profit organization, IEEE is the world's largest technical professional organization dedicated to advancing technology for the benefit of humanity. © Copyright 2024 IEEE - All rights reserved. Use of this web site signifies your agreement to the terms and conditions.

Analysis of news sentiments using natural language processing and deep learning

  • Open access
  • Published: 30 November 2020
  • Volume 36 , pages 931–937, ( 2021 )

Cite this article

You have full access to this open access article

  • Mattia Vicari 1 &
  • Mauro Gaspari 2  

11k Accesses

13 Citations

1 Altmetric

Explore all metrics

This paper investigates if and to what point it is possible to trade on news sentiment and if deep learning (DL), given the current hype on the topic, would be a good tool to do so. DL is built explicitly for dealing with significant amounts of data and performing complex tasks where automatic learning is a necessity. Thanks to its promise to detect complex patterns in a dataset, it may be appealing to those investors that are looking to improve their trading process. Moreover, DL and specifically LSTM seem a good pick from a linguistic perspective too, given its ability to “remember” previous words in a sentence. After having explained how DL models are built, we will use this tool for forecasting the market sentiment using news headlines. The prediction is based on the Dow Jones industrial average by analyzing 25 daily news headlines available between 2008 and 2016, which will then be extended up to 2020. The result will be the indicator used for developing an algorithmic trading strategy. The analysis will be performed on two specific cases that will be pursued over five time-steps and the testing will be developed in real-world scenarios.

Similar content being viewed by others

research paper on sentiment analysis using machine learning

Artificial intelligence in Finance: a comprehensive review through bibliometric and content analysis

Salman Bahoo, Marco Cucculelli, … Jasmine Mondolo

research paper on sentiment analysis using machine learning

A systematic review of fundamental and technical analysis of stock market predictions

Isaac Kofi Nti, Adebayo Felix Adekoya & Benjamin Asubam Weyori

research paper on sentiment analysis using machine learning

Sentiment Analysis in Social Media Data for Depression Detection Using Artificial Intelligence: A Review

Nirmal Varghese Babu & E. Grace Mary Kanaga

Avoid common mistakes on your manuscript.

1 Introduction

Stock forecasting through NLP is at the crossroad between linguistics, machine learning, and behavioral finance (Xing et al. 2018 ). One of the main NLP techniques applied on financial forecasting is sentiment analysis (Cambria 2016 ) which concerns the interpretation and classification of emotions within different sources of text data. It is a research area revived in the last decade due to the rise of social media and cheap computing power availability (Brown 2016 ). Like products and services, market sentiments influence information flow and trading, thus trading firms hope to profit based on forecasts of price trends influenced by sentiments in financial news (Ruiz-Martínez et al. 2012 ). Is it possible to find predictive power in the stock market’s behavior based on them? It seems to be the case in the work “On the importance of text analysis for stock market prediction” by Lee and MacCartney ( 2014 ) that shows, based on text, an improved predictability in the performance of a security. Intuitively, the cause of the stocks’ fluctuation can be the aggregated behavior of the stockholders, who will act based on news (Xing et al. 2018 ). Although the predicting models reported in the literature have not been able to profit in the long run, many theories and meaningful remarks have been made from the financial markets’ data (Xing et al. 2018 ). One of the biggest differences between market sentiment problems and linguistics ones is that the ladder has some guarantee of having some type of structures (Perry 2016 ). There are many models that have been proposed and used in the recent years, each with their positive and negative aspects. Specifically, overly complicated models generally have poor performance, while simpler linear models rely on strong hypotheses, for example, a Gaussian distribution, which does not always apply in real-world cases (Xing et al. 2018 ). Deep learning seems to be the most fit for this purpose since it has the ability to analyze a great amount of data that NLP needs to understand context and grammatical structures. In this paper, we investigate this scenario, exploring DL for forecasting the market sentiment using news headlines. We assume a basic linguistic framework for preprocessing (stopword, lowercasing, removing or numbers and other special characters), and we reject the common assumption that “positive financial sentiment” = “positive words” and vice versa. The reason is that we do not know if that is the case, and we want the model to learn freely by itself. Taking this assumption into consideration, we are going to test two scenarios, both based on the news published today: case A tries to forecast the movement of the Dow Jones Industrial Average (DJIA) in the next four individual days; case B focuses on time intervals from today to the next 4 days. We discuss the obtained results, and we conclude the paper with a road map for the future.

2 Background

2.1 deep learning.

Supervised ML models create mechanisms that can look at examples and produce generalizations (Goldberg 2017 ). Deep learning (DL) Footnote 1 is a function that imitates the mechanisms of the human brain for finding patterns. Since our case is a binary classification problem (does the DJIA go up or down?), to test our results we used both a binary cross-entropy loss function:

And an accuracy metric: Footnote 2

They are both useful in different ways since the first is used in the training phase, while accuracy is intuitive as long as the classes considered are balanced, like in our case. The basic structure of a neural network (NN) is the neuron: it receives the signal, decides whether to pass the information or not, and sends it to the next neuron. Mathematically, the neuron structure takes some input values × and their relative weights w which are both initialized randomly, and thanks to an “activation function”, Footnote 3 which was the real game-changer (Thanaki 2018 ), the neurons have the ability to spot non-linear behaviors, that is why they are often addressed as being “Universal Function Approximators”. Since searching over the set of all the possible functions is a difficult task, we need to restrict our scope of action using smaller sets, but by doing so, we added an “inductive bias” (b) that must be taken into account. The most commonly used function for this purpose has the form (Goldberg 2017 ):

The parameters Footnote 4 have the purpose to minimize the loss function over the training set and the validation set (Goldberg 2017 ). How can NN learn? Through an optimization method that, thanks to a first-order iterative algorithm called gradient descent, can minimize the error function of our model until a local minimum is reached. Footnote 5 Backpropagation is the central mechanism of the learning process which calculates the gradient of the loss function and, by making extensive use of the chain rule, distributes it back through the layers of the NN and adjusts the weights of the neurons for the next iteration. The learning rate used during backpropagation starts with a value of 0.001 and is based on the adaptive momentum estimation (Adam), a popular learning-rate optimization algorithm. Traditionally, the Softmax function is used for giving probability form to the output vector (Thanaki 2018 ) and that is what we used. We can think of the different neurons as “Lego Bricks” that we can use to create complex architectures (Goldberg 2017 ). In a feed-forward NN, the workflow is simple since the information only goes…forward (Goldberg 2017 ). However, when humans read a book, for example, they comprehend every section, sentence or word taking into account what they saw previously (Olah 2015 ); therefore, a feed-forward NN is not fit for our purposes because it cannot “remember”. Recurrent neural networks (RNN), on the other hand, can catch the sequential nature of the input and can be thought of as multiple copies of the same network, each passing a message to a successor (Olah 2015 ). A well-known drawback of standard RNN is the vanishing gradients’ problem that can be dramatically reduced using, as we did, a gating-based RNN architecture called long short-term memory Footnote 6 (LSTM).

2.2 NLP and vectorization

Natural language processing (NLP) is a field of artificial intelligence (AI) focused on finding interactions between computers and the human language. Linguistics can be a slippery field for humans, and its intrinsic ambiguity is what makes NLP even more problematic for machines (Millstein 2017 ) since complexity can appear on many levels: morphological, lexical, syntactic, or semantic. Data preprocessing is a crucial method used to simplify raw text data. In NLP, words are the features Footnote 7 used to find sentiment, based on their frequency in a database (Velay and Daniel 2018 ). The goal of the language model that is to assign probabilistic sentiment to sentences by trying to capture its context, but to do so, the so-called Markov assumption Footnote 8 is necessary. We use encoding for creating word-embeddings, which are tools used to vectorize words into feature-vectors that a NN can use (Millstein 2017 ). Word-embeddings are representations of documents where vectors with small distances represent words with closely related meanings. These structures allow us to perform math on texts. With the recent advances in DL, word-embeddings are formed with more accuracy, and they make it easier to compute semantic similarities (Xing et al. 2018 ). Unfortunately, distributional methods can fall victim of different corpus biases, which can range from cultural to thematic: a common saying is that “Words are similar if used in similar contexts,” but linguistics is more complicated than it looks. Footnote 9 Each model has its pros and cons (Velay and Daniel 2018 ), the difference stays in the user’s ability to have control over its dataset (Goldberg 2017 ).

3 Learning trading indicators on news

A trading indicator is a call for action to buy/sell an asset given a specific condition. When it comes to short-term market behaviors, we are trying to profit on the investors’ “gut-feeling,” but since this phenomenon is something that cannot be unequivocally defined, we must reduce human judgment as much as possible by letting the algorithm learn directly. As shown in Fig.  1 , we will start by seeing the chosen dataset. Right after, we will analyze which preprocessing operations have been implemented to ease the computational effort for the model. Then we will see all the components of the DL model put in place and ultimately we will present the results with a real-case scenario.

figure 1

Model structure

When working with neural networks, we encounter some limitations that might affect our results (Thanaki 2018 ): the dimension of the data set, our computing-power availability and the complexity of our model. Just like in the work of Vargas et al. ( 2017 ), the dataset is based on news headlines; specifically from the DJIA Database which comes from Kaggle, which contains 25 daily news with economic content from 2008 to 2016 scraped by the author from the most upvoted by the community on Reddit WorldNews: https://www.kaggle.com/aaron7sun/stocknews/home .

Given the fact that the database stretches over almost a decade and contains, for each day considered, a conspicuous amount of news with inherent economic content, we decided that it represents a plausible research instrument. The database was labeled based on whether the DJIA increased or decreased over each time step considered.

4 The studied model

The focus is on aggregate market indicators and two cases are considered, namely cases A and B as shown, respectively, in Figs.  2 and 3 .

figure 2

The T0 event, common in both instances, analyzes if, based on the news published today, today’s Adjusted closing price is higher than today’s opening price. While, based on the news published today, case A tries to forecast the movement of the DJIA in individual days, case B focuses on time intervals. After defining these market indicators, the preprocessing phase is crucial to reduce the number of independent variables, namely the word tokens, that the algorithms need to learn. At this stage, the news strings need to be merged to represent the general market indicator, from which stopwords, numbers and special elements (e.g. hashtags, etc.) were removed. In addition, every word has been lowercased and only the 3000 most frequent words have been taken into consideration and vectorized into a sequence of numbers thanks to a tokenizer. Furthermore, the labels are transformed into a categorical matrix with as many columns as there are classes, for our case two. The NN Footnote 10 presented in Fig.  1 starts with an embedding layer, which is the input of the model, whose job is to receive the two-dimensional matrix and output a three-dimensional one, which is randomly initialized with a uniform distribution. Then this 3D-matrix is sent to the hidden layer made of LSTM neurons whose weights are randomly initialized following a Glorot Uniform Initialization, which uses an ELU activation function and dropout. Finally, the output layer is composed of two dense neurons and followed by a softmax activation function. Once the model’s structure has been determined, it needs to be appropriately compiled using the ADAM optimizer for backpropagation, which provides a flexible learning rate to the model.

As shown in Fig.  4 , the database is then divided into training and validation set with an 80/20 split and evaluated by the binary cross-entropy and accuracy metrics that we previously discussed. Moreover, the training set is split into small pieces called batches (which, for instance, have a dimension of 64 for the T0 case) that are given to the computer one by one for 25 iterations in the training set and two epochs to ease the computational effort when updating the weights.

figure 4

Structure of training, validation and testing sets with DJIA labels

Table 1 shows the level of accuracy obtained in this experiment relative to the validation sets:

Consistent with previous studies (Velay and Daniel 2018 ), we immediately notice that the accuracy is particularly low since, in both cases, our peaks stick around 58%, which is slightly higher than the flip of a coin.

Besides, in both versions of the model, the highest accuracy appears in the T0 case, behavior that suggests that forecasting attempts within shorter time periods should be preferred, confirming existing literature on the topic (Souma et al. 2019 ; Sohangir et al. 2018 ). Nevertheless, we can notice the tendency that the accuracy gradually decreases in case A, while case B shows a less evident decrease that ends with a final increase in T4. Given the low accuracy level, we asked ourselves: How would this model behave in a practical application? Thus, we chose specific news related to major political events that, from our perspective, might have affected the global markets. For this reason, the first testing case looks at major political events that might have caused relevant shifts in the balance of the world:

Start of Trump’s formal impeachment inquiry.

Large crowds of protesters gathered in Hong Kong.

Boris Johnson becomes prime minister.

Protests for George Floyd’s murder explode.

COVID-19 was declared a global pandemic by the WHO.

In the second testing category, we decided to look closer at Trump’s presidency. The events chosen were:

The Summit in Singapore between Trump and Kim Jong-Un.

Trump signs tariffs on steel and aluminum.

Trump formally announced US withdrawal from the Paris Agreements.

Trump signs a big tax cut that was beneficial to big corporations.

Trump starts the government shutdown to build the wall.

To be consistent with the training and validation set, we manually retrieved from the “News” section of Google 25 news focused on the date of every major event listed above. By comparing each prediction with how the DJIA behaved, we notice that the results are not significant because the results obtained did not show any consistent pattern. Therefore, we agree with Arora et al. ( 2019 ) when stating that financial information is extremely unpredictable and that the task of predicting stock movements remains open. For this reason, we thought it could be interesting to keep studying how such a model would behave over longer periods of time by feeding it with more data in the training phase. In fact, for image recognition experiments, DL learning is known for having better performance the more data it sees; therefore, we will follow the same path for financial forecasting. Thus, we decided to extend the original data set until August 2020 and test the same scenario using the same methodology. We retrieved 25 most upvoted daily news from the sub-Reddit/r/worldnews/, just like the author originally did, thanks to the data submissions available in the Pushshift data collections: https://www.files.pushshift.io/reddit/submissions/ .

To keep our results comparable, we kept the same NN structure as in the previous case. The results of the experiment using this extended data set in reported in Table 2 .

On the one hand, for the extended case A, the outcome is mixed and there is no added benefit to our initial model. On the extended case B, on the other hand, we notice an even worse forecasting performance. In addition, as in the previous test for individual news, the results obtained did not show any relevant pattern and are not significant. Unlike in the image-recognition field, which is known for improving its performance whenever a more extensive data-set is fed to the deep learning model, for our financial forecasting case we did not obtain the same performance improvement. Why despite increasing the dataset did we get worse results? We analyzed the datasets for the T0 case and the extended T0 case deeper.

The original dataset T0 gives us 2267 trading dates for the DJIA and the DJIA closed higher than the open on 1212 times and closed lower on 1055 days. This means that an algorithm which would guess “DJIA increase” every single day which would result in

True positives: 1212

True negatives: 0

False positives: 1055

False negatives: 0

At the same time, the extended T0 case shows us that the DJIA went up 1625 days and down 1367 times. Applying again an algorithm which would guess “DJIA increase” every single day would have these characteristics:

True positives: 1625

False positives: 1367

Therefore, we would have an accuracy of around 53% for the original T0 case and of around 54% for the extended T0 case, values that are both extremely close to the results we obtained through an optimized DL model. Thus, we presume that the algorithm did not learn anything other than the bias in the data. Further paths worth exploring would be testing DL models with different databases, that could be:

Using news more focused on a specific security or newspaper.

Testing cases right after the news comes out.

Exploiting a dataset with more news per day.

Exploring a wider variety of NLP methods in the preprocessing phase.

By removing the assumptions “positive financial sentiment” = “positive words”, forecasting sentiment becomes particularly difficult and an elaborated model such as DL the indicator appears weak (Sohangir et al. 2018 ). In fact, the closer we get into situations where human discretion is involved, the more unstable a model becomes (Goldfarb et al. 2018 ).

6 Discussion and future roadmap

Forecasting consists in grabbing data that we have to generate new information: this is a field that economics has been studying for years, and perhaps by combining the old logic of decision theory and DL methods, we may have a better understanding from a different perspective (Goldfarb et al. 2018 ). The benefit of DL is that it does not require lots of human effort to effectively provide outputs that were previously considered impossible and a human exclusive. Indeed, AI did not give us more intelligence, but more prediction capacity.

In fact, the innovation of this technology is not the math behind it that existed for decades, but the combination of this knowledge with the Big Data environments that we can access nowadays (Goldfarb et al. 2018 ). Economics says that by eliminating every obstacle to trading leads to market efficiency. Is it always true? Would the introduction of market agents using trading strategies based on DL and sentiment analysis increase market efficiency? The answer is “it depends”. By introducing a new investing technology we could indeed increase market liquidity and therefore stock markets’ efficiency as a whole. However, the stock market is a complex ecology of interacting players, all with their own strategies and things can go wrong very fast. Let us take a step back and analyze the situation with a few examples from the past that might help us better understand what to expect tomorrow (Buchanan 2012 ):

First, the belief that financial innovations such as derivatives could help us reach more market stability and efficiency dominated the last decades (Buchanan 2012 ). However, those instruments, passed a threshold, profoundly endangered the whole system in 2008 (Buchanan 2012 ).

Second, the innovation of High-Frequency Trading (HFT) has improved the stock market as well by reducing trading costs, enhance liquidity, making markets faster and more reactive in calm times, but they do the exact opposite in troubled ones; exactly when the market would need it the most (Buchanan 2012 ).

Third, hedge funds typically borrow money to attract investors and increase their profits. However, right before the “Quant meltdown” of August 2007 it was clear the strategies used became too similar, causing their margins to decrease. Therefore, to keep being appealing to investors, managers were slowly forced to increase leverage until it was unsustainable and everything collapsed (Buchanan 2012 ).

The belief that the world is becoming more efficient and stable than ever thanks to financial innovation (Buchanan 2012 ) can be misleading. Then why do we always seem surprised when such crises appear? (Buchanan 2012 ) It is not a coincidence that excessive efficiency can compromise stability (Buchanan 2012 ). Efficiency means doing more with less, while stability implies the opposite: some extra room to absorb a hit. Given the current hype for AI, these interconnected markets will have access to new ways to invest at incredible speed. In such a context, trading strategies based on sentiment may be particularly prone to the “herd behavior” assuming that these algorithms must train on datasets containing lots of news that may turn out to be extremely similar (Buchanan 2012 ). In this context, the hypothesis of a “splash crash” ranging across many asset classes does not seem impossible (Buchanan 2012 ). The global financial crisis has revealed the need to drastically rethink how we regulate the financial system to get ready for when the next recession will strike (Buchanan 2012 ). Innovation cannot be stopped, but acknowledging its limitations can help us find the best ways to reduce its weaknesses. The fight is rapidly shifting from humans towards machines, but they are playing with real businesses and people, and we must counteract effectively and rapidly in that regard (Buchanan 2012 ).

7 Conclusion

In conclusion, by removing the assumptions “positive financial sentiment” =  “positive words”, financial forecasting becomes particularly difficult and even a DL model does not show better results than the flip of a coin for any of the cases or time-intervals studied. We acknowledge that there are many other paths that could potentially improve our results, but increasing the database size over a longer period of time, common practice in image-recognition problems, does not look like one of them. Despite that, we believe that deep learning methods could potentially be the cause of a relevant danger, due to its proneness towards the “herd behavior”, for the financial system as a whole and thus must be handled carefully.

Also known as a deep neural network or deep neural learning.

Where TP means’True Positives’, TN means’True Negatives’, FP means’False Positive’ and FN means’False negatives’.

The Activation Function chosen is the Exponential Linear Unit (ELU) since it shows better performance than its Rectified Linear Unit (ReLU) predecessor Hochreiter et al. ( 2016 ) by reducing the vanishing gradient problem and by accepting negative values.

Where x is the input, W are the weights, b is the inductive bias. Right after are explained the dimensions of the vectors that have been enumerated. These are all cases that we will go over later on but, for example: x in our case is a vector whose dimension represents a matrix whose dimensions are represented by the product between the maximum length of the sentences and the random initialization created using the embedding layer for each batch. The notation named ‘ d out ’ represents the dimension of the following layer that will receive the output of the current one.

For the sake of completeness, it does not tell us the minimum error immediately but shows us the proper direction we need to follow Thanaki ( 2018 ).

To avoid overfitting of the LSTM structure, we adopted a common regularization method called "Dropout".

Features in a model based on AI are individual measurable properties used as input to obtain an output Thanaki ( 2018 ); therefore, any attribute can be a feature as long as it is measurable and useful.

Which states that the probability of the next word is calculated by looking at the previous one.

The clearest example is polysemy.

The number of neurons, dropout and epochs named below represent the common T0 case, but have been fine-tuned for each other case using the "babysitting method".

Arora A, Datta A, Ding V (2019) Using news titles and financial features to predict intraday movements of the DJIA. https://www.web.stanford.edu/class/archive/cs/cs224n/cs224n.1194/reports/custom/15744462.pdf . Accessed 30 Sept 2020

Brown ED (2016) Analysis of Twitter messages for sentiment and insight for useinstockmarketdecisionmaking. Amazon Digital Services LLC, Seattle

Google Scholar  

Buchanan M (2012) Forecast: what physics, meteorology and natural sciences can tell us about economics. Bloomsbury, London

Cambria E (2016) Affective computing and sentiment analysis. IEEE Intell Syst 31(2):102–107

Article   Google Scholar  

Goldberg Y (2017) Neural network methods for natural language processing. Morgan Claypool publishers, San Rafael

Book   Google Scholar  

Goldfarb A, Agrawal A, Gans J (2018) Prediction machines: the simple economics of artificial intelligence. Harvard Business Review Press, Cambridge

Hochreiter S, Clevert DA et al (2016) Fast and accurate deep network learning by exponential linear units. https://arxiv.org/abs/1511.07289

Lee S, MacCartney J (2014) On the importance of text analysis for stock price prediction. In: Language resources and evaluation conference (LREC)

Millstein F (2017) Natural language processing with python: natural language processing using NLTK. Packt Publishing, Birmingham

Olah C (2015) Understanding LSTM networks. http://www.colah.github.io/posts/2015-08-Understanding-LSTMs/ . Accessed 30 Sept 2020

Perry T (2016) Deep learning the stock market. https://www.medium.com/@TalPerry/deep-learning-the-stock-market-df853d139e02 . Accessed 30 Sept 2020

Ruiz-Martínez JM, Valencia-García R, García-Sánchez F (2012) Semantic-based sentiment analysis in financial news. In: Joint Proceedings of the 1st International Workshop on Finance and Economics on the Semantic Web (FEOSW 2012), 5th International Workshop on REsource Discovery (RED 2012) and 7th International Workshop on Semantic Business Process Management (SBPM 2012) in conjunction with 9th Extended Semantic Web Conference (ESWC 2012). CEUR. CEUR Workshop Proceedings. 862:38–51.  http://ceur-ws.org/Vol-862/FEOSWp4.pdf . Accessed 30 Sept 2020

Sohangir S, Wang D et al (2018) Big data: deep learning for financial sentiment analysis. J Big Data. https://doi.org/10.1186/s40537-017-0111-6

Souma W, Vodenska I, Aoyama H (2019) Enhanced news sentiment analysis using deep learning methods. J Comput Soc Sci 2:33–46

Thanaki J (2018) Python natural language processing: advanced machine learning and deep learning techniques for natural language processing. CreateSpace Independent Publishing Platform, South Carolina

Vargas MR, de Lima BSLP, Evsukoff AG (2017) Deep learning for stock market prediction from financial news articles. In: 2017 IEEE international conference on computational intelligence and virtual environments for measurement systems and applications (CIVEMSA), pp 60–65

Velay M, Daniel F (2018) Using NLP on news headlines to predict index trends. Lusis https://arxiv.org/abs/1806.09533

Xing FZ, Cambria E, Welsch RE (2018) Natural language based financial forecasting: a survey. Artif Intell Rev 50:49–73. https://doi.org/10.1007/s10462-017-9588-9

Download references

Acknowledgements

This paper has been inspired from my master thesis that can be found here: https://www.github.com/Mattia9494/Sentiment-analysis-with-AI .

Open access funding provided by Alma Mater Studiorum - Università di Bologna within the CRUI-CARE Agreement.

Author information

Authors and affiliations.

University of Bologna, Bologna, Italy

Mattia Vicari

Department of Computer Science and Engineering, University of Bologna, Bologna, Italy

Mauro Gaspari

You can also search for this author in PubMed   Google Scholar

Corresponding author

Correspondence to Mattia Vicari .

Additional information

Publisher's note.

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/ .

Reprints and permissions

About this article

Vicari, M., Gaspari, M. Analysis of news sentiments using natural language processing and deep learning. AI & Soc 36 , 931–937 (2021). https://doi.org/10.1007/s00146-020-01111-x

Download citation

Received : 29 October 2020

Accepted : 02 November 2020

Published : 30 November 2020

Issue Date : September 2021

DOI : https://doi.org/10.1007/s00146-020-01111-x

Share this article

Anyone you share the following link with will be able to read this content:

Sorry, a shareable link is not currently available for this article.

Provided by the Springer Nature SharedIt content-sharing initiative

  • Deep learning
  • Machine learning
  • Natural language processing
  • Trading signals
  • Sentiment analysis
  • Trading strategies
  • Find a journal
  • Publish with us
  • Track your research

IMAGES

  1. (PDF) Sentiment Analysis in Data of Twitter using Machine Learning

    research paper on sentiment analysis using machine learning

  2. Machine learning and lexicon-based approaches to sentiment analysis

    research paper on sentiment analysis using machine learning

  3. Architecture of Sentiment Analysis Using Machine Learning Algorithms

    research paper on sentiment analysis using machine learning

  4. Best Sentiment Analysis In Machine Learning In 2022

    research paper on sentiment analysis using machine learning

  5. Figure 1 from Drug Recommendation System based on Sentiment Analysis of

    research paper on sentiment analysis using machine learning

  6. Sentiment Analysis

    research paper on sentiment analysis using machine learning

VIDEO

  1. How to review machine learning paper, Sentiment Analysis (Af Soomaali)

  2. Natural Language Processing Based Sentiment Analysis using Machine learning with Deployment

  3. C16 YouTube comments sentiment Analysis using Machine Learning

  4. Text Classification|Sentiment Analysis Example|Python Libraries explanation|Naive Byes Classifiers

  5. guiding functional verification regression analysis using machine learning and big data methods

  6. sentiment analysis part 2

COMMENTS

  1. Sentiment analysis using machine learning: Progress in the machine

    Sentiment analysis using NLP based feed forward neural network. ... In this research paper, an effective machine learning based DFFNN is developed to classify the sentiments. The emotions and sentiments of a person identified from the comments and other activities on social media. The retrieved sentiments plays significant role in the promotion ...

  2. (PDF) Sentiment Analysis: Machine Learning Approach

    Sentiment analysis is a growing area of research in NLP which aims to investigate subjective opinions (Pang and Lee, 2008). Machine learning-based and semantic-based approaches are the two main ...

  3. Machine learning and deep learning for sentiment analysis across

    (b) The main approach is based on machine learning or deep learning. (c) The research studies sentiment analysis either on sentences or documents written in a single language. Finally, 44 studies were included in the state-of-the-art against the two main approaches of sentiment analysis across languages, especially CLSA and MSA.

  4. Optimization of sentiment analysis using machine learning classifiers

    This paper exploits four machine learning classifiers for sentiment analysis using three manually annotated datasets. The mean of 29 epochs of experimentation recorded in Table 4 shows that OneR is more precise in terms of percentage of correctly classified instances. On the other hand, Naïve Bayes exhibits faster learning rate and J48 reveals ...

  5. Sentiment analysis using deep learning techniques: a ...

    With the exponential growth of social media platforms and online communication, the necessity of using automated sentiment analysis techniques has significantly increased. Deep learning techniques have emerged in extracting complex patterns and features from unstructured text data, which makes them a powerful tool for sentiment analysis. This research article presents a comprehensive review of ...

  6. Sentiment Analysis Using Machine Learning Algorithms

    Abstract and Figures. This work aims at building a classifier able of predicting the polarity of a comment while using Machine Learning (ML) algorithms. Our work is essentially divided into three ...

  7. SENTIMENT ANALYSIS USING MACHINE LEARNING: A REVIEW

    Abstract. Sentiment analysis is a field of Research under analytics. This can use computational techniques by reading raw data to make sense. This is called analysis. By using sentiment analysis ...

  8. Classification, detection and sentiment analysis using machine learning

    These findings show how machine learning algorithms can categorise news items based on their nature of the content, identify false news, and gauge the mood of media stories. 7. Conclusion and future work. The work has effectively shown that machine learning algorithms may be used for news classification, fake news detection, and sentiment analysis.

  9. Sentiment Analysis using Machine Learning and Deep Learning

    In this paper, we have taken into account various methods to perform sentiment analysis. Sentiment Analysis has been performed by using Machine Learning Classifiers. Polarity-based sentiment analysis, and Deep Learning Models are used to classify user's tweets as having 'positive' or 'negative' sentiment.

  10. Sentiment Analysis by Using Supervised Machine Learning and Deep

    Sentiment analysis is a popular text classification task in the data mining domain. In this research, we develop a text classification machine learning solution that can classify customer reviews into positive or negative class by predicting the overall sentiment of the review text on a document level. For this purpose, we carry out the ...

  11. Systematic reviews in sentiment analysis: a tertiary study

    Cambria state that approaches for sentiment analysis and affective computing can be divided into the following three categories: knowledge-based techniques, statistical approaches (e.g., machine learning and deep learning approaches), and hybrid techniques that combine the knowledge-based and statistical techniques.Sentiment analysis models can adopt different pre-processing methods and apply ...

  12. A Literature Review on Application of Sentiment Analysis Using Machine

    This paper presents the common techniques of analyzing sentiment from a machine learning perspective. In light of this, this literature review explores and discusses the idea of Sentiment analysis by undertaking a systematic review and assessment of corporate and community white papers, scientific research articles, journals, and reports.

  13. Survey of Sentiment Analysis Using Deep Learning Techniques

    This paper presents a detailed review of deep learning techniques used in Sentiment Analysis. Sentiment analysis is one of the most researched areas in natural language processing. Natural language processing has a wide range of applications like voice recognition, machine translation, product review, aspect oriented product analysis, sentiment analysis and text classification like email ...

  14. An empirical research on sentiment analysis using machine learning

    Their reviews on social media have a significant impact on customers for making effective and optimal decisions for buying products or using services. In sentiment analysis, most of the used approaches are based on machine learning techniques. In this paper, the well-known methods of machine learning are reviewed and compared against each other.

  15. Sentiment analysis using Twitter data: a comparative application of

    Consequently, applying machine learning approaches to sentiment analysis can accurately extract text features without being restricted by lexicons. It is important to note that this paper only collects the opinions of people in England on Twitter about Covid-19; thus, the result should be interpreted by considering this limitation.

  16. Sentiment analysis using machine learning for business intelligence

    Abstract: This paper suggests use of sentiment analysis classification as an effective method for examining textual data coming from variety of resources on internet. Sentiment analysis is a method of data mining that evaluates textual data consuming machine learning techniques. Due to tremendous expanse of opinions of users, their reviews, feedbacks and suggestions available over the web ...

  17. A Literature Review on Application of Sentiment Analysis Using Machine

    Sentiment analysis is a technique of machine learning that senses polarities such as positive or negative thoughts within the text, full documents, paragraphs, lines, or subsections. Machine ...

  18. Movie recommendation and sentiment analysis using machine learning

    How can machine learning help movie lovers find their favorite films and express their opinions? This article presents a novel approach to movie recommendation and sentiment analysis using various techniques such as collaborative filtering, content-based filtering, and natural language processing. The article also discusses the challenges and future directions of this research field.

  19. A review on sentiment analysis and emotion detection from text

    This review paper provides understanding into levels of sentiment analysis, various emotion models, and the process of sentiment analysis and emotion detection from text. ... Rai P (2019) Sentiment analysis using machine learning classifiers: evaluation of performance. In: Proceedings of the 2019 IEEE 4th international conference on computer ...

  20. PDF Sentiment Analysis of Text Using Machine Learning Models

    In this section we cite the various relevant past literature that use the sentiment analysis and machine learning technique. 1. Ali Hasan, Sana Mohin et al [1] their paper includes a hybrid approach for doing sentiment analysis. Hybrid approach includes sentiment analyzer and machine learning technique. This paper also provides comparison of

  21. (PDF) Sentiment Analysis Using Deep Learning

    This paper aims to study the efficiency of various seq2seq deep learning architectures for the solution of toxic speech classification and performing efficient sentiment analysis using unilingual ...

  22. A survey on sentiment analysis methods, applications, and challenges

    The rapid growth of Internet-based applications, such as social media platforms and blogs, has resulted in comments and reviews concerning day-to-day activities. Sentiment analysis is the process of gathering and analyzing people's opinions, thoughts, and impressions regarding various topics, products, subjects, and services. People's opinions can be beneficial to corporations, governments ...

  23. Design Amharic Text Sentiment Analysis Model Using Machine Learning

    Sentiment analysis is a type of natural language processing for tracking the attitude of the public about a particular product, service or topic. It is also highly challenging as natural language processing research topic, and covers many novel sub-problems. Now business organizations and academics are putting forward their efforts to find the best system for sentiment analysis. The focus of ...

  24. Survey on sentiment analysis: evolution of research methods ...

    Sentiment analysis, one of the research hotspots in the natural language processing field, has attracted the attention of researchers, and research papers on the field are increasingly published. Many literature reviews on sentiment analysis involving techniques, methods, and applications have been produced using different survey methodologies and tools, but there has not been a survey ...

  25. Sentiment Analysis on News Headlines of Nation's ...

    Furthermore, this paper compares the performances of various machine learning approaches namely, Random Forest Classifier, Logistic Regression, XGBoost classifier, Support Vector Machine and ...

  26. Analysis of news sentiments using natural language ...

    Stock forecasting through NLP is at the crossroad between linguistics, machine learning, and behavioral finance (Xing et al. 2018).One of the main NLP techniques applied on financial forecasting is sentiment analysis (Cambria 2016) which concerns the interpretation and classification of emotions within different sources of text data.It is a research area revived in the last decade due to the ...