Next Article in Journal
A Review of Tabular Data Synthesis Using GANs on an IDS Dataset
Previous Article in Journal
The Mitigators of Ad Irritation and Avoidance of YouTube Skippable In-Stream Ads: An Empirical Study in Taiwan
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Tweet Sentiment Classification Approach Using a Hybrid Stacked Ensemble Technique

School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing 100083, China
*
Author to whom correspondence should be addressed.
Information 2021, 12(9), 374; https://doi.org/10.3390/info12090374
Submission received: 20 August 2021 / Revised: 6 September 2021 / Accepted: 8 September 2021 / Published: 14 September 2021

Abstract

:
With the extensive availability of social media platforms, Twitter has become a significant tool for the acquisition of peoples’ views, opinions, attitudes, and emotions towards certain entities. Within this frame of reference, sentiment analysis of tweets has become one of the most fascinating research areas in the field of natural language processing. A variety of techniques have been devised for sentiment analysis, but there is still room for improvement where the accuracy and efficacy of the system are concerned. This study proposes a novel approach that exploits the advantages of the lexical dictionary, machine learning, and deep learning classifiers. We classified the tweets based on the sentiments extracted by TextBlob using a stacked ensemble of three long short-term memory (LSTM) as base classifiers and logistic regression (LR) as a meta classifier. The proposed model proved to be effective and time-saving since it does not require feature extraction, as LSTM extracts features without any human intervention. We also compared our proposed approach with conventional machine learning models such as logistic regression, AdaBoost, and random forest. We also included state-of-the-art deep learning models in comparison with the proposed model. Experiments were conducted on the sentiment140 dataset and were evaluated in terms of accuracy, precision, recall, and F1 Score. Empirical results showed that our proposed approach manifested state-of-the-art results by achieving an accuracy score of 99%.

1. Introduction

Advances in internet technology and the continuous development of web 2.0 is resulting in the production of a substantial amount of data daily. The availability of a plethora and variety of social media platforms increased the connectivity among social media users which changes the prevalent viewpoint of socialization, personalization, and networking. For the fourth quarter of 2020, an estimated number of 1.8 billion users were active on Facebook each day [1]. This is in addition to Facebook ancillary services like Instagram, WhatsApp, and messenger each of which have active users amounting to 1 billion on a monthly basis [2]. Similarly, according to third-party analysis, other social media platforms such as iMessage owned by Apple, WeChat by Tencent, and YouTube by Google, which is no longer exclusive, are now members of the 1 billion-per-month-active-user-club. Furthermore, 75% of internet users now regularly use at least one social media platform [3]. From a purely technical standpoint, increased accessibility has provided new opportunities and challenges by encouraging users to share their views, emotions, and opinions, in addition to consuming services [4,5]. One of the fast-growing and impactful social media networks is Twitter, on which users can read, post, and update short text messages termed as ‘tweets’ which enable Twitter users to communicate their views, opinions, and sentiments about a particular entity. These sentiment-bearing tweets play a vital role in many areas, for instance, social media marketing [6], academics [7], and election campaign news [6].
Sentiment analysis aims at categorizing and determining the polarity of a subjective text at phrase, sentence, or document level [8]. It has a variety of applications in various fields, including e-commerce, politics, entertainment, and health care to name a few. For instance, sentiment analysis can help companies to track customer perceptions about their products, it can also assist customers in selecting the best product based on public opinion [9]. While sentiment analysis has various applications and implementations in a variety of fields, it comes with diverse issues and challenges related to Natural Language Processing (NLP). Recent research related to sentiment analysis is still afflicted by technical and theoretical complexity which limit its inclusive accuracy in sentiment detection [10]. Hussein et al. [11] investigated the challenges of sentiment analysis and their effects on the accuracy of the results. The experimental results substantiated that accuracy is a considerable challenge in the conduction of sentiment analysis. It also demonstrates that accuracy is affected by several challenges, such as handling sarcasm, negation, abbreviations, domain dependence, bi-polar words, etc.
As for sentiment analysis of tweets, the key task is to classify the divergence between opinion-bearing tweets as either negative or positive. Sentiment analysis of tweets comes with its challenges. People tend to use informal language in their tweets, which might heighten the risk of not detecting the overall sentiment of the text [12]. Some tweets are short text, which may carry little contextual information which provides inadequate indications of sentiment [13]. Understanding the sentiments of tweets containing acronyms and abbreviations is also an immense challenge for the computer. Attributable to these challenges, a growing interest of researchers is classifying the sentiments of tweets.
Sentiments of tweets can be investigated using three approaches: (I) a machine learning (ML) approach which utilizes learning models for classification of sentiments; (II) a rule-based approach which uses either corpus-based sentiment lexicons, publicly available sentiment lexicons, or lexical dictionary for extraction of sentiments; and (III) a hybrid approach which combines the ML approach and rule-based approach. Consequently, deep learning approaches, as integrated into a large number of researches, have shown their significance in sentiment analysis [14], computer vision [15], and speech recognition [16]. Consequently, the authors of [17] showed that integrating ConvBiLSTM, a deep learning model, produced more effective and robust results in analyzing sentiments of tweets. They utilized a convolution neural network (Conv) for extraction of local features and bidirectional long short-term memory (BiLSTM) to capture the long-distance dependencies and classified tweets with 91.13% accuracy. In line with the above, another research integrated deep learning models in the sentiment classification of tweets which resulted in more accurate results as compared to conventional ML models [18].
The Sentiment140 dataset is utilized in this study which contains 1.6 million tweets, among which 800,000 tweets are negative tweets and 800,000 tweets are annotated as positive. Tweets in this dataset were originally annotated by considering the emoticons, for instance, tweets with happy emoticons were considered as positive and tweets containing sad emoticons were considered negative [19]. This study proposes that annotating tweets using a lexical dictionary generates more correlated features for more accurate sentiment classification of tweets. For this purpose, the current study proposes a novel approach that leverages the benefits of a lexical dictionary from the rule-based approach and learning models from the ML approach for the sentiment classification of tweets.
Key contributions of this study are summarized as follows:
  • This study explores the viability of the implementation of a lexical dictionary and evaluates the potency of a stacked ensemble for the sentiment classification of tweets.
  • A lexical dictionary, namely TextBlob, is integrated for sentiment annotation of tweets. TextBlob returns a float value within a range of “+1.0” and “−1.0” which represents the sentiment orientation of the text. Here, “+1.0” corresponds to positive, and “−1.0” corresponds to negative sentiments. We set the threshold value to “0” which indicates that output values greater than “0” will be regarded as positive tweets and vice versa.
  • Three feature engineering approaches are integrated and evaluated in this study including term frequency-inverse document frequency (TF-IDF), bag of words (BOW), and a union of BOW and TF-IDF.
  • A novel stacked ensemble of the ML model, logistic regression (LR), and a deep learning model, LSTM, is proposed for sentiment classification of tweets. LR works best with binary classification tasks; on the other hand, LSTM is the best choice for remembering the long-term dependencies of larger datasets. Thus, the proposed stacked ensemble harnesses the proficiency of combining the predictions made by three LSTMs as base learners using LR as a meta learner.
  • A diverse range of experimentation is carried out in this study to compare the performance of the proposed approach with conventional state-of-the-art ML models including random forest (RF), AdaBoost (ADB), and logistic regression (LR). Moreover, this study also compares the performance of models using original sentiments of tweets with sentiments extracted by TextBlob.
  • We also compare the performance of our proposed approach with correlated studies carried out on the sentimnet140 dataset for the sentiment classification of tweets.
The remainder of the paper is organized as follows: Section 2 explores sentiment analysis-related work which gives a brief description of previous studies. Section 3 briefly describes the dataset along with preprocessing techniques utilized to create clean data. It also explains the techniques and algorithms utilized in this research to conduct experiments. Section 4 presents a detailed discussion and analysis of the results. Section 5 is comprised of the conclusion and future direction.

2. Related Work

In the field of text classification, there is a wide scope for analyzing sentiments, and many researchers have studied the mechanism of sentiment analysis by identifying emotions contained in the text [20,21]. Ankit and Saleena [22] carried out Twitter sentiment analysis by integrating an ensemble of Naïve Bayes (NB), Random Forest (RF), Support Vector Machine (SVM), and Logistic Regression (LR) models with BOW as a feature extraction technique. The authors proposed a two-fold study in which they first predicted the sentiment score of the tweet and, in the second phase, they predicted the polarity of the tweet based on sentiment score. They utilized four datasets including sentiment 140, HCR (Health Care Reforms), the Frist GOP debate Twitter Sentiment dataset, and Twitter sentiment analysis dataset for analysis of the proposed approach. The results showed that the proposed ensemble learning classifier performs better than the stand-alone classifiers.
Onan et al. [23] proposed a multi-objective weighted voting ensemble classifier for text sentiment classification. Their proposed system incorporates Bayesian Logistic Regression (BLR), Linear Discriminant Analysis (LDA), NB, LR, and SVM as base learners whose performance in terms of sensitivity and security determines the weighted adjustment. Different classification tasks which include sentiment analysis, software defect prediction, spam filtering, credit risk modeling, and semantic mapping suggest that their proposed system outperforms the conventional ensemble learning models. The highest accuracy of 98.86% is achieved in the software defect detection task on a dataset containing details of laptops.
Rustam et al. [24] proposed a voting classifier (VC) for the sentiment analysis of tweets. VC comprises logistic regression (LR) and an SGDC (stochastic gradient descent classifier) which produces prediction under soft voting. In their study, they classified the tweets into three classes (positive, negative, and neutral). Different ML classifiers were also tested on the “twitter-airline-sentiment” dataset. Their study investigated the role of feature extraction techniques like TF, TF-IDF, and word2vec on classification accuracy. LSTM, a deep learning model, was also used and it achieved an accuracy lower than ML models. The accuracy achieved by the voting classifier is 78.9% and 79.1% with TF and TF-IDF feature extraction.
Umer et al. [25] conducted sentiment analysis of tweets using an ensemble of a Convolutional Neural Network (CNN) and LSTM. As an ML classifier does not perform well on the vast amount of data, to overcome this limitation, they advised use of a Deep Learning-based ensemble system. They evaluated their proposed approach on three different datasets. They integrate feature extraction methods such as word2vec and TF-IDF. Results showed that the CNN-LSTM achieved higher accuracy than other classifiers. They also compared the performance of the (CNN-LSTM) proposed model with the other deep learning models which authenticated the proposed approach.
Stjanovski et al. [26] used the deep CNN approach to perform experiments on sentiment analysis on Twitter data. The proposed CNN was trained on the top most pre-trained word embeddings derived from large text corpora using unsupervised learning, which was further used with the dropout layer, softmax layer and two fully connected layers, and multiple varying windowed filters. The results show that the pre-trained word vectors are very effective on Twitter corpora for the unsupervised learning phase. They used the Twitter 2015 dataset and achieve an F1 Score of 64.85%.
Jianqiang et al. [27] suggested a deep learning-based system to classify tweets into negative and positive sentiments. The authors named the system global vector (Glove) depth CNN (DCNN). For sentiment features, the authors concatenated the pre-trained N-gram features and word embedding features as feature vectors. Moreover, they captured contextual features by using a recurrent structure and used CNN for the representation of text. Their proposed system achieved the highest accuracy of 85.97% on the STSGd dataset.
Santos et al. [28] recommended a deep convolutional neural network that uses character level to sentence level information to deploy sentiment classification for short texts. They used two datasets in their study; the maximum accuracy they have achieved was 86.4% on the ST’s corpus.
Ishaq et al. [29] advocated a deep neural network-based model for hotel review sentiments given by the guests of the hotel. The authors evaluated their proposed approach in terms of binary class classification and multi-class classification including 3 classes and 10 classes. The results showed that a maximum accuracy of 97% is achieved by LSTM on binary class classification.
Sentiment classification using deep learning models is highly impacted by the structure of the data under consideration. In this regard, three CNN-based and five RNN-based deep neural networks were employed and compared in a study to exploit significant implications for the development of a maximized sentiment classification system. The study concluded that, the larger the training data size, the higher the accuracy of the model. They also investigated the character-level and word-level input structure of the data on the models which showed that a word-level structure makes the model learn the hidden patterns more accurately as compared to the character-level structure of input data [30].
Consequently, a hybrid sentiment classification model leveraging the benefits of word embedding techniques along with deep learning models is proposed in a study [31]. The authors combined the FastText embedding with character embedding which are fed as an input to the proposed hybrid of CNN and BiLSTM which achieved the highest accuracy score of 82.14%.
Another study investigated the deep learning model CNN-LSTM for Twitter sentiment analysis [32]. Their method first utilized unlabeled data to pre-train word embeddings with the subset of data, along with distant supervision and fine-tuning. Their proposed system is based on the number of ensembles of CNN and LSTM networks used in the classification of the tweets. They used the SemEval-2017 twitter dataset for evaluation of the proposed approach. Using an ensemble of 10 CNN and 10 LSTM networks, they achieved an accuracy of 74.8%.

3. Materials and Methods

This study aims to classify the sentiments of tweets by proposing a two-fold method. It first focuses on extracting the sentiments of the tweets using a lexical dictionary, and then it classifies the tweets into positive and negative. The proposed approach involves the usage of various techniques which are briefly described in the following section.

3.1. Dataset

Sentiment140 is the dataset utilized for carrying out the diverse range of experiments in this study and was acquired from Kaggle which is a public repository for benchmark datasets [33]. The dataset consists of 1.6 million tweets which were extracted using Twitter search API. It is a well-balanced dataset comprising 0.8 million positive and 0.8 million negative tweets. The tweets in this dataset are labeled as 0 which corresponds to negative sentiment and 4 which corresponds to positive sentiment. Manually annotating the tweets would have been a labor-intensive and time-consuming task due to the quantity. The authors of the dataset annotated the tweets by considering the emoticons’ noise for the prediction of the tweet as positive or negative. The dataset consists of six features which are listed and described in Table 1. A few samples of tweets from the dataset are shown in Table 2.

3.2. Data Pre-Processing

Data extracted from online platforms are largely unstructured or semi-structured such that they contain unnecessary data which are insignificant for the analysis. This makes pre-processing of the data a very important step in cleaning the data of redundant and noisy material. It also impacts the performance of the ML models as stated by the authors of [24]. An effective pre-processing method can reduce the size of the featured set extracted from the dataset from 30% to 50%, which leaves behind only significant features which are highly correlated with the target value. Furthermore, a large dataset necessitates more time for training, and stop words, punctuations, numeric values, and data that are not correlated with the analysis reduce the accuracy of the prediction. Thus, pre-processing is essential to save computing resources and makes it easier for the ML models to train more effectively, providing precise prediction [34].
Pre-processing of data integrates various steps such as data cleaning which involves removing numbers, usernames, punctuations, stop words, lower case transformation, and stemming which are described as follows:

Data Cleaning

Numbers and punctuation do not impact the sentiment of the tweet thus making them unnecessary for sentiment classification of tweet [35]. Similarly, usernames are nouns that are not relevant for the sentiment classification of text. Along the lines of this, stop words which refer to the most common words in the text, only add computer overhead and are not valuable for the text analysis. In this step, we removed numbers, punctuation, usernames, and stop words from the tweets.
Lower Case Transformation
After the cleaning of unnecessary data, the text tweets are converted to lower case. Machine learning models are case sensitive; therefore, the same words with the upper or lower case will be considered as different words. For instance, “Boy” and “boy” will be treated as two different words by statistical models as they will count the occurrence of each word separately [36]; this impairs the efficiency of the classifier if case normalization is not carried out.
Stemming
Stemming involves the conversion of words into their root forms by deleting affixes from the words [24]. For instance, small, smaller, and smallest are variations of the root word “small” with the same meaning. By the process of stemming, the complexity of the textual feature is reduced, which enhances the learning ability of the classifier. Sample tweets from the dataset before and after pre-processing are illustrated in Table 3.

3.3. TextBlob

TextBlob is a publicly available lexical dictionary that offers a simple API for carrying out natural language processing (NLP) tasks [37]. It is a python library that integrates two modules such as NaiveBayesAnalyzer (a classifier trained on a corpus of movie reviews) and PatternAnalyzer (integrates pattern libraries). In this study, we incorporated the PatternAnalyzer for carrying out sentiment annotation tasks. TextBlob outputs float values of the sentiment score of the text along with its subjectivity score. In this study, we only included sentiment scores in our experiments. Sentiment scores range from +1.0 to −1.0, where +1.0 refers to positive sentiment and −1.0 refers to negative sentiment. The authors of [38] integrated TextBlob in carrying out sentiment classification of user reviews. The study concluded that the feature set generated by TextBlob boosted the performance of ML models.
TextBlob extracted a total of positive tweets and the rest were regarded as negative tweets by TextBlob. Table 4 shows the number of positive and negative tweets predicted by TextBlob in comparison to the original sentiments. We used positive and negative tweets for experiments, as we discard neutral tweets to make the dataset binary and make a fair comparison of models on the original dataset. Table 5 shows the sample tweets with their corresponding original sentiment and sentiments predicted by TextBlob.

3.4. Feature Extraction Techniques

Feature extraction is a method of developing meaningful features or vectors from the textual data for ML models to understand more effectively [39]. As stated by Heaton [40], feature extraction can uplift the efficacy of the ML models. The current study incorporates three feature extraction techniques including TF-IDF, BOW, Feature union (TF-IDF + BOW) which are described as follows:

3.4.1. Bag of Words (BOW)

BOW is a feature extraction technique that models textual data by describing the frequency of occurrence of words in a document. BOW involves two main things, including a dictionary of familiar words and a measure of the existence of those words in the document irrespective of their place in the text. The vocabulary size of BOW is depended on the number of words in the document. It is a collection of words and features, where every feature is assigned a value that represents the occurrence of that feature [41].

3.4.2. Term Frequency Inverse Document Frequency (TF-IDF)

This is a feature extraction technique that is used to extract features from data. TF-IDF is widely used for Text analysis and music information retrieval [42]. In TF-IDF each term in the document is given a weight based on its term frequency (TF) and inverse document frequency (IDF) [24,43]. The term with the highest weight score is considered an important term [44]. It computes the weight of each term by using the formula below:
W i , j = T F i , j ( N D f , t )
Here N is the total number of documents in the corpus, D f , t is the number of documents containing the term t, and T F i , j is the number of occurrences of term t in a document d.

3.4.3. Feature Union

Feature union concatenates the feature sets extracted by several feature extraction techniques resulting in a larger feature set for the training of machine learning classifiers. In this study, the union of TF-IDF and BOW feature sets is integrated into a single feature set on the criteria of end-to-end concatenation. The resulting concatenated set provides the advantage of autonomous fitting of data to each concatenated vectorizer, which in this study is CountVectorizer() and TfidfVectorizer(). The feature union can be computed in this study as:
( T F I D F + B O W ) = ( F e a t u r e s e t o f B O W ) U ( F e a t u r e s e t o f T F I D F )

3.5. Machine Learning Models

Three ML models are integrated in this study to conduct sentiment classification of tweets including RF, ADB, and RF. ML models undergo the training and testing phase to carry out classification tasks. Training of these models is conducted by integrating different hyper-parameter settings for each model which were optimized under the criteria of the hit and trial method. The optimized hyper-parameter for each model is shown in Table 6.

3.5.1. Random Forest (RF)

RF is a tree-based classifier in which every tree is the result of a random vector from the input vector. Firstly, RF develops a forest by producing multiple decision trees on random features. Then it aggregates voting from all decision trees to predict the class labels for test data. Votes from a decision tree with a low error rate are given a higher weight and vice versa. By using decision trees with low error rates, this intern lowers the chances of wrong prediction [45]. In this study, we accumulated a random state of 42 for the bootstrapped samples. We also integrated 300 n_estimators which refer to the number of trees generated in the forest with each tree having a max_depth of 300 as shown in Table 6.

3.5.2. Logistic Regression (LR)

LR is a statistical-based classifier that is mostly used for the analysis of binary data in which one or more variables are used to find the results. It is also used for the evaluation of the probability of class association [46]. LR gives better results when the target class is categorical. It produces the affiliation among the categorical dependent variable and one or more independent variables by approximating probabilities using an LR sigmoid function. A logistic function or logistic curve is a common "S" shaped slope or sigmoid curve as, illustrated in Equation (3).
σ ( x ) = 1 1 + e x
where,
  • σ (x) = output in the range of 0 and 1;
  • x = input;
  • e = base of nature log.
While conducting the experiments, the random state for LR was set to “42” with the “liblinear” solver as it handles L1 regularization. Since this study is concerned with binary classification, multi_class is therefore set to “ovr” and “C” is set to “3.0” for more optimized results as shown in Table 6.

3.6. AdaBoost (ADB)

The Adaboost classifier is an ensemble learning approach that uses a boosting technique for the training of weak learners (decision trees). ADB is an acronym for adaptive boosting. ADB is very significant and popular as it pioneers an algorithm that could adapt to weak learners [47]. ADB combines the number of “weak learners” and trains them recursively on duplicates of the original data set, while all comparative weak learners focus on the difficult data outliners or data points [48]. Like the metadata model, it takes N copies of weak learners and trains them on the same feature set but with different weights assigned to them. The major difference between ADB and RF is that ADB uses the boosting method while RF uses the bagging method and ADB is exactly the weighted combination of N weak learners. In ADB high weighted data points are used to identify the outlier and the same function is performed by the gradient boosting algorithm using gradients in the loss function [48]. Hyperparameters of ADB were optimized as shown in Table 6. The random state is set to 42 with 300 maximum depth and 300 maximum estimators at which the learning procedure is terminated.

3.7. Long Short-Term Memory (LSTM)

LSTM is a deep learning model which is a variant of a recurrent neural network (RNN) with the capability of preserving information for a long period, which assists in back-propagation [49]. LSTM is comprised of memory units called cells which allow information to be preserved, edited, and updated. The cell determines the information that is needed to be stored, updated, or removed via gates which open and close depending upon the signal received by the cell. Each LSTM is initiated with an embedding layer of 300 input length with ReLU and sigmoid as an activation function, which enables the model to learn complex data. To reduce the complexity and overfitting of the proposed model, neurons are removed randomly with a dropout rate of 0.2. This study corresponds to binary classification; thus, the loss function for LSTM is set to “binary- cross-entropy” and am "adam" optimizer is used to handle the complex problem such as categorization of tweet’s sentiments. Each LSTM is trained on 30 epochs.

3.8. Proposed Framework of LR-LSTM Model

This study focuses on the sentiment classification of tweets by integrating a stacked approach to construct an ensemble of LR and LSTM. Stacking is an ensemble of heterogeneous base learners and a meta learner which uses the output prediction of base learners as an input and then produces final predictions [50]. Each base-learner is trained similarly as k-fold cross-validation where each fold consisted of m/k number of training samples where m is the number of total records in the dataset and k is the number of folds. Training of base learners is carried out on k − 1 folds, whereas one-fold is used for validation. Base learners produce n number of predictions for each instance of data for m-fold which results in an m/k × n matrix. Afterwards, the meta learner is trained on this matrix and makes final predictions. The proposed stacked ensemble model integrates 3 LSTMs as base-learners, which will create individual predictions on the training data. These predictions will be treated as training data for the meta learner. The architecture of the proposed LR-LSTM model is shown in Figure 1.

3.9. Proposed Methodology

This study aims at investigating the sentiment of tweets by integrating a lexical dictionary along with a stacked ensemble model. Dataset “Sentiment140” is utilized for the evaluation of the proposed approach. It consists of 1.6 million tweets among which 50% are positive and 50% are negative tweets. The tweets in this study are reannotated using TextBlob which resulted in positive and negative tweets which are further compared with the original sentiments of the tweets. The comparison shows that TextBlob annotated the tweets with more efficacy as compared to the original sentiment annotations. Afterward, the data are preprocessed to transform the raw data into useful data by removing data that are irrelevant for the sentiment analysis. Preprocessed data are then split into training and testing sets with a 70:30 ratio. The proposed LR-LSTM model is then trained on the training set and evaluated on the testing set in terms of accuracy, precision, recall, and F1 Score. The proposed methodology is illustrated in Figure 2.

3.10. Performance Evaluation Criteria

Evaluation parameters are used to evaluate the performance of models including precision, F1 Score, recall, and accuracy [51]. These are the commonly used evaluation metrics.

3.10.1. Accuracy

Accuracy is the measure of correctly predicted instances from total instances. It has the highest value of 1 and lowest value of 0 and is calculated by the following formula:
Accuracy = N u m b e r o f c o r r e c t p r e d i c t i o n s T o t a l n u m b e r o f p r e d i c t i o n s
For binary classification, accuracy can be calculated as follows:
Accuracy = TP + TN TP + TN + FP + FN
where FN, FP, TN, and TP show false negative, false positive, true negative, and true positive are defined as follows [52]:
  • False-negative (FN): Incorrectly predicted negative instances.
  • False-positive (FP): Incorrectly predicted positive instances.
  • True negative (TN): Correctly predicted negative instances.
  • True Positive (TP): Correctly predicted positive instances.

3.10.2. Precision

Precision is the veracity of the predicting model. Precision refers to the percentage of instances predicted as positive and that are actually positive. It can be computed as:
Precision = TP TP + TN

3.10.3. Recall

A recall is the completeness of the classifier. It describes the percentage of correctly predicted instances from the positive class. Recall can be computed by the following formula:
Recall = TP TP + FN

3.10.4. F1 Score

F1 Score is the harmonic mean between precision and recall, in other words, F1 Score conveys the balance between precision and recall. Like another score, it provides a float value within the range of 0 and 1.
F 1 Score = Precision Recall Precision + Recall

4. Results and Analysis

This section contains detailed experimental results along with an analysis of the results. A diverse range of experiments was conducted involving several ML models to evaluate the performance of these models with three different feature extraction techniques including BOW, TF-IDF, and BOW + TF-IDF. Experiments were conducted by integrating original sentiments as well as sentiments extracted by TextBlob. The sole purpose of carrying out a variety of experiments is to acquire the highest accuracy pertaining to the sentiment classification of the sentiment140 dataset. In this section, we have compared our proposed approach with previous studies conducted on the sentiment140 dataset.

4.1. Experimental Results of ML Models with Original Sentiment of the Sentiment140 Dataset

We first illustrate the experimental results of ML models trained on features extracted by TF-IDF with original sentiments from the dataset, which are shown in Table 7. Statistical model LR outperformed two other models in terms of evaluation metrics. It acquired the highest accuracy of 0.75 when integrated with features extracted by TF-IDF with a precision of 0.76, recall of 0.76, and of F1 Score 0.75, whereas the tree-based models such as ADB and RF acquired 0.73 accuracies. RF yielded the lowest precision when analyzing sentiments of the dataset by integrating TF-IDF features. From the results, it can be seen that, for original sentiments and TF-IDF features, RF performed the worst. Conversely, it can be observed that RF predicted a negative class with the highest precision of 0.77 and lowest recall of 0.70, whereas, in terms of positive class, LR remains the leading ML model with the highest precision of 0.73 and lowest recall of 0.77.
Similarly, with features extracted by BOW, LR yielded the highest accuracy of 0.74 along with the highest precision, recall, and F1 Score of 0.76, 0.75, and 0.75 as compared to other ML models as shown in Table 8. This shows the effectiveness of LR in classifying the sentiments of tweets. On the other hand, RF acquired the lowest accuracy of 0.73 and ADB acquired 0.74 accuracy. Despite showing the highest accuracy, LR was not able to provide optimized results in the prediction of the negative class, as RF leads with 0.78 precision and lowest recall of 0.63. As for the positive class, LR outperformed other ML models with the highest precision of 0.73 and the lowest recall of 0.78, although the F1 Score, which is the harmonic mean of precision and recall, remains the same i.e., 0.75 in the prediction of the positive class.
In the case of the feature union, it can be observed that LR acquired the highest accuracy score of 0.78 with 0.78 precision, 0.76 recall, and F1 Score as shown in Table 9. The stable values of precision, recall, and F1 Score show the efficacy of LR when trained with features extracted by the feature union. ADB on the contrary did not perform well, whereas RF acquired an accuracy score of 0.76. In terms of the positive class, LR yielded the highest precision, recall, and F1 Score of 0.78, 0.76, 0.76 respectively; in the case of the negative class LR outperformed the other models.

4.1.1. Comparative Analysis of ML Models with Original Sentiments Using TF-IDF, BOW and BOW + TF-IDF

Figure 3 shows the comparison between the performance of ML models using three different feature extraction techniques when original sentiments of the dataset were integrated. It can be observed that the feature union, i.e., BOW + TF-IDF boosted the performance of LR. Moreover, a boost in the performance of ADB and RF can also be noted with the feature union, showing that the features extracted by the union of BOW and TF-IDF are more correlated with the target sentiments as compared to features extracted by BOW and TF-IDF individually. It also creates a larger feature set for the models to train, thus enhancing the performance of the models. On the contrary, models including RF and ADB did not quite perform well with features extracted by TF-IDF, whereas the performance of LR remained the same with TF-IDF and BOW.

4.1.2. Experimental Results of Proposed LR-LSTM with Original Sentiments

The performance of ML models varies with feature extraction techniques, thus leaving room for improvement. To enhance the accuracy of sentiment classification of tweets, this study proposes a stacked ensemble model LR-LSTM. The proposed model does not require any feature extraction technique as LSTM is a deep learning approach that has the capability of extracting features automatically. LR is trained on features extracted by LSTM. The experimental results of the proposed LR-LSTM with original sentiments as the target value are shown in Table 10. It can be observed that our proposed model outperformed the conventional state-of-the-art models in terms of accuracy, precision, recall, and F1 Score. Proposed LR-LSTM acquired a maximum accuracy of 0.81 which shows the effectiveness of the proposed stacked ensemble model.

4.2. Experimental Results of ML Models with TextBlob Sentiment

Table 11 shows that the highest accuracy score of 0.95 is yielded by LR through integration when trained with features extracted by TF-IDF and given the sentiments extracted by TextBlob. This shows that TextBlob sentiments are in more correlation with the feature set extracted by TF-IDF. Similarly, an improvement in the performance of ML models, including ADB and RF, shows the efficacy of using TextBlob sentiments. In terms of positive class, LR yielded the highest precision and lowest recall as compared to other ML models, whereas, in the case of the negative class, ADB is the leading ML model in terms of the highest precision of 0.96. Overall, it can be observed that LR performed well with TextBlob sentiments when trained on features extracted by TF-IDF.
Experiments conducted using features extracted by BOW with TextBlob sentiments as target values resulted in comparatively better performance in the case of LR shown in Table 12. Concerning the BOW features, LR outperformed the tree-based model RF and boosting model ADB by achieving a 0.97 accuracy score. LR also performed well in terms of other evaluation parameters. While ADB yielded an accuracy score of 0.92, RF performed the worst with a 0.82 accuracy score.
Table 13 shows experimental results of ML models when trained on features extracted by the feature union. The results showed that LR surpassed other ML models by achieving an accuracy of 0.98 with similar precision, recall, and F1 Score.

4.2.1. Comparative Analysis of ML Models with TextBlob Sentiments Using TF-IDF, BOW and BOW + TF-IDF

Figure 4 shows that the highest accuracy is achieved by LR with BOW + TF-IDF using TextBlob sentiments, while the performance of ADB remained the same with three feature extraction techniques. BOW + TF-IDF creates a large feature set for the models to train, thus resulting in better performance of models. On the other hand, RF performed poorly in comparison to the other two ML models. From this, we can observe that LR, due to its statistical structure, transcended in classifying sentiments of tweets. LR not only quantifies the coefficient size but also provides the direction of association (negative or positive) of the record under analysis. This makes LR more efficient as compared to other ML models in this study.

4.2.2. Experimental Results of Proposed LR-LSTM with TextBlob Sentiments

Stacking is a powerful solution for combining the learning models. From the above discussion, it can be observed that LR with its efficacy has surpassed other ML models in classifying sentiments of tweets. This provides the basis of our proposed model LR-LSTM. LSTM works well with long time dependencies, giving us an edge in experimental results. From Table 14, it can be observed that our proposed model LR-LSTM achieved state-of-the-art results by classifying sentiments of tweets extracted by TextBlob. LR-LSTM does not require extraction of features separately and thus it is a time-efficient method. It acquired an accuracy of 0.99 with similar precision, recall, and F1 Score showing the robustness and effectiveness of the proposed model.

4.3. Impact of TextBlob Sentiments on Classifiers

Experimental results illustrated that the classifier efficiently predicted sentiments of tweets extracted by TextBlob. LR performed well with its capability of predicting model coefficients as a measure of feature importance. On the contrary, its performance was limited with original sentiments as the target value. This shows that TextBlob annotates the tweets which are more correlated with its textual features. This shows the effectiveness of the proposed methodology. ADB also performed comparatively better, but its sensitivity to outliners limited its performance as compared to LR. RF performed poorly due to its continuous approximation, but its performance was boosted with TextBlob sentiments. On the other hand, our proposed LR-LSTM showed empirical results with the highest accuracy score of 0.99 which is an optimized accuracy score showing the robustness of the proposed model.
From Figure 5 it can be observed that performance of all classifiers was enhanced with the sentiments extracted by TextBlob, as compared to their performance with the original sentiments. This shows that sentiments labeled by TextBlob are more relevant to the features of the tweets. TextBlob assigns a sentiment score to words with greater clarification in relation to its PatternAnalyzer property, which results in better learning for the classifiers and hence better performance.

4.4. Comparative Analysis of Proposed LR-LSTM with Deep Learning Models

The performance of the proposed LR-LSTM model was compared with several deep learning models including the gated recurrent unit (GRU), convolutional neural network (CNN), and long short-term memory (LSTM) to validate the effectiveness of the model. GRU is a modified version of a recurrent neural network (RNN) which deals with the problem of the vanishing gradient of a standard RNN. CNN, on the other hand, has the ability to extract the textual features from the input data in a direct manner without the requirement of preprocessing tasks. CNN has three main components including a convolution layer, pooling layer, and dense layer to carry out predictive tasks. LSTM makes a prediction based on individual time steps of the sequential data. Considering the aforementioned structures of CNN and LSTM, we also employed a combined CNN–LSTM model. Hyperparameter settings for each deep learning model are presented in Table 15.
Extensive experiments were conducted using the sentiment140 dataset combined with TextBlob sentiments for the training and testing of the deep learning models. Experimental results reveal that performance of the deep learning models is comparatively lower than the proposed LR-LSTM classifier which shows the effectiveness of this study. Table 16 presents the performances of deep learning models in comparison with the proposed approach. The results disclose that the highest accuracy score of 0.96 is achieved by LSTM as compared to other deep learning models. This analysis also supports the integration of LSTM as a base learner in the proposed LR-LSTM model.

4.5. Comparative Analysis of Proposed Study with Correlated Studies

A considerable amount of research has been carried out on the benchmark sentiment140 dataset. In this section, we compare our proposed approach to a few state-of-the-art approaches proposed in previous studies to carry out sentiment classification of tweets in the sentimnet140 dataset. Previous studies are summarized in Table 17 which shows that our proposed system exceeded in performance in comparison to previous studies which shows the potency of our proposed approach.

5. Conclusions

This study proposes a novel approach by integrating a lexical dictionary along with a stacked ensemble of three LSTMs and LR which is aimed at enhancing the performance of sentiment analysis of tweets from the sentiment140 dataset. The study suggests that sentiments extracted from TextBlob are in more correlated with the textual features of tweets as compared to the original sentiments. Training of classifiers was carried out on a 70% training set and tested on a 30% testing set. No feature extraction was required in terms of our proposed approach, contrarily, ML models including RF, LR, and ADB required extracted features for which three feature extraction techniques including BOW, TF-IDF, and BOW + TF-IDF were used. Classification of tweets is performed using the proposed model and above-mentioned ML classifiers with original sentiments and TextBlob sentiments. Conventional ML models revealed scant performance in classifying sentiments of tweets given the original sentiments, but their performance enhanced with TextBlob sentiments, thus revealing that there is a high level of association between tweets and sentiments extracted by TextBlob as compared to the original sentiments. The proposed LR-LSTM model is adapted for optimized results which outperformed other conventional models with a maximum accuracy of 99%, precision of 99%, recall of 99%, and F1 Score of 99%, respectively, which shows the efficacy and feasibility of the proposed model.
Modification of the proposed LR-LSTM model can be a future direction. Furthermore, preprocessing techniques including POS tagging can further improve the accuracy of the model. Moreover, this research can be extended to sarcasm detection, fake review detection, fake advertisement classification, spam email detection, and many more. Additionally, word embeddings can be added to the model.

Author Contributions

Conceptualization, B.G. and A.W.; data curation, B.G. and A.W.; formal analysis, B.G. and D.Z.; funding acquisition, A.W.; investigation, D.Z., D.Z. and A.W.; methodology, B.G., A.W. and D.Z.; project administration, B.G.; resources, A.W.; software, B.G. and D.Z.; supervision, D.Z.; visualization, D.Z.; writing—original draft, B.G.; writing—review and editing, A.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Statista. Available online: https://www.statista.com/statistics/346167/facebook-global-dau/ (accessed on 6 September 2021).
  2. Statista. Available online: https://www.statista.com/statistics/272014/global-social-networks- (accessed on 6 September 2021).
  3. You, Q.; Bhatia, S.; Luo, J. A picture tells a thousand words—About you! User interest profiling from user generated visual content. Signal Process. 2016, 124, 45–53. [Google Scholar] [CrossRef] [Green Version]
  4. Persia, F.; D’Auria, D. A survey of online social networks: Challenges and opportunities. In Proceedings of the 2017 IEEE International Conference on Information Reuse and Integration (IRI), San Diego, CA, USA, 4–6 August 2017; pp. 614–620. [Google Scholar]
  5. Khattak, A.M.; Batool, R.; Satti, F.A.; Hussain, J.; Khan, W.A.; Khan, A.M.; Hayat, B. Tweets classification and sentiment analysis for personalized tweets recommendation. Complexity 2020, 2020, 8892552. [Google Scholar] [CrossRef]
  6. Crisci, A.; Grasso, V.; Nesi, P.; Pantaleo, G.; Paoli, I.; Zaza, I. Predicting TV programme audience by using twitter based metrics. Multimed. Tools Appl. 2018, 77, 12203–12232. [Google Scholar] [CrossRef] [Green Version]
  7. McConnell, J. Twitter and the 2016 US Presidential Campaign: A Rhetorical Analysis of Tweets and Media Coverage. Master’s Thesis, New York University, New York, NY, USA, 2015. [Google Scholar]
  8. Coletta, L.F.; da Silva, N.F.; Hruschka, E.R.; Hruschka, E.R. Combining classification and clustering for tweet sentiment analysis. In Proceedings of the 2014 Brazilian Conference on Intelligent Systems, Sao Paulo, Brazil, 18–22 October 2014; pp. 210–215. [Google Scholar]
  9. Dhelim, S.; Ning, H.; Aung, N.; Huang, R.; Ma, J. Personality-Aware Product Recommendation System Based on User Interests Mining and Metapath Discovery. IEEE Trans. Comput. Soc. Syst. 2020, 8, 86–98. [Google Scholar] [CrossRef]
  10. Cambria, E.; Das, D.; Bandyopadhyay, S.; Feraco, A. Affective computing and sentiment analysis. In A Practical Guide to Sentiment Analysis; Springer: Berlin, Germany, 2017; pp. 1–10. [Google Scholar]
  11. Hussein, D.M.E.D.M. A survey on sentiment analysis challenges. J. King Saud Univ.-Eng. Sci. 2018, 30, 330–338. [Google Scholar] [CrossRef]
  12. Agüero-Torales, M.M.; Salas, J.I.A.; López-Herrera, A.G. Deep learning and multilingual sentiment analysis on social media data: An overview. Appl. Soft Comput. 2021, 107, 107373. [Google Scholar] [CrossRef]
  13. Munandar, D.; Rozie, A.F.; Arisal, A. A multi domains short message sentiment classification using hybrid neural network architecture. Bull. Electr. Eng. Inform. 2021, 10, 2181–2191. [Google Scholar] [CrossRef]
  14. Schmidhuber, J. Deep learning in neural networks: An overview. Neural Netw. 2015, 61, 85–117. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  15. Campos, V.; Jou, B.; Giro-i Nieto, X. From pixels to sentiment: Fine-tuning CNNs for visual sentiment prediction. Image Vis. Comput. 2017, 65, 15–22. [Google Scholar] [CrossRef] [Green Version]
  16. Marasek, K. Deep belief neural networks and bidirectional long-short term memory hybrid for speech recognition. Arch. Acoust. 2015, 40, 191–195. [Google Scholar]
  17. Tam, S.; Said, R.B.; Tanriöver, Ö.Ö. A ConvBiLSTM Deep Learning Model-Based Approach for Twitter Sentiment Classification. IEEE Access 2021, 9, 41283–41293. [Google Scholar] [CrossRef]
  18. Chugh, A.; Sharma, V.K.; Kumar, S.; Nayyar, A.; Qureshi, B.; Bhatia, M.K.; Jain, C. Spider monkey crow optimization algorithm with deep learning for sentiment classification and information retrieval. IEEE Access 2021, 9, 24249–24262. [Google Scholar] [CrossRef]
  19. Go, A.; Bhayani, R.; Huang, L. Twitter sentiment classification using distant supervision. CS224N Proj. Rep. Stanf. 2009, 1, 2009. [Google Scholar]
  20. Vo, H.T.; Lam, H.C.; Nguyen, D.D.; Tuong, N.H. Topic classification and sentiment analysis for Vietnamese education survey system. Asian J. Comput. Sci. Inf. Technol. 2016, 6, 27–34. [Google Scholar]
  21. Sarkar, S.; Seal, T. Sentiment Analysis-An Objective View. J. Res. Vol. 2016, 2, 26–29. [Google Scholar]
  22. Saleena, N.; others. An ensemble classification system for twitter sentiment analysis. Procedia Comput. Sci. 2018, 132, 937–946. [Google Scholar]
  23. Onan, A.; Korukoğlu, S.; Bulut, H. A multiobjective weighted voting ensemble classifier based on differential evolution algorithm for text sentiment classification. Expert Syst. Appl. 2016, 62, 1–16. [Google Scholar] [CrossRef]
  24. Rustam, F.; Ashraf, I.; Mehmood, A.; Ullah, S.; Choi, G.S. Tweets classification on the base of sentiments for US airline companies. Entropy 2019, 21, 1078. [Google Scholar] [CrossRef] [Green Version]
  25. Umer, M.; Ashraf, I.; Mehmood, A.; Kumari, S.; Ullah, S.; Sang Choi, G. Sentiment analysis of tweets using a unified convolutional neural network-long short-term memory network model. Comput. Intell. 2021, 37, 409–434. [Google Scholar] [CrossRef]
  26. Stojanovski, D.; Strezoski, G.; Madjarov, G.; Dimitrovski, I. Twitter sentiment analysis using deep convolutional neural network. In Proceedings of the International Conference on Hybrid Artificial Intelligence Systems, Bilbao, Spain, 22–24 June 2015; pp. 726–737. [Google Scholar]
  27. Zhao, J.; Gui, X.; Zhang, X. Deep convolution neural networks for twitter sentiment analysis. IEEE Access 2018, 6, 23253–23260. [Google Scholar]
  28. Dos Santos, C.; Gatti, M. Deep convolutional neural networks for sentiment analysis of short texts. In Proceedings of the COLING 2014, the 25th International Conference on Computational Linguistics: Technical Papers, Dublin, Ireland, 23–29 October 2014; pp. 69–78. [Google Scholar]
  29. Ishaq, A.; Umer, M.; Mushtaq, M.F.; Medaglia, C.; Siddiqui, H.U.R.; Mehmood, A.; Choi, G.S. Extensive hotel reviews classification using long short term memory. J. Ambient. Intell. Humaniz. Comput. 2020, 1–11. [Google Scholar] [CrossRef]
  30. Seo, S.; Kim, C.; Kim, H.; Mo, K.; Kang, P. Comparative study of deep learning-based sentiment classification. IEEE Access 2020, 8, 6861–6875. [Google Scholar] [CrossRef]
  31. Salur, M.U.; Aydin, I. A novel hybrid deep learning model for sentiment classification. IEEE Access 2020, 8, 58080–58093. [Google Scholar] [CrossRef]
  32. Cliche, M. BB_twtr at SemEval-2017 task 4: Twitter sentiment analysis with CNNs and LSTMs. arXiv 2017, arXiv:1704.06125. [Google Scholar]
  33. Sentiment 140. Available online: https://www.kaggle.com/kazanova/sentiment140 (accessed on 30 September 2020).
  34. Kalra, V.; Aggarwal, R. Importance of Text Data Preprocessing & Implementation in RapidMiner; ICITKM: Singapore, 10 November 2017; pp. 71–75. [Google Scholar]
  35. Dhelim, S.; Ning, H.; Aung, N. ComPath: User interest mining in heterogeneous signed social networks for Internet of people. IEEE Internet Things J. 2020, 8, 7024–7035. [Google Scholar] [CrossRef]
  36. Yang, S.; Zhang, H. Text mining of Twitter data using a latent Dirichlet allocation topic model and sentiment analysis. Int. J. Comput. Inf. Eng. 2018, 12, 525–529. [Google Scholar]
  37. Loria, S. textblob Documentation. Release 0.15 2018, 2, 269. [Google Scholar]
  38. Saad, E.; Din, S.; Jamil, R.; Rustam, F.; Mehmood, A.; Ashraf, I.; Choi, G.S. Determining the Efficiency of Drugs under Special Conditions from Users’ Reviews on Healthcare Web Forums. IEEE Access 2021, 9, 85721–85737. [Google Scholar] [CrossRef]
  39. Bocca, F.F.; Rodrigues, L.H.A. The effect of tuning, feature engineering, and feature selection in data mining applied to rainfed sugarcane yield modelling. Comput. Electron. Agric. 2016, 128, 67–76. [Google Scholar] [CrossRef]
  40. Heaton, J. An empirical analysis of feature engineering for predictive modeling. In Proceedings of the SoutheastCon 2016, Norfolk, VA, USA, 30 March 2016; pp. 1–6. [Google Scholar]
  41. Hu, X.; Downie, J.S.; Ehmann, A.F. Lyric text mining in music mood classification. Am. Music 2009, 183, 2–209. [Google Scholar]
  42. Yu, B. An evaluation of text classification methods for literary study. Lit. Linguist. Comput. 2008, 23, 327–343. [Google Scholar] [CrossRef]
  43. Robertson, S. Understanding inverse document frequency: On theoretical arguments for IDF. J. Doc. 2004, 60, 503–520. [Google Scholar] [CrossRef] [Green Version]
  44. Zhang, W.; Yoshida, T.; Tang, X. A comparative study of TF* IDF, LSI and multi-words for text classification. Expert Syst. Appl. 2011, 38, 2758–2765. [Google Scholar] [CrossRef]
  45. Da Silva, N.F.; Hruschka, E.R.; Hruschka Jr, E.R. Tweet sentiment analysis with classifier ensembles. Decis. Support Syst. 2014, 66, 170–179. [Google Scholar] [CrossRef]
  46. Kleinbaum, D.G.; Klein, M.; Pryor, E.R. Logistic Regression: A Self-Learning Text; Springer: New York, NY, USA, 2002. [Google Scholar]
  47. Zhang, Y.; Zhang, H.; Cai, J.; Yang, B. A weighted voting classifier based on differential evolution. In Abstract and Applied Analysis; Hindawi: London, UK, 2014; Volume 2014. [Google Scholar]
  48. Freund, Y.; Schapire, R.E. A decision-theoretic generalization of on-line learning and an application to boosting. J. Comput. Syst. Sci. 1997, 55, 119–139. [Google Scholar] [CrossRef] [Green Version]
  49. Fan, D.; Sun, H.; Yao, J.; Zhang, K.; Yan, X.; Sun, Z. Well production forecasting based on ARIMA-LSTM model considering manual operations. Energy 2021, 220, 119708. [Google Scholar] [CrossRef]
  50. Divina, F.; Gilson, A.; Goméz-Vela, F.; García Torres, M.; Torres, J.F. Stacking ensemble learning for short-term electricity consumption forecasting. Energies 2018, 11, 949. [Google Scholar] [CrossRef] [Green Version]
  51. Rupapara, V.; Rustam, F.; Shahzad, H.F.; Mehmood, A.; Ashraf, I.; Choi, G.S. Impact of SMOTE on Imbalanced Text Features for Toxic Comments Classification using RVVC Model. IEEE Access 2021, 9, 78621–78634. [Google Scholar] [CrossRef]
  52. Han, J.; Pei, J.; Kamber, M. Data Mining: Concepts and Techniques; Elsevier: Amsterdam, The Netherlands, 2011. [Google Scholar]
  53. Iqbal, N.; Chowdhury, A.M.; Ahsan, T. Enhancing the performance of sentiment analysis by using different feature combinations. In Proceedings of the 2018 International Conference on Computer, Communication, Chemical, Material and Electronic Engineering (IC4ME2), Rajshahi, Bangladesh, 8–9 February 2018; pp. 1–4. [Google Scholar]
  54. Korovkinas, K.; Garšva, G. Selection of intelligent algorithms for sentiment classification method creation. In Proceedings of the International Conference on Information Technologies, Tomsk, Russia, 18–20 January 2018; Volume 2145, pp. 152–157. [Google Scholar]
  55. Dang, N.C.; Moreno-García, M.N.; De la Prieta, F. Sentiment analysis based on deep learning: A comparative study. Electronics 2020, 9, 483. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Architecture of the proposed LR-LSTM model.
Figure 1. Architecture of the proposed LR-LSTM model.
Information 12 00374 g001
Figure 2. Architecture of the proposed methodology.
Figure 2. Architecture of the proposed methodology.
Information 12 00374 g002
Figure 3. Performance comparison of ML models using three different feature extraction techniques.
Figure 3. Performance comparison of ML models using three different feature extraction techniques.
Information 12 00374 g003
Figure 4. Performance comparison of ML models using three different feature extraction techniques.
Figure 4. Performance comparison of ML models using three different feature extraction techniques.
Information 12 00374 g004
Figure 5. Accuracy score of classifiers with TextBlob sentiments and original sentiments (LR + TF-IDF refers to an LR model trained on features extracted by TF-IDF; the same goes for other labels).
Figure 5. Accuracy score of classifiers with TextBlob sentiments and original sentiments (LR + TF-IDF refers to an LR model trained on features extracted by TF-IDF; the same goes for other labels).
Information 12 00374 g005
Table 1. Dataset Description.
Table 1. Dataset Description.
FeaturesDescription
targetSentiment label corresponding to each tweet. (0 = negative, 4 = positive)
idTweet ID
dateDate on which the tweet was posted.
flagThe query (lyx). If there is no query, then this value is NO QUERY.
userUsername of the person who posted the tweet.
textText written in the tweet.
Table 2. Tweet sample from the dataset.
Table 2. Tweet sample from the dataset.
No.TargetText
10@Kenichan I dived many times for the ball. Managed to save 50% The rest go out of bounds
20@octolinz16 It counts, Idk why I did either. you never talk to me anymore
34@ElleCTF I would like to call you at night after dinner
Table 3. Sample tweets after preprocessing.
Table 3. Sample tweets after preprocessing.
Sr.Sample TweetsPre-Processed Tweets
1@ehfu Want to accept credit cards ? Credit approved no checks do it now 1234want accept credit card credit approve check
2@rtew Newyork will go long in investment as other cities going longnewyork go long investment other city go long newyork
Table 4. TextBlob sentiment ratio.
Table 4. TextBlob sentiment ratio.
SentimentCount
Positive686,142
Negative319,374
Neutral594,484
Table 5. Sample of tweets with corresponding original sentiments and sentiments extracted by TextBlob.
Table 5. Sample of tweets with corresponding original sentiments and sentiments extracted by TextBlob.
Sr.TweetsOriginalTextBlob
2I need a hugNegativePositive
3laying in bed with no voice..NegativePositive
4I’m tired now so... I’m going to bed...goodnight...?PositiveNegative
Table 6. Machine Learning Models Hyper-parameter Settings.
Table 6. Machine Learning Models Hyper-parameter Settings.
ModelsHyper-Parameter Settings
RFmax_depth = 300, random_state = 42, n_estimators = 300
LRC = 3.0, multi_class = ‘ovr’, solver = ‘liblinear’, random_state = 42
ADBrandom_state = 42, n_estimators = 300
Table 7. Experimental results of ML models with original sentiments using TF-IDF.
Table 7. Experimental results of ML models with original sentiments using TF-IDF.
ClassifierAccuracyClassPrecisionRecallF1 Score
LR0.75negative0.760.720.73
positive0.730.770.75
macro avg0.760.760.75
ADB0.73negative0.760.630.70
positive0.690.820.75
macro avg0.730.730.73
RF0.73negative0.770.700.72
positive0.690.820.75
macro avg0.730.730.73
Table 8. Experimental results of ML models with original sentiments using BOW.
Table 8. Experimental results of ML models with original sentiments using BOW.
ClassifierAccuracyClassPrecisionRecallF1 Score
negative0.770.710.74
LR0.74positive0.730.780.75
macro avg0.760.750.75
negative0.780.650.72
ADB0.74positive0.700.820.75
macro avg0.740.720.73
negative0.790.630.70
RF0.73positive0.700.820.76
macro avg0.710.700.70
Table 9. Experimental results of ML models with original sentiments using a feature union.
Table 9. Experimental results of ML models with original sentiments using a feature union.
ClassifierAccuracyClassPrecisionRecallF1 Score
negative0.790.760.76
LR0.78positive0.780.760.76
macro avg0.780.760.76
negative0.770.640.70
ADB0.73positive0.710.840.74
macro avg0.730.730.72
negative0.780.760.73
RF0.76positive0.750.750.77
macro avg0.760.760.75
Table 10. Experimental results of LR-LSTM with original sentiments.
Table 10. Experimental results of LR-LSTM with original sentiments.
AccuracyClassPrecisionRecallF1 Score
negative0.820.800.80
0.80positive0.810.810.80
Macro avg0.810.800.90
Table 11. Experimental results of ML models with TextBlob sentiments using TF-IDF.
Table 11. Experimental results of ML models with TextBlob sentiments using TF-IDF.
ModelAccuracyClassPrecisionRecallF1 Score
negative0.950.950.95
LR0.95positive0.950.950.95
avg0.950.950.95
negative0.960.870.92
ADB0.92positive0.890.970.93
avg0.920.920.92
negative0.950.740.83
RF0.84positive0.790.960.87
avg0.870.850.85
Table 12. Experimental results of ML models with TextBlob sentiments using BOW.
Table 12. Experimental results of ML models with TextBlob sentiments using BOW.
ModelAccuracyClassPrecisionRecallF1 Score
negative0.970.980.98
LR0.97positive0.970.980.97
avg0.970.980.97
negative0.970.870.92
ADB0.92positive0.890.970.93
avg0.930.920.92
negative0.990.650.78
RF0.82positive0.740.990.85
avg0.870.820.82
Table 13. Experimental results of ML models with TextBlob sentiments using the feature union.
Table 13. Experimental results of ML models with TextBlob sentiments using the feature union.
ModelAccuracyClassPrecisionRecallF1 Score
negative0.980.980.98
LR0.98positive0.980.980.98
avg0.980.980.98
negative0.910.870.91
ADB0.92positive0.880.970.92
avg0.900.920.92
negative0.980.730.84
RF0.86positive0.780.990.88
avg0.880.860.86
Table 14. Experimental results of LR-LSTM with TextBlob sentiments.
Table 14. Experimental results of LR-LSTM with TextBlob sentiments.
AccuracyClassPrecisionRecallF1 Score
negative0.990.990.98
0.99positive0.990.980.99
Macro avg0.990.990.99
Table 15. Deep Learning Models’ Hyperparameter Settings.
Table 15. Deep Learning Models’ Hyperparameter Settings.
ModelsHyperparameter Settings
LSTMembedding=5000, dropout=0.5, dense_layer=3, activation=’softmax’, loss=’categorical_crossentropy’, optimizer=’adam’, epoch=100, batch_size=16, LSTM(100)
GRUembedding=5000, dropout=0.5, dense_layer=3, activation=’softmax’, loss=’categorical_crossentropy’, optimizer=’adam’, epoch=100, batch_size=16, GRU(256), SimpleRNN(128)
CNNembedding=5000, dropout=0.5, dense_layer=3, activation=’softmax’, loss=’categorical_crossentropy’, optimizer=’adam’, epoch=100, batch_size=16, Conv1D(128, 5, activation=’relu’), MaxPooling1D(pool_size=4)
CNN-LSTMembedding=5000, dropout=0.5, dense_layer=3, activation=’softmax’, loss=’categorical_crossentropy’, optimizer=’adam’, epoch=100, batch_size=16, Conv1D(128, 5, activation=’relu’), MaxPooling1D(pool_size=4), LSTM(100)
Table 16. Experimental results of Deep learning models using TextBlob sentiments.
Table 16. Experimental results of Deep learning models using TextBlob sentiments.
ClassifiersAccuracyPrecisionRecallF1 Score
GRU0.950.950.950.95
LSTM0.960.960.960.96
CNN0.930.930.930.93
CNN–LSTM0.930.930.930.93
Table 17. Comparative Analysis of the Proposed Study with Correlated Studies.
Table 17. Comparative Analysis of the Proposed Study with Correlated Studies.
RefYearProposed MethodologyClassifier with Highest AccuracyAccuracy
[53]2018Analyzed the impact of various features such as combining bigram features with unigrams, unigrams and word features without stopwords, bigrams with word features without stopwords, and the highest weighted unigrams with the highest weighted bigrams, on the performance of ML classifiers.NB with unigram and bigram features.88%
[22]2018Calculating sentiment score of a tweet and then classifying based on its sentiment score with a majority voting ensemble model provides comparatively better results.Ensemble of NB, RF, LR and SVM75.81%
[54]2018An ensemble of two ML models can perform comparatively better in classifying sentiments of tweets as comparison to individual classifiers.LR-SVM81.83%
[55]2020Deep learning models perform better with word embedding in comparison to TF-IDF features.RNN with word Embedding82.8%
Proposed2021Sentiments extracted by lexical dictionary are more correlated with the textual features, thus enhancing the performance of a stacked ensemble model.LR-LSTM99%
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Gaye, B.; Zhang, D.; Wulamu, A. A Tweet Sentiment Classification Approach Using a Hybrid Stacked Ensemble Technique. Information 2021, 12, 374. https://doi.org/10.3390/info12090374

AMA Style

Gaye B, Zhang D, Wulamu A. A Tweet Sentiment Classification Approach Using a Hybrid Stacked Ensemble Technique. Information. 2021; 12(9):374. https://doi.org/10.3390/info12090374

Chicago/Turabian Style

Gaye, Babacar, Dezheng Zhang, and Aziguli Wulamu. 2021. "A Tweet Sentiment Classification Approach Using a Hybrid Stacked Ensemble Technique" Information 12, no. 9: 374. https://doi.org/10.3390/info12090374

APA Style

Gaye, B., Zhang, D., & Wulamu, A. (2021). A Tweet Sentiment Classification Approach Using a Hybrid Stacked Ensemble Technique. Information, 12(9), 374. https://doi.org/10.3390/info12090374

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop