0% found this document useful (0 votes)
3 views

An Optimized Crossover Framework for Social Media Sentiment Analysis

The document presents an optimized crossover framework for social media sentiment analysis, utilizing deep learning techniques. It outlines a four-phase approach involving pre-processing, feature extraction, feature selection, and sentiment classification, with a focus on enhancing classification accuracy through the Self Improved Honey Badger Algorithm (SI-HBA). The proposed model aims to efficiently classify sentiments as positive, negative, or neutral by integrating optimized Bi-LSTM and Deep Belief Network (DBN) methodologies.

Uploaded by

sssamant
Copyright
© © All Rights Reserved
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
3 views

An Optimized Crossover Framework for Social Media Sentiment Analysis

The document presents an optimized crossover framework for social media sentiment analysis, utilizing deep learning techniques. It outlines a four-phase approach involving pre-processing, feature extraction, feature selection, and sentiment classification, with a focus on enhancing classification accuracy through the Self Improved Honey Badger Algorithm (SI-HBA). The proposed model aims to efficiently classify sentiments as positive, negative, or neutral by integrating optimized Bi-LSTM and Deep Belief Network (DBN) methodologies.

Uploaded by

sssamant
Copyright
© © All Rights Reserved
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 30

Cybernetics and Systems

An International Journal

ISSN: (Print) (Online) Journal homepage: https://www.tandfonline.com/loi/ucbs20

An Optimized Crossover Framework for Social


Media Sentiment Analysis

Surender Singh Samant, Vijay Singh, Arun Chauhan & Jagadish Dasarahalli
Narasimaiah

To cite this article: Surender Singh Samant, Vijay Singh, Arun Chauhan & Jagadish Dasarahalli
Narasimaiah (2022): An Optimized Crossover Framework for Social Media Sentiment Analysis,
Cybernetics and Systems, DOI: 10.1080/01969722.2022.2146849

To link to this article: https://doi.org/10.1080/01969722.2022.2146849

Published online: 23 Nov 2022.

Submit your article to this journal

View related articles

View Crossmark data

Full Terms & Conditions of access and use can be found at


https://www.tandfonline.com/action/journalInformation?journalCode=ucbs20
CYBERNETICS AND SYSTEMS: AN INTERNATIONAL JOURNAL
https://doi.org/10.1080/01969722.2022.2146849

An Optimized Crossover Framework for Social Media


Sentiment Analysis
Surender Singh Samanta, Vijay Singha, Arun Chauhana, and Jagadish
Dasarahalli Narasimaiahb
a
Department of Computer Science and Engineering, Graphic Era (Deemed to be University),
Dehradun, India; bDepartment of Electronics and Communication Engineering, IIIT Dharwad,
Dharwad, India

ABSTRACT KEYWORDS
Using deep learning, a new sentiment analysis model is B-BoW; DBN with transfer
designed in our article. The original data (input) is first pre-proc- learning; optimized Bi-
essed by stemming, stop-word removal, and tokenization. The LSTM; sentiment analysis;
SI-HBA; social media;
projected technique includes 4 phases: "pre-processing, feature T-TFIDF
extraction, feature selection, and sentiment classification." The
characteristics “Bigram-BoW (B-BoW), Threshold Term Frequency-
Inverse Document Frequency (T-TFIDF), Unigram, and N-Gram”
are then retrieved from the pre-processed data. Utilizing the
self-improved Honey Badger Algorithm, the best features out of
the chosen features will be chosen (SI-HBA). The basic Honey
Badger Algorithm (HBA) has been conceptually improved by this
SI-HBA model. The review classification will then be conducted
via the proposed optimized crossover framework, which is con-
structed by hybridizing the optimized Bi-Long Short-Term
Memory (Bi-LSTM) and Deep Belief Network (DBN) trained with
Transfer learning, respectively. The SI-HBA model’s optimally
chosen features are used to train the hybrid classifier within the
optimized crossover architecture. A self-improved Honey Badger
Algorithm is used to fine-tune the weight of the Bi-LSTM classi-
fier to improve the classification performance of the gathered
reviews (SI-HBA). The final results will indicate whether the
reviews are mostly good, negative, or neutral. The proposed
sentiment classification model is then validated by a compari-
son analysis.

Introduction
The amount of people using the internet has dramatically increased
recently, with the number of users doubling every year (Alfrjani, Osman,
and Cosma 2019; Khan, Qamar, and Bashir 2016; Wadawadagi and Pagi
2019; Ruz, Henrıquez, and Mascare~ no 2020). This growth has led to a rise
in the popularity of user-generated content on particular goods and serv-
ices. After the emergence of social media, People are free to express their

CONTACT Surender Singh Samant [email protected] Department of Computer Science and


Engineering, Graphic Era (Deemed to be University), Dehradun, Uttarakhand, India.
ß 2022 Taylor & Francis Group, LLC
2 S. S. SAMANT ET AL.

consensus sentiments and speculations, share thoughts, socialize, and com-


municate on various platforms. Facebook, Twitter, and the Internet Movie
Database are just a few examples of such platforms (Khan, Qamar, and
Bashir 2016). “Opinion mining, also known as sentiment analysis”, has
recently attracted a lot of interest from specialists since it is useful for
understanding many facets of client-created content and sentimental ana-
lysis. Additionally, by identifying the market pattern, decision-making in
the business sector can be aided by sentiment analysis or opinion mining.
(Xu et al. 2019; Li, Li, and Jin 2020; Wu et al. 2019; Yang et al. 2020;
Wang et al. 2021). There are many ways to represent the feelings, but the
polarities of positive, negative, and neutral are the most common. A favor-
able opinion denotes joy, zeal, and consideration for the subject. Similar to
how a negative opinion is associated with a negative emotion like grief,
hate, violence, discrimination. Neutrality is the quality of not endorsing
any position or side in a dispute. (Valdivia, Luzı
on, and Herrera 2017)
In general, Sentiment analysis (Zhai et al. 2020; Wang, Niu, and Yu
2020) is a sub-classification of text mining that makes use of Natural lan-
guage processing (NLP) as well as data mining techniques (Khan, Qamar,
and Bashir 2016). It is additionally isolated into “polarity classification or
opinion detection”. In the area of polarity classification, Sentiment analysis
is much more appropriate for motion pictures, items, or client surveys
(Khan, Qamar, and Bashir 2016; Aydin and G€ ung€or 2020; Smetanin 2020;
Zhou, Jin, and Huang 2020). According to Khan, Qamar, and Bashir
(2016), Obiedat et al. (2021), Alattar and Shaalan (2021), Silva et al. (2022),
and Januario et al. (2022), the methodologies used to conduct assessment
research can be split into four categories: supervised, semi-supervised,
unsupervised, and lexicon-based machine learning. The supervised algo-
rithms’ data training and testing depend on the domain-specific labeled set.
Support Vector Machine (SVM), Naive Bayes (NB), Neural network (NN),
K-nearest neighbor (kNN), and decision tree algorithms are among models
that incorporate these techniques (Wadawadagi and Pagi 2019). For
unsupervised algorithms, training doesn’t require labeled training data.
Models incorporate methodologies dependent on “emoticons, bag of words,
and lexicons” (Khan, Qamar, and Bashir 2016; Phan et al. 2020; Mehanna
and Mahmuddin 2021; Sehar et al. 2021; Liang, Ganeshbabu, and Thorne
2020). Further, there are "dictionary-based and corpus-based categories" in
the lexicon-based sentiment classification. When no emotions are
expressed, a sentiment is considered neutral in SA. It is important to note
that due to their lack of knowledge, impartial opinions are rarely taken
into account in SA assignments.
Several sentiment analysis systems have been put out in the field. The
crowd explicit sentiment analysis (CESA) paradigm can be used to any
CYBERNETICS AND SYSTEMS: AN INTERNATIONAL JOURNAL 3

language dataset and is appropriate for social media articles. But in this case,
the polarity of the emotions was not taken into account. The suggested
method can also be improved by using web-based learning algorithms and
up-to-date preparatory data (Khan, Qamar, and Bashir 2016). The BabelNet
depended on the knowledge-enhanced meta-classifier. The meta-learning tech-
nique assisted with acquiring the most appropriate outcomes across the area
and accomplished high execution (Kastrati, Imran, and Kurti 2020; Yin et al.
2020). Although many "supervised learning techniques" for sentiment categor-
ization have been put forth, obtaining labeled corpora for each area is incred-
ibly difficult and is a major barrier to their practical use. Conversely,
universally applicable assumption techniques don’t need any labeled informa-
tion, although their significance level isn’t very high. (Aziz and Starkey 2020;
Gao et al. 2019; Deng et al. 2019; Zhang et al. 2019; Gu, Xu, and Luo 2020).
Therefore, method improvements for deep learning can be used to tackle
these challenges. Since the public is free to voice their opinions or discuss any
topic on blogs, online social networks, e-commerce websites, forums, etc., the
development of web technologies has recently fueled the growth in user-gener-
ated data (Chan et al. 2022).
This work’s main contribution is:

 To extract the Bigram-BoW (B-BoW), Threshold Term Frequency-


Inverse Document Frequency (T-TFIDF) features from the pre-proc-
essed data for better sentiment analysis
 To choose the optimal features with the new Self Improved Honey
Badger Algorithm (SI-HBA), to minimize the complexity of the model
and to lessen the time consumption.
 To introduce a new optimized crossover framework for precise senti-
ment classification. The optimized crossover framework includes the Bi-
LSTM and DBN trained with Transfer learning.
 The weight of Bi-LSTM is fine-tuned with the new SI-HBA model to
enhance the solutions’ classification accuracy and convergence speed.

This article’s remaining sections are organized as follows: The literature


on sentiment analysis is included in Section “Literature Review”. Sections
“Proposed Sentiment Classification Model: An Overview”, “Pre-Processing”,
“Feature Extraction”, and “Feature Selection” address the pre-processing,
feature extraction, feature selection, and optimal crossover frameworks for
sentiment classification. Discussion of the outcomes obtained using the
projected model is found in Section “Sentiment Classification with
Optimized Crossover Framework”. Section “Result and Discussion” pro-
vides a conclusion to this article.
4 S. S. SAMANT ET AL.

Literature Review
Alfrjani, Osman, and Cosma (2019) introduced a “Hybrid Semantic
Knowledgebase-Machine Learning technique” for mining views and sum-
marizing general opinions on a multi-point scale at the domain feature
level. This "Hybrid Semantic Knowledgebase-Machine Learning" technique
was shown to be suitable for expanding the database of semantic features
with higher-level precision since it improved the recall and accuracy of the
retrieved domain characteristics in a trial evaluation.
Khan, Qamar, and Bashir (2016) suggested a semi-supervised system for
sentiment analysis based on "Multi-objective model selection (MOMS)". To
increase classification accuracy and learn feature weights, they used a MOMS
procedure and a support vector machine. Comparing the results of the pro-
posed method to past studies on this issue, better results are obtained.
Wadawadagi & Pagi (Wadawadagi and Pagi 2019) studied the use of lin-
guistic kernels in "subjectivity identification, opinion extraction, and polar-
ity classification" along with the addition of a hierarchical framework
centered on the Document object model (DOM) tree. The "opinion polar-
ity" in the returning emotional material was categorized in the last stage
using the fine-grained linguistic kernels.
Ruz, Henrıquez, and Mascare~ no (2020) have discussed the focus of senti-
ment analysis during big events, including social movements or natural dis-
asters. By using the Bayes factor technique to dynamically modify the
quantity of edges provided by the training examples in the Bayesian net-
work classifier, they were able to produce a more realistic network. Given a
high number of training instances, the results demonstrated the benefit of
employing the Bayes factor measure and its superior prediction results
when compared to SVM and RF.
Nagamanjula and Pethalakshmi (2020) developed a novel method for identi-
fying a vocabulary set specific to Twitter Sentiment Analysis (TSA) in 2018.
They demonstrated that the "Twitter Specific Lexicon Set (TSLS)" was minimal
and, more crucially, portable across domains. A set of vectorized tweets are pro-
vided by this method, which machine learning algorithms can use as input.
Kang, Ahn, and Lee (2018) designed a text classification sentiment ana-
lysis method based on text-based hidden Markov models (TextHMMs).
The proposed model did not employ a predetermined lexicon sentiment,
but rather a word arrangement in training texts. They have studied textual
patterns to build ensemble TextHMMs. Moreover, when determining the
"hidden variables in TextHMMs," the semantic cluster data was taken into
consideration. They demonstrated that this method was superior to several
usual strategies in studies using a benchmark informational index and that
it might potentially organize verified attitudes.
CYBERNETICS AND SYSTEMS: AN INTERNATIONAL JOURNAL 5

According to Symeon et al. (Pandey, Rajpoot, and Saraswat 2017), pre-


processing models seem to be the first step in text classification, and adopting
proper pre-processing models can boost classification performance. Finally,
these schemes were classed according to their performance based on the
results obtained. Experiments were also conducted, demonstrating that the
devised technique outperformed traditional systems in terms of accuracy.
Ali, Kwak, and Kim (2016) established a robust sentiment classification
strategy with SVM and Fuzzy domain ontology (FDO). This framework
compiles information from several surveys on amenities and features. The
SVM detects Inn emphasizes surveys while channeling out irrelevant sur-
veys. FDO was also used to calculate the polarities phrase in each attribute.
The combination of FDO and SVM increased the accuracy rate of evalu-
ation, presumption word retrieval, and specificity of evaluation mining.
Araque, Sanchez-Rada, and Iglesias (2022) the framework is designed with
researchers and practitioners in mind, making it easier to replicate earlier sen-
timent models and providing solutions for frequent problems. This is accom-
plished by layering various abstractions on well-known libraries like NLTK
and scikit-learn. By using understandable Python code, GSITK enables users
to create intricate sentiment pipelines. The Open Source framework has been
utilized effectively in a number of research projects and competitions.

Proposed Sentiment Classification Model: An Overview


Recent history has seen a huge rise in the usage of social media, and senti-
ment analysis of reviews, comments, and opinions from social media has
gained popularity in the research community. Sentiment analysis is a sort
of data analysis that accurately portrays a person’s opinions about current
events, topics, products, and tweet reviews. Sentiment research helps you
understand societal trends and client opinions better. Therefore, there is a
requirement for accurate sentiment analysis.
In this study, a novel deep-learning-based sentiment analysis strategy is
created. "(a) pre-processing, (b) feature extraction, (c) feature selection, and
(d) sentiment classification" are the four main stages of the estimated tech-
nique. The suggested work’s architecture is shown in Figure 1. The planned
sentiment categorization model’s steps are illustrated below:
Let the collected data be denoted as Dinp :

Step 1- Initially, the collected raw data Dinp is pre-processed via tokenization,
stop-word removal, and stemming. The pre-processed data is denoted as Dpre :
Step 2- Then, from the pr-processed data Dpre , the features like Bigram-
BoW (B-BoW) g pBoW , Threshold- TFIDF (T-TFIDF) g pTFIFD , Unigram
6 S. S. SAMANT ET AL.

Figure 1. Block diagram of the proposed sentiment classification model.

g uni and N-Gram g ngram are extracted. These features are together denoted
as G ¼ g pBoW þ g pTFIFD þ g uni þ g ngram
Step 3- From the choosen features G, the most optimal features are picked
by utilizing the SI-HBA. This SI-HBA model is the conceptual improve-
ment of standard HBA. The selected optimal features are pointed as
Gopt follows.
Step 4- As a last step, the review classification is carried out using the sug-
gested optimal crossover framework, which was created by fuzing together
optimized Bi-LSTM and DBN that have each been trained using transfer
learning. The DBN and optimized Bi-LSTM are given training by using
best possible feature set of the SI-HBA model. The classification perform-
ance of the acquired reviews is enhanced by fine-tuning the weight of the
Bi-LSTM classifier using a self-improved Honey Badger Algorithm (SI-
HBA).Whether a good, negative, or neutral sentiment is indicated by the
reviews that were gathered will be shown in the final results.
CYBERNETICS AND SYSTEMS: AN INTERNATIONAL JOURNAL 7

Figure 2. Pre-processing stage.

Pre-Processing
The cleansing and prepping of the text for categorization is known as pre-
processing the data. Noise and ambiguous sections such as HTML tags,
javascript, and ads are common in electronic writings. Furthermore, most
phrases in the language have very little bearing on the overall direction of
the document. Retaining such phrases increases the problem’s complexity,
making categorization increasingly complicated as each phrase in text is
considered a uni-dimensional construct. The premise behind appropriately
pre-processing information is that reducing distortion in the text would
increase the effectiveness of the classification and quicken the categoriza-
tion, allowing for real-time sentiment analysis. Tokenization, stop-word
erasure, and stemming are used in our article to perform the pre-process-
ing. Figure 2 diagrammatically depicts this phase.
The pre-processing phase’s steps are illustrated below:

Step 1- Input data initially enters the tokenization phase. The tokens Dtoken
are first extracted from Dinp : In any NLP pipeline, tokenization is indeed
the preliminary stage. It has a significant impact on the remainder of
the pipelines.
Step 2- From Dtoken , the stop words are removed. Stop words are often
screened out before a natural language is processed. These are the most
frequent words in any language (“articles, prepositions, pronouns, conjunc-
tions, and so on”), and they don’t add anything to the text. Stop words in
English include "the," "a," "an," "so," and "what.” The data acquired after
stop word removal is denoted as Dstop :
Step 3- Then, the stemming is carried out upon Dstop : Whenever ’fluff’ let-
ters (not words) are eliminated from a phrase, the "stem form" is gathered
collectively. The terms ’play,’ ’playing,’ and ’plays,’ for example, imply the
same thing. Rather than having them as separate terms, we can group
them underneath the phrase ’play’. The outcome acquired after stemming
is the pre-processed data, and it is denoted as Dpre :
8 S. S. SAMANT ET AL.

Figure 3. Extracted features from pre-processed data.

Feature Extraction
From the pr-processed data Dpre , the features like Bigram-BoW (B-BoW),
Threshold Term Frequency-Inverse Document Frequency (T-TFIDF), Unigram,
and N-Gram are extracted. This phase is diagrammatically shown in Figure 3.

Bigram-Bag-of-Words
The Bag of Words (BoW; Chen, Yap, and Chau 2011) model is the sim-
plest form of text representation in numbers. The Bag-of-Words counts the
total occurrence of the most frequently utilized words in the document.
However, the Bag-of-Words suffers from the drawbacks like: “(a) If the
new sentences contain new words, then our vocabulary size would increase
and thereby, the length of the vectors would increase too and (b)
Additionally, the vectors would also contain many 0 s, thereby resulting in
a sparse matrix (which is what we would like to avoid)”. In this proposed
study, a proposed Bigram-BoW (B-BoW) model is presented. The Bigram-
BoW (B-BoW) makes the list of all the words in Dpre :

 Then, score the words in each document based on their frequency using
the bigram (instead of unigram that is being followed in standard BoW).
 The frequency of each word in Dpre is computed, and the positive
weight is computed as per Eq. (1).

PCi
K
i¼1 ti
B¼ (1)
Ct

Here, Ct denotes the keyword t in Dpre and ti is the frequency of the


word t in Dpre : In addition, Kti is the position weight coefficient of ti: The
CYBERNETICS AND SYSTEMS: AN INTERNATIONAL JOURNAL 9

Table 1. Value allocation for words.


Noun 1
adjective 0.75
verb 0.5
others 0.3

position words are allocated as shown in Table 1. The extracted feature is


denoted as g pBoW

Threshold- TFIDF (T-TFIDF)


The TF-IDF (Dalaorao, Sison, and Medina 2019) idea has been employed
for text search and retrieval. A mathematical metric called "term frequency-
inverse document frequency" demonstrates how crucial a phrase is to a
document inside a corpus or collection. The traditional TF-IDF has a num-
ber of shortcomings, including: I Because it is based on the bag-of-words
(BoW) model, it ignores factors such as text position, semantics, co-occur-
rences across texts, and other factors. (ii) Semantics are still not captured
(e.g. as compared to topic models, word embeddings). As a result, we’ve
included a new feature extraction phase based on T-TFIDF. The newly pro-
posed T-TFIDF characteristics may be stated mathematically as Eq. (2).
Thrðtft  id:ft Þ
f WTFIDF ¼ qffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi
P (2)
t2d½ t
tf  idft 2

Here, tft and idft denotes the term frequency and inverse document fre-
quency, respectively. In addition, Thr indicates the threshold value that is
fixed between [-1,þ1]. Here, the most extreme negative is represented by
1, and the most extreme positive is represented by þ1. In addition, d, t
denotes the count of files in the corpus and count of documents containing
the words, respectively. The extracted feature is denoted as g pTFIFD :

Unigram
The features based on unigrams are retrieved using a dictionary created
using user-defined parameters. It is assumed that the term’s appearance is
independent to the word before it. The extracted unigram based feature
from Dpre is pointed as g uni :

N-Gram
Text N-grams are commonly employed in text mining and natural language
processing. An n-gram is a continuous series of n components from a given
sample of text. N-gram may be calculated mathematically using Eq. (3).
10 S. S. SAMANT ET AL.

Figure 4. Solution encoding.

g Ngram ¼ XðN1Þ (3)

Here, X denotes the count of words in the sentence and N (N-gram; i.e.
for unigram N ¼ 1, for bigram N ¼ 2 as well). The extracted unigram based
feature from Dpre is pointed as g Ngram : The extracted features are together
denoted as G ¼ g pBoW þ g pTFIFD þ g uni þ g ngram :

Feature Selection
SA-HBA
The development of a new SI-HBA model resolves the optimization
issue that is the focus of the current research. This SI-HBA model was
created using the conventional HBA (Hashim et al. 2022) model, which
is based on the feeding habits of honey badgers. The "digging phase"
and the "honey phase" are the two main phases that the suggested SI-
HBA model covers. The input to SI-HBA is the weight of BI-LSTM
W1, 2, ::::Q and the extracted features G: The solution encoding is shown
in Figure 4.
Initialization: the population of the solutions is initialized. Here, N
denotes the population size and D is the dimension. This is shown in Eq.
(4) and Eq. (5), respectively.
2 3
p11 p12 :::: p1D
6 p21 p22 :::: p2D 7
6 7
P¼6 6 ::::::::::::::::::: 7 (4)
7
4 pN1 pN2 pN3 pND 5

pi ¼ LBi þ r1  ðUBi  LBi Þ (5)


The upper and lower boundaries of the solution iare indicated by the symbols
UBi and LBi , respectively. In addition, is a number generated at random
from 0 to 1. itr stands for the current iteration, while maxitr stands for the
maximum iteration.

Step 1- Using Eq. (6), determine the search agent’s fitness. The main goal
of this study is to reduce categorization errors as much as possible. This
may be expressed mathematically in Eq.(6).
CYBERNETICS AND SYSTEMS: AN INTERNATIONAL JOURNAL 11

obj ¼ minðCerror Þ (6)


Here, Cerror are the classification errors.

Step 2- The best position is saved as pprey and assigned fitness to fitprey
Step 3- Check the termination criterion: While itr  maxitr : If this condition
is satisfied, then processed to Step 4.
Step 4- Use the newly projected equation (Eq. (7) and Eq. (8)) to update
the decreasing factor of the density factor a: The density factor governs
time-varying randomness to guarantee a successful transition from explor-
ation to exploitation. To limit randomness over time, update a lowering
factor that lowers with iteration.
 
itrðradÞ
a ¼ C  exp (7)
maxitr

Here, a new variable rad is introduced to avoid the solution from getting
trapped into the local optima.
Q
rad ¼ (8)
2p

Here, Q is the circumference of the prey and rad is the radius between
the honey badger (search agent) and prey (target: optimal weight or opti-
mal feature) and constant C ¼ 2:
Step 5- For i ¼ 1 to N, compute the intensity of the solutions Ii using the
newly projected expression (proposed) given in Eq. (9)- Eq. (11). The
prey’s concentration strength and the distance between it and the honey
badger determine the intensity. Ii is the prey’s scent strength; if the smell
is strong, the motion will be quick, and vice versa, according to the
Inverse Square Law.
 
S
Ii ¼ rand2  Mean  (9)
4pD2i
rand2 is a number between 0 and 1 that is created at random. In order to
determine the best answers, the mean is computed amongst the gathered
solutions. As a result, the solutions’ convergence speed improves.
S ¼ ðpitr pitrþ1 Þ2 (10)
Di ¼ pprey pi (11)
The distance between the prey and the ith search agent is denoted by Di ,
while the source strength or concentration strength is denoted by S:
Step 6- Generate a random number r between 0 to 1.
12 S. S. SAMANT ET AL.

Step 7- If1 r<0:5, then update the position of pnew in the digging phase using
Eq. (12). During the digging phase, a honey badger adopts a cardioid shape.

¼ pnew ¼ pprey þ flag  b  I  pprey þ flag  rand3  a  Di  Cosð2prand4Þ


 ½1 cos ð2prand5Þ
(12)

Here, pprey denotes the prey’s position, which is the best search agent so
far, and flag denotes the flag that changes the search direction.
Step 8- elseIf1 r  0:5, then update the position of pnew using the honey stage as
shown in Eq. (13). The scenario in which a honey badger follows a honeyguide
bird to a beehive is recreated here.

pnew ¼ pprey þ flag  rand7  a  Di (13)

The random value rand7 is generated between 0 to 1.


Step 9- End if1
Step 10- Compute new position and assign it to pnew .
Step 11- If 2fnew  fi , then set pi ¼pnew and fi ¼ fnew
Step 12- End if 2
Step 13- If 3fnew  fprey , then set pprey ¼ pnew and fprey ¼ fnew
Step 14- End if 3
Step 15- Perform arithmetic crossover
Step 16- End for
Step 17- End while
Step 18- Return pprey (optimal feature and optimal weight of Bi-LSTM)

The selected optimal feature is denoted as Gopt : The feature selection phase
is shown in Figure 5.

Figure 5. Optimal feature selection.


CYBERNETICS AND SYSTEMS: AN INTERNATIONAL JOURNAL 13

Figure 6. Optimized crossover framework for sentiment classification.

Sentiment Classification with Optimized Crossover Framework


The review classification is undergone via the proposed optimized crossover
framework, constructed by hybridizing the optimized Bi-LSTM and DBN
trained with Transfer learning, respectively. The SI-HBA model’s optimally
chosen features are used to train the DBN and optimized Bi-LSTM. The
classification performance of the collected reviews is enhanced by fine-tun-
ing the weight of the Bi-LSTM classifier using a self-improved Honey
Badger Algorithm (SI-HBA). Whether a good, negative, or neutral senti-
ment is indicated by the reviews that were gathered will be shown in the
final results. Figure 6 depicts the built-in optimized crossover framework.

Bi-LSTM
A particular kind of artificial RNN utilized in deep learning is the Bi-
LSTM (Rajeyyagari 2020). The LSTM’s advantage is that it can analyze a
stream of data as opposed to just one data point. A typical Bi-LSTM unit
consists of a cell, an input gate, an output gate, and a forget gate. The three
gates control how information enters and leaves the cell, and the cell
retains values for an extended length of time. Memory blocks are a unique
type of special unit found in the Bi-recurrent LSTM’s hidden layer. Self-
connections are placed in memory cells to save the network’s temporal
state within the memory blocks. The gates also function as particular multi-
plicative units for managing data flow. The two gates in the Bi-LSTM
architecture are an input gate and an output gate. The output gate regulates
the activity flow in the unused portions of the network, while the input
14 S. S. SAMANT ET AL.

gate regulates the input activation flows into the memory cell. There is also
a forget gate in the memory block. In order to adaptively forget or reset
the memory of the cell, the forget gate scales the internal state of the cell
before adding it to the cell (as input) via the cell’s self-recurrent link. E
and H represent the hidden and cell states of the LSTM, respectively. it is
trained with Gopt
In addition, for the time instant t, outt and Inpt stand for the forget gate
and input gate, respectively. The Bi-LSTM model is mathematically mod-
eled as per Eq. (14)- Eq. (17), respectively.
 
opt
Inpt ¼ r WeInp Gt þ RInp :Ht1 þ BInp (14)
 opt

Fort ¼ r WeFor Gt þ RFor :Ht1 þ BFor (15)
 opt

Ct ¼ r WeC Gt þ RC :Ht1 þ BC (16)
 opt

outt ¼ r Weout Gt þ Rout :Ht1 þ Bout (17)
The notation r denotes the gate’s activation function (sigmoid). In add-
ition, WeInp , WeFor , WeC and Weout denotes the input weight metrics.
opt
RInp , RFor , RC and Rout are recurrent weight matrices. In addition, Gt is
the input and Ht1 is the output at the previous time t1: Moreover, BInp ,
BFor , BC and Bout points to the bias vector.
The notation r denotes the gate’s activation function (sigmoid). In add-
ition, WeInp , WeFor , WeC and Weout denotes the input weight metrics.
opt
RInp , RFor , RC and Rout are recurrent weight matrices. In addition, Gt is
the input and Ht1 is the output at the previous time t1: Moreover, BInp ,
BFor , BC and Bout points to the bias vector.
The outcome acquired from Bi-LSTM is denoted as OBiLSTM :

DBN with Transfer Learning


Typically, Smolensky began DBN (Ayesha 2021) in 1986. There are many
multiple levels within the DBM. The input layers have visible neurons,
whereas the output layers have hidden neurons. DBNs, built from RBMs,
are among the essential instruments for deep learning. RBMs have excellent
training mechanisms, making them ideal as DBN key components. The
input neurons and the hidden neurons are also interconnected. The DBN
uses the stochastic neuron model to provide an accurate output for a given
input. Moreover, the DBM of the Boltzman network creates output prob-
abilistically. The DBN can construct a strong probabilistic model using the
training data. Nonetheless, determining the best structure for a particular
set of applications is difficult. As a result, the DBN with optimal transfer
learning is automatically introduced to determine its architecture. As a
result, learning progresses more quickly and the model become more
CYBERNETICS AND SYSTEMS: AN INTERNATIONAL JOURNAL 15

accurate. The following are the steps taken in the proposed Transfer learn-
ing based Deep belief network (DBN):

 Pre-train the network after initializing the basic DBN structure with a
single hidden layer.
 Transmit learning is used to move the knowledge of the neurons and
hidden layers to the newly inserted neurons and hidden layers.

RBMs, or probabilistic graphical models in stochastic neural networks,


are capable of learning a probability distribution across a collection of
inputs. RBMs are Boltzmann machines with the added constraint that their
neurons can form a bipartite graph. The bipartite graph is one in which
two distinct vertices V1 (visible units) and V2 (hidden units). There seems
to be no link across nodes inside the same set, and all these two sets have
a symmetric relationship. it is trained with Gopt
The visible and concealed units in a typical RBM are represented by bin-
ary values. A Bernoulli-Bernoulli RBM is the name given to this type of
RBM. A continuous probability distribution with two likely outcomes is the
Bernoulli distribution, denoted by n ¼ 0, 1, respectively. If n ¼ 1, the real
value happens with probability p, and if n ¼ 0, the false case occurs with
likelihood q ¼ 1p, wherein 0<p<1 is the true value.
RBM seems to be an energy-based paradigm with n viewable units and
m concealed units. The visible and hidden unit states are represented by
vectors vis and hid, respectively. According to Eq. (18), an RBM system’s
energy is defined; given a set of states ðvis, hidÞ:
The state of the ith visible unit is vis, while the state of the jth hidden
unit is hid: wij denotes the weights of visible and hidden unit connec-
tions. Bias weights (offsets) Ai for visible units and Bj for hidden units
are indeed available. After defining the parameters, the joint probability
distribution of ðvis, hidÞ is calculated in terms of the energy function as
per Eq. (18).
Xn X
m n X
X m
Eðvis, hidÞ ¼  Ai visi  Bj hidj  visi wij hidj (18)
i¼1 j¼1 i¼1 j¼1

Once the parameters are established, using Eq (19) and (20), we can
determine the joint probability distribution of (v, h) in terms of the energy
function, respectively.
1
probðvis, hidÞ ¼ eEðvis, hisÞ (19)
Z
X
Z¼ eEðvis, hidÞ (20)
vis, hid
16 S. S. SAMANT ET AL.

where Z is a constant for normalization. The activation states of every hid-


den unit are independent of one another whenever the state of a visible
unit is known. Following that, the likelihood of activation of the jth hidden
unit is computed as per Eq. (21).
X
pðhisj ¼ 1jvisÞ ¼ aðBj þ wij Þ (21)
ivisi

A logistical sigmoid activation function is aðxÞ ¼ 1þe1ðxÞ Likewise, the


activation states of the each visible unit are conditionally independent a
certain hidden state and the probability of the ith visible units of visible
state given hidden state is derived as per Eq. (22).
X
pðvisj ¼ 1jhidÞ ¼ aðAI þ wij Þ (22)
jvhidj

The following formula is used to compute a log-likelihood of training


data concerning weight W: This is shown in Eq. (23).
@ log pðvisÞ
¼ hvisihisj idata hvisihisj imodel0 (23)
@Wij

In the data or model distribution, h:idata and h:imodel denote anticipated


values. In log-likelihood-based training data, the training criteria for con-
nection weights are per Eq. (24).
 0
DWij ¼ - hvisihisj idata  hvisihisj imodel (24)

The learning rate is denoted by the symbol -: We could easily gather an


undisturbed sample of hvisihisj idata since there are no obvious ties in the
hidden layer of an RBM. Nevertheless, because the hvisihisj imodel needs expo-
nentially duration, computing an undisturbed sample is challenging. A
quick learning method known as Contrastive Divergence (CD) was devel-
oped to solve this problem.
After the hidden units’ states have been selected, a "reconstruction" is
created by setting each vis to 1 with such a likelihood determined by
Eq.(25).
 0
DWij ¼ - hvisihisj idata  hvisihisj irecon (25)

It is a decent approximation to hvisihisj imodel because hvisihisj idata is the


arithmetic mean over all data input for every updating and hvisihisj irecon is
indeed the arithmetic mean over reconstruction. The outcome acquired is
denoted as ODBN : The ultimate outcome is the mean of ODBN and OBiLSTM
CYBERNETICS AND SYSTEMS: AN INTERNATIONAL JOURNAL 17

Result and Discussion


Simulation Procedure
Python has been used to implement the proposed sentiment analysis meth-
odology. The evaluations were done at several LPs, namely 60, 70, 80, and
90. The anticipated model has been contrasted with the current models like
SVM, RF, MLP, DT, DCNN (Jianqiang, Xiaolin, and Xuejun 2018), GA
(Iqbal et al. 2019), DOA, DOXA, SSOA, BOA, and HBA, respectively. The
evaluation has been made regarding “positive, negative and miscellaneous
measures”. Both positive performance (specificity, sensitivity, precision, and
accuracy) and negative performance (false positive rate (FPR), false negative
rate (FNR), and other metrics) are assessed for the proposed classifier and
approach (F1-score and MCC). The suggested model’s positive parameters’
specificity, sensitivity, precision, and accuracy ought to be kept as high as
possible. On the other side, negative indicators like the FPR and FNR
should be kept as low as possible.

Dataset Description
The dataset for assessment is taken from: https://analyticsdrift.com/top-15-
datasets-for-sentiment-analysis-with-significant-citations/. Thirty percent of
the data gathered are used for testing, and the leftover seventy percent are
used for training. The Yelp’13 and IMDB reviews with “segment-level
polarity labels” (positive/neutral/negative) were combined to create the
SPOT sentiment analysis dataset, which contains 197 reviews. Sentences
and Elementary Discourse Units are the two levels of granularity where
annotations have been gathered (EDUs). Totally 2591 reviews are consid-
ered for this study. The dataset is perfect for testing approaches aimed at
segment-level and fine-grained sentiment prediction.

Performance Analysis on Testing Accuracy


For each variation in the LP, The highest accuracy readings were obtained
using the predicted model. Tables 2–5 show the results that were achieved,
respectively. The projected model has shown the greatest accuracy for all
LP variations. The anticipated model had a maximum accuracy of 93.54%
at the 90th LP. Additionally, the projected model’s highest sentiment cat-
egorization accuracy at the 80th LP was 91.82806, which is enhanced over
SVM ¼ 88.78921, RF ¼ 87.48917743, MLP ¼ 88.36221, DT ¼ 89.0586098,
DCNN (Jianqiang, Xiaolin, and Xuejun 2018) ¼87.497721, GA (Iqbal et al.
2019) ¼88.83073, DOA ¼ 86.323975728, DOXA ¼ 88.575749, SSOA ¼
89.41952, BOA ¼ 89.3989505 and HBA ¼ 89.42981. The introduction of a
new, optimized crossover classifier framework for sentiment classification is
18
S. S. SAMANT ET AL.

Table 2. Performance Analysis of the projected sentiment analysis model with optimized crossover classifier framework: Accuracy.
DOA þ DOXA þ SSOA þ BOA þ HBA þ SI-HBA þ
DCNN (Jianqiang, crossover crossover crossover crossover crossover crossover
Percentage Xiaolin, and GA (Iqbal classifier classifier classifier classifier classifier classifier
(%) SVM RF MLP DT Xuejun 2018) et al. 2019) framework framework framework framework framework framework
60 77.5501 79.28002 86.37899 85.6866 86.24184 84.71725 81.37901 82.71071 81.22962 80.41529 87.19045 89.407
70 86.42203 84.81105 88.22619 87.12184 87.36478 85.64309 83.79588 86.72254 84.56748 85.03198 89.10113 91.062
80 88.78922 87.48918 88.36221 89.05861 87.49772 88.83073 86.32398 88.57575 89.41953 89.39895 89.42981 91.82806
90 91.79274 91.29498 92.06983 91.95122 87.6092 91.43474 89.35776 91.56226 92.0797 92.06983 92.11916 93.54201
Table 3. Performance Analysis of the projected sentiment analysis model with optimized crossover classifier framework: Precision.
DCNN
(Jianqiang, DOA þ DOXA þ SSOA þ BOA þ HBA þ SI-HBA þ
Xiaolin, and crossover crossover crossover crossover crossover crossover
Percentage Xuejun GA (Iqbal classifier classifier classifier classifier classifier classifier
(%) SVM RF MLP DT 2018) et al. 2019) framework framework framework framework framework framework
60 76.55309 75.67735 86.09524 84.89108 72.61913 82.29053 81.8116 80.1628 79.02274 77.9492 88.28955 90.33242
70 86.34398 78.14407 87.44712 86.22608 86.55079 83.84125 82.68269 81.58098 80.18023 79.28348 88.4889 90.95782
80 88.12229 87.07446 88.45157 87.84766 86.65414 87.80949 87.49627 87.84232 85.91789 88.24308 90.58501 92.10971
90 90.50345 89.40628 90.031 90.36564 87.96859 90.53813 90.30596 89.64969 90.35713 90.34009 92.33536 93.96676
CYBERNETICS AND SYSTEMS: AN INTERNATIONAL JOURNAL
19
20
S. S. SAMANT ET AL.

Table 4. Performance Analysis of the projected sentiment analysis model with optimized crossover classifier framework: F1-score.
DCNN
(Jianqiang, DOA þ DOXA þ SSOA þ BOA þ HBA þ SI-HBA þ
Xiaolin, crossover crossover crossover crossover crossover crossover
Percentage and Xuejun GA (Iqbal classifier classifier classifier classifier classifier classifier
(%) SVM RF MLP DT 2018) et al. 2019) framework framework framework framework framework framework
60 75.67568 78.14371 85.69158 84.48282 86.12615 83.63879 81.87609 80.16492 79.02176 77.95276 88.06303 90.44404
70 76.4992 85.99196 87.25889 86.03081 87.24895 86.79919 82.68003 81.58285 80.1509 79.28311 88.27502 90.74501
80 87.54668 88.55807 87.50559 86.84115 87.41319 89.80916 87.452 87.83297 85.72281 86.18824 90.36066 91.9837
90 89.77078 92.05005 89.24815 90.01946 88.03171 92.17497 90.29453 89.56859 90.35123 90.33235 92.28726 93.90249
Table 5. Performance analysis of the projected sentiment analysis model with optimized crossover classifier framework: MCC.
DCNN DOA þ DOXA þ SSOA þ BOA þ HBA þ SI-HBA þ
(Jianqiang, crossover crossover crossover crossover crossover crossover
Percentage Xiaolin, and GA (Iqbal classifier classifier classifier classifier classifier classifier
(%) SVM RF MLP DT Xuejun 2018) et al. 2019) framework framework framework framework framework framework
60 75.73362 72.54056 75.64814 74.27443 76.33068 74.18818 75.53932 72.98386 72.06794 77.26195 79.0119 81.02263
70 80.4513 77.85612 79.326 80.99051 80.34718 80.53436 81.58991 80.02691 81.71374 81.67248 81.73436 83.52166
80 86.49555 85.48857 87.0571 86.81664 80.7728 85.77107 87.0571 87.15718 87.07713 86.03018 87.19718 90.09583
90 91.84366 91.3741 92.23413 91.99994 81.55799 91.68726 92.1756 90.76053 91.47058 92.23413 92.23413 93.6981
CYBERNETICS AND SYSTEMS: AN INTERNATIONAL JOURNAL
21
22 S. S. SAMANT ET AL.

the main factor contributing to the improvement in sentiment classification


accuracy for planned method which has been shown to have the maximum
precision, specificity, and sensitivity for any change in the LP. The esti-
mated approach had the best precise readings at the 60th, 70th, 80th, and
90th LPs: 90.3%, 90.95%, 92.1%, and 93.96%. This improvement over the
predicted model is primarily attributable to the careful selection of the opti-
mal features for training the new, enhanced crossover classifier framework.
The anticipated model also identified the largest F1 score. The suggested
technique’s F1-score at the 90th LP is 93.90248, which is better than SVM
¼ 89.77, RF ¼ 92.05, MLP ¼ 89.25, DT ¼ 90.02, DCNN (Jianqiang,
Xiaolin, and Xuejun 2018) ¼88.03, GA (Iqbal et al. 2019) ¼92.17, DOA ¼
90.29, DOXA ¼ 89.57, SSOA ¼ 90.35, BOA ¼ 90.33 and HBA ¼ 92.29.
The highest MCC, 93.6%, was reported by the predicted model at the 90th
LP. The reduction in classification errors can be attributed in large part to
the extraction of the most relevant variables for sentiment classification.

Ablation Study
The anticipated model has received approval in 4 distinct fields: (a)pro-
posed model with conventional BOW þ Threshold- TFIDF (T-TFIDF),
Unigram and N-Gram þ optimal feature selection þ optimized crossover
classifier framework); (b) proposed model with conventional TF-
IDF þ Bigram-BoW (B-BoW), Unigram and N-Gram þ optimal feature
selection þ optimized crossover classifier framework); (c) proposed model
without optimal feature selection þ feature selection (Bigram-BoW (B-
BoW), Threshold- TFIDF (T-TFIDF), Unigram and N-Gram) þ optimized
crossover classifier framework; (d) proposed model with feature selection
(Bigram-BoW (B-BoW), Threshold- TFIDF (T-TFIDF), Unigram and N-
Gram) þ optimized crossover classifier framework þ optimal feature selec-
tion. With standard BOW model, standard TF-IDF model, and without
optimal feature selection, the projected model can only provide the accur-
acy level as 89.2%, 90.3%, and 87.5% when assessing the acquired out-
comes. But when a better BOW, TF-IDF, and feature selection are used,
the projected model has improved dramatically. Table 6 displays the results
that were obtained.

Performance Analysis of Training Accuracy


Figure 7 depicts the analysis on training accuracy. At 90th percentage the
accuracy is ( 0.98) when compared to the other existing method like
SVM, RF, MLP, DT, DCNN, GA(, DOA, DOXA, SSOA, BOA and HBA.
Therefore, when compared to the existing approach, the accuracy,
Table 6. Ablation study of the projected sentiment classification model.
Proposed work (improved
proposed model with proposed model without optimal feature selection þ
conventional BOW þ optimized proposed model with conventional feature selection þ optimized improved feature extraction þ
crossover classifier TF-IDF þ optimized crossover classifier optimized crossover
Metrices framework crossover classifier framework framework classifier framework )
specificity 0.889823381 0.900336417 0.874264087 0.913372582
sensitivity 0.894643719 0.90601951 0.877207392 0.919694371
precision 0.913900756 0.923375364 0.899292691 0.934706815
npv 0.865970943 0.878449072 0.847620018 0.894011113
mcc 0.782166025 0.804086989 0.749188627 0.830889595
fpr 0.110176619 0.099663583 0.125735913 0.086627418
fnr 0.105356281 0.09398049 0.122792608 0.080305629
F1-score 0.904169715 0.914615107 0.88811276 0.927139826
accuracy 0.892554452 0.903575788 0.875916448 0.917002417
CYBERNETICS AND SYSTEMS: AN INTERNATIONAL JOURNAL
23
24 S. S. SAMANT ET AL.

Figure 7. Analysis on training (a) Accuracy (b) Precision (c) F-Measure and (d) MCC.

precision, F-measure, and MCC of proposed model hold the greatest val-
ues. The improvement of the suggested method is thus demonstrated.

Convergence Analysis
Using a new SA-HBA model, the optimization problem with the sentiment
classification model was resolved. The SA-HBA model is used to select the best
characteristics and change the Bi-weight LSTM’s in order to increase classifica-
tion accuracy. The fitness function behind the SA-HBA model minimizes clas-
sification errors. Therefore, it should record the minimal cost function to
exhibit it as the best approach in terms of convergence speed. Interestingly,
due to improvements made to the HBA model, The estimated scheme will
adjust for changes in the count of iterations has consistently returned the func-
tion with the lowest cost. (updating of density factor as well). Moreover, at the
100th iteration, the HBA model can record the cost function as 1.09. But, with
SA-HBA, the enhanced convergence of 1.07 has been recorded. Thus, it is clear
that with self-adaptation within standard optimization, the solutions’ conver-
gence rate can be boosted. The findings are presented in Figure 8.

Analysis on Loss Measure


Figure 9 depicts the loss measure analysis. A loss function evaluates the
accuracy with which your prediction model can forecast the desired result.
CYBERNETICS AND SYSTEMS: AN INTERNATIONAL JOURNAL 25

Figure 8. Convergence analysis of the projected SI-HBA model.

Figure 9. Loss measure analysis.

Table 7. Computation time analysis.


Methods Time (sec)
SVM 97.47
RF 104.37
MLP 196.07
DT 131.36
DCNN 219.143
GA 263.128
DOA 258.25
DOXA 175.57
SSOA 286.20
BOA 127.13
HBA 104.57
SI-HBA 94.65

Loss results from an inaccurate prediction. In other terms, loss represents


how poorly a single case was predicted by the model. If the model’s fore-
cast is inaccurate, loss will be greater; if the model’s forecast is accurate,
loss will be smaller.
26 S. S. SAMANT ET AL.

Analysis on Computation Time


The findings of the computational time are illustrated in Table 7. The com-
plexity (computation time) of the developed method is 2.97%, 10.26%,
71.51%, 38.78%, 80.15%, 78.01%, 72.84%, 85.49%, 88.37%, 34.31% and
10.48% better than the SVM, RF, MLP, DT, DCNN, GA, DOA, DOXA,
SSOA, BOA, and HBA. The improvement of the provided model over the
other typical model is thus demonstrated.

Conclusion
This study created a new sentiment analysis model based on deep learning.
Here, the SI-HBA model’s ideally chosen features are used to train the
hybrid classifier within an optimized crossover framework. By means of a
SI-HBA, for improvising the classification results of the gathered reviews,
the weights of the Bi-LSTM classifier is adjusted. The final results will indi-
cate whether the reviews are mostly good, negative, or neutral. The pro-
posed sentiment classification model is then validated by a comparison
analysis. At the 90th LP, the predicted strategy had the maximum accuracy,
93.54%. Moreover, at 80th LP, the planned technique has observed as the
largest sentiment classification accuracy as 91.82806, that is superior to
SVM ¼ 88.78921, RF ¼ 87.48917743, MLP ¼ 88.36221, DT ¼ 89.0586098,
DCNN (Jianqiang, Xiaolin, and Xuejun 2018) ¼87.497721, GA (Iqbal et al.
2019) ¼88.83073, DOA ¼ 86.323975728, DOXA ¼ 88.575749, SSOA ¼
89.41952, BOA ¼ 89.3989505 and HBA ¼ 89.42981. Designing a new opti-
mized crossover classifier architecture for sentiment classification is the
main factor contributing to the projected model’s improved accuracy in
sentiment classification.

References
Alattar, F., and K. Shaalan. 2021. Using artificial intelligence to understand what causes
sentiment changes on social media. IEEE Access 9:61756–67. doi:10.1109/ACCESS.2021.
3073657.
Alfrjani, R., T. Osman, and G. Cosma. 2019. A hybrid semantic knowledgebase-machine
learning approach for opinion mining. Data & Knowledge Engineering 121:88–108. doi:
10.1016/j.datak.2019.05.002.
Ali, F., K. S. Kwak, and Y. G. Kim. 2016. Opinion mining based on fuzzy domain ontology
and Support Vector Machine: A proposal to automate online review classification.
Applied Soft Computing 47:235–50. doi:10.1016/j.asoc.2016.06.003.
Araque, O., J. F. Sanchez-Rada, and C. A. Iglesias. 2022. GSITK: A sentiment analysis
framework for agile replication and development. SoftwareX 17:100921. doi:10.1016/j.
softx.2021.100921.
CYBERNETICS AND SYSTEMS: AN INTERNATIONAL JOURNAL 27

Aydin, C. R., and T. G€ ung€or. 2020. Combination of recursive and recurrent neural net-
works for aspect-based sentiment analysis using inter-aspect relations. IEEE Access 8:
77820–32. doi:10.1109/ACCESS.2020.2990306.
Ayesha, H. 2021. Race detection using mutated salp swarm optimization algorithm based
DBN from face shape features. Multimedia Research 4 (2):879–905.
Aziz, A. A., and A. Starkey. 2020. Predicting supervise machine learning performances for
sentiment analysis using contextual-based approaches. IEEE Access 8:17722–33. doi:10.
1109/ACCESS.2019.2958702.
Chan, J. Y. L., K. T. Bea, S. M. H. Leow, S. W. Phoong, and W. K. Cheng. 2022. State of
the art: a review of sentiment analysis based on sequential transfer learning. Artificial
Intelligence Review:1–32.
Chen, T., K. H. Yap, and L. P. Chau. 2011. From universal bag-of-words to adaptive bag-
of-phrases for mobile scene recognition. In 2011 18th IEEE International Conference on
Image Processing (825–8). IEEE.
Dalaorao, G. A., A. M. Sison, and R. P. Medina. 2019. Integrating collocation as tf-idf
enhancement to improve classification accuracy. In 2019 IEEE 13th International
Conference on Telecommunication Systems, Services, and Applications (TSSA) (pp.
282–285). IEEE. doi:10.1109/TSSA48701.2019.8985458.
Deng, D., L. Jing, J. Yu, and S. Sun. 2019. Sparse self-attention LSTM for sentiment lexicon
construction. IEEE/ACM Transactions on Audio. Speech, and Language Processing 27
(11):1777–90.
Gao, Y., J. Liu, P. Li, and D. Zhou. 2019. CE-HEAT: an aspect-level sentiment classification
approach with collaborative extraction hierarchical attention network. IEEE Access 7:
168548–56. doi:10.1109/ACCESS.2019.2954590.
Gu, T., G. Xu, and J. Luo. 2020. Sentiment analysis via deep multichannel neural networks
with variational information bottleneck. IEEE Access 8:121014–21. doi:10.1109/ACCESS.
2020.3006569.
Hashim, F. A., E. H. Houssein, K. Hussain, M. S. Mabrouk, and W. Al-Atabany. 2022.
Honey Badger Algorithm: New metaheuristic algorithm for solving optimization prob-
lems. Mathematics and Computers in Simulation 192:84–110. doi:10.1016/j.matcom.2021.
08.013.
Iqbal, F., J. M. Hashmi, B. C. Fung, R. Batool, A. M. Khattak, S. Aleem, and P. C. Hung.
2019. A hybrid framework for sentiment analysis using genetic algorithm based feature
reduction. IEEE Access 7:14637–52. doi:10.1109/ACCESS.2019.2892852.
Januario, B. A., A. E. D. O. Carosia, A. E. A. da Silva, and G. P. Coelho. 2022. Sentiment
analysis applied to news from the Brazilian stock market. IEEE Latin America
Transactions 20 (3):512–8. doi:10.1109/TLA.2022.9667151.
Jianqiang, Z., G. Xiaolin, and Z. Xuejun. 2018. Deep convolution neural networks for twit-
ter sentiment analysis. IEEE Access 6:23253–60. doi:10.1109/ACCESS.2017.2776930.
Kang, M., J. Ahn, and K. Lee. 2018. Opinion mining using ensemble text hidden Markov
models for text classification. Expert Systems with Applications 94:218–27. doi:10.1016/j.
eswa.2017.07.019.
Kastrati, Z., A. S. Imran, and A. Kurti. 2020. Weakly supervised framework for aspect-
based sentiment analysis on students’ reviews of MOOCs. IEEE Access 8:106799–810.
doi:10.1109/ACCESS.2020.3000739.
Khan, F. H., U. Qamar, and S. Bashir. 2016. Multi-objective model selection (MOMS)-
based semi-supervised framework for sentiment analysis. Cognitive Computation 8 (4):
614–28. doi:10.1007/s12559-016-9386-8.
28 S. S. SAMANT ET AL.

Li, Z., R. Li, and G. Jin. 2020. Sentiment analysis of danmaku videos based on naïve bayes
and sentiment dictionary. IEEE Access 8:75073–84. doi:10.1109/ACCESS.2020.2986582.
Liang, H., U. Ganeshbabu, and T. Thorne. 2020. A dynamic Bayesian network approach
for analysing topic-sentiment evolution. IEEE Access 8:54164–74. doi:10.1109/ACCESS.
2020.2979012.
Mehanna, Y. S., and M. B. Mahmuddin. 2021. A semantic conceptualization using tagged
bag-of-concepts for sentiment analysis. IEEE Access 9:118736–56. doi:10.1109/ACCESS.
2021.3107237.
Nagamanjula, R., and A. Pethalakshmi. 2020. A novel framework based on bi-objective
optimization and LAN2FIS for Twitter sentiment analysis. Social Network Analysis and
Mining 10 (1):1–16. doi:10.1007/s13278-020-00648-5.
Obiedat, R., D. Al-Darras, E. Alzaghoul, and O. Harfoushi. 2021. Arabic aspect-based senti-
ment analysis: A systematic literature review. IEEE Access 9:152628–45. doi:10.1109/
ACCESS.2021.3127140.
Pandey, A. C., D. S. Rajpoot, and M. Saraswat. 2017. Twitter sentiment analysis using
hybrid cuckoo search method. Information Processing & Management 53 (4):764–79. doi:
10.1016/j.ipm.2017.02.004.
Phan, H. T., V. C. Tran, N. T. Nguyen, and D. Hwang. 2020. Improving the performance
of sentiment analysis of tweets containing fuzzy sentiment using the feature ensemble
model. IEEE Access 8:14630–41. doi:10.1109/ACCESS.2019.2963702.
Rajeyyagari, S. 2020. Automatic speaker diarization using deep LSTM in audio lecturing of
e-Khool platform. Journal of Networking and Communication Systems 3 (4):17–25.
Ruz, G. A., P. A. Henrıquez, and A. Mascare~ no. 2020. Sentiment analysis of Twitter data
during critical events through Bayesian networks classifiers. Future Generation Computer
Systems 106:92–104. doi:10.1016/j.future.2020.01.005.
Sehar, U., S. Kanwal, K. Dashtipur, U. Mir, U. Abbasi, and F. Khan. 2021. Urdu sentiment
analysis via multimodal data mining based on deep learning algorithms. IEEE Access 9:
153072–82. doi:10.1109/ACCESS.2021.3122025.
Silva, H., E. Andrade, D. Ara ujo, and J. Dantas. 2022. Sentiment analysis of tweets related
to SUS before and during COVID-19 pandemic. IEEE Latin America Transactions 20
(1):6–13. doi:10.1109/TLA.2022.9662168.
Smetanin, S. 2020. The applications of sentiment analysis for Russian language texts:
Current challenges and future perspectives. IEEE Access 8:110693–719. doi:10.1109/
ACCESS.2020.3002215.
Valdivia, A., M. V. Luzı on, and F. Herrera. 2017. Neutrality in the sentiment analysis prob-
lem based on fuzzy majority. In 2017 IEEE international conference on fuzzy systems
(FUZZ-IEEE) (1–6). IEEE.
Wadawadagi, R. S., and V. B. Pagi. 2019. A multi-layer approach to opinion polarity classi-
fication using augmented semantic tree kernels. Journal of Experimental & Theoretical
Artificial Intelligence 31 (3):349–67. doi:10.1080/0952813X.2018.1549108.
Wang, Y., G. Huang, J. Li, H. Li, Y. Zhou, and H. Jiang. 2021. Refined global word embed-
dings based on sentiment concept for sentiment analysis. IEEE Access 9:37075–85. doi:
10.1109/ACCESS.2021.3062654.
Wang, L., J. Niu, and S. Yu. 2020. SentiDiff: combining textual information and sentiment
diffusion patterns for Twitter sentiment analysis. IEEE Transactions on Knowledge and
Data Engineering 32 (10):2026–39. doi:10.1109/TKDE.2019.2913641.
Wu, J., K. Lu, S. Su, and S. Wang. 2019. Chinese micro-blog sentiment analysis based on
multiple sentiment dictionaries and semantic rule sets. IEEE Access 7:183924–39. doi:10.
1109/ACCESS.2019.2960655.
CYBERNETICS AND SYSTEMS: AN INTERNATIONAL JOURNAL 29

Xu, G., Z. Yu, H. Yao, F. Li, Y. Meng, and X. Wu. 2019. Chinese text sentiment analysis
based on extended sentiment dictionary. IEEE Access 7:43749–62. doi:10.1109/ACCESS.
2019.2907772.
Yang, L., Y. Li, J. Wang, and R. S. Sherratt. 2020. Sentiment analysis for e-commerce prod-
uct reviews in chinese based on sentiment lexicon and deep learning. IEEE Acces 8:
23522–30. doi:10.1109/ACCESS.
Yin, F., Y. Wang, J. Liu, and L. Lin. 2020. The construction of sentiment lexicon based on
context-dependent part-of-speech chunks for semantic disambiguation. IEEE Access 8:
63359–67. doi:10.1109/ACCESS.2020.2984284.
Zhai, G., Y. Yang, H. Wang, and S. Du. 2020. Multi-attention fusion modeling for senti-
ment analysis of educational big data. Big Data Mining and Analytics 3 (4):311–9. doi:10.
26599/BDMA.2020.9020024.
Zhang, B., D. Xu, H. Zhang, and M. Li. 2019. STCS lexicon: spectral-clustering-based
topic-specific Chinese sentiment lexicon construction for social networks. IEEE
Transactions on Computational Social Systems 6 (6):1180–9. doi:10.1109/TCSS.2019.
2941344.
Zhou, J., S. Jin, and X. Huang. 2020. ADeCNN: An improved model for aspect-level senti-
ment analysis based on deformable CNN and attention. IEEE Access 8:132970–9. doi:10.
1109/ACCESS.2020.3010802.

You might also like