Aspect Based Sentiment Analysis Using A Novel Ensemble Deep Network
Aspect Based Sentiment Analysis Using A Novel Ensemble Deep Network
Corresponding Author:
Abraham Rajan
Department of Computer Science and Engineering, Christ University
Bangalore, India
Email: [email protected]
1. INTRODUCTION
Through blogs, discussion forums, social networking sites, and e-commerce websites, users have
expressed their opinions on a wide range of topics, including products, individuals, and organizations, during
the past several years. This swapping of ideas is made possible by the fast development of web applications
and the broad availability of inexpensive internet connections, both of which generate massive amounts of
online data. Sentiment analysis uses this large quantity of data that is used to analyses major challenges. Before
the widespread use of the internet as a resource for decision-making, due to the widespread use of the internet
and the world wide web, it is now possible to read the opinions and perspectives of anonymous individuals
from a range of cultures and places. Many users' shopping opinions are impacted by internet reviews by various
other shoppers [1].
Systems that analyses user sentiment will automatically summarize user reviews, which may help a
customer, make an opinion. Major advantages of sentiment analysis systems include scalability, the ability to
summarize massive volumes of text, real-time analysis, and the capability to deliver fast assessments, and
consistent criteria, which is the result of automation and lacks bias compared to user judgment. Organizations,
whether public or private, require a sentiment analysis system; it is feasible to supplement standard
recommendation systems using sentiment analysis. It assists the owners of the organization in understanding
the sentiments of customers towards their items, however, this is suitable for market research and competition
analysis, politics, public policy development, and legal research are further uses [2]. Although sentiment
analysis may be undertaken at various levels, aspect-level sentiment analysis is the most useful. Researchers
have conducted an extensive study on the identification of explicit sentiments that result in the generation of
many alternatives.
Due to its complexity, embedded characteristic recognition receives less attention; this is
simultaneously a substantial portion of a sentence, which has various parameters, to detect sentiment analysis
in understanding the concept. Sentiment analysis research focuses on making predictions at the phrase or
document level by assessing the overall sentiment towards the entire phrase or content [2]–[4]. The opinion is
based on the perhaps inaccurate assumption that the available data determines a single opinion towards a
specific issue. In consequence, the need for aspect-based sentiment analysis (ABSA) is a method for
differentiating more complex aspect-level thoughts and feelings, which has surged over the past decade [5],
[6]. The target achieved for opinion mining on ABSA shifts its focus from being a phrase or document on a
whole to being an entity or a particular characteristic of an entity. For example, an entity may be an
online-sold product, with its features including its price and size. Refer to the following example review to
understand the words used in aspect-level sentiment analysis: "the display quality of the Samsung TV is
outstanding." The term outstanding in this context denotes a favorable opinion of the Samsung TV's display,
which is the element of the Samsung TV that is being discussed. As illustrated in Figure 1, the three methods
that may be employed to perform aspect-level sentiment analysis are aspect identification, sentiment evaluation
concerning the specific aspect, and sentiment aggregation.
Context-based methods and syntax-based methods are the two main areas of focus in the existing
research for ABSA based on deep neural networks [5], [6]. To gather aspect word representations in a phrase,
context-based approaches often employ convolutional neural networks (CNN) or recurrent neural networks in
conjunction with attention processes. Graph convolutional networks (GCN) are frequently used to represent
phrases in syntax-based models. Recent research on ABSA has relied widely on deep neural models due to
their capability to capture the semantic features of specific aspects [1]–[4]. In the existing research on ABSA
that uses deep neural networks, context-based techniques [5], [6] and syntax-based methods [7]–[9] are
distinguished into two groups. Context-based algorithms frequently combine CNN or recurrent neural networks
with attention mechanisms to collect aspect word representations in a sentence, in syntax-based models, GCN
are typically employed to represent sentences.
Finding several aspect-level sentiment components, including sentiment polarities, opinion terms,
aspect words, and aspect categories, is the focus of this research work. Present ABSA models that have been
developed to achieve better performance in a variety of tasks, which focuses on training and testing data, which
is developed from the same distribution (e.g., the same domain or the same language). When the data
distribution changes, it is usually important to retrain the ABSA model to maintain its performance. Huge
quantities of labelled data are often too expensive or even impossible to collect for ABSA applications that
require aspect-level annotations. Incorporating various other methods for building ABSA-based models, which
generalize across various domains and languages, is cross-domain transfer or cross-lingual transfer, which
requires adapting the trained model to unknown domains or unknown languages.
− Convert the words of the input sentence into word-vectors using the optimized bidirectional encoder
representations from transformers (BERT) model.
− To extract the context-based word representation using the optimized graph neural network (GNN) model
with convolutions over the word-vector embedding.
− To develop a BERT-GNN model that analyses the ABSA of the input sentence.
Aspect based sentiment analysis using a novel ensemble deep network (Abraham Rajan)
1670 ISSN: 2252-8938
− To develop a proposed novel ensemble deep network (EDN) for an ABSA model for optimized BERT over
GNN with convolution.
This particular research is organized as follows: first section of the research starts with background
work and significance of sentiment analysis in natural language processing (NLP); section concludes with
research motivation and contribution. Second section of the research discuss various existing deep learning
protocol along with its shortcoming; third section presents mathematical modelling of EDN. Thereafter, EDN
is evaluated in fourth section.
2. RELATED WORK
ABSA has been one of the key area of research in sentiment analysis, thus there has been plenty of
research in past decade. Traditional approach lacks the deep understanding of the common sense among the
sentences. This section discuss several deep learning mechanism for sentiment analysis. Huang and Carley [10]
have developed a novel, reliable CNN for aspect-level sentiment analysis, and integrating this aspect data into
a CNN with the aspect-level features results in developing a high-accuracy model. Long short-term memory
(LSTM) networks are difficult to analyses and have low temporal efficiency, whereas CNN can only allocate
local semantic data. The sparse attention-based convolutional neural network (SA-SDCCN) proposed by [11]
deals with semantic and sentiment embedding that is added to an embedding tensor at the multichannel
embedding layer, which leads to accurate representations of the input sequence.
According to Wang et al. [12], a novel model known as unified position-aware convolutional neural
network (UP-CNN), which initially proposed an aspect detection network with prior knowledge to
accommodate the absence of aspect positions in ACSA, and has proposed an aspect mask to construct
aspect-aware contextual representations to accommodate CNN in ABSA. Consequently, several NLP
applications have successfully integrated GCNs, by using grammatical links in aspect-level sentiment analysis
tasks, Zhang et al. [7] have developed the first aspect-specific graph convolutional network (ASGCN). Initially,
the context information is collected using a network with bidirectional long short-term memory (Bi-LSTM),
then aspect-based word information is collected using the masked approach. The ABSA approach developed
by [13] is based on the convolution of dependency trees.
Zhang and Qian [14] introduced a bidirectional GCN model consisting of a hierarchical syntactic
network and a lexical graph for each phrase based on syntactic links and word co-occurrence data. Based on a
dependency network, Tang et al. [15] suggested a modified double-transformer structure for aspect sentiment
categorization that focuses on essential sentence components. Using relational graph-based attention networks,
Wang et al. [16] have modified the dependent trees. The previously published GCN-based models perform
well on the ABSA; however, they lacked sentiment data across context-related factors. In addition to this the
interactive sentiment correlation between aspects recovered by GCNs on the dependency tree with external
sentiment-based knowledge and integrate with MHSA to generate semantic information between context and
aspects.
According to Zhang et al. [17], transformer-based semantic-primary knowledge transferring (TSPKT)
network for aspect-term sentiment analysis across domains makes use of semantic-primary information to
facilitate the transfer of knowledge across multiple domains. Before semantic-primary information is collected
from the S-Graph, it is generated using external semantic lexicons. The second postulate is that Aoa graphormer
learns syntactically meaningful aspect-term phrases. Thirdly, the standard by LSTM classifier is developed by
adding a knowledge-aware memory unit (KAMU) that holds the whole corpus of semantic primary input.
According to Hu et al. [18], this approach focuses on the above-mentioned two issues by combining aspect and
task-level regularizations. Aspect-level regularizations limit the attention-based weights to reduce noise. Sparse
regularization is more suited for sentences with a single aspect than orthogonal regularization for sentences
with various aspects. Here they have proposed a task-level regularization by adding aspect-level category
identification as an orthogonal auxiliary task to generate sentiment-dominant features.
3. PROPOSED METHODOLOGY
The proposed novel EDN for ABSA model for optimized BERT over GNN with convolutions.
Figure 2 shows the proposed architecture. The pooling layer here is responsible to extract features of similar
size by selecting the maximum number from each feature vector. Further, this connected layer fine-tunes the
model considering the sentiments taken from the previous layer. Finally determining the sentiments from the
softmax layer.
− Transfer the words of the input sentence into words using the optimized BERT model.
− To extract the context-based word representation using the optimized GNN model over the word-vector
embedding.
− Build a BERT-GNN model with convolutions that analyses the ABSA of the input sentence.
Consider a given set of sentiments D = {d1 , d2, … . . , dn} for l ∈ L, let
Sen = {sen1 , sen2, … . . , senn} be a set of sentences for k sentiments. For sen ∈ Sen be the set of words
sen = {l1 , l2, … . . , ln} for one sentence denoted assen. In the proposed architecture for a novel EDN for
aspect-based sentiment, analysis the input is given in the form of sentiments, which is then converted into a
sentence–word-vector. This is further fed into the optimized BERT model where it is converted into word
embedding by the optimized BERT model. The optimized LSTM model here extract the context-word
representations, the context-based words are fed into the GNN with convolution, which converts the node
matrices into the Attention mechanism wherein the further bias is reduced while, processing the GNN model
with convolutions.
1, if Sxy is positive
G: (g xy ) = { 0, if Sxy is neutral (1)
−1, if Sxy is negative
Aspect based sentiment analysis using a novel ensemble deep network (Abraham Rajan)
1672 ISSN: 2252-8938
jx = (L→ j j → jx + L→ j j ← jx + ej ) (5)
Here L is the weight matrix, by considering the Lg→ j amidst the input given and the forward hidden vectors, jx
for representing the context. jx+1 is the hidden vector represented for jx and jk+1 = 0. jx−1 for the previously
hidden vector for jx and j0 = 0.
− Step 3: the context-based word-vector matrix denoted as G = {g1 , g 2 , … , g k }∈ Nk∗pj , here pj depicts the
size of the vector in the matrix G.
1, if Nx , Ny ∈ N, and vxy ∈ V
Fxy = { 1, if Nx = Ny , (6)
0, else
Additionally, this graph SGNN has a feature matrix embedded with nodes denoted as Y = [G] ∈ N|N|∗pj , where
each row is denoted as Gx which is responsible for the context-based representation of word node Nx ∈ V.
− Step 2: generation of node-embedding
Representing the nodes: the matrices are generated as P ∈ N|N|∗pj and F ∈ N |N|∗pj are given as input into a
GNN network with convolution shown to generate the nodes represented as (7):
Here b depicts the number of layers in the GNN network with convolutions layer, P ∈ N |N|∗ph ; G (0) = Y, here
ϑ is a non-linear activation function as ReLU. L(b) ∈ N|N |∗ph for the transformation layer in the b − th layer.
q(b) is the bias associated with GNN layers associated with convolution, respectively:
K xx = ∑y Fxy (9)
|N|
Ix = ∑x=1 ϑx Hx (10)
Here,
|x−initiate|
, x ≤ initiate
k
wx = {0, initiate ≤ x ≤ terminate (12)
|x−terminate|
, x > terminate
k
Here initiate and terminate determines the starting and ending phase of the sentiment in the sentence.
ux = ReLU(RꭈMx:x+|N|−1 + q) (13)
Here x = [1, |N|] determines the order of the node representation M; ꭈ is the convolution operator, the
activation function is denoted by ReLU and q is the bias term. The feature vectors are constructed by node
representations as shown in (14):
Y = SM(H M . u + q) (16)
Here H M ∈ Lb∗|N| and q ∈ H b is the weight matrix and bias of the layer. b is the number of sentiment phases.
Training the optimized GNN model with convolution to reduce the entropy loss values.
Aspect based sentiment analysis using a novel ensemble deep network (Abraham Rajan)
1674 ISSN: 2252-8938
Here q denotes the sentiment classes for the training model, yk denotes the distribution of true sentiment for
the k − th class and yk ^ denotes the distribution of the sentiment to be predicted for the k − th class, δ denotes
the regulation co-efficient, and γ is the parameter set from the previous layer.
4. PERFORMANCE EVALUATION
The task of solving ABSA is more complex than conventional text-level sentiment analysis. It focuses
on recognizing the characteristics or features of an item described in a text and the perspective has taken on
each characteristic. It has been proven that applying optimized BERT to ABSA as a sentence-pair classification
task and developing an auxiliary sentence yields better results than the existing state-of-art models and baseline
models. Further, we have evaluated our proposed ensemble deep network proposed system (EDN-PS) model
with existing state-of-art methods and baseline models in terms of accuracy and F1-score.
− RAM [25]: it increases resistance to irrelevant data by using a method called multiple attention to identify
sentiment features from a distance.
− MGAN [26]: employs a combination of coarse- and fine-grained attention to capture the interactions
between aspects and contexts at the word level and an aspect alignment loss to explain the aspect-level
interactions between aspects that share a shared context.
− CDT [13]: the BiLSTM-learned embedding are amplified using GCN after learning the feature
representation of sentences with BiLSTM. GCN directly adjusts a phrase's dependency tree.
− DMMN-SDCM [27]: integrates information from semantic analysis into the memory network and assigns
an auxiliary task to learn the sentiment distribution of the entire phrase to offer the necessary background
knowledge for sentiment classification of the target aspect.
− TD-BERT [28]: a max-pooling layer and an FCL are added to the BERT model to create the necessary
sentiment classification model.
− AEN-BERT [29]: uses attention-based encoding to model the link between context and target, examines
the issue of inaccurate labels, and suggests label smoothing regularization. The idea also applies to Bert's
prior condition.
− BERT-PT [30]: after training BERT, a technique is introduced to compensate for this lack of domain and
task-related knowledge and to improve the training's results by utilizing the remaining quantity of data.
− BERT-BASE [31]: in all stages, attempts are made to simultaneously condition on both left and right
contexts to pre-train deep bidirectional representations from the unlabeled text. The pre-trained BERT
model may be enhanced to produce existing models for a variety of purposes with just one additional output
layer. We employ ABSA for multilingual learning.
− RepWalk [32]: finds the context words that have a great impact on the ability of aspect words to predict the
sentiment by randomly searching the grammatical graph. To determine the importance of each word in a
phrase using grammatical knowledge.
− SK-GCN (BERT) [33]: combines two modelling strategies for both GCN's syntactic dependency trees and
common sense information to specifically enhance phrase representation.
− SAGAT-BERT [34]: the BERT language model's external pre-training data and understanding of syntax
perception are applied to the dependent tree structure using the graph attention network, which aids in
replicating the interaction between the context and aspects.
− ASGCN-BERT [7]: uses the syntactic dependent structure of the phrase to solve the long-standing multi-
word reliability issue of ABSA.
− R-GAT+BERT [16]: by using this methodology and pruning the traditional dependency parse tree, a unified
aspect-oriented dependency tree structure depending on the desired aspect is produced. It is also
recommended to encode a novel tree structure for sentiment prediction using a relational graph attention
network (R-GAT).
− LCFS-BERT-CDM [1]: this method recommends employing syntactic relative distance to reduce the
impact of unrelated words with unstable syntactic ties to aspects to improve the aspect-based sentiment
classifier’s accuracy. To improve the efficiency of the aspect extractor, it also incorporates part of speech
embedding, syntactic relation embedding, and context embedding (e.g., BERT and Roberta).
4.3.1 Accuracy
The accuracy is evaluated for various existing state-of-art models and baseline models and a graph is
plotted against the proposed model we can conclude that the proposed EDN-ABSA model performs better in
comparison with the existing system. Figure 3 shows the accuracy comparison graphically. Table 2 shows the
accuracy comparison. The accuracy value for the baseline method for ASGCN-DT is 71.53 and whereas for
IAN is 72.5 and the maximum value among the baseline methods is MAN methodology, which gives an
accuracy value of 76.56. Whereas the accuracy value for the state-of-art-model for a methodology for
ATAE-LSTM is 68.64 and for the RepWalk methodology, it is 74.4 and the maximum value for the
state-of-art-model at different times is TD-BERT value is 76.7, whereas the proposed EDN-PS model
outperforms other existing state-of-art and baseline models with an accuracy value of 78.5.
Aspect based sentiment analysis using a novel ensemble deep network (Abraham Rajan)
1676 ISSN: 2252-8938
4.4. F1-score
The F1-score is evaluated for various existing state-of-art models and baseline models and a graph is
plotted against the proposed model we can conclude that the proposed EDN-PS model performs better in
comparison with the existing models. Figure 4 shows the F1-score comparison graphically. Table 3 shows the
F1-score comparison. The F1-score value for the baseline method for ASGCN-DT is 69.68, which is the least,
and whereas for MemNet is 69.9 and the maximum value among the baseline methods is R-GAT-BERT
methodology, which gives an accuracy value of 74.88. Whereas the F1-score value for the state-of-art-model
at different times for methodology for MemNet is 69.9, and for SAGAT-BERT methodology it is 74.17 and
the maximum value for the state-of-art-model at different times is R-GAT-BERT value is 74.88, whereas the
proposed EDN-PS model outperforms other existing state-of-art and baseline models with an F1-score value
of 76.6.
5. CONCLUSION
ABSA, a method of text analysis, divides data into groups and evaluates the sentiment connected to
each group. ABSA may assess customer feedback by linking different sentiments with specific qualities of an
item or service. The existing research is focused on the correlation between the sentiment polarity of various
aspects and the local context. The essential deep correlations between the global context and aspects are not
focused on attention. The proposed EDN-ABSA is compared with various state-of-art models and baseline
models and a graph is plotted for accuracy and F1-score for ACL Twitter Dataset whereas the proposed
EDN-PS model outperforms other existing state-of-art and baseline models with an accuracy value of 78.5 and
F1-score value of 76.6.
REFERENCES
[1] M. H. Phan and P. O. Ogunbona, “Modelling context and syntactical features for aspect-based sentiment analysis,” in Proceedings
of the 58th Annual Meeting of the Association for Computational Linguistics, 2020, pp. 3211–3220, doi: 10.18653/v1/2020.acl-
main.293.
[2] N. Majumder, S. Poria, A. Gelbukh, M. S. Akhtar, E. Cambria, and A. Ekbal, “IARM: inter-aspect relation modeling with memory
networks in aspect-based sentiment analysis,” in Proceedings of the 2018 Conference on Empirical Methods in Natural Language
Processing, 2018, pp. 3402–3411, doi: 10.18653/v1/D18-1377.
[3] Y. Wang, M. Huang, X. Zhu, and L. Zhao, “Attention-based LSTM for aspect level sentiment classification,” in Proceedings of the
2016 Conference on Empirical Methods in Natural Language Processing, 2016, pp. 606–615, doi: 10.18653/v1/D16-1058.
[4] D. Ma, S. Li, X. Zhang, and H. Wang, “Interactive attention networks for aspect-level sentiment classification,” in Proceedings of
the Twenty-Sixth International Joint Conference on Artificial Intelligence, Aug. 2017, pp. 4068–4074, doi: 10.24963/ijcai.2017/568.
[5] B. Xu, X. Wang, B. Yang, and Z. Kang, “Target embedding and position attention with LSTM for aspect based sentiment analysis,”
in Proceedings of the 2020 5th International Conference on Mathematics and Artificial Intelligence, Apr. 2020, pp. 93–97, doi:
10.1145/3395260.3395280.
[6] L. Xu, L. Bing, W. Lu, and F. Huang, “Aspect sentiment classification with aspect-specific opinion spans,” in Proceedings of the
2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), 2020, pp. 3561–3567, doi:
10.18653/v1/2020.emnlp-main.288.
[7] C. Zhang, Q. Li, and D. Song, “Aspect-based sentiment classification with aspect-specific graph convolutional networks,” in
Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint
Conference on Natural Language Processing (EMNLP-IJCNLP), 2019, pp. 4568–4578, doi: 10.18653/v1/D19-1464.
[8] X. Hou et al., “Graph ensemble learning over multiple dependency trees for aspect-level sentiment classification,” in Proceedings
of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language
Technologies, 2021, pp. 2884–2894, doi: 10.18653/v1/2021.naacl-main.229.
[9] R. Li, H. Chen, F. Feng, Z. Ma, X. Wang, and E. Hovy, “Dual graph convolutional networks for aspect-based sentiment analysis,”
in Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint
Conference on Natural Language Processing (Volume 1: Long Papers), 2021, pp. 6319–6329, doi: 10.18653/v1/2021.acl-long.494.
[10] B. Huang and K. Carley, “Parameterized convolutional neural networks for aspect level sentiment classification,” in Proceedings
of the 2018 Conference on Empirical Methods in Natural Language Processing, 2018, pp. 1091–1096, doi: 10.18653/v1/D18-1136.
[11] C. Gan, L. Wang, Z. Zhang, and Z. Wang, “Sparse attention based separable dilated convolutional neural network for targeted
sentiment analysis,” Knowledge-Based Systems, vol. 188, p. 104827, Jan. 2020, doi: 10.1016/j.knosys.2019.06.035.
[12] X. Wang, F. Li, Z. Zhang, G. Xu, J. Zhang, and X. Sun, “A unified position-aware convolutional neural network for aspect based
sentiment analysis,” Neurocomputing, vol. 450, pp. 91–103, Aug. 2021, doi: 10.1016/j.neucom.2021.03.092.
[13] K. Sun, R. Zhang, S. Mensah, Y. Mao, and X. Liu, “Aspect-level sentiment analysis via convolution over dependency tree,” in
Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint
Conference on Natural Language Processing (EMNLP-IJCNLP), 2019, pp. 5679–5688, doi: 10.18653/v1/D19-1569.
[14] M. Zhang and T. Qian, “Convolution over hierarchical syntactic and lexical graphs for aspect level sentiment analysis,” in
Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), 2020, pp. 3540–3549, doi:
10.18653/v1/2020.emnlp-main.286.
[15] H. Tang, D. Ji, C. Li, and Q. Zhou, “Dependency graph enhanced dual-transformer structure for aspect-based sentiment
classification,” in Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, 2020, pp. 6578–6588,
doi: 10.18653/v1/2020.acl-main.588.
[16] K. Wang, W. Shen, Y. Yang, X. Quan, and R. Wang, “Relational graph attention network for aspect-based sentiment analysis,” in
Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, 2020, pp. 3229–3238, doi:
10.18653/v1/2020.acl-main.295.
[17] B. Zhang, X. Fu, C. Luo, Y. Ye, X. Li, and L. Jing, “Cross-domain aspect-based sentiment classification by exploiting domain-
invariant semantic-primary feature,” IEEE Transactions on Affective Computing, vol. 14, no. 4, pp. 3106–3119, Oct. 2023, doi:
10.1109/TAFFC.2023.3239540.
[18] M. Hu, S. Zhao, H. Guo, and Z. Su, “Hybrid regularizations for multi-aspect category sentiment analysis,” IEEE Transactions on
Affective Computing, vol. 14, no. 4, pp. 3294–3304, Oct. 2023, doi: 10.1109/TAFFC.2023.3236948.
[19] L. Dong, F. Wei, C. Tan, D. Tang, M. Zhou, and K. Xu, “Adaptive recursive neural network for target-dependent Twitter sentiment
classification,” in Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics (Volume 2: Short
Papers), 2014, vol. 2, pp. 49–54, doi: 10.3115/v1/P14-2009.
[20] D. Tang, B. Qin, and T. Liu, “Aspect level sentiment classification with deep memory network,” in Proceedings of the 2016
Conference on Empirical Methods in Natural Language Processing, 2016, pp. 214–224, doi: 10.18653/v1/D16-1021.
[21] H. Peng, Y. Ma, Y. Li, and E. Cambria, “Learning multi-grained aspect target sequence for Chinese sentiment analysis,” Knowledge-
Based Systems, vol. 148, pp. 167–176, May 2018, doi: 10.1016/j.knosys.2018.02.034.
[22] B. Huang, Y. Ou, and K. M. Carley, “Aspect level sentiment classification with attention-over-attention neural networks,” in Social,
Cultural, and Behavioral Modeling, Cham: Springer, 2018, pp. 197–206, doi: 10.1007/978-3-319-93372-6_22.
[23] Q. Xu, L. Zhu, T. Dai, and C. Yan, “Aspect-based sentiment classification with multi-attention network,” Neurocomputing,
vol. 388, pp. 135–143, May 2020, doi: 10.1016/j.neucom.2020.01.024.
[24] N. Liu and B. Shen, “Aspect-based sentiment analysis with gated alternate neural network,” Knowledge-Based Systems, vol. 188,
p. 105010, Jan. 2020, doi: 10.1016/j.knosys.2019.105010.
[25] P. Chen, Z. Sun, L. Bing, and W. Yang, “Recurrent attention network on memory for aspect sentiment analysis,” in Proceedings of
the 2017 Conference on Empirical Methods in Natural Language Processing, 2017, pp. 452–461, doi: 10.18653/v1/D17-1047.
[26] F. Fan, Y. Feng, and D. Zhao, “Multi-grained attention network for aspectlevel sentiment classification,” in Proceedings of the 2018
Conference on Empirical Methods in Natural Language Processing, 2018, pp. 3433–3442, doi: 10.18653/v1/D18-1380.
[27] P. Lin, M. Yang, and J. Lai, “Deep mask memory network with semantic dependency and context moment for aspect level sentiment
classification,” in Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence, Aug. 2019,
pp. 5088–5094, doi: 10.24963/ijcai.2019/707.
Aspect based sentiment analysis using a novel ensemble deep network (Abraham Rajan)
1678 ISSN: 2252-8938
[28] Z. Gao, A. Feng, X. Song, and X. Wu, “Target-dependent sentiment classification with BERT,” IEEE Access, vol. 7,
pp. 154290–154299, 2019, doi: 10.1109/ACCESS.2019.2946594.
[29] Y. Song, J. Wang, T. Jiang, Z. Liu, and Y. Rao, “Targeted sentiment classification with attentional encoder network,” in Artificial
Neural Networks and Machine Learning – ICANN 2019: Text and Time Series, Cham: Springer, 2019, pp. 93–103, doi:
10.1007/978-3-030-30490-4_9.
[30] H. Xu, B. Liu, L. Shu, and P. S. Yu, “BERT post-training for review reading comprehension and aspect-based sentiment analysis,”
in Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human
Language Technologies, 2019, pp. 2324–2335.
[31] J. Devlin, M. W. Chang, K. Lee, and K. Toutanova, “BERT: Pre-training of deep bidirectional transformers for language
understanding,” Arxiv-Computer Science, pp. 1–16, 2019.
[32] Y. Zheng, R. Zhang, S. Mensah, and Y. Mao, “Replicate, Walk, and Stop on Syntax: An Effective Neural Network Model for
Aspect-Level Sentiment Classification,” Proceedings of the AAAI Conference on Artificial Intelligence, vol. 34, no. 5,
pp. 9685–9692, Apr. 2020, doi: 10.1609/aaai.v34i05.6517.
[33] J. Zhou, J. X. Huang, Q. V. Hu, and L. He, “SK-GCN: modeling syntax and knowledge via graph convolutional network for aspect-
level sentiment classification,” Knowledge-Based Systems, vol. 205, p. 106292, Oct. 2020, doi: 10.1016/j.knosys.2020.106292.
[34] L. Huang, X. Sun, S. Li, L. Zhang, and H. Wang, “Syntax-aware graph attention network for aspect-level sentiment classification,”
in Proceedings of the 28th International Conference on Computational Linguistics, 2020, pp. 799–810, doi:
10.18653/v1/2020.coling-main.69.
BIOGRAPHIES OF AUTHORS