Knowledge Enhanced Graph Convolutional Networks For Arabic Aspect Sentiment Classification
Knowledge Enhanced Graph Convolutional Networks For Arabic Aspect Sentiment Classification
https://doi.org/10.1007/s13278-023-01166-w
ORIGINAL ARTICLE
Abstract
Aspect sentiment classification (ASC) is a sub-task of aspect-based sentiment analysis (ABSA) that aims at identifying the
sentiment polarity toward a specific aspect in a given text or sentence. Most existing research on Arabic ABSA adopted
rule-based or machine learning-based methods, with little attention to deep learning techniques. Additionally, the major-
ity of these deep learning-based models relied on attention mechanisms to capture the interaction between the context and
aspect words. However, attention-based methods are generally inefficient in extracting the syntactic dependencies between
contextual tokens and aspects. Therefore, we introduce a combined model that incorporates an Arabic BERT model with
graph convolutional network and local context focus layers to capture syntactic dependencies relevant to a specific aspect
while emphasizing the contribution of semantic-related tokens related to this aspect. We also integrate affective common-
sense knowledge into the graph networks to capture the sentiment-related dependencies between contextual words and the
specific aspect. The experimental results on an Arabic hotel dataset show that the proposed method outperforms the baseline
and related work models and achieves a state-of-the-art accuracy score of 92.77% in Arabic ASC. The achieved results show
the effectiveness of the proposed model in enhancing the aspect-specific sentiment representations, which can be promising
for future research in this field.
Keywords Aspect sentiment classification · BERT · Graph convolutional networks · Affective commonsense knowledge ·
Local context focus · Arabic
13
Vol.:(0123456789)
6 Page 2 of 14 Social Network Analysis and Mining (2024) 14:6
Arabic is one of the most spoken languages in the world, The reminder of the paper is structured as follows: Sect. 2
with more than 422 million speakers (Omar and Abd El- overviews related work to Arabic ASC, GCN, and Sentic-
Hafeez 2023). According to the Internet World Stats, it is Net; Sect. 3 describes the proposed model in detail; Sect. 4
ranked as the fourth most used language on the internet. presents the experimental setup; Sect. 5 discusses the eval-
The Arabic online content is growing rapidly (Alqurashi uation results; finally, the conclusion and future research
2023), enabling Arabic sentiment analysis (ASA) to get lots directions are provided in Sect. 6.
of attention from researchers in recent years. However, most
existing ASA work has focused on the document and sen-
tence levels, with little attention to the ABSA tasks (Ben-
2 Related work
soltane and Zaki 2023a).
Moreover, most existing Arabic ABSA papers have
This section is divided into three main subsections. The first
adopted rule-based or traditional machine learning classifi-
subsection overviews related work to Arabic ASC using
ers to handle these tasks with limited work on deep learning
deep learning models. The second subsection introduces
models (Bensoltane and Zaki 2023b). Besides, most of the
graph-based studies for ASC. Since the lack of Arabic ASC
proposed Arabic ABSA deep learning-based methods used
methods using GCN, papers that targeted the English lan-
traditional word embedding models for word vector repre-
guage are overviewed. Finally, the third subsection discusses
sentations, which provide static embedding vectors for each
studies that have used SenticNet knowledge base, primarily
word independently of the context in which it occurs. Fur-
for handling the ASC task.
thermore, most of these deep learning-based methods were
implemented based on attention mechanisms, which are
inefficient in extracting syntactical dependencies between 2.1 Arabic aspect sentiment classification
contextual tokens and aspect words (Huang et al. 2023a, b).
Therefore, the aim of this paper is to overcome the afore- Most existing ASA work focused mainly on the document or
mentioned shortcomings by proposing a combined model sentence levels with limited studies on ABSA. Furthermore,
that incorporates an Arabic BERT model, called AraBERT deep learning models are under-discovered in this task for
(Antoun et al. 2020), along with graph convolutional net- the Arabic language (Bensoltane and Zaki 2023a).
work (GCN) and local context focus (LCF) layers to cap- One of the earlier studies adopting a deep learning
ture the syntactically related words to the specific aspect method for Arabic ASC is Ruder et al. (2016). The sys-
based on the syntactic dependency tree of the sentence tem was submitted to the SemEval workshop (Pontiki et al.
while focusing on local context features that are relevant 2016) and was ranked 1st in Arabic. The model fed both
to the given aspect term. Moreover, we explore the use of the aspect embedding and the input sequence embeddings
an affective commonsense knowledge base to enhance the into a convolutional neural network (CNN) model for sen-
dependency graphs of the sentence. The main contributions timent polarity identification. Different variants of adding
of this study are: the aspect vector to the model were examined, and the best
results were achieved by concatenating each aspect vector
• We introduce a combined model that incorporates the with the word vector before the convolution. The proposed
strengths of the AraBERT model with GCN and LCF model achieved an overall enhancement of 11.8% over the
layers to further enhance the task of ASC on Arabic data- baseline model. Nevertheless, one main limitation of this
set. To our knowledge, this is the first work to use this model is the use of random word embeddings to represent
combination to handle this task in Arabic. input text, which negatively affects its ability in encoding
• We exploit the syntax-related context information using context and semantic features.
GCN over the dependency tree of the sentence to enhance The authors in Al-Smadi et al. (2019) implemented their
this task, which is, to our knowledge, the first time to use model based on bidirectional long short-term memory (BiL-
GCN for the Arabic ASC task. STM) to enhance the ASC task on the same Arabic hotel
• We investigate the use of an affective commonsense dataset from SemEval 2016 task 5. They followed the work
knowledge to capture the contextual affective dependen- of (Wang et al. 2016) by encoding the opinions targets as
cies related to the specific aspect. As far as we know, features to force the attention layer to focus on the impor-
this is the first time to examine the use of an affective tant parts of the sentence. Experiments showed that the pro-
knowledge base for this task in Arabic. posed method obtained similar results to the previous model
• Extensive experiments on an Arabic reference dataset (accuracy = 82.6% vs. 82.7%). Although the proposed model
show that the proposed model outperforms the baseline utilizes an attention mechanism to capture the importance of
and related work models and achieves a state-of-the-art context words to a specific aspect term, the context vectors
accuracy score. must encode both the aspect and sentiment information and
13
Social Network Analysis and Mining (2024) 14:6 Page 3 of 14 6
the alignment scores are uniformly applied across all feature of 15% over the baseline model. Nevertheless, it is unable to
dimensions, irrespective of their differences. capture syntax information and long-range word dependen-
An attempt to enhance the previous results was proposed cies, which are crucial for the ASC task.
in Abdelgwad et al. (2021). The authors followed the work A recent study of Fadel et al. (2023) adapted the LCF-
of Ma et al. (2017) to implement an interactive attention ATEPC model (Yang et al. 2021) to handle extract aspect
network (called IAN-BGRU) but using bidirectional gated terms and identify their sentiment polarities at the same
recurrent units (BiGRU) instead of BiLSTM. Two BiGRU time. The authors employed an Arabic BERT model as a
layers were first applied to extract both aspect and context shared layer for Arabic contextual text representation. They
representations, followed by two attention mechanisms to then used multi-head self-attention and LCF mechanism to
compute attention vectors. The proposed model achieved an model the interactive information between the context and
overall improvement over the model of Ruder et al. (2016). a specific aspect word. The proposed model achieved an
However, the main limitation of this model is the use of a accuracy score of 91.5% in ASC and an F1 score of 75.94%.
coarse-grained attention mechanism since they rely on aver- This method considers the word counts between two words
age vectors to determine attention weights for contexts, and as their semantic-relative distance, disregarding the mutual
as a result, they cannot capture word-level interactions. syntactic associations. Nonetheless, crucial sentiment words
Al-Dabet et al. (2021) have also sought to enhance the may be linked to the target aspect terms through grammatical
previous results by exploiting the success of memory net- rules, even if they have a considerable semantic separation.
works and attention mechanisms. They first implemented a
memory module using a stacked bidirectional independent 2.2 Graph convolutional network
long short-term memory network (IndyLSTM) with three
layers, followed by a recurrent attention module to help GCN has recently been successively applied in many NLP
the model focus on the sentence’s key parts. The proposed tasks, such as named entity recognition (Madan et al.
model achieved more than 3% improvement compared to 2023), text classification (Zhao et al. 2022a, b), and ques-
the IAN-BGRU model. Although this model achieved bet- tion answering (Yusuf et al. 2022). In addition, GCN-based
ter results than other attention-based models thanks to the models have enabled the detection of syntactically relevant
combination of multiple attentions with IndyLSTM layers, words to the given aspect by capturing long-term syntactic
it primarily focuses on the semantic information of the sen- dependencies through convolutional operation, which has
tence and ignores the syntactic structure knowledge, which shown promising results in ASC. One of the first papers
may increase the noise introduced by attention mechanisms. investigating the use of GCN for the ASC task is Zhang
Unlike the previous models that exploited context-inde- et al. (2019). The authors implemented a GCN layer over
pendent word embeddings to initialize the neural network the dependency tree of the sentence to capture syntactical
models, the authors in Abdelgwad et al. (2022) investigated information and long-range multi-word dependencies. The
using contextualized word embeddings from the BERT authors in Zhao et al. (2020) used a GCN-based model to
model to enhance the Arabic ASC task. They used BERT’s learn the sentiment dependencies between multiple aspects
sentence pair input architecture to handle this task by accept- in one sentence. Zhang and Qian (2020) first used a global
ing the sentence as the first input and the aspect as the sec- lexical graph for encoding word co-occurrence informa-
ond input. The BERT-based model was then fine-tuned on tion at the corpus level. They then build a concept hier-
the downstream task with a linear classification layer for the archy on each of the syntactic and lexical graphs for dif-
sentiment polarity prediction. The proposed model achieved ferentiating various types of dependency relations or word
enhanced results compared with previous models on the co-occurrence relations. Zhao et al. (2022a, b) introduced
Arabic hotel reviews (accuracy = 89.51%). Despite BERT’s a GCN model with multiple weight mechanisms, called
effectiveness in capturing contextual and semantic features MWM-GCN. Unlike conventional methods, MWM-GCN
from text, it does not possess explicit capabilities for mod- utilizes a dynamic weight alignment mechanism to ensure
eling sentence syntactic structures and dependencies. BERT’s WordPiece unit aligns with word-level dependen-
Al-Smadi et al. (2023) has exploited a multilingual uni- cies, improving its performance. An aspect-aware weight
versal sentence encoder (MUSE) with a gated recurrent unit mechanism regulates the flow of information to the aspect
model to improve the previous results for the Arabic ASC during graph convolution. Additionally, it incorporates an
task. Instead of word or character-level embeddings, the aspect-oriented dynamic loading layer and multi-head self-
MUSE model provides sentence-level embeddings. Besides, attention to enhance the model’s ability to capture syntax
The BiGRU layers were concatenated using global max and semantics for classification. A recent work of Huang
pooling and global average pooling to reduce the dimensions et al. (2023a, b) employed a conditional random field (CRF)
of the previous layers while maintaining the important fea- chain to identify the opinion span of a word specific to a
tures. The proposed model achieved an overall enhancement particular aspect. It then incorporates the context within this
13
6 Page 4 of 14 Social Network Analysis and Mining (2024) 14:6
opinion span into a multilayer graph convolutional network that impact sentiment polarity, including negations, modi-
by utilizing an enhanced position decay function. Finally, it fiers, and other related terms.
computes vector expressions of global nodes to predict senti- Despite the success of the previous GCN-based models in
ment polarity labels that are specific to the aspect. handling the ASC task, this method is under-discovered in
Arabic ABSA. Therefore, the aim of this study is to enhance
the Arabic ASC task by exploiting the syntactic information
2.3 SenticNet knowledge base using GCN over the dependency tree of the sentence. We
also investigate the enhancement of the graph dependencies
SenticNet is one of the most used affective knowledge bases, using affective knowledge from BabelSenticNet. Moreover,
which is a publicly available lexical resource that provides our proposed model integrates an LCF layer to reduce the
several thousand commonsense concepts along with their influence of irrelevant words to a specific aspect.
sentiment scores (Bisio et al. 2017). Different versions have
been released for SenticNet, primarily focused on Eng-
lish, such as SenticNet 2 (Cambria et al. 2012), SenticNet 3 Research methodology
3 (Cambria et al. 2014), and SenticNet 4 (Cambria et al.
2016). In addition, a multilingual version, called BabelSen- 3.1 Problem definition
ticNet, has been provided to handle 40 languages (i.e., Ara-
bic), which we use in this study. SenticNet has been utilized The aim of the ASC task is to determine the sentiment
in many tasks such as sarcasm detection (Du et al. 2022), polarity of a specific aspect word in a given text by captur-
emotion recognition (Chen et al. 2023), and text classifica- ing aspect-related sentiment information from the context,
tion (Wu 2023). Recently, many studies have used the affec- which can be formulated as follows:
tive knowledge from SenticNet to enhance the sentiment Given a sentence S = [W1, W2,…, Wn] of n words and
features representation for handling the ASC task. Zhou an aspect term x = [x1, x2,…, xt], which contains t words
et al. (2020) implemented two strategies to model the syn- (1 ≤ t ≤ n). The aim of this task is to predict the sentiment
tactic dependency tree and commonsense knowledge graph polarity (positive, negative, or neutral) regarding the aspect
to enrich the representation of a sentence toward a specific term x for S.
aspect. In the first strategy, two GCN models were employed
to independently encode the syntactic graph and the knowl- 3.2 Model overview
edge graph. In the second strategy, a combined syntactic and
knowledge GCN model was constructed to represent both As illustrated in Fig. 1, our combined model is composed of
the syntactic and knowledge graphs jointly. The authors in different components: two independent AraBERT layers for
Liang et al. (2022) proposed a novel solution for construct- local and global contextual features encoding, an LCF layer
ing the graph by enhancing the dependency graph of the to focus on local contextual words, an enhanced GCN based
sentence using affective knowledge from SenticNet. The on BabelSenticNet to exploit the affective dependencies
novel affective-enhanced graph model takes into account between contextual words and specific aspects, an interac-
both the relationships among contextual words and aspect tive learning layer to learn the profound correlation between
words, as well as the affective information between opinion context and targeted aspects, and finally an output layer to
words and the aspect. Liu et al. (2023) introduced a dual- predict the sentiment polarity toward the aspect word.
gated graph convolutional network (DGGCN) model that
integrates GCN into the gating mechanism to amplify the 3.2.1 AraBERT layers
interaction between the context and the aspect word while
consolidating the affective attributes linked to the specified Pre-trained language models (PTMs) have achieved state-
aspect. Moreover, including contextual affective knowledge of-the-art in many NLP tasks, including ABSA (Abas et al.
in the graph networks helps DGGCN better understand sen- 2020; Zhao and Yu 2021). Many PTMs models were pro-
timent-related data. Gu et al. (2023) proposed a GCN model vided in the last few years, such as BERT (Devlin et al.
(called EK-GCN) that incorporates external knowledge like 2019) and XLNET (Yang et al. 2019). BERT is a deeply
SenticNet and part-of-speech information. It employed a bidirectional model based on transformers. Unlike tradi-
Part-of-Speech Matrix to capture the influence of words tional word embedding models such GloVe (Pennington
like negations and modifiers, a Sentiment Lexicon to assign et al. 2014) and Word2Vec (Mikolov et al. 2013), BERT
sentiment scores to words, and a Word–Sentence Interaction can provide different vector representations to the same word
Network (WSIN) to filter relevant sentence information. The based on the context in which it appears.
proposed model can effectively overcome the challenge of In this paper, two independent AraBERT layers are
capturing edge labels and enhance the significance of words employed to provide two types of contextual features. One
13
Social Network Analysis and Mining (2024) 14:6 Page 5 of 14 6
{
layer generates global context features by accepting a sen- Z, SRDi ≤ 𝛼
tence pair as input (Xg = [CLS]sentence [SEP]aspect term Ki =
O, SRDi > 𝛼 (4)
[SEP]) to help detect the semantic relationship between the
contextual sentence and the aspect. The other layer provides [ ]
local context features by taking the context sentence as input M = K1m , K2m , K3m , … Knm (5)
(Xl = [CLS]sentence [SEP]). The outputs of the AraBERT
layers are annotated as follows: BlCDM = Bl ⋅ M (6)
( )
Bg = AraBERTglobal Xg (1) M is a feature masking matrix and Ki denotes the mask
vectors for each token. m is the input sequence length,
( ) whereas α is the SRD threshold. O and Z are the zeros and
Bl = AraBERTlocal Xl (2)
the ones vectors, respectively. “·” denotes the vector dot
product operation.
3.2.2 LCF layer Unlike CDM that totally ignores the features of non-local
context, CDW keeps less-semantic-relative context features
Following the work of Zeng et al. (2019), we integrated but de-emphasizes them according to their SRD to the aspect
an LCF layer to help the model focus on the sentimental term:
information contained in the neighbor words to the aspect {
term. This layer uses the semantic-relative distance (SRD) Z SRDi ≤ 𝛼
Ki = SRDi −𝛼 (7)
to detect the local context of the targeted aspect. SRD is n
⋅ Z SRDi > 𝛼
measured by counting words between context tokens and
aspect terms as follows: [ ]
Q = K1w , K2w , K3w , … Knw (8)
⌊ ⌋
l
SRDi = ||i − Pa || − (3)
2
BlCDW = Bl ⋅ Q (9)
where i represents the location of the context word, Pa is
the average position of the aspect term, and l represents the where Q is the weight matrix and Ki denotes the weight vec-
length of the aspect. S RDi denotes the SRD between a spe- tor for each less-semantic-relative contextual word.
cific aspect and the i-th context word.
Two architectures are evaluated in this paper to focus 3.2.3 Knowledge‑enhanced GCN layers
on local contexts, namely, context-features dynamic mask
(CDM) and context-features dynamic weighting (CDW). For the ASC task, it is crucial to capture syntactical infor-
CDM masks out less-semantic context features learned mation and mine word dependencies between words from a
by the AraBERT layer for which the SRD to aspect words given text. Therefore, we applied a multilayer GCN over the
is greater than a pre-defined threshold. The local context syntactic dependency tree of the sentence. First, the graph
output is calculated as follows: of convolutional networks over the dependency tree of each
13
6 Page 6 of 14 Social Network Analysis and Mining (2024) 14:6
Table 1 Example of affective Word Score Ui is the degree of the i-th token in the tree.
words and their corresponding
scores from BabelSenticNet ( سيءBad) − 0.605
3.2.4 Interactive learning layer
( رائعAmazing) 0.793
( صعبDifficult) − 0.539
In this layer, a multi-head self-attention (MHSA) is applied
( جميلBeautiful) 0.721
on the concatenation (Bconc) of local context, global context,
( مقززDisgusting) − 0.439
and the final output of GCN layers to improve the model’s
ability to learn the profound correlation between context and
aspect words:
sentence is constructed. Figure 2 illustrates an example of
a dependency tree. Then, the adjacency matrix of the graph Bconc
FCL
= W conc ⋅ Bconc + bconc (18)
is derived as follows:
{ ( )
1, if wi , wj contains dependency BILconc = MHSA Bconc
FCL (19)
Ai,j = (10)
0, 0 otherwise
Wconc and bconc are the weight and bias vectors of the fully
connected layer (FCL), respectively.
Inspired by the work of Liang et al. (2022), the represen-
tation of the adjacency matrix is enhanced by incorporating 3.2.5 Output layer
the affective information between context tokens and aspect
words using BabelSenticNet as follows: The sentiment polarity is predicted by pooling out the output
( ) of the interactive layer and feed it into a SoftMax layer:
Ei,j = Ai,j × Si,j + Ri,j + 1 (11) ( )
Bconc
pool
= POOL BILconc (20)
( ) ( )
Si,j = BabelSenticNet wi + BabelSenticNet wj (12) � �
exp Bconc
pool
{
1, if wi or wj is an aspect term Y=∑ � � (21)
Ri,j = (13) C
exp Bconc
0, 0 otherwise j=1 pool
where BabelSenticNet (wi) and BabelSenticNet (wj) are where C denotes the number of sentiment classes and Y is
the affective scores from BabelSenticNet database of the the predicted value of aspect sentiment polarity.
words wi and wj, respectively. Examples of affective scores
of some Arabic words are illustrated in Table 1.
The graph is then fed into the GCN layers to learn the 4 Experimental setup
affective dependencies related to a given aspect. Each node
in the l-th GCN layer is updated based on its neighboring 4.1 Dataset
nodes:
( ) The experiments were conducted on an Arabic dataset
vli = relu Ẽi gl−1
i
wl
+ b l
(14) from the hotel domain (Al-Smadi et al. 2016; Pontiki et al.
2016). The data was selected from hotel reviews collected
( ) by ElSahar and El-Beltagy (2015). The annotation scheme
gl−1
i
= F vl−1
i (15) followed the SemEval 2016 task 5 guidelines. Figure 3
illustrates an example of an annotated review from the
where vli denotes the product of current GCN layer. gl−1 is
i used dataset. Each review consists of one or multiple
the representation of i-th token evolved from the previous
sentences, and each sentence consists of a list of tuples
13
Social Network Analysis and Mining (2024) 14:6 Page 7 of 14 6
Table 2 Distribution size of the Arabic hotel dataset in terms of train- Table 4 Experimental hyper-parameters
ing and testing sets
Hyper-parameter Value
Reviews Sentences Tuples
Batch size 16
Training set 1839 4802 10,509 Max sequence length 128
Testing set 452 1227 2604 Dropout 0.2
Optimizer Adam
L2 regularization 1e−5
Table 3 Number of aspects per sentiment class in the training and Learning rate for BERT-based models 2e−5
testing sets Learning rate for non-BERT models 1e−3
Positive Negative Neutral Loss function Cross-entropy
13
6 Page 8 of 14 Social Network Analysis and Mining (2024) 14:6
The results with “†” are retrieved from the original papers. The results with “*” are obtained by imple-
menting the models based on the published source codes using the same Arabic dataset. The best results
are marked in bold and the unreported values are represented with “–”
4.4 Baseline and related work LCF-BERT (Zeng et al. 2019): Uses a local context focus
mechanism with an SRD threshold to help the model focus
In order to comprehensively evaluate the performance of on local context words and alleviate the negative influence
our model, we compare it with the following baseline and of unrelated sentiment tokens.
related work models: ASGCN-BERT: The BERT version of ASGCN-DG.
Baseline (Pontiki et al. 2016): Is an SVM-based classifier SenticGCN-BERT (Liang et al. 2022): The BERT version
provided by the SemEval competition. N-gram features were of SenticGCN.
used to train an SVM classifier with a linear kernel. It is worth mentioning that AOA, LCF-BERT, and
AB-LSTM-PC (Al-Smadi et al. 2019): The aspect embed- ASGCN-DG were implemented based on a publicly avail-
ding is combined with each word embedding, and then, an able repository,1 while SenticGCN was implemented using
attention mechanism is applied to help the model to focus the official published source code.2
on context words related to the aspect term.
IAN-GRU (Abdelgwad et al. 2021): Employs BiGRU to
provide the hidden states of targets and context, followed by 5 Results and discussion
two associated attention networks to learn the interactions
of aspects and their related context words. 5.1 Comparative analysis
MBRA (Al-Dabet et al. 2021): A memory module is
implemented using a stack of bidirectional independent The experimental results are illustrated in Table 5. They
long short-term memory with three layers, followed by a show that the AraBERT-based models achieve better results
recurrent attention mechanism to handle long and complex than AraVec-based models. This can be justified thanks to
sentences. the ability of AraBERT model to represent semantic fea-
AOA (Huang et al. 2018): This model adopts an atten- tures better than context-independent word embeddings.
tion-over-attention module to capture the interaction rela- Additionally, AraBERT uses a WordPiece tokenizer that
tionships between aspects and context sentences explicitly. splits unknown words into a set of sub-words, allowing bet-
ASGCN-DG (Zhang et al. 2019): This model uses GCN ter handling of out of vocabulary (OOV) issues, especially
on the dependency tree of a sentence to leverage the syn- in morphologically rich languages like Arabic.
tactical information and word dependencies to resolve the Besides, the graph-based (i.e., ASGCN-DG and Sen-
problem of multi-word dependency for the ASC task. ticGCN) methods have outperformed attention-based models
SenticGCN (Liang et al. 2022): Similar to ASGCN, but (i.e., AB-LSTM-PC, IAN, AOA, and MBRA). This proves
enhances the graph of the sentence using affective informa-
tion from BabelSenticNet.
BERT-Linear-pair (Abdelgwad et al. 2022): Fine-tuned 1
https://github.com/songyouwei/ABSA-PyTorch
BERT with a sentence pair input for ASC task. 2
https://github.com/BinLiang-NLP/Sentic-GCN
13
Social Network Analysis and Mining (2024) 14:6 Page 9 of 14 6
Table 6 Evaluation results of ablation study context words is more beneficial than down-weighting their
Model Accuracy (%) Macro F1 signals for the ASC task.
score (%)
13
6 Page 10 of 14 Social Network Analysis and Mining (2024) 14:6
5.5 Impact of fine‑tuning
13
Social Network Analysis and Mining (2024) 14:6 Page 11 of 14 6
Table 7 Results of case study based on four different sentences, randomly selected from the test dataset
Reviews Aspect Models Predicted Label True label
AR: المبنى جميل والغرف ال بأس بها ( الغرفrooms) BERT-Linear-pair Positive Neutral
EN: The building is beautiful and the rooms are okay LCF-BERT Neutral
ASGCN-BERT Neutral
SenticGCN-BERT Neutral
ArBLCSNGCN (ours) Neutral
AR: إذا كنت تود التعرض للسرقة فعليك بهذا المكان ( المكانlocation) BERT-Linear-pair Positive Negative
EN: If you want to be robbed, you should visit this place LCF-BERT Positive
ASGCN-BERT Negative
SenticGCN-BERT Negative
ArBLCSNGCN (ours) Negative
AR: الغداء افضل بكثير, و رغم أن الفطور عادي,االنترنت عادة جيدة الغداء BERT-Linear-pair Neutral Positive
EN: The internet is usually good, and although breakfast is normal, (lunch) LCF-BERT Positive
lunch is much better
ASGCN-BERT Neutral
SenticGCN-BERT Positive
ArBLCSNGCN (ours) Positive
ً اإلفطار بطيء، والتكييف يعمل،الفندق جيد! الغرف نظيفة ومريحة
AR: قليال ولكن جيد اإلفطار BERT-Linear-pair Positive Neutral
EN: The hotel is good! The rooms are clean and comfortable, the air- (breakfast) LCF-BERT Negative
conditioning works, the breakfast is a little slow but good
ASGCN-BERT Positive
SenticGCN-BERT Positive
ArBLCSNGCN (ours) Neutral
13
6 Page 12 of 14 Social Network Analysis and Mining (2024) 14:6
13
Social Network Analysis and Mining (2024) 14:6 Page 13 of 14 6
interpersonal relation into context-aware emotion recognition. Pontiki M, Galanis D, Papageorgiou H, Androutsopoulos I, Manand-
Appl Intell 53(4):4201–4217 har S, Al-Smadi M, Al-Ayyoub M, Zhao Y, Qin B, De Clercq O
Costola M, Hinz O, Nofer M, Pelizzon L (2023) Machine learning (2016). Semeval-2016 task 5: aspect based sentiment analysis.
sentiment analysis, COVID-19 news and stock market reactions. International workshop on semantic evaluation
Res Int Bus Finance 64:101881 Rodríguez-Ibánez M, Casánez-Ventura A, Castejón-Mateos F,
Devlin J, Chang M-W, Lee K, Toutanova K (2019) BERT: pre-train- Cuenca-Jiménez P-M (2023) A review on sentiment analysis
ing of deep bidirectional transformers for language understand- from social media platforms. Expert Syst Appl 119862
ing. In: Proceedings of the 2019 conference of the north Amer- Ruder S, Ghaffari P, Breslin JG (2016) INSIGHT-1 at SemEval-2016
ican chapter of the association for computational linguistics: task 5: deep learning for multilingual aspect-based sentiment
human language technologies, vol 1 (Long and Short Papers) analysis. Proc SemEval 330–336
Du Y, Li T, Pathan MS, Teklehaimanot HK, Yang Z (2022) An effec- Soliman AB, Eissa K, El-Beltagy SR (2017) AraVec: a set of Ara-
tive sarcasm detection approach based on sentimental context bic word embedding models for use in Arabic NLP. Procedia
and individual expression habits. Cognit Comput 1–13 Comput Sci 117:256–265. https://doi.org/10.1016/j.procs.2017.
ElSahar H, El-Beltagy SR (2015) Building large arabic multi-domain 10.117
resources for sentiment analysis Wang Y, Huang M, Zhu X, Zhao L (2016) Attention-based LSTM
Fadel AS, Abulnaja OA, Saleh ME (2023) Multi-task learning model for aspect-level sentiment classification. In: Proceedings of
with data augmentation for arabic aspect-based sentiment analy- the 2016 conference on empirical methods in natural language
sis. Comput Mater Contin 75(2) processing
Gu T, Zhao H, He Z, Li M, Ying D (2023) Integrating external Wu M (2023) Commonsense knowledge powered heterogeneous
knowledge into aspect-based sentiment analysis using graph graph attention networks for semi-supervised short text clas-
neural network. Knowl-Based Syst 259:110025. https://doi.org/ sification. Expert Syst Appl 120800
10.1016/j.knosys.2022.110025 Xu C, Luo X, Wang D (2022) MCPR: a Chinese product review data-
Huang B, Ou Y, Carley KM (2018) Aspect level sentiment classifi- set for multimodal aspect-based sentiment analysis. Cognitive
cation with attention-over-attention neural networks. In: Inter- Computing—ICCC 2022, Cham
national conference on social computing, behavioral-cultural Yang Z, Dai Z, Yang Y, Carbonell J, Salakhutdinov RR, Le QV
modeling and prediction and behavior representation in mod- (2019) XLNet: generalized autoregressive pretraining
eling and simulation for language understanding. Adv Neural Inf Process Syst
Huang B, Zhang J, Ju J, Guo R, Fujita H, Liu J (2023a) CRF-GCN: 32:5753–5763
an effective syntactic dependency model for aspect-level senti- Yang H, Zeng B, Yang J, Song Y, Xu R (2021) A multi-task learning
ment analysis. Knowl-Based Syst 260:110125. https://doi.org/ model for Chinese-oriented aspect polarity classification and
10.1016/j.knosys.2022.110125 aspect term extraction. Neurocomputing 419:344–356
Huang H, Asemi A, Mustafa MB (2023) Sentiment analysis in Yang G, Xu Y, Tu L (2023) An intelligent box office predictor based
E-commerce platforms: a review of current techniques and on aspect-level sentiment analysis of movie review. Wirel Netw
future directions. IEEE Access 29(7):3039–3049
Kastrati Z, Arifaj B, Lubishtani A, Gashi F, Nishliu E (2020) Aspect- Yusuf AA, Chong F, Xianling M (2022) Evaluation of graph convo-
based opinion mining of students’ reviews on online courses. lutional networks performance for visual question answering on
In: Proceedings of the 2020 6th international conference on reasoning datasets. Multimed Tools Appl 81(28):40361–40370
computing and artificial intelligence Zeng B, Yang H, Xu R, Zhou W, Han X (2019) LCF: a local context
Li X, Zhang J, Du Y, Zhu J, Fan Y, Chen X (2023) A novel deep focus mechanism for aspect-based sentiment classification. Appl
learning-based sentiment analysis method enhanced with Emo- Sci 9(16):3389
jis in microblog social networks. Enterp Inf Syst 17(5):2037160 Zhang M, Qian T (2020) Convolution over hierarchical syntactic and
Liang B, Su H, Gui L, Cambria E, Xu R (2022) Aspect-based senti- lexical graphs for aspect level sentiment analysis. In: Proceed-
ment analysis via affective knowledge enhanced graph convo- ings of the 2020 conference on empirical methods in natural
lutional networks. Knowl-Based Syst 235:107643 language processing (EMNLP)
Liu H, Wu Y, Li Q, Lu W, Li X, Wei J, Liu X, Feng J (2023) Enhanc- Zhang C, Li Q, Song D (2019) Aspect-based sentiment classification
ing aspect-based sentiment analysis using a dual-gated graph with aspect-specific graph convolutional networks. In: Proceed-
convolutional network via contextual affective knowledge. Neu- ings of the 2019 conference on empirical methods in natural
rocomputing 553:126526. https://d oi.o rg/1 0.1 016/j.n eucom. language processing and the 9th international joint conference
2023.126526 on natural language processing (EMNLP-IJCNLP)
Ma D, Li S, Zhang X, Wang H (2017) Interactive attention networks Zhao A, Yu Y (2021) Knowledge-enabled BERT for aspect-based
for aspect-level sentiment classification. IJCAI sentiment analysis. Knowl-Based Syst 107220
Madan M, Rani A, Bhateja N (2023) Applications of named entity Zhao P, Hou L, Wu O (2020) Modeling sentiment dependencies
recognition using graph convolution network. SN Comput Sci with graph convolutional networks for aspect-level sentiment
4(3):266. https://doi.org/10.1007/s42979-023-01739-8 classification. Knowl-Based Syst 193:105443
Mikolov T, Chen K, Corrado G, Dean J (2013) Efficient estimation Zhou J, Huang JX, Hu QV, He L (2020) SK-GCN: modeling syntax
of word representations in vector space. Preprint http://arxiv. and knowledge via graph convolutional network for aspect-level
org/abs/1301.3781 sentiment classification. Knowl-Based Syst 205:106292
Omar A, Abd El-Hafeez T (2023) Quantum computing and machine Zhao H, Xie J, Wang H (2022a) Graph convolutional network based
learning for Arabic language sentiment classification in on multi-head pooling for short text classification. IEEE Access
social media. Sci Rep 13(1):17305. https://d oi.o rg/1 0.1 038/ 10:11947–11956
s41598-023-44113-7 Zhao Z, Tang M, Tang W, Wang C, Chen X (2022b) Graph convo-
Pennington J, Socher R, Manning CD (2014) Glove: global vectors lutional network with multiple weight mechanisms for aspect-
for word representation. In: Proceedings of the 2014 conference based sentiment analysis. Neurocomputing 500:124–134.
on empirical methods in natural language processing (EMNLP) https://doi.org/10.1016/j.neucom.2022.05.045
13
6 Page 14 of 14 Social Network Analysis and Mining (2024) 14:6
Zhao Y, Zhang L, Zeng C, Lu W, Chen Y, Fan T (2023) Construction Springer Nature or its licensor (e.g. a society or other partner) holds
of an aspect-level sentiment analysis model for online medical exclusive rights to this article under a publishing agreement with the
reviews. Inf Process Manage 60(6):103513 author(s) or other rightsholder(s); author self-archiving of the accepted
manuscript version of this article is solely governed by the terms of
Publisher's Note Springer Nature remains neutral with regard to such publishing agreement and applicable law.
jurisdictional claims in published maps and institutional affiliations.
13