Skip to main content
PLOS One logoLink to PLOS One
. 2024 Jun 27;19(6):e0301508. doi: 10.1371/journal.pone.0301508

CKG: Improving ABSA with text augmentation using ChatGPT and knowledge-enhanced gated attention graph convolutional networks

Yapeng Gao 1, Lin Zhang 1,*, Yangshuyi Xu 1
Editor: Toqir Rana2
PMCID: PMC11210807  PMID: 38935678

Abstract

Aspect-level sentiment analysis (ABSA) is a pivotal task within the domain of neurorobotics, contributing to the comprehension of fine-grained textual emotions. Despite the extensive research undertaken on ABSA, the limited availability of training data remains a significant obstacle that hinders the performance of previous studies. Moreover, previous works have predominantly focused on concatenating semantic and syntactic features to predict sentiment polarity, which inadvertently severed the intrinsic connection. Several studies have attempted to utilize multi-layer graph convolution for the purpose of extracting syntactic characteristics. However, this approach has encountered the issue of gradient explosion. This paper investigates the possibilities of leveraging ChatGPT for aspect-level text augmentation. Furthermore, we introduce an improved gated attention mechanism specifically designed for graph convolutional networks to mitigates the problem of gradient explosion. By enriching the features of the dependency graph with a sentiment knowledge base, we strengthen the relationship between aspect words and the polarity of the contextual sentiment. It is worth mentioning that we employ cross-fusion to effectively integrate textual semantic and syntactic features. The experimental results substantiate the superiority of our model over the baseline models in terms of performance.

1. Introduction

In recent years, with the pervasive utilization of social media and online reviews, individuals have been afforded greater opportunities to articulate their emotions on the internet [1]. These expressions of emotion not only involve overall sentiment polarity determination but also encompass specific aspects of emotions. This presents new challenges for neurorobotics in understanding textual emotions at a more intricate level. Therefore, research on ABSA has become increasingly important. ABSA is a branch of sentiment analysis that aims to identify and analyze the sentiment orientation towards specific aspects or targets in text. As shown in the Fig 1, in product reviews, users may have different sentiment evaluations for different aspects of the product, such as performance, design, price, and so on. Understanding these aspect-level sentiment information can help companies gain insights into consumer satisfaction with different aspects, thereby improving products and providing better user experiences.

Fig 1. Example of aspect-level sentiment analysis.

Fig 1

However, ABSA faces several challenges. Aspect identification is a critical issue because aspects are often specific entities, attributes, or viewpoints that are contextually related [2]. Furthermore, ABSA needs to consider the sentiment expressions associated with different aspects in the text, such as positive, negative, or neutral sentiments. The challenge in this multi-label classification problem lies in accurately capturing the sentiment information related to aspects. Additionally, sentiment expressions in the text may be implicit or ambiguous, requiring context understanding and inference.

Most of the early work on ABSA utilized neural network models to extract sentiment information from the given aspect in context. Although earlier models based on temporal models [36] and attention mechanisms [7, 8] have achieved certain effectiveness, they still face several challenges, including the scarcity of sufficient training data, imbalanced data distribution, and limitations in the generalization and practical applicability of existing models. Consequently, the analysis of syntax has gradually gained increasing attention, as shown in the Fig 2. In addition, existing methods still need improvement in handling grammatical and semantic learning abilities, as well as effectively linking aspect words, context, and multiple network features.

Fig 2. Dependency tree of a sentence.

Fig 2

The scarcity of training data and the imbalanced distribution of data present significant challenges in ABSA. Obtaining domain-specific data is often a complex task, resulting in limited training data size. This limitation can undermine the generalization ability of the models, hindering their performance across diverse scenarios. Furthermore, the uneven distribution of data introduces instability in model performance when confronted with various aspect and sentiment expressions. Moreover, existing methods exhibit limitations in their grammatical and semantic learning abilities. Traditional attention mechanisms excel at extracting semantic information but struggle with handling grammatical nuances effectively. Conversely, methods solely reliant on grammar dependency trees fail to capture the contextual information of aspect words adequately. As a result, the model’s understanding capability within complex contexts becomes restricted. Addressing these limitations is imperative to enhance the grammatical and semantic learning abilities of ABSA models.

To tackle the aforementioned challenges, this study presents the CKG (Aspect-Level Sentiment Analysis with Text Augmentation using ChatGPT and Knowledge-Enhanced Gated Attention Graph Convolutional Network) model. The CKG model employs a graph convolutional neural network to analyze sentences, leveraging sentiment knowledge from SenticNet to amplify the influence of aspect words on sentiment polarity within the sentence’s dependency graph. Furthermore, the model incorporates gate attention mechanisms to update nodes. In comparison to conventional graph neural networks, the CKG model exhibits enhanced efficiency and superior performance. Additionally, we streamline the attention mechanism by employing relative position encoding to extract semantic information, thereby effectively handling the relative relationships of positional information within sequences.

Furthermore, before the final sentiment prediction output, we employ cross-fusion to integrate the final outputs of the sequence module and the graph convolution module. Compared to previous concatenation or summation methods, this approach not only improves the model’s performance but also enhances its interpretability. In summary, CKG not only pays attention to the syntactic dependency of texts but also effectively combines the semantic features of neighboring nodes, thereby enhancing the model’s understanding of text semantics and grammatical structures.

The main contributions of our work can be summarized as follows:

  • A method is proposed to expand text data based on the original dataset using ChatGPT.

  • A graph convolutional network with gate attention mechanisms is introduced, which enhances the aspect features of the original sentence by incorporating syntactic dependency trees and an external knowledge base. This ensures the full consideration of the connection between the original sentence and aspect words, avoiding the loss of information.

  • We propose a CKG model, which integrates syntactic and semantic features through cross-fusion, and its effectiveness in ABSA tasks has been demonstrated on five benchmark datasets.

2. Related works

2.1 Aspect-based sentiment analysis

Aspect-based sentiment analysis (ABSA) is an important research direction in the field of sentiment analysis. Its goal is to identify and analyze the sentiment inclination towards specific aspects or targets in the text. This task is more fine-grained, judging the sentiment polarity of the text based on specific aspect words. Compared to document-level or sentence-level sentiment analysis, ABSA is more reasonable and practical [9]. The key to ABSA is to accurately identify and analyze the sentiment inclination towards specific aspects or targets in the text. ABSA needs to focus on aspect words and sentiment words in the text and understand the relationship between them. When conducting sentiment analysis, it is necessary to consider the semantic information of the context and the complexity of the sentiment expression. Traditional machine learning algorithms require manual design and selection of features, such as word bag models and n-gram features. The selection of these features may be influenced by subjective factors and cannot fully capture the semantic information of the text. In addition, these methods require a large amount of annotated data to train the model, and the acquisition cost of annotated data is high. When the dataset changes, the performance of the methods is also greatly affected [10]. Therefore, at the current stage, this technology is basically only used as a research reference.

With the advancement of deep learning technology, models based on Convolutional Neural Networks (CNN) and Recurrent Neural Networks (RNN) have been widely used in this task. These models can learn richer feature representations from the text and capture the complex relationship between sentences and aspects. In addition, the development of word embedding techniques, such as Word2Vec [11], PV [12], and GloVe [13], has also promoted the progress of deep learning in sentiment analysis tasks. Long Short-Term Memory (LSTM), a variant of Recurrent Neural Network (RNN), is widely used in sentiment analysis tasks, especially in solving the problem of long-term dependencies in time sequences. However, traditional LSTM has some limitations, such as being difficult to train in parallel and ineffective in handling the joint relationship between context information and aspect words in ABSA, ignoring their intrinsic connection.

2.2 Attention mechanism in aspect-based sentiment analysis

In recent years, attention mechanisms have received considerable attention in ABSA research because they can capture the relationship between aspects and sentiments and place them in a context [14]. Attention mechanisms allow models to focus on relevant parts of the text when generating sentiment predictions, effectively handling long and context-rich sentences. By assigning different weights to each word or phrase, attention mechanisms can emphasize the most informative aspects, improving overall sentiment classification performance.

Some studies have explored the application of attention mechanisms in ABSA and achieved good results. For example, researchers have proposed using self-attention mechanisms to capture the importance of different words in aspect-related contexts [15]. This technique allows the model to assign higher weights to words related to specific aspects, thereby improving ABSA. In addition, hierarchical attention mechanisms [16] have been used to capture sentiment information at different granularity levels. By assigning attention weights at the word level and aspect level, it can effectively capture the sentiment expressed for each aspect and aggregate the information for final sentiment prediction. This approach has shown better performance in capturing subtle differences in fine-grained sentiment.

Despite the notable achievements of attention mechanisms in ABSA, several challenges persist that necessitate attention and resolution. One such challenge pertains to the limited interpretability of attention weights. While attention mechanisms offer insights into the significance of different words or phrases, elucidating the precise reasoning process underlying the assigned weights often proves arduous. Techniques aimed at enhancing the interpretability of attention weights can significantly contribute to understanding the decision-making process of the model. It is noteworthy that the resilience of attention mechanisms against noise and ambiguous inputs remains an open question. Noise inputs encompassing spelling errors, grammatical inaccuracies, or informal language usage can potentially impede the accurate allocation of attention weights and hinder the capture of relevant sentiment information. Research dedicated to developing attention mechanisms resilient to such noise is indispensable for practical applications.

2.3 GNN in aspect-based sentiment analysis

To address the limitations of attention mechanisms in ABSA, research based on Graph Neural Networks (GNN) has received widespread attention in the ABSA field. GNN has the ability to model complex relationships and capture contextual information [17], making it a significant application in ABSA. GNN provides a powerful framework for integrating syntactic and semantic information, allowing models to better understand the sentiments expressed for different aspects.

Some studies have explored the application of GNN in ABSA and achieved promising results. Researchers have proposed methods that represent text data as a graph, where aspects and related sentiments are represented as nodes, and their relationships are modeled as edges. By leveraging GNN [18], these models can capture the interactions between aspects and sentiments, thereby improving sentiment analysis performance. Zhang et al. [19] have developed the Tree Communication Model (TCM), which uses graph convolutional networks and recurrent graph neural networks to construct a tree structure based on syntactic parsing results. This model achieves better performance in text sentiment classification by capturing the relationships between nodes in a more comprehensive way. Building on this idea, Zhou et al. [20] have improved the syntax tree structure by learning a tree structure centered on aspects to shorten the distance between aspects and corresponding opinion words, showing good results through experiments. In order to further enhance the impact of aspect words on graph network construction and their relevance to the context, some works [21, 22] have attempted to introduce external knowledge bases to increase the importance of aspect words in constructing graph networks. However, although progress has been made in applying GNN in ABSA, there are still some unresolved research questions, such as how to effectively integrate external knowledge into graph networks, how to effectively apply attention mechanisms to improve the understanding of graph networks, and how to avoid the problem of gradient explosion during training.

In this paper, we propose a Graph Convolutional Network (GAGCN) with gate-controlled attention mechanisms that focus the network’s attention on nodes crucial for ABSA, thereby improving the model’s performance. Furthermore, we introduce SenticNet [23], a sentiment knowledge base, to enhance the weights of aspect nodes and their related words, ensuring that the model’s attention is focused on aspect words and their corresponding sentiment tendencies. Additionally, this method can effectively handle multi-label tasks and achieve sentiment analysis for multiple aspects by assigning different attention weights to different aspects. Experimental results show that the proposed model, which combines sentiment knowledge enhancement and an adaptive attention mechanism, performs better.

2.4 Overview of ChatGPT

The rapid development of artificial intelligence has brought many exciting technological breakthroughs, one of which is ChatGPT (Chat-based Generative Pre-trained Transformer) [24]. ChatGPT is a deep learning-based natural language processing model with a wide range of applications and profound impact. It can understand and generate natural language text, making human-computer interaction more natural and fluent. ChatGPT and related models have achieved remarkable results in natural language processing tasks through pre-training and fine-tuning methods. By learning language patterns and semantic information from large-scale corpora, they can generate accurate, coherent, and meaningful answers.

As an innovative natural language processing technology, ChatGPT has a wide range of applications in many fields [25]. First, it plays an important role in virtual assistants and customer services. ChatGPT can interact with users in real-time as an intelligent conversational system, providing personalized information and answering questions. In the field of education, ChatGPT can serve as a personalized learning companion, offering question answering, learning materials, and guidance to improve learning outcomes and engagement [26]. ChatGPT is also applied in areas such as intelligent customer support, automated text generation, information retrieval, and smart homes, providing people with convenient and intelligent services.

The emergence of ChatGPT brings significant value and impact. It provides a new way of human-computer interaction, allowing users to have more freedom in conversation with computers and enjoy personalized services and customized experiences. Intelligent answering and recommendation systems of ChatGPT can provide users with accurate and high-quality information, improving the efficiency and accuracy of information retrieval. Additionally, ChatGPT provides an important tool for researchers in language learning and natural language processing, promoting the development and innovation in these fields.

Despite the significant progress of ChatGPT, there are also negative evaluations and potential issues [27]. Due to its training on large-scale datasets, ChatGPT may have problems related to information bias and implicit bias. Furthermore, ChatGPT may encounter difficulties in handling complex questions and understanding context, leading to inaccurate answers or lack of logical reasoning. The application of ChatGPT has also raised concerns about ethics and privacy, such as privacy protection of user data and potential risks of misuse. This paper mainly uses ChatGPT for text data augmentation based on the original dataset and does not involve sensitive issues.

3. Methods

In this section, we provide a detailed explanation of the proposed graph attention convolutional network, which utilizes ChatGPT for data augmentation and incorporates sentiment knowledge enhancement and gate mechanism. We also provide an introduction to the various details of the overall model.

3.1 Overview

As shown in the Fig 3, our proposed model consists of two main components: (1) learning contextual representations using BiLSTM or BERT, which takes the embedding matrix of each sentence as input and outputs the contextual feature representation of the sentence. (2) Feature extraction of the syntactic dependencies of the sentences, where the enhanced representations of the sentences are inputted through the sentiment knowledge base and further feature extraction is performed using the graph attention convolutional network with incorporated gate mechanism. (3) The sentence features obtained from (1) and aspect features are enhanced using the self-attention mechanism with relative positional encoding. The graph information outputted by GAGCN is also fused to extract important sentiment dependencies related to specific aspects. Previous methods that solely relied on sequence networks or attention-based methods to analyze global semantics and aspect features failed to effectively link aspect words with underlying syntactic features of the original text. In addition, most models that employed graph convolutional neural networks only analyzed the syntactic dependencies and overlooked the aspect-sentiment connections contained in the semantic information of the text. Our proposed approach addresses these two shortcomings and strengthens the influence of aspects on the sentence sentiment, thereby improving the accuracy and interpretability of sentiment polarity output.

Fig 3. Overall architecture diagram of the CKG model.

Fig 3

3.2 Definition

Aspect-level sentiment classification tasks typically involve preparing an input sequence for the model consisting of a context sequence and an aspect sequence. This input format enables the model to learn the relationship between the context and the aspect term. Assuming that the context sequence containing the input aspect term is denoted by s = {w0, w1, …, wn}, it contains n + 1 words, including the target aspect term. The target aspect sequence is represented as st={w0t,w1t,,wmt} and is a subsequence of s composed of m(m ≥ 1) words. The objective of aspect-based sentiment analysis is to determine the sentiment polarity related to the aspect based on the given sentence and aspect term.

3.3 Data augmentation using ChatGPT

ChatGPT is developed based on the research of GPT [28], GPT-2 [29], and GPT-3 [30]. Its core idea is to incorporate a reinforcement learning model into GPT-3 to fine-tune the model’s output to be more reasonable, accurate, truthful, and harmless [31]. During the pre-training phase of ChatGPT, an unsupervised task is performed where input samples are denoted as X = {x1, x2, …, xn}, xi = (s1, s2, …, sm). The Transformer model is trained to obtain corresponding tokens and their associated positional encodings. Representing the trainable parameters as θ, the ultimate goal of the pre-training phase is to maximize the following Eq 1:

L(xi)=i=1mlogP(si|s1,,si-1;θ) (1)

After pre-training, a reinforcement learning model with human feedback is used to fine-tune the pre-trained model. Human experts are involved in constructing and evaluating samples with questions and their corresponding answers. These prompted samples are utilized for further training of the model. In addition to fine-tuning, ChatGPT incorporates an additional training of a reward model. This reward model combines the question-answer pairs set by human experts with the model’s predicted results and assigns reward scores accordingly, as shown in the following Eq 2:

loss(θr)=E(x,yw,yl)Dc[log(σ(rθr(x,yw)-rθr(x,yl)))] (2)

Here, θr represents the parameters of the reward model. x denotes the prompt, while yw represents the preferred completion among the given alternatives in yl. Dc denotes the dataset used for human expert comparisons. Based on the reward model, ChatGPT is fine-tuned using the Proximal Policy Optimization (PPO) strategy, ultimately completing the training of the final model.

There are several models available for text data augmentation [3234]. However, ChatGPT is particularly suitable for the data augmentation task in this study [35]. This is because ChatGPT combines manually labeled training samples by human experts, resulting in generated content that is more genuine and aligned with human language conventions. Additionally, under the supervision of the reinforcement learning model, the generated content is more reliable and of higher quality. Furthermore, the training of ChatGPT based on a large-scale corpus enhances the diversity of data generation, which is beneficial for expanding the data in this study. Fig 4 presents a real case of text generation by ChatGPT, demonstrating its high-quality output that ensures overall semantic fluency and naturalness of sentences, while also satisfying the diversity requirements of the data augmentation task. Based on ChatGPT, the algorithmic procedure for our data augmentation is outlined as this Algorithm 1.

Fig 4. ChatGPT-generated text examples incorporating aspect words based on the original text.

Fig 4

Algorithm 1 The framework of ChatGPT for augmenting text in the original dataset.

Input: base dataset D and novel Dataset Dn

Output: Dn

Definition: POS, NEU, NEGD represent the collections of texts in the original dataset labeled as positive, neutral, and negative, respectively, containing aspect words. αPOS, βNEU, γNEG represent individual texts within these collections. count() denotes a counting function, and M is a manually set parameter. ChatGPTaug(x, y) represents the data augmentation process using ChatGPT, which takes two inputs:x ≥ 1, xN+ denotes the number of texts to be generated, and yD represents an individual text.

1: M = 3000

2: for α in POS and count(POS) < M do

3:  DaugPQS=ChatGPT_aug(Mcount(pos),α)

4: end for

5: for β in NEU and count(NEU) < M do

6:  DaugNEU=ChatGPT_aug(Mcount(NEU),β)

7: end for

8: for γ in NEG and count(NEG) < M do

9:  DaugNEG=ChatGPT_aug(Mcount(NEG),γ)

10: end for

11: Dn=DaugPOSDaugNEUDaugNEGD

12: return Dn

The function ChatGPT_aug() in Algorithm 1 represents the data augmentation of training text data using ChatGPT. It is worth noting that the prompt configuration has a significant impact on the generation effectiveness of ChatGPT. In Fig 5, we illustrate how our carefully designed prompt effectively guides ChatGPT in performing the data augmentation task.

Fig 5. Process flow diagram for text augmentation using prompts in ChatGPT.

Fig 5

This paper [35] have employed cosine similarity and transRate to evaluate the quality of generated text. However, we argue that these approaches are not suitable for ABSA. Firstly, when proper prompts are provided and the generation process is carried out iteratively with a limited amount of text, ChatGPT exhibits commendable text generation quality. With the support of large language models, ChatGPT can creatively generate text based on the original input. It may produce text that meets the requirements of ABSA, but differs significantly in similarity to the original sentence. For example, it may change aspect words or alter sentence expressions, resulting in a low cosine similarity Table 3. Therefore, we adopt human expert to assess the quality of the generated text, as it provides a more reliable and accurate measure. To ensure the effectiveness of the augmented data and the feasibility for model training, we randomly selected 50 instances of generated data. These instances were then mixed with 50 instances of real data from the dataset. The combined set was distributed to five human experts specializing in ABSA for scoring tests. Each expert was required to assign a score ranging from -1 to 3 (integer values only) to indicate the suitability of each sentence for ABSA. The specific scoring rules were as Table 1.

Table 1. The scoring criteria for evaluating the applicability of text data generated by ChatGPT.

Score Rules
-1 logic inconsistent, semantics unclear, no aspect words
0 logic inconsistent, semantics ambiguous, no aspect words
1 logic indistinct, semantics ambiguous, no aspect words
2 logic distinct, semantics ambiguous, contain aspect words
3 logic is consistent, semantics clear, contain aspect words

3.4 Embedding module

The Embedding layer serves as the fundamental building block of the model architecture, and in this study, two methods are employed for text encoding. The first method leverages a Bi-LSTM to compute the textual representation by utilizing word vectors derived from the input text. Conversely, the second method incorporates the utilization of the pre-trained BERT model for text encoding. BERT utilizes a transformer structure, which encompasses a self-attention mechanism facilitating bidirectional information flow. By employing a Masked Language Model (MLM) during pre-training, BERT effectively considers both preceding and succeeding context, thereby comprehensively capturing semantic contexts and enhancing the accuracy of information extraction.

We map each word to an m-dimensional word embedding through an embedding table XRm×|N|, resulting in an embedding matrix X=[x1,x2,,xa1,xa2,,xak,,xn]. Here, xaiRm represents the word embedding of the aspect term wai. The dimensionality of the word vectors is denoted by m, n represents the sentence length, and |N| denotes the vocabulary size. We derive the embedding lookup table from pre-trained embeddings such as GloVe or BERT [36], and fine-tune it during the training process.

CKG conducts experimental comparisons between the two approaches. The LSTM-based model demonstrates lower accuracy compared to BERT, but requires significantly less training time. On the other hand, the BERT-based model exhibits significantly improved accuracy at the cost of longer training time and higher computational expenses.

3.5 Knowledge-enhanced GCN with attention and gated skip-connection (GAGCN)

The graph structure incorporates syntactic encoding as input. To encode syntactic information, the SpaCy (The spaCy toolkit is used to parse the dependency tree of the sentence: https://spacy.io/.) dependency parser is used to compute syntactic dependencies. By providing the syntactic structure, we capture richer information about the sentence structure and construct the adjacency matrix DRn×n according to Eq 3:

Di,j={1ifwi,wjcontainsdependency0otherwise} (3)

To leverage the relationships between words, our graph structure adopts an undirected dependency graph. Inspired by [37], we enhance the adjacency matrix with sentiment information using SenticNet. First, we need to calculate the SenticNet scores between nodes, which can be represented by the following Eq 4:

Si,j=SenticNet(wi,wj) (4)

Where, SenticNet(wi, wj) ∈ [−2, 2] represents the combined sentiment score of wi and wj after being computed by SenticNet. This Eq 4 ensures that the model tends to favor words with higher sentiment scores, capturing more accurate sentiment polarity. It is worth noting that if wi does not appear in SenticNet or is a neutral word, SentNet(wi) = 0. Next, based on the computed S(i, j), we can enhance the previously constructed adjacency matrix with sentiment information using the following Eq 5:

Ai,j=Di,j×(Si,j+Ti,j+1) (5)
Ti,j={1ifwiorwjisanaspectword0otherwise} (6)

Using the adjacency matrixAi,j for syntactic encoding, GAGCN takes the input hidden state vectors H as initial node representations in the syntactic graph. Then, the GAGCN module obtains the graph representation of the syntax as Hsyn={h1syn,h2syn,,hnsyn}, where the symbol hisynRdu represents the hidden representation of the i-th node. The update equation for the representation of the i-th node in the l-th layer is as Eq 7:

hisynl=ϕ(j=1nAi,jF(hjsynl-1+natti)Wl+bl) (7)

Wl represents the weight matrix and symbol bl represents the bias term. For aspect nodes, we use the {ha1syn,ha2syn,,hansyn} to represent their hidden representations. F is the position-aware function referenced from [38], and θ denotes the relu activation function. Based on this equation, assuming there exists a node Hil in the l-th layer with three neighboring nodes H1l, H2l, and H3l, we can simplify and rewrite the above equation by omitting the bias term as Eq 8:

sHil+1=ϕ(HilWl+H1lWl+H2lWl+H3lWl) (8)

Based on this equation, we introduce the attention mechanism as shown in the Fig 6 to update the node states as Eq 9:

Hil+1=ϕ(αiilHilWl+αi1lH1lWl+αi2lH2lWl+αi3lH3lWl) (9)

Where αijl represents the attention score for the j-th node influencing the update of the i-th node in the l-th layer. The general attention calculation can be simply represented by the Eq 10:

αijl=ϕ(HilWl,HjlWl) (10)

From the above equation, we can deduce the attention weights between each node and its neighboring nodes. However, for ABSA, inspired by [39], our focus should be on the influence of aspect terms on the sentence rather than the attention weights between adjacent nodes. Therefore, we can evaluate the attention coefficients by calculating the coupling between aspect terms and other words.

αijl=tanh((HilWl)Cl,(HjlWl)T) (11)

Fig 6. The arrows represent the information propagation between nodes, and the different colors of the arrows indicate the varying importance of adjacent nodes to the aspect term node ai.

Fig 6

It is worth noting that Zhang in [38] employed Aspect-aware Attention to retrieve important features semantically related to the aspect terms from the hidden state vectors. Accordingly, attention weights based on retrieval were assigned to each context word. Inspired by this, we propose the aspect term weight anchoring method. This method utilizes aspect term-specific masking to mask out non-aspect words learned by the final GCN layer’s output, while keeping the representation of aspect terms unchanged. This is done to construct our coupling matrix. Since we enhance aspect terms in the graph network with SenticNet, it is necessary to anchor the representation of aspect terms to prevent the model from weakening the sentiment representation of aspect terms during the training process. This is shown in the following equation:

h˙t={h˙tifηtη+la0otherwise} (12)

Where h˙i represents the representation of the i-th word obtained through GAGCN, ι is the starting index of the aspect term in the sentence, and la is the length of the aspect term. From this, we can obtain the final representation of the anchored aspect term masking:

C˙masksynl={0,,h˙η,,h˙η+la-1,,0} (13)

Furthermore, we utilize a multi-head attention mechanism to describe the interactions between nodes using K different channels:

Hil=tanh(1Kk=1KjN(i)αij,klHjlWl) (14)

As is well known, when multiple graph convolutional layers are added, the model can better understand long-distance information. However, issues such as gradient explosion may arise, which can impact the performance of the model. Therefore, we introduce a gate mechanism in skip connections to accurately propagate historical information when updating the hidden states, as shown in the Fig 7. This gate mechanism determines the forget rate and update rate, ensuring precise information transfer.

Hi,gscl+1=γiHil+1+(1-γi)Hil (15)
γi=σ(Gγ,1Hil+1+Gγ,2Hil+bγ) (16)

Where Gγ,1, Gγ,2 and bγ are trainable parameters, ⊙ represents element-wise matrix multiplication, and σ denotes the sigmoid activation function.

Fig 7. The difference between skip-connection and gated skip-connection.

Fig 7

Finally, in order to effectively connect the semantic module and the syntactic module features and fully consider all information, we employ cross fusion [40] as the feature fusion method, calculated through the following equation:

Hsyn=f(HsynW1(X)T)X (17)
Hsem=f(XW2(Hsyn)T)Hsyn (18)

Where W1 and W2 are trainable parameters, and f represents the softmax activation function. After this operation, both Hsyn and Hsem fully consider the features of syntax and semantics. Then, the final fusion feature r is calculated by applying average pooling (ap) to these features:

hasyn=ap(ha1synha2syn,,hamsyn) (19)
hasem=ap(ha1semha2sem,,hamsem) (20)
r=[hasyn;hasem] (21)

3.6 Output layer

The CKG will extract features from the global structural information, semantic information, and the content of the target aspect. The final representation r is then sent to a linear layer and passed through the softmax function to generate the probability distribution y for the given aspect a:

y=softmax(Wpr+bp) (22)

Here, Wp and bp represent the learnable weights and biases, respectively. The function softmax() denotes the softmax function, which enables us to learn the final distribution of emotions as output.

3.7 Model training

We employ the standard gradient descent algorithm to optimize and update the parameters of the proposed model. The objective of training the model is to minimize the cross-entropy loss through L2 regularization:

loss=-i=1sj=1cy^logyij+λΘ2 (23)

Here, S represents the number of training samples, and C denotes the number of classes. y^ represents the correct distribution of emotions. Θ signifies all trainable parameters. λ is the coefficient for the L2 regularization term.

4. Experiments

In this section, we will introduce the dataset used to evaluate the model’s performance, provide an overview of the experimental parameters, and conclude with a comparative analysis of the models.

4.1 Datasets and augmentation with ChatGPT

We trained and validated our model on five publicly available benchmark models: Twitter, proposed by Dong [41], which contains Twitter post data, and four others (LAP14, REST14, REST15, REST16) from the lap and restaurants domains of SemEval 2014 task 4 [42], the restaurants domain of SemEval 2015 task 12 [43], and the restaurants domain of SemEval 2016 task 5 [44]. Based on these five datasets, we performed data augmentation operations of varying scales. The aim was to balance the distribution of different sentiment polarities and avoid data inconsistencies, thereby improving the training effectiveness and robustness of the model. Through the algorithm 1, we obtained the augmented datasets, as shown in Table 2. Table 3 presents examples of the augmented data. Upon observation, we found that the quality of the generated data is high, to the extent that aspect terms can be replaced without contradicting the final sentiment polarity.

Table 2. The data distribution of the dataset.

Here, “⇒” represents data augmentation performed using ChatGPT.

Dataset Division Positive Neural Negative
Twitter Train 1507⇒3014 3016 1528⇒3056
Test 597 190 38
Lap14 Train 994⇒3976 464⇒3248 870⇒3480
Test 341 169 128
REST14 Train 2164⇒4328 637⇒3185 807⇒3228
Test 728 169 128
REST15 Train 1178⇒3534 50⇒3000 382⇒3056
Test 439 35 328
REST16 Train 1620⇒3240 88⇒3080 709⇒3545
Test 597 38 190

Table 3. Examples of the original text after data augmentation using ChatGPT is as follows.

We implement the calculation of cosine similarity using the sklearn package [45].

Original Sentence Text augmented via ChatGPT Cosine Similarity
The plot of this movie is excellent. The cinematography of this movie is excellent, capturing the plot perfectly. 0.758
The character development in this movie is excellent, enhancing the overall plot. 0.571
The pacing of this movie is excellent, keeping the plot engaging from start to finish. 0.619
The service at this restaurant is terrible, but the food is delicious. Despite the terrible service, the ambiance and decor of this restaurant create a pleasant dining experience. 0.323
While the service may be lacking, the presentation of the food is exceptional and adds to the overall dining experience. 0.384
Although the service leaves much to be desired, the extensive menu and the quality of the food make up for it. 0.286
The screen of this phone is clear, but the performance and battery life are disappointing. Despite the disappointing performance and battery life, the clear screen of this phone provides an enjoyable visual experience. 0.617
The screen clarity of this phone is commendable, but the disappointing performance and battery life hinder the overall user experience. 0.692
While the screen is clear and vibrant, the disappointing performance and short battery life limit the phone’s functionality. 0.630

In order to evaluate the suitability of the generated samples for ABSA tasks, we conducted an evaluation of the generated data based on the assessment rules outlined in section 3.3, which is shown in Table 1. Both the original samples and the generated samples were randomly selected and mixed in a shuffled manner. Five human experts were tasked with scoring each sample based on the specified criteria. The scoring results are presented in the Fig 8.

Fig 8. The statistical representation of the scoring results by human experts.

Fig 8

The experts (p1 to p5) are represented on the x-axis, while the y-axis represents the number of votes corresponding to each score level. Due to the significant disparity in the data distribution, we have applied the operation log(y) + 1 to the y-axis during plotting. However, the values indicated on the graph still represent the original y-values.

From Fig 8, it is evident that the majority of human experts considered these data to meet the requirements of ABSA. These 100 samples were a combination of real and generated data, further demonstrating the effectiveness of data augmentation using ChatGPT for training models in ABSA. Interestingly, certain experts assigned scores of -1, 0, and 2 to specific samples. We extracted these samples and compiled them in Table 4.

Table 4. Several examples with lower scores.

Sentence Score
The bookshelf sprouted wings and flew through the library, while the fish swam through the pages of a novel. -1
The pencil drew swirling galaxies, while the teacup recited mathematical equations in a hushed tone. 0
The plot of this novel is so complex, it’s as simple and understandable as a clear glass of water. 2
The quality of this product is truly indescribable, it’s like a leaping deer, full of stability and reliability. 2
The coffee mug floated through the sky, carrying dreams and memories on the wings of a butterfly. 2

Upon analyzing the original sentences, it becomes apparent that these sentences express abstract semantics or logic, making them less suitable for ABSA. However, they possess an artistic quality in terms of their sentence structure. After analysis and adjustment, we concluded that this discrepancy can be attributed to the “memory” function of ChatGPT, which requires further improvement. Generating the entire batch of texts at once can cause ChatGPT to “forget” the initial requirements. Therefore, it is advisable to generate the data in batches. Through our experiments, we found that controlling the number of generated texts to be within 20 entries or less yields better results.

4.2 Models for comparison

To evaluate the performance of our proposed CKG model, we compared it with other existing models:

  • SVM [46], a classical model used for addressing ABSA.

  • TD-LSTM [3], a model that utilizes an improved version of recurrent neural networks to extract text features.

  • MemNet [47], a model that explicitly considers contextual information and utilizes a memory network architecture.

  • ATAE-LSTM [48], a model that introduces an aspect-sentence attention mechanism to incorporate aspects and sentences into LSTM, thereby effectively considering the crucial information in the text.

  • IAN [4], a method that optimizes the attention mechanism by proposing an interactive attention mechanism, which takes into account the feature representation of aspect words and context, ensuring comprehensive consideration of both aspects in the text.

  • GCAE [49], a model based on convolutional neural networks and gating mechanisms, selectively outputs sentiment features based on given aspects or entities. The model structure is lightweight and straightforward.

  • ASGCN-DT [38], a model that constructs a directed graph based on dependency trees obtained from sentence analysis. It combines graph convolutional networks to extract syntactic features, thereby integrating semantic and dependency relationship analysis for sentiment information.

  • ASGCN-DG [38], a model that shares the same core approach as ASGCN-DT but differs in the construction of the graph. Instead of a directed graph, ASGCN-DG utilizes an undirected graph.

  • AOA [50], a model that introduces an attention-over-attention mechanism to learn the interaction between aspect words and sentences, focusing on the important parts of the sentence that have a significant impact on the overall sentiment.

  • TransCap [9], a model that introduces a transfer capsule network. It leverages aspect routing methods to encapsulate sentence-level semantic representations from aspect-level and document-level data into semantic capsules, thus enhancing the influence of aspect sentiment on semantics.

  • BERT [36], an application of the original BERT model in ABSA, utilizing positional encoding as the input for text.

  • AEN-BERT [51], a model that introduces label smoothing regularization and utilizes attention-based encoders.

  • MWGCN-BERT [52], a model that addresses the issue of long-distance dependencies by combining the LCG method to create a locally contextualized weighted adjacency graph.

  • EK-GCN [53], a model that introduces external knowledge to compensate for the inability of syntactic dependency trees to capture fine-grained labels. It also designs a word-sentence interaction network that fully considers aspect information.

  • Dual-GCN [54], an improved model that enhances the dependency parser and jointly considers both syntactic structure and semantic relevance.

  • T-GCN [55], a method that explicitly leverages dependency types in ABSA using a type-aware graph convolutional network. It utilizes attention to differentiate between different edges in the graph.

  • Sentic-GCN [23], a model that incorporates SenticNet to enhance the dependency graph, considering the dependency relationship between contextual words and aspect words, as well as the sentiment information between opinion words and aspect words.

  • KGAN [21], proposes a knowledge graph attention network that utilizes a hierarchical fusion module to capture sentiment feature representations from multiple perspectives.

4.3 Evaluation metrics

The evaluation metrics for the model include Accuracy (Acc) and macro-averaged F1-score. Accuracy is a commonly used evaluation measure, which represents the ratio of correctly predicted samples to the total number of samples. Macro-averaged F1-score (F1) is a performance metric for multi-class classification problems. It assesses the performance of the model for each individual class and calculates the average as an overall performance indicator for the model:

Acc=PtrueTotal (24)
F1=1Ni=1N(2×Pi×RiPi+Ri) (25)

4.4 Overall performance comparison

We present the comparison results of our proposed CKG model with some existing excellent ABSA models in Table 5. The data used in the table is sourced from the reported results in the original papers. Any missing data is indicated with a “-”. Based on the displayed results in the table, our model demonstrates excellent performance across all five datasets.

Table 5. Main experimental results of five datasets.

“Acc” represents accuracy, “F1” represents Macro-F1 score. The best results are shown in bold and second best underlined. The experimental results of other models are partly from the original paper and partly verified through reproducing the open-source code.

Model TWITTER LAP14 REST14 REST15 REST16
Acc F1 Acc F1 Acc F1 Acc F1 Acc F1
SVM 63.40 63.30 70.49 - 80.16 - - - - -
TD-LSTM 70.62 69.01 71.48 68.43 78.11 66.73 78.80 61.71 83.77 71.20
MemNet 68.52 66.71 72.34 64.25 79.13 66.41 80.12 67.46 82.58 67.32
ATAE-LSTM - - 68.70 63.93 77.20 67.02 78.48 60.53 83.77 61.71
IAN - - 72.05 67.38 79.26 70.09 78.54 52.65 84.74 55.21
GCAE - - 69.14 68.71 77.28 62.45 77.56 56.03 83.70 62.69
ASGCN-DT 71.53 69.68 74.14 69.24 80.86 72.19 79.34 60.78 88.69 66.64
ASGCN-DG 72.15 70.40 75.55 71.05 71.05 80.77 79.89 61.89 88.99 67.48
AOA - - 72.62 67.52 79.97 70.42 78.17 57.02 87.50 66.21
TransCap - - 73.87 70.10 79.29 70.85 - - - -
BERT - - 77.59 73.28 84.11 76.68 83.48 66.18 90.10 74.16
AEN-BERT - - 79.93 76.31 83.12 73.76 - - - -
MWGCN-BERT 75.00 74.30 79.78 76.68 86.612 80.18 85.61 73.32 82.05 79.21
EK-GCN 75.84 74.57 78.46 76.54 83.96 74.93 - - 89.36 69.32
Dual-GCN 76.37 75.44 82.59 79.34 88.47 82.92 81.73 65.05 89.29 68.08
T-GCN 78.03 77.31 81.97 78.71 87.41 82.23 86.00 72.81 92.97 80.07
Sentic-GCN - - 82.12 79.05 87.20 82.50 82.72 65.86 90.52 74.53
KGAN 79.97 79.39 82.66 78.98 87.15 82.05 86.21 74.20 92.34 81.31
CKG 77.95 77.20 83.41 79.41 89.24 83.31 88.92 76.42 92.87 81.95

Specifically, the traditional analytical methods exhibit relatively poorer performance. LSTM encoding methods improved with attention mechanisms show some enhancements, but they still fall short compared to models that incorporate GCN due to the absence of syntactic dependency knowledge. When we focus on the models that include GCN, we observe significant improvements in performance, particularly evident in the REST15 and REST16 datasets. Compared to non-BERT-based baseline models, our model performs exceptionally well, reaffirming the effectiveness of our approach. Introducing the BERT-base semantic encoding module leads to significant improvements in our model compared to other models. Experimental results show that our model performs competitively, except for a relatively lower performance on the TWITTER dataset, outperforming the other four datasets. Overall, in ABSA tasks, the CKG+BERT model demonstrates outstanding performance.

4.5 Ablation study

To validate the effectiveness of each component in our proposed CKG model, we conducted ablation experiments and present the results in Table 6. We can observe that the models incorporating the “BERT” encoding consistently outperform those using “LSTM”. Additionally, the models without attention or gate mechanisms exhibit unsatisfactory performance across all datasets, indicating the significant improvement in model performance with the inclusion of attention mechanisms and the gate-augmented GCN (“GAGCN”) for regular GCN. Furthermore, it is noteworthy that models with only attention mechanisms or only gate mechanisms perform worse than the models that incorporate both.

Table 6. Experimental results of ablation study on four datasets.

“a” represents the addition of attention mechanism in the regular graph convolution, while “ga” represents the incorporation of both gate and attention mechanisms in the regular graph convolution.

Model TWITTER LAP14 REST14 REST15 REST16
Acc F1 Acc F1 Acc F1 Acc F1 Acc F1
CKG LSTM w/o a 74.12 72.25 77.09 73.01 82.14 73.07 82.35 65.77 89.88 71.04
CKG LSTM w/o ga 73.69 71.01 73.81 69.51 80.83 70.03 81.10 60.45 88.74 67.36
CKG LSTM w/o a+ga 73.21 70.62 73.49 68.07 80.65 69.76 80.01 58.14 87.58 64.32
CKG LSTM 75.83 73.02 79.80 75.57 84.40 76.01 84.27 70.50 91.02 71.89
CKG BERT w/o a 75.77 73.88 81.37 76.12 87.13 79.78 82.51 71.73 91.58 71.88
CKG BERT w/o ga 76.61 74.76 80.04 75.64 87.96 81.97 88.32 75.64 91.72 75.62
CKG BERT w/o a+ga 76.27 74.40 80.33 75.33 87.28 79.41 86.87 70.31 90.08 74.71
CKG BERT 77.95 77.20 83.41 79.41 89.24 83.31 88.92 76.42 92.87 81.95

Based on the above analysis, we can conclude that our proposed enhancements effectively integrate semantic and syntactic features. The model, when combined with the sentiment knowledge base, demonstrates outstanding performance and merits attention in ABSA tasks.

4.6 Effect of the GCN layer number

The impact of the number of GCN layers on the focus of aspect features with varying degrees of local contextual information is often observed. We investigated the influence of the number of GCN layers on model performance, as shown in Fig 9. The evaluation was conducted using ACC and F1 scores, with the GCN layers ranging from 1 to 7. We tested the models on the five datasets using either the BiLSTM with attention mechanism as the semantic module or the BERT-base encoding. The graph convolution modules used in this study were based on the proposed GAGCN network.

Fig 9. The ACC and F1 values obtained by training and testing with different CKG layer settings after applying ChatGPT data augmentation on the five datasets (Twitter, LAP14, REST14, REST15, and REST16).

Fig 9

It is apparent that there exists a notable disparity in the outcomes when employing BERT-base as the semantic extraction module compared to the absence thereof. Under the same number of GCN layers, models using BERT-base as the encoding module outperform those using BiLSTM. When not using the BERT-base module, the model performs best with 1 to 2 GCN layers. One possible reason is that the dimensionality of the data encoded by BiLSTM is smaller, and GCN can learn the syntactic features with fewer layers. When BERT-base is used as the sentence encoding module, on the one hand, the model has a better understanding of semantics, and on the other hand, the input dimensionality of the syntax module is larger. Therefore, the model performs best with 3 to 4 GCN layers. One common observation for both encoding methods is that the model’s performance tends to decline when the number of layers exceeds a certain threshold, indicating overfitting due to the complexity of the model.

4.7 Attention visualization

By visualizing the attention weights, we can intuitively observe how the CKG model focuses on different sentiment words to extract specific aspect-level sentiment features. As shown in Fig 10, darker colors indicate higher attention weights given by the model to those words. From the upper subgraph of Fig 10, we can observe that a general attention model exhibits some level of attention towards the aspect words in the sentence. However, it also assigns weights to other irrelevant words. Moreover, when there are multiple aspect words in a single sentence, the model fails to distinguish them significantly. On the other hand, the lower subgraph demonstrates that the CKG model, after adjustment, exhibits clear attention towards each aspect word and the corresponding words that influence the sentiment polarity. It assigns less attention to other words. Based on the aforementioned observations, it can be deduced that the model proposed in this study exhibits formidable prowess in addressing Aspect-Based Sentiment Analysis (ABSA) and presents commendable interpretability.

Fig 10. The attention visualizations of typical samples.

Fig 10

4.8 Case study

Through Table 7, we can observe the model’s performance during actual predictions. Examples 1 to 3 are selected from the test set of the dataset, while examples 4 and 5 are comment-like statements generated by ChatGPT. By examining examples 1, 2, 4, and 5, we can see that our model accurately predicts the sentiment tendencies corresponding to different aspect words in sentences that contain multiple aspects. Additionally, our model demonstrates robustness by accurately predicting sentiment in the simulated comment-like statements that align with human conventions. Example 3 showcases our model’s ability to accurately predict sentiment for aspect words associated with neutral emotions, further highlighting its excellent performance.

Table 7. The scoring criteria for evaluating the applicability of text data generated by ChatGPT.

Sample Sentence Sentiment Predict
1 love the drinks, esp lycheemartini, and the food is also VERY good pos, pos, pos pos, pos, pos
2 i liked the atmosphere very much but the food was not worth the price pos, neg pos, neg
3 the decor is rustic, traditional japanese. neu neu
4 The hotelroom was spacious and clean, but the Wi-Fi connection was slow and unreliable. pos, neg pos, neg
5 The movie had breathtaking visual effects, but the plot was confusing and hard to follow. pos, neg pos, neg

5. Conclusion

In this paper, we use ChatGPT to enhance text data and we modify the network architecture to improve the model’s understanding of fine-grained emotions. Building upon the traditional “semantic & syntactic” analysis paradigm, we further improve it by introducing a cross-fusion mechanism that effectively integrates semantic and syntactic features. Additionally, we introduce a gated attention mechanism in the graph convolutional network to improve the performance of the syntactic feature extraction module and alleviate the potential issue of gradient explosion associated with multi-layer graph convolutions. We also incorporate an external knowledge base to enhance the characteristics of aspect words, ensuring that the model maintains its focus on aspect words. This enhancement significantly improves the accuracy and interpretability of the model’s predictions. In summary, our work makes considerable advancements in ABSA for neurorobotics. In the future, we will continue to explore the potential of ChatGPT, focusing on expanding and enriching large-scale datasets for ABSA. Furthermore, we plan to refine and simplify the model architecture while delving into the domain of multilingual analysis.

Data Availability

All relevant data are within the paper and its Supporting Information files. The following are the original dataset sources: LAP14 and REST14 [https://aclanthology.org/S14-2004], REST15 [https://aclanthology.org/S15-2082], REST16 [https://aclanthology.org/S16-1002], and X (Twitter) [https://aclanthology.org/P14-2009].

Funding Statement

The work has supported by the Shanghai Sailing Program [21YF1416700]. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

References

  • 1. Zhang W, Li X, Deng Y, Bing L, Lam W. A survey on aspect-based sentiment analysis: Tasks, methods, and challenges. IEEE Transactions on Knowledge and Data Engineering. 2022;. [Google Scholar]
  • 2.Li Z, Li X, Wei Y, Bing L, Zhang Y, Yang Q. Transferable End-to-End Aspect-based Sentiment Analysis with Selective Adversarial Learning. In: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP); 2019. p. 4590–4600.
  • 3.Tang D, Qin B, Feng X, Liu T. Effective LSTMs for Target-Dependent Sentiment Classification. In: Proceedings of COLING 2016, the 26th International Conference on Computational Linguistics: Technical Papers; 2016. p. 3298–3307.
  • 4.Ma D, Li S, Zhang X, Wang H. Interactive attention networks for aspect-level sentiment classification. In: Proceedings of the 26th International Joint Conference on Artificial Intelligence; 2017. p. 4068–4074.
  • 5. Zeng B, Yang H, Xu R, Wu Z, Han X. LCF: A Local Context Focus Mechanism for Aspect-Based Sentiment Classification. Applied Sciences. 2019;9(16):3389. doi: 10.3390/app9163389 [DOI] [Google Scholar]
  • 6.Wang B, Shen T, Long G, Zhou T, Chang Y. Eliminating Sentiment Bias for Aspect-Level Sentiment Classification with Unsupervised Opinion Extraction. In: Conference on Empirical Methods in Natural Language Processing. Association for Computational Linguistics; 2021.
  • 7.Xu H, Liu B, Shu L, Yu P. BERT Post-Training for Review Reading Comprehension and Aspect-based Sentiment Analysis. In: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. vol. 1; 2019.
  • 8. Su J, Tang J, Jiang H, Lu Z, Ge Y, Song L, et al. Enhanced aspect-based sentiment analysis models with progressive self-supervised attention learning. Artificial Intelligence. 2021;296:103477. doi: 10.1016/j.artint.2021.103477 [DOI] [Google Scholar]
  • 9.Chen Z, Qian T. Transfer capsule network for aspect level sentiment classification. In: Proceedings of the 57th annual meeting of the association for computational linguistics; 2019. p. 547–556.
  • 10. Lin Y, Meng Y, Sun X, Han Q, Kuang K, Li J, et al. BertGCN: Transductive Text Classification by Combining GNN and BERT. In: Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021; 2021. p. 1456–1462. [Google Scholar]
  • 11. Mikolov T, Sutskever I, Chen K, Corrado GS, Dean J. Distributed representations of words and phrases and their compositionality. Advances in neural information processing systems. 2013;26. [Google Scholar]
  • 12.Le Q, Mikolov T. Distributed representations of sentences and documents. In: International conference on machine learning. PMLR; 2014. p. 1188–1196.
  • 13.Pennington J, Socher R, Manning CD. Glove: Global vectors for word representation. In: Proceedings of the 2014 conference on empirical methods in natural language processing (EMNLP); 2014. p. 1532–1543.
  • 14. Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, et al. Attention is all you need. Advances in neural information processing systems. 2017;30. [Google Scholar]
  • 15. Xu Q, Zhu L, Dai T, Yan C. Aspect-based sentiment classification with multi-attention network. Neurocomputing. 2020;388:135–143. doi: 10.1016/j.neucom.2020.01.024 [DOI] [Google Scholar]
  • 16.Cheng J, Zhao S, Zhang J, et al. Aspect-level sentiment classification with heat (hierarchical attention) network[C]//Proceedings of the 2017 ACM on Conference on Information and Knowledge Management. 2017: 97–106.
  • 17.Kumar V S, Alemran A, Karras D A, et al. Natural language processing using graph neural network for text classification[C]//2022 international conference on knowledge engineering and communication systems (ICKES). IEEE, 2022: 1–5.
  • 18.Huang L, Sun X, Li S, et al. Syntax-aware graph attention network for aspect-level sentiment classification[C]//Proceedings of the 28th international conference on computational linguistics. 2020: 799–810.
  • 19.Zhang Y, Zhang Y. Tree communication models for sentiment analysis[C]//Proceedings of the 57th annual meeting of the association for computational linguistics. 2019: 3518–3527.
  • 20.Zhou Y, Liao L, Gao Y, et al. To be Closer: Learning to Link up Aspects with Opinions[C]//Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing. 2021: 3899–3909.
  • 21. Zhong Q, Ding L, Liu J, et al. Knowledge graph augmented network towards multiview representation learning for aspect-based sentiment analysis[J]. IEEE Transactions on knowledge and data engineering, 2023. [Google Scholar]
  • 22. Lu Q, Sun X, Sutcliffe R, et al. Sentiment interaction and multi-graph perception with graph convolutional networks for aspect-based sentiment analysis[J]. Knowledge-Based Systems, 2022, 256: 109840. doi: 10.1016/j.knosys.2022.109840 [DOI] [Google Scholar]
  • 23.Cambria E, Havasi C, Hussain A. Senticnet 2: A semantic and affective resource for opinion mining and sentiment analysis[C]//Twenty-Fifth international FLAIRS conference. 2012.
  • 24. Deng J, Lin Y. The Benefits and Challenges of ChatGPT: An Overview[J]. Frontiers in Computing and Intelligent Systems, 2022, 2(2): 81–83. doi: 10.54097/fcis.v2i2.4465 [DOI] [Google Scholar]
  • 25. George A S, George A S H. A Review of ChatGPT AI’s Impact on Several Business Sectors[J]. Partners Universal International Innovation Journal, 2023, 1(1): 9–23. [Google Scholar]
  • 26. Baidoo-Anu D, Owusu Ansah L. Education in the Era of Generative Artificial Intelligence (AI): Understanding the Potential Benefits of ChatGPT in Promoting Teaching and Learning[A]. Rochester, NY, 2023. [Google Scholar]
  • 27. Ray P P. ChatGPT: A comprehensive review on background, applications, key challenges, bias, ethics, limitations and future scope[J]. Internet of Things and Cyber-Physical Systems, 2023, 3: 121–154. doi: 10.1016/j.iotcps.2023.04.003 [DOI] [Google Scholar]
  • 28. Radford A, Narasimhan K, Salimans T, Sutskever I, et al. Improving language understanding by generative pre-training. 2018;. [Google Scholar]
  • 29. Radford A, Wu J, Child R, Luan D, Amodei D, Sutskever I, et al. Language models are unsupervised multitask learners. OpenAI blog. 2019;1(8):9. [Google Scholar]
  • 30. Brown T, Mann B, Ryder N, Subbiah M, Kaplan JD, Dhariwal P, et al. Language models are few-shot learners. Advances in neural information processing systems. 2020;33:1877–1901. [Google Scholar]
  • 31. Ouyang L, Wu J, Jiang X, Almeida D, Wainwright C, Mishkin P, et al. Training language models to follow instructions with human feedback. Advances in Neural Information Processing Systems. 2022;35:27730–27744. [Google Scholar]
  • 32.Belinkov Y, Bisk Y. Synthetic and Natural Noise Both Break Neural Machine Translation. In: International Conference on Learning Representations; 2018.
  • 33.Coulombe C. Text data augmentation made simple by leveraging nlp cloud apis. arXiv preprint arXiv:181204718. 2018;.
  • 34.Niu T, Bansal M. Adversarial Over-Sensitivity and Over-Stability Strategies for Dialogue Models. In: Proceedings of the 22nd Conference on Computational Natural Language Learning; 2018. p. 486–496.
  • 35.Dai H, Liu Z, Liao W, Huang X, Cao Y, Wu Z, et al. AugGPT: Leveraging ChatGPT for Text Data Augmentation. arXiv preprint arXiv:230213007. 2023;.
  • 36. Kenton J D M W C, Toutanova L K. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding[C]//Proceedings of NAACL-HLT. 2019: 4171–4186. [Google Scholar]
  • 37.Ma Y, Peng H, Cambria E. Targeted aspect-based sentiment analysis via embedding commonsense knowledge into an attentive LSTM. In: Proceedings of the AAAI conference on artificial intelligence. vol. 32; 2018.
  • 38.Zhang C, Li Q, Song D. Aspect-based Sentiment Classification with Aspect-specific Graph Convolutional Networks[C]//Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP). 2019: 4568–4578.
  • 39.Ryu S, Lim J, Hong SH, Kim WY. Deeply learning molecular structure-property relationships using attention-and gate-augmented graph convolutional network. arXiv preprint arXiv:180510988. 2018;.
  • 40.Dozat T, Manning C D. Deep biaffine attention for neural dependency parsing[J]. arXiv preprint arXiv:1611.01734, 2016.
  • 41.Li Dong, Furu Wei, Chuanqi Tan, Duyu Tang, Ming Zhou, and Ke Xu. 2014. Adaptive Recursive Neural Network for Target-dependent Twitter Sentiment Classification. In Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), pages 49–54, Baltimore, Maryland. Association for Computational Linguistics.
  • 42.Maria Pontiki, Dimitris Galanis, John Pavlopoulos, Harris Papageorgiou, Ion Androutsopoulos, and Suresh Manandhar. 2014. SemEval-2014 Task 4: Aspect Based Sentiment Analysis. In Proceedings of the 8th International Workshop on Semantic Evaluation (SemEval 2014), pages 27–35, Dublin, Ireland. Association for Computational Linguistics.
  • 43.Maria Pontiki, Dimitris Galanis, Haris Papageorgiou, Suresh Manandhar, and Ion Androutsopoulos. 2015. SemEval-2015 Task 12: Aspect Based Sentiment Analysis. In Proceedings of the 9th International Workshop on Semantic Evaluation (SemEval 2015), pages 486–495, Denver, Colorado. Association for Computational Linguistics.
  • 44.Maria Pontiki, Dimitris Galanis, Haris Papageorgiou, Ion Androutsopoulos, Suresh Manandhar, Mohammad AL-Smadi, et al. 2016. SemEval-2016 Task 5: Aspect Based Sentiment Analysis. In Proceedings of the 10th International Workshop on Semantic Evaluation (SemEval-2016), pages 19–30, San Diego, California. Association for Computational Linguistics.
  • 45. Virtanen P, Gommers R, Oliphant T E, et al. SciPy 1.0: fundamental algorithms for scientific computing in Python[J]. Nature methods, 2020, 17(3): 261–272. doi: 10.1038/s41592-020-0772-5 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 46. Kiritchenko S, Zhu X, Cherry C, et al. NRC-Canada-2014: Detecting Aspects and Sentiment in Customer Reviews[C]//Proceedings of the 8th International Workshop on Semantic Evaluation (SemEval 2014). Dublin, Ireland: Association for Computational Linguistics, 2014: 437–442. [Google Scholar]
  • 47.Tang D, Qin B, Liu T. Aspect Level Sentiment Classification with Deep Memory Network[C]//Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing. 2016: 214–224.
  • 48.Wang Y, Huang M, Zhu X, et al. Attention-based LSTM for Aspect-level Sentiment Classification[C]//Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing. Austin, Texas: Association for Computational Linguistics, 2016: 606–615.
  • 49.Xue W, Li T. Aspect Based Sentiment Analysis with Gated Convolutional Networks[C]//Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). Melbourne, Australia: Association for Computational Linguistics, 2018: 2514–2523.
  • 50. Huang B, Ou Y, Carley K M. Aspect Level Sentiment Classification with Attention-over-Attention Neural Networks[C]//Thomson R, Dancy C, Hyder A, et al. Social, Cultural, and Behavioral Modeling. Cham: Springer International Publishing, 2018: 197–206. [Google Scholar]
  • 51. Song Y, Wang J, Jiang T, et al. Attentional Encoder Network for Targeted Sentiment Classification[M]. 2019: 93–103. [Google Scholar]
  • 52. Yu B, Zhang S. A novel weight-oriented graph convolutional network for aspect-based sentiment analysis[J]. The Journal of Supercomputing, 2022. [Google Scholar]
  • 53. Gu T, Zhao H, He Z, et al. Integrating external knowledge into aspect-based sentiment analysis using graph neural network[J]. Knowledge-Based Systems, 2023, 259: 110025. doi: 10.1016/j.knosys.2022.110025 [DOI] [Google Scholar]
  • 54. Li R, Chen H, Feng F, et al. DualGCN: Exploring Syntactic and Semantic Information for Aspect-Based Sentiment Analysis[J]. IEEE Transactions on Neural Networks and Learning Systems, 2022: 1–15. [DOI] [PubMed] [Google Scholar]
  • 55.Tian Y, Chen G, Song Y. Aspect-based Sentiment Analysis with Type-aware Graph Convolutional Networks and Layer Ensemble[C]//Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. Online: Association for Computational Linguistics, 2021: 2910–2922.

Decision Letter 0

Toqir Rana

3 Nov 2023

PONE-D-23-24706Improving Aspect-Based Sentiment Analysis with Affective Knowledge Enhancements and Graph Adaptive Attention MechanismPLOS ONE

Dear Dr. Zhang,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Please submit your revised manuscript by Dec 18 2023 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Toqir Rana, Ph.D.

Academic Editor

PLOS ONE

Journal requirements:

When submitting your revision, we need you to address these additional requirements.

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

2. Please note that PLOS ONE has specific guidelines on code sharing for submissions in which author-generated code underpins the findings in the manuscript. In these cases, all author-generated code must be made available without restrictions upon publication of the work. Please review our guidelines at https://journals.plos.org/plosone/s/materials-and-software-sharing#loc-sharing-code and ensure that your code is shared in a way that follows best practice and facilitates reproducibility and reuse.

3. We note that you have stated that you will provide repository information for your data at acceptance. Should your manuscript be accepted for publication, we will hold it until you provide the relevant accession numbers or DOIs necessary to access your data. If you wish to make changes to your Data Availability statement, please describe these changes in your cover letter and we will update your Data Availability statement to reflect the information you provide.

4. Please include captions for your Supporting Information files at the end of your manuscript, and update any in-text citations to match accordingly. Please see our Supporting Information guidelines for more information: http://journals.plos.org/plosone/s/supporting-information.

5. We are unable to open your Supporting Information file [reference.bib]. Please kindly revise as necessary and re-upload.

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Partly

Reviewer #2: Yes

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: I Don't Know

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: No

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: The author proposed method for Improving Aspect-Based Sentiment Analysis with Affective Knowledge Enhancements and Graph Adaptive Attention Mechanism mainly used Graph Convolutional Neural Networks (GAN). The experimental analysis is not well explained, and the proposed method is worthy for investigation. The paper is lack of small issue which should be consider for improving the manuscript.

1. In Abstract section not proper clarification about Problem statement

2. Author mainly focused Dataset, features, and metrics in whole paper, so that it is missing Methods clarification properly, kindly re organized paper.

3. The author must focus on the small typos, uses of punctuation and English level throughout the manuscript.

4. I will suggest to author to use proper table for comparing techniques with some relevant parameters.

5. I will suggest author to use more exiting method for comparison of the proposed method to show the result.

6. Need to modify references also, most of the references are too old.

7. Author Should add Future scope of this Article.

Reviewer #2: 1. There are no results in the abstract. Abstract must contain the important results.

2. There are typos and other language issues in the text. For example, “food”. There must be opening and closing quotes, whereas it contains only closing quotes on both sides of the word food.

3. Line 173. It should be Figure 3, not just 3.

4. Line 199, “it contains n words.” This is wrong, it contains n + 1 words, since the subscript is starting from 0 not 1.

5. The formulas at line 232 have no citation. Their source must be cited in case the authors themselves have not come up with them.

6. Equation 9, there must be space between contains and dependency.

7. The conclusion of the study must be furnished with the concrete results.

8. A discussion section needs to be added in the study to compare and contrast your results with the published works.

9. What are the limitations of your study?

10. Future work needs to be added in this work.

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: Yes: ZULFIKAR ALI ANSARI

Reviewer #2: Yes: Dr Nadeem Iqbal

**********

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2024 Jun 27;19(6):e0301508. doi: 10.1371/journal.pone.0301508.r002

Author response to Decision Letter 0


7 Jan 2024

Dear Editor,

Thank you for your letter and the reviewers’ comments concerning our manuscript. We sincerely appreciate receiving your letter and the reviewers' comments regarding our manuscript titled "Improving Aspect-Based Sentiment Analysis with Affective Knowledge Enhancements and Graph Adaptive Attention Mechanism"(PONE-D-23-24706R1). This research paper is now entitled “CKG: Improving Aspect-Level Sentiment Analysis with Text Augmentation using ChatGPT and Knowledge-Enhanced Gated Attention Graph Convolutional Networks”. Your valuable comments have been instrumental in guiding us towards the revision and improvement of our paper. We have thoroughly examined each comment and made the necessary revisions accordingly. We have reconstructed the article and virtually rewritten the entire paper. By the way, we have deleted the “reference.bib” file.

We would like to express our gratitude once again for your patience and guidance throughout this process.

Best regards,

Yapeng Gao, Lin Zhang, Yangshuyi Xu

Reviewer's Responses to Questions

Dear Reviewer #1,

Thank you for reviewing our paper and providing valuable feedback. We appreciate your constructive comments and suggestions. We have almost completely rewritten the entire article. We have carefully considered each point raised and have made the following revisions to address your concerns:

1. We have rewritten the abstract section of the original manuscript, summarizing the issues present in the field of aspect-level sentiment analysis from previous research and our corresponding solutions in the revised Abstract section.

2. We have reorganized the paper to provide a more balanced explanation of the proposed method, giving appropriate attention to the methods used.

3. We have systematically reviewed and addressed the improper word usage and grammar errors throughout the article. It is worth mentioning that we have almost completely rewritten the entire paper.

4. In section 4.4 "Results," we have included Table 5, which compares the performance of our model with some other models in the field of aspect-level sentiment analysis. This table provides a comprehensive comparison of our model's performance. In section 4.5 "Ablation Study," we have presented Table 6, which visually presents the results of our model's ablation experiments. Additionally, we have provided explanations regarding the configuration of model-related layers and encoding choices.

5. To enhance the comparison of the proposed method, we have incorporated more exciting methods and presented the results accordingly. In the section 4.4 "Results" of the paper, we have compared our model with traditional and recent mainstream models, using commonly accepted evaluation metrics in the field of aspect-level sentiment analysis. We have analyzed the comparative results and provided insights based on these evaluation standards, which have achieved a consensus in the field. In the section 4.7 "Attention Visualization" and the section 4.8 "Case Study", we have further elaborated on the effectiveness of our model using a combination of images and textual explanations. These sections provide visual and descriptive evidence to support the efficacy of our model.

6. We have revised the references and included more recent sources to ensure the relevance and up-to-dateness of the cited literature.

7. In section 5 "Conclusion", we have supplemented the future research directions.

Once again, we thank you for your valuable feedback, which has significantly improved the quality of our paper. We hope that the revisions adequately address your concerns and improve the overall clarity and presentation of our work.

Dear Reviewer #2,

Thank you for your thorough review of our paper. We appreciate your valuable feedback and have carefully considered each of your points. We have almost completely rewritten the entire article. We have made the following revisions to address your concerns:

1. Based on your feedback, we have restructured the Abstract section, highlighting the existing shortcomings in the past research and our research focus. We have ensured a clear and concise correspondence between them.

2. We apologize for such a basic mistake. We have rectified this error and revised the article.

3. We have conducted a detailed review of the image labels throughout the entire document, ensuring that each image is numbered clearly and accurately.

4. You are absolutely right, and I apologize for the mistake. We have meticulously derived all the formulas in the document, ensuring their accuracy. The corrected formulas can be found on line 246.

5. We have provided the sources for all the formulas that were not originally proposed by us.

6. We have adjusted the expression of this formula (Eq3) and reviewed other text-containing formulas to ensure their accuracy.

7. We have furnished the conclusion with concrete results to provide a more comprehensive summary of our study.

8. Firstly, in the Results section (4.4), we have provided a demonstration of the effectiveness of our experimental results. Secondly, in sections 4.5 to 4.8, we have discussed and analyzed our ablation experiments and the improvement achieved by our model. We have also included examples of practical applications.

9. The usage limit of ChatGPT can indeed affect our research efficiency. As mentioned in the paper, the "expert review of text generation quality" required communication and incurred a certain time cost. Additionally, even with specified evaluation metrics, human subjectivity can still influence the final judgment. Therefore, we are actively considering an effective and quantifiable evaluation method. In the Conclusion section (5), we elaborate on this matter.

10. In the Conclusion section of the paper, we have provided an exposition on the future research directions.

We appreciate your feedback, which has significantly helped improve the quality and clarity of our paper. We hope that the revisions adequately address your concerns and enhance the overall contribution of our work. Thank you once again for your valuable input.

Attachment

Submitted filename: Response to Reviewers.docx

pone.0301508.s001.docx (18.4KB, docx)

Decision Letter 1

Toqir Rana

25 Jan 2024

PONE-D-23-24706R1CKG: Improving Aspect-Level Sentiment Analysis with Text Augmentation using ChatGPT and Knowledge-Enhanced Gated Attention Graph Convolutional NetworksPLOS ONE

Dear Dr. Zhang,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Please submit your revised manuscript by Mar 10 2024 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Toqir Rana, Ph.D.

Academic Editor

PLOS ONE

Journal Requirements:

Please review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the rebuttal letter that accompanies your revised manuscript. If you need to cite a retracted article, indicate the article’s retracted status in the References list and also include a citation and full reference for the retraction notice.

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #1: All comments have been addressed

Reviewer #2: (No Response)

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: Yes

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: Yes

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: All changes have been made. There is no need for further changes from my end. Now manuscript is OK to publish.

Reviewer #2: 1. Page 10. ABSA. This abbreviation has been used but I could not find its full form. It must be given in full form first time. Then its abbreviated version will be used. In the same way, other abbreviations must be treated.

2. There are language and other grammar issues in the manuscript which must be addressed.

3. Page 15. Equations must be cited if authors have not developed them.

4. Algorithm 1. “nd”. I think it is “and”

5. Line 285. “algorithm” must be changed with “Algorithm”. Since a specific algorithm is being referred to.

6. Line 306. Table1. It needs space.

7. Page 17. Equation 3. “D” should be made italic. All the variables being used must be italicized.

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: Yes: ZULFIKAR ALI ANSARI

Reviewer #2: Yes: Dr Nadeem Iqbal

**********

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2024 Jun 27;19(6):e0301508. doi: 10.1371/journal.pone.0301508.r004

Author response to Decision Letter 1


28 Feb 2024

Dear Editor,

Thank you for your letter and the reviewers’ comments concerning our manuscript. We sincerely appreciate receiving your letter and the reviewers' comments regarding our manuscript titled " CKG: Improving ABSA with Text Augmentation using ChatGPT and Knowledge-Enhanced Gated Attention Graph Convolutional Networks" (PONE-D-23-24706R1). Your valuable comments have been instrumental in guiding us towards the revision and improvement of our paper. We have thoroughly examined each comment and made the necessary revisions accordingly.

We would like to express our gratitude once again for your patience and guidance throughout this process.

Best regards,

Yapeng Gao, Lin Zhang, Yangshuyi Xu

Reviewer's Responses to Questions

Dear Reviewer #1,

Thank you for your recognition of our research. We are confident that with your meticulous academic approach, you will undoubtedly be able to generate further achievements.

Dear Reviewer #2,

Thank you once again for your valuable feedback. Your meticulous attention to detail has deeply touched us. Based on your suggestions, we have made the following revisions:

1. We have included an explanation of the abbreviation ABSA in the abstract, and replaced the existing content throughout the entire manuscript while also verifying other abbreviations.

2. We have made modifications to the grammar and expressions in the manuscript, and have indicated these changes in the "Revised Manuscript with Track Changes."

3. We have indicated the source of the equations and acknowledge your reminder.

4. Thank you for your attention to detail. We have made the necessary correction to the erroneous expression, replacing "nd" with "and".

5. We have corrected the initial term "algorithm" to "Algorithm" and reviewed other similar usage for any potential errors.

6. At the corresponding locations, we have populated the empty spaces.

7. Based on your feedback, we have made adjustments to all the formulas by converting them into italics format.

Thank you for acknowledging our research work, which motivates us to pursue further investigations. With your guidance, we believe this paper can yield more remarkable outcomes and reach a wider scholarly audience. Wishing you continued success in your endeavors.

Attachment

Submitted filename: Response to Reviewers_2.docx

pone.0301508.s002.docx (16.5KB, docx)

Decision Letter 2

Toqir Rana

18 Mar 2024

CKG: Improving Aspect-Level Sentiment Analysis with Text Augmentation using ChatGPT and Knowledge-Enhanced Gated Attention Graph Convolutional Networks

PONE-D-23-24706R2

Dear Dr. Zhang,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice will be generated when your article is formally accepted. Please note, if your institution has a publishing partnership with PLOS and your article meets the relevant criteria, all or part of your publication costs will be covered. Please make sure your user information is up-to-date by logging into Editorial Manager at http://www.editorialmanager.com/pone/ and clicking the ‘Update My Information' link at the top of the page. If you have any questions relating to publication charges, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Toqir Rana, Ph.D.

Academic Editor

PLOS ONE

Additional Editor Comments (optional):

Authors are requested to incorporate changes highlighted by the reviewer#2 before the final submission.  

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #2: All comments have been addressed

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #2: Yes

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #2: Yes

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #2: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #2: Yes

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #2: Following are my reviews:

1. Algorithm 1 inputs are D and Dn. but this algorithm is using many other data structures like POS. How are they accessible to this algorithm?

2. It would be nice if abstract is furnished with the important results, this study has found.

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #2: Yes: Dr Nadeem Iqbal

**********

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    Attachment

    Submitted filename: Response to Reviewers.docx

    pone.0301508.s001.docx (18.4KB, docx)
    Attachment

    Submitted filename: Response to Reviewers_2.docx

    pone.0301508.s002.docx (16.5KB, docx)

    Data Availability Statement

    All relevant data are within the paper and its Supporting Information files. The following are the original dataset sources: LAP14 and REST14 [https://aclanthology.org/S14-2004], REST15 [https://aclanthology.org/S15-2082], REST16 [https://aclanthology.org/S16-1002], and X (Twitter) [https://aclanthology.org/P14-2009].


    Articles from PLOS ONE are provided here courtesy of PLOS

    RESOURCES