Skip to main content
PLOS One logoLink to PLOS One
. 2020 Jul 15;15(7):e0235796. doi: 10.1371/journal.pone.0235796

A transition-based neural framework for Chinese information extraction

Wenzhi Huang 1,2, Junchi Zhang 2,*, Donghong Ji 1,*
Editor: Jie Zhang3
PMCID: PMC7363078  PMID: 32667950

Abstract

Chinese information extraction is traditionally performed in the process of word segmentation, entity recognition, relation extraction and event detection. This pipelined approach suffers from two limitations: 1) It is prone to introduce propagated errors from upstream tasks to subsequent applications; 2) Mutual benefits of cross-task dependencies are hard to be introduced in non-overlapping models. To address these two challenges, we propose a novel transition-based model that jointly performs entity recognition, relation extraction and event detection as a single task. In addition, we incorporate subword-level information into character sequence with the use of a hybrid lattice structure, removing the reliance of external word tokenizers. Results on standard ACE benchmarks show the benefits of the proposed joint model and lattice network, which gives the best result in the literature.

Introduction

The detection of entity mentions, relations and events are three fundamental tasks in information extraction, which can benefit many downstream tasks such as question answering [1, 2], reading comprehension [3, 4], and stock prediction [5, 6]. Intuitively, these three subtasks are closely correlated in the sense that entity mentions are core components connecting relations and events, while the extraction of relations and events can help improve the accuracy of entity results. For example, consider a Chinese text in Fig 1 which contains three entities “华视(Huashi)”, “大厅(hall)”, “影迷(fans)” and an event trigger “蜂拥而来(crowd)”. If an information extraction model aware of partially identified the event role types Destination between “大厅(hall)”影迷(fans)” and Aartifact between “影迷(fans)”-“蜂拥而来(crowd)”, this result could thus be served as a knowledge source to imply that there probably has a physical location relation (PHYS) from “影迷(fans)” to long-range entity “大厅(hall)”. In turn, the directed type PHYS could improve the confidence of a model to identify that “大厅(hall)” is the Destination and “影迷(fans)” is the Artifact, instead of the other way around. Therefore, it is beneficial to treat this output structure as a whole without leaving aside strong connections among subtasks.

Fig 1. Example sentence from ACE05 dataset.

Fig 1

Previous studies have showed that joint learnings of entities and relations [710], entities and events [1113] can lead to better extraction performance than pipelined methods [1417]. Due to joint learnings are effective at integrating interactive information between tasks and alleviating the problem of error propagation, there has been work inferencing all subtasks using one single model such as perceptron-based structural predictions [18], contextualized span representations [19], two-channel neural networks [20]. Despite the progress of existing efforts, they still follow a pipelined framework by first predicting entities and event triggers from texts, and then making assignments of relations and arguments to entity-entity pairs and entity-trigger pairs, respectively. To address this issue and perform all subtasks in an integrated graph, we propose a novel transition-based framework [2123] for Chinese information extraction. The transition-based methods tackle the complex joint search space with a state-transition process, which has been used for structured prediction tasks for NLP, including syntactic parsing [23], entity recognition [24], semantic role labeling [25]. In this paper, we construct the output graph in Fig 1 incrementally from left to right by designing eight transition actions for entity, relation and event recognitions. In this process, all actions can be executed alternately in an interleaving order, which is inspired by the human reading process [26].

Besides, modeling Chinese sentences is challenging compared to languages with explicit word boundaries (e.g. English) [27]. It is a dilemma to decide whether to segment sentences explicitly before involving downstream tasks or directly use the raw characters. The first comes to the consequences of inaccurate boundary cuttings that may bring in noises, whereas the latter is incapable of taking advantage of semantic features of word units, resulting in degradation of the overall performance. To overcome this difficulty, we make use of a lattice architecture [28] that simultaneously consider both character and word features, leveraging the gate controllers in Long Short Term Memory networks (LSTM) [29] to automatically select the most useful sources for the information extraction subtasks. Different from [28] where off-the-shelf word tokenizers are used, we obtain the most frequent words by using a byte-pair encoding [30] to circumvent segmentation errors.

In addition to the need of encoding words and characters simultaneously, other characteristics of Chinese relation extraction disallow existing transition systems to be applied to our task directly. For example, more than 50% of our Chinese information extraction dataset contains relations or roles between nested named entities and triggers which no existing work deals with. We design a novel transition system that handles such situations.

On standard ACE 2005 benchmarks, our model achieves F1-scores of 81.7%, 50.8%, 68.8% on named entity recognition, relation extraction and event trigger detection, respectively, which are the best in the literature. The main contributions of this paper can be summarized as follows:

  1. We propose a novel transition-based system for extracting nested entities, relations, and events in a unified network.

  2. We empirically show that lattice LSTM with byte-pair encoding is useful for Chinese information extraction.

  3. Experimental results demonstrate that our model significantly outperforms previous methods of using additional knowledge and the state-of-the-art neural models.

Task definition

According to the Automatic Content Extraction (ACE) 2005 evaluation program, we briefly present the terminologies that relevant to entity relation extraction and event detection. There are four terms involved:

  1. Entity mentions: An entity is an object or set of objects that belong to a semantic category, such as Person, Location. An entity mention is a reference to an entity, usually a noun phrase (NP). We consider the standard PER, ORG, GPE, LOC, FAC, VEH, WEA entity types plus ACE VALUE and TIME expressions as in [11].

  2. Semantic relation: An particular interests relation that is held between two entity mentions. For example, there is a PART-WHOLE relation between “华视(Huashi)” and “大厅(hall)”. ACE2005 defines 6 main relation types which are PHYS, PER-SOC, PART-WHOLE, ORG-AFF, ART and GEN-AFF.

  3. Event Trigger: The keywords that most clearly expresses the occurrence of an event, so that identifying events is equivalent to recognizing trigger words. Event triggers can be verbs, nominalizations, and occasionally adjectives, for instance, the word “蜂拥而来(crowd)” triggers a Transport event in Fig 1. Following previous studies [31, 32], we treat 33 event subtypes as separated categories and ignore their hierarchical structure.

  4. Event Argument: Event arguments are entities that participate as specific roles in the event mention. An argument reflects attributes that an entity carried in an event generally indicates place and time, and specifically have certain values (e.g. CRIME, ARTIFACT). We collapse 8 time-related types into one as in [12], which results in a total of 29 role subtypes.

Formally, given an input sentence represented as a sequence of tokens C = c1, c2, …cn, we concentrate on two tasks:

  1. Relation extraction (RE) task involves predicting a set of entity mentions E and a set of semantic relations R between recognized entity pairs. Note that there may exist nested entities, such as “香港展览中心(Hong Kong exhibition center)” and “香港(Hong Kong)” share common characters prefix. Each relation rR can be represented as a tuple (es, eo, lso), where es and eo refer to a subject entity and object entity, respectively. lso is a semantic relation category assigned to es and eo. An extra None label is added to represent no relation holds for the entity pair.

  2. Event detection (ED) task involves detecting event trigger sets T and event arguments A. In particular, each token wi will be distinguished to be a true or pseudo trigger mention (i.e. Trigger Identification) and will be further assigned to an event type label ti if wi is a positive trigger mention (i.e. Trigger Type Detection). Then, for each trigger ti, participant arguments are assigned to this event trigger by predicting an argument role aij for all candidate entity mentions ejE in the same sentence (i.e. Argument Role Classification). Depart from most prior work on event extraction, we consider the realistic setting where gold entity labels are not available. Instead, we prepare argument candidates using predicted entity mentions from RE task.

Methods

Transition system

The transition-based framework is an effective algorithm that builds structure output incrementally in a step by step process. In particular, a transition system has two key components: (1) transition states and (2) a set of transition actions. In this work, we define each transition state as a tuple s = (σ, δ, λ, e, β, E, R), where σ is a stack storing processed elements (an element can be either an entity or an event trigger), δ is a stack containing elements that are temporarily popped out of σ but will be pushed back later, e is a stack holding words popped off β, λ is a variable storing the current element recognized by e, and β is a buffer containing unprocessed tokens. R is a set of relations or argument role triples that have been extracted. E is a set of elements that have been recognized. A is a stack storing action histories.

On the other hand, transition actions are used to control how a transition state advance by one step, which should be well designed to ensure that entities, relations and events are extracted in an proper order and covers all possible output graphs. To meet this purpose, we develop eight types of transition actions, including LEFT-*, RIGHT-*, SHIFT, DUAL-SHIFT, DELETE, ELEMENT-SHIFT, ELEMENT-GEN, ELEMENT-BACK. The first four actions are used to extract relations and argument roles, while the last four actions are utilized to recognize named entities. We summarize transition actions as follows:

  • LEFT-* pops the top element ej from σ and push it onto δ. It also add a relation or argument role type l between λ(ei) and σ(ej), assuming the directed edge is from ei to ej.

  • RIGHT-* pops the top element ej from σ and push it onto δ. It also add a relation or argument role type l between element λ(ei) and σ(ej) which is similar to LEFT-*, but the direction is opposite from ej to ei.

  • SHIFT pops all the elements from δ back to σ and move element ei in λ to the first element of σ.

  • DUAL-SHIFT is similar to DUAL-SHIFT but additionally copies the element words in λ and pushes it onto β, in order to handle situations where a word is a trigger and also the first word of an entity.

  • DELETE simply removes the left-most word off buffer β.

  • ELEMENT-SHIFT pops the front word off β and moves it to the front of e

  • ELEMENT-GEN summarizes all the words in e and generates an entity label or an event trigger label.

  • ELEMENT-BACK moves all but the last word from e back to β. It is designed to tackle overlapping entities and triggers.

To ensure a valid output, each action needs to satisfy certain preconditions. For example, actions that extract relations or argument roles can only be conducted after the action ELEMENT-GEN is performed. Table 1 shows the full list of preconditions.

Table 1. Preconditions of transition actions.

Transitions Preconditions of transition actions
LEFT-* (λ ≠ ψ)∧(σ ≠ [])∧(jT)∧(iE)
RIGHT-* (λ ≠ ψ)∧(σ ≠ [])∧(jE)∧(iT)
SHIFT (λ ≠ ψ)∧(σ = [])
DUAL-SHIFT (λ ≠ ψ)∧(σ = [])∧(jT)
DELETE (∃jβ)∧(e = [])
ELEMENT-SHIFT (λ = ψ)∧(∃jβ)
ELEMENT-GEN (λ = ψ)∧(e ≠ [])∧(jE)
ELEMENT-BACK (λ = ψ)∧(e ≠ [])∧(jE)

Neural transition-based model

The network architecture of our proposed transition method is shown in Fig 2. For a given Chinese text, we first utilize lattice-lstm [28] to encode unigram and bigram characters. Then these hidden representations are fed into a decoder layer to generate a transition state with the guidance of the previous step transition action. At last, all structure features are summarized to predict possible actions for the next transition state.

Fig 2. Our encoder-decoder model for joint entities, relations and events extraction.

Fig 2

Embedding encoder layer

As presented in Fig 1, for each input Chinese character ci, we represent it as a concatenation of the corresponding character unigram embeddings and character bigram embeddings:

xi=cici,i+1 (1)

where ⊕ represents concatenate operation. The unigram embeddings are initialized by taking the last layer output of the pre-trained bi-directional language models, termed as BERT [33]. Specially, we use the Chinese version of the base BERT parameters, which is provided by huggingface project(https://huggingface.co/bert-base-chinese).

In addition to character-level features, we also build word-level segments by adopting the Byte Pair Encoding (BPE) [34] algorithm to encode the subword information. The original idea of BPE is to iteratively compress data by merging the most frequent pair of bytes in a sequence as a new byte. Here, we construct the most frequent subwords using Chinese Gigaword corpus (https://catalog.ldc.upenn.edu/LDC2003T09). The subword vector wb,e that starts from index b to index e is initialized with pre-trained word2vec method [35].

Lattice LSTM encoder layer

We represent input sentences with a lattice structure balancing word and character features dynamically with the gate mechanism [28]. At each step, lattice lstm takes a word embedding wb,e and a character embedding xi as input, first calculate three types of word gates and the cell state as follows:

[ib,efb,ec˜b,e]=[σσtanh](Ws[wb,ehb]+bs)cb,e=fb,ecbc+ib,ec˜b,e (2)

where hb is hidden vector started at b. σ is Sigmoid function. Ws and bs are model parameters. cb,e is the memory cell of the shortcut path starting from character cb to character ce.

An additional gate is used to obtain the cell state of the character ci:

ib,i=σ(Wg[xicb,i]+bg) (3)

Then input gate ib,i and ii should be normalized to ensure their sum equals to 1:

αb,i=exp(ib,i)exp(ii)+ib,iIiexp(ib,i)αi=exp(ii)exp(ii)+ib,iIiexp(ib,i) (4)

The final forward lattice LSTM representation hi of character ci is calculated as:

[oific˜i]=[σσtanh](W[xihi-1]+b)ii=1-fiPci=cb,iCiαb,icb,i+αic˜ihi=oitanh(ci) (5)

where W and b are the model parameters. Due to Eq 5 has a more complex memory calculation step compared to traditional LSTM [29], which can take account of all the matched input subwords. Therefore, Lattice lstm is more suitable for Chinese texts to integrate character and word semantics simultaneously.

Finally, to obtain a bi-directional representation hi=hihi for character ci, the backward lattice LSTM vector hi is computed using Eqs 25 in the reversed order of input sequence.

Transition state representation

Given transition State s = (σ, δ, λ, e, β, E, R), we encode each component as follows:

  1. For unprocessd characters C = ci, c2, …cn, the buffer β is first represented by pushing all lattice LSTM outputs H = h1, h2, …hn onto a vanilla stack the reversed order.

  2. For σ, δ, e and A, which involve popping out top elements, we use Stack LSTM [23] for encoding. In the initial state, the stacks σ, δ, e and A all has an zero-vector indicating an empty-stack. The calculation process of all stacks are similar, we thus take σ as an example:
    σi=StackLSTM[e1,e2,...em] (6)
    where ei denotes the representation of the entities recognized at time step i. Note that an element ei can be either an entity or a event trigger.

Action decoder layer

For a given sentence, we find the best transition sequence by taking the action with maximum probability at every step. Each action is decided according to the current state encoding. Denote the feature vector for the transition state at t as gt, which consists of the current σ, δ, e, A, λ, β, as shown in Fig 2. We have representation of the state as:

gt=[st;dt;et;at;λt;ht] (7)

A feed-forward neural network with tanh activation function is adopted to transform feature gt to action prediction space:

mt=tanh(Wmgt+bm) (8)
p(zt|mt)=exp(uztmt+bz)zßν(S,A)exp(uzßmt+bz) (9)

where Wm, uz and bm, bz are trainable parameters. The set ν(S, A) represents the set of valid candidate actions. zt is the current predicted transition action.

For network training, we minimize the negative log-likehood of the corresponding gold action:

L(θ)=-1Ttlogp(zt|mt;θ) (10)

where T is the size of the gold action sequence.

Experiments

Experimental settings

Dataset

We perform experiments on the publicly available dataset: ACE 2005 Chinese corpus (https://catalog.ldc.upenn.edu/ldc2006t06). The dataset contains 633 documents, which were collected from Newswires (NW), Broadcast News (BN), Weblog (WL). Due to that there is relatively little study on joint all subtasks for Chinese, we randomly split the ACE 2005 dataset into 8:1:1, with 507 documents for training, 63 documents for validation and 63 documents for testing. In addition, for named entities, we take nested entities that contain over-lapping spans into consideration. For relations and argument roles, we use a “None” type to represent there is no semantic connections hold between entities and event triggers. Our code and sample test results are available (https://github.com/zjcalva/chinese_information_extraction).

Evaluation metrics

We use Precision (P), Recall (R) and F-Measure (F1) on both entities, relations and events. An entity is considered correct if we can identify its head and the entity type correctly. A relation instance is regarded as correct when its relation type and the head offsets of two corresponding entities and entity types are both correct. An event trigger is considered correct if both its offset and event subtype is correct. An argument role is regarded as correct when argument offset, role type and the corresponding triggers are both correct.

Pre-processing

To represent input Chinese sentences, we use the same character unigram embeddings, bigram embeddings and word embeddings with [36], which pretrain those embeddings using word2vec [35] on Chinese Gigaword corpus. The vocabulary of subword is constructed with 150000 merge operations in byte-pair encoding. To accelerate the subwords construction process of lattice network inputs, trie structure [37] is used. Note that all the embeddings are fine-tuned during training.

Hyper-parameter settings

All the hyper-parameters are tuned by selecting the best model with early stopping using the evaluation results on the validation set. Specifically, we set the embedding sizes of character unigram,bigram and subword all to 50 dimensions. Dropout [38] is set to 0.33 on both the character input and the subword input to prevent overfitting. Adam [39] is applied to optimize the model parameters, with an initial learning rate of 0.015 and a decay rate of 0.05. The batch size and the lattice lstm hidden dimension are set to 32 and 150, respectively.

Development results

To examine the influence of several key model components, we report entity and relation extraction results on the ACE2005 validation set.

The influence of lattice LSTM

Table 2 shows the influence of the encoder layer by comparing the following models:

Table 2. The influence of lattice LSTM results on the ACE2005 test set.
Models Entity Recognition Relation Classification
P R F1 P R F1
Transition 75.3 71.0 73.1 55.7 39.0 45.3
Transition—BERT 74.5 68.6 71.5 44.8 39.8 42.2
Transition + Bigram 78.1 70.7 74.2 58.0 41.4 45.5
Transition + Lattice (Word) 83.5 79.2 81.3 55.3 48.2 49.9
Transition + Lattice (Subword) 84.2 79.3 81.7 56.5 48.8 50.8
  1. Transition: using a standard LSTM to encode the character embedding of the input sentence, in which BERT embeddings are incorporated.

  2. Transition-BERT: BERT embeddings are removed from Transition to see the effect of the pre-trained bidirectional language model.

  3. Transition+Bigram: in addition to unigram raw characters, this model also leverages adjacent characters in a sentence to form bigram features, such as “蜂(bee)” can be represented as “蜂拥(swarming)”.

  4. Transition+Lattice (word): the proposed hybrid model that integrates characters and words in a lattice networkï¼?where Chinese words are obtained by applying Jieba tokenizer.

  5. Transition+Lattice (subword): similar to Transition+Lattice (word), but subwords are built by the byte-pair encoding algorithm.

With raw unigram input, baseline Transition yeilds 73.1% and 45.3% F1-scores on entities and relations, respectively. The F1-scores are decreased to 71.5% and 42.2% by removing BERT embeddings to the baseline model, indicating the effectiveness of the bi-directional language model pre-trained on a large amount of plain texts. Compared with the only unigram features, Transition+Bigram gives slightly higher F1-scores, which are 74.2% on entities and 45.5% on relations, respectively. The results indicate that bigram information will promote the performance of the baseline model, but the increase in F1-score is not significant.

Finally, although we take multiple strategies to combine character features, we can see that lattice lstm network gives the most high performance on both entity and relation extractions. In particular, subword-based lattice model achieves 81.7% F1-scores on entities and 50.8% on relations, respectively. The results reveal two conclusions: 1) Lattice lstm is effective to integrate character and word sequence information in a hybrid framework; 2) Lattice+Subword works better than Lattice+Word indicating the usefulness of allevating segmentation errors. With above observations, we take subword-based lattice lstm model for subsequent experiments.

Joint model vs pipeline model

To evaluate the advantages of the joint model (Lattice-transition-joint), we construct a pipelined version (Lattice-transition-pipeline) by first performing entity recognition with relation and event-related transition actions removed from our joint model including “LEFT-*”, “RIGHT-*”, “SHIFT”, “DUAL-SHIFT”, then modify some actions for detecting relations and events in the sentence independently.

Table 3 lists the results on the ACE2005 test set. The first observation is that the pipelined model is a strong baseline, which gives the F1-score of 79.6% on entity recognition and 47.2% on relation extraction. However, by treating entities, relations, and events as a single task, the joint transition model outperforms the piplined model by 2.1% on entities and 3.6% on relations. We attribute these improvements as twofold: 1) Modeling the dependencies of entities and the semantically related tasks are effective for reducing error propagation. 2) Our joint model is able to capture label correlations between tasks not only at the encoding part but also at the decoding stage, which is, in turn, also beneficial for entity recognition.

Table 3. Joint model and pipeline model results on the ACE2005 test set.
Models Entity Recognition Relation Classification
P R F1 P R F1
Lattice-transition-pipeline 81.0 78.3 79.6 44.2 51.4 47.2
Lattice-transition-joint 84.2 79.3 81.7 56.5 48.8 50.8

Comparison results

Baselines To demonstrate the effectiveness of various kind of models, we implement several advanced approaches as baselines for comparison:

  • Word-Tree-Structure [40] is a word-based joint learning model, where word segments are obtained by applying Jieba tokenizer (https://github.com/fxsjy/jieba). In particular, a bidirectional LSTM network is leveraged as the shared encoder. We first extract entities and event triggers as two sequence labeling tasks with BILOU output scheme. Then we construct the tree structure of input word sequences with Stanford dependency parsing (https://nlp.stanford.edu/software/stanford-dependencies.shtml) to obtain shortest dependency paths between entities and event triggers. Finally, recognized element type embeddings and bi-lstm vectors along the shortest dependency path are concatenated as input for relation and argument role classifications.

  • Char-BERT-pipeline [33] learns extractions of entities, relations and events in a pipelined process. It employs pre-trained Chinese BERT-base embeddings as a contextualized character encoder which is fine-tuned during model training. To predict task labels, we simply add a linear transformation layer on top of individual BERT outputs and use softmax function to normalize label vectors. It has been shown that BERT is a powerful representation method, which contains hierarchical lexical, syntactic and semantic knowledges [41]. Hence, we believe it is a strong baseline for comparison.

Analysis Table 4 shows the results of different types of approaches. The first observation is that despite the end-to-end learning process of Word-Tree-Structure, it performs relatively poor on both entity recognition (72.6%) and relation extraction (39.3%). One possible reason is that it copes with subtasks only on word-level without taking advantage of character-level information. On the other hand, it suffers from noise parsing results incorporated by off-the-shelf word segmenters and dependency parsers, which is prone to affect downstream applications.

Table 4. Entity and relation results compared to previous systems on the ACE2005 test set.

Models Entity Recognition Relation Classification
P R F1 P R F1
Word-Tree-Structure 79.0 67.2 72.6 43.6 35.7 39.3
Char-BERT-pipeline 77.6 73.1 75.3 45.0 42.3 43.6
Lattice-transition-joint 84.2 79.3 81.7 56.5 48.8 50.8

Second, Char-BERT-pipeline shows a noticeable improvements (2.7% on entities and 4.3% on relations) compared to Word-Tree-Structure even with a separated extraction process. This result demonstrates that hierarchical information contained implicitly in the pre-trained deep language model is a key factor to attain a feasible performance.

The final observation is that our model lattice-transition-joint significantly outperforms the existing methods on both two tasks. In particular, for entity recognition, our model achieves 9.1% improvements compared to [40] and 6.4% improvements compared to BERT-based model [33]. This result indicates the importance to exploit long-distance and cross-task dependencies between entities, relations and events. By allowing entity information propagate through transition states, our model reaches the best 50.8% F1-scores on relation extraction, demonstrating it is beneficial to use the partial identified graph as a source of features for more informed constructions of output structures.

Event extraction results

In addition to extracting entities and relations, our transition methods also extract event triggers and argument roles. We compare their results with our pipeline method as well as previous work. As it can be observed from the Table 5, by incorporating the relation information generated in state-transition process, Lattice-transition-joint can improve the performance than the pipelined method Lattice-transition-pipeline by 1.8% and 5.3% on event trigger detection and argument role classification, respectively. This demonstrates that relation attributes truly reinforce the event recognition. And incremental joint learning is capable of delivering the features of relations to the event modeling process. Additionally, our joint model significantly outperforms Word-Tree-Structure and Char-BERT-pipeline with a performance gain of about 3% F1-scores on event triggers and 7% F1-scores on argument roles. We attribute the failure of previous approaches to error propagation from the upstream task and the incapacity of handing nested entities.

Table 5. Event trigger and argument role results on ACE2005 test set.

Models Event Trigger Detection Argument Role Classification
P R F1 P R F1
Word-Tree-Structure 67.7 58.3 62.6 49.2 39.8 44.5
Char-BERT-pipeline 62.3 68.9 65.4 48.5 51.5 50.0
Lattice-transition-pipeline 62.3 72.2 66.9 52.8 52.0 52.4
Lattice-transition-joint 65.5 72.5 68.8 56.2 59.2 57.7

Related work

Entity recognition [28, 42], relation extraction [14, 43] and event detection [44, 45] are fundamental tasks in NLP, which have drawn much attention in recent years. For English, joint methods include integer linear programming models [46, 47], feature-based structured learning models [7, 48] and neural network models [13, 40, 49]. Methods considering all subtask simultaneously includes contextualized span representations [19], interactive two-channel neural networks [20]. The most closed work to ours is [18], which jointly decodes all subtasks by designing two types of decoding actions in the structured perceptron algorithm. However, they heavily rely on manually designed indicator features struggling to capture sufficient discriminative information.

For Chinese, there have been word-based [50] and character-based [28, 51] models for Chinese named entity recognition. Peng and Dredze [52] propose an integrated model to process multi-task learning that allows for joint training learned representations. By integrating latent word information into character-based LSTM-CRF, Zhang and Yang [28] propose lattice LSTM representation for mixed characters and lexicon words. For relation and event extraction, there have been kernel-based methods [8, 43], feature-based methods [8, 31, 32] and neural network methods [5355]. All of the models require segmentation first and thus suffer from the potential issue of error propagation. To our knowledge, we are the first to design a neural network to jointly recognize Chinese named entities, relations and events without the need of word segmentation.

Conclusion

We proposed a transition-based framework that treats all Chinese subtasks of entity recognition, relation extraction and event detection as a single task without leaving aside mutual benefits among tasks. In particular, we designed eight transition actions to ensure all subtasks are identified in a proper order. In addition, to unburden the need of word segmentation before information extraction, we adopt a lattice network to integrate subword features into character-level sequences. Experimental results on the standard ACE2005 benchmark show that our model achieves superior performance over both word-based joint learning method and character-based pipelined BERT network, reaching a state-of-the-art result on Chinese information extraction.

Acknowledgments

We would like to thank the anonymous reviewers for their many valuable comments and suggestions.

Data Availability

ACE2005 data used in our paper is owned by Linguistic Data Consortium, The Trustees of the University of Pennsylvania(https://www.ldc.upenn.edu/). Please refer to https://catalog.ldc.upenn.edu/LDC2006T06 for the data details. To facilitate following researches, we provide our code and sample data at https://github.com/zjcalva/chinese_information_extraction, which is added in the paper (Section Experiments). e have provide sample dataset (about 3% of full dataset) in this URL. To fully reproduce our study, readers need to replace "test_sample.txt" with full dataset obtained from LDC. The authors confirm that they had no additional access privilege, and that they accessed the LDC data in the same manner described above.

Funding Statement

Science Foundation of China (No. 61772378), the Social Science Foundation of Ministry of Education of China (No. 18JZD015), the Natural Science Foundation of Hubei Province (No. 2012FFA088), and the National Key Research and Development Program of China (No. 2017YFC1200500). The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

References

  • 1.Lin Y, Ji H, Liu Z, Sun M. Denoising distantly supervised open-domain question answering. In: Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers); 2018. p. 1736–1745.
  • 2.Wang W, Yang N, Wei F, Chang B, Zhou M. Gated self-matching networks for reading comprehension and question answering. In: Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers); 2017. p. 189–198.
  • 3.Hoang L, Wiseman S, Rush AM. Entity Tracking Improves Cloze-style Reading Comprehension. In: Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing; 2018. p. 1049–1055.
  • 4.Hu M, Peng Y, Wei F, Huang Z, Li D, Yang N, et al. Attention-Guided Answer Distillation for Machine Reading Comprehension. In: Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing; 2018. p. 2077–2086.
  • 5.Makrehchi M, Shah S, Liao W. Stock prediction using event-based sentiment analysis. In: 2013 IEEE/WIC/ACM International Joint Conferences on Web Intelligence (WI) and Intelligent Agent Technologies (IAT). vol. 1. IEEE; 2013. p. 337–342.
  • 6. Ding X, Zhang Y, Liu T, Duan J. Using structured events to predict stock price movement: An empirical investigation. In: EMNLP; 2014. [Google Scholar]
  • 7.Miwa M, Sasaki Y. Modeling joint entity and relation extraction with table representation. In: Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP); 2014. p. 1858–1869.
  • 8.Zhang J, Ouyang Y, Li W, Hou Y. A novel composite kernel approach to Chinese entity relation extraction. In: International Conference on Computer Processing of Oriental Languages. Springer; 2009. p. 236–247.
  • 9. Zheng S, Hao Y, Lu D, Bao H, Xu J, Hao H, et al. Joint entity and relation extraction based on a hybrid neural network. Neurocomputing. 2017;257:59–66. 10.1016/j.neucom.2016.12.075 [DOI] [Google Scholar]
  • 10. Zhang J, Zhang Y, Ji D, Liu M. Multi-task and multi-view training for end-to-end relation extraction. Neurocomputing. 2019;364:245–253. 10.1016/j.neucom.2019.06.087 [DOI] [Google Scholar]
  • 11. Yang B, Mitchell T. Joint Extraction of Events and Entities within a Document Context. In: NAACL-HLT; 2016. [Google Scholar]
  • 12. Nguyen TM, Nguyen TH. One for All: Neural Joint Modeling of Entities and Events. In: AAAI; 2019. [Google Scholar]
  • 13.Zhang J, Qin Y, Zhang Y, Liu M, Ji D. Extracting entities and events as a single task using a transition-based neural model. In: Proceedings of the 28th International Joint Conference on Artificial Intelligence. AAAI Press; 2019. p. 5422–5428.
  • 14.GuoDong Z, Jian S, Jie Z, Min Z. Exploring various knowledge in relation extraction. In: Proceedings of the 43rd annual meeting on association for computational linguistics. Association for Computational Linguistics; 2005. p. 427–434.
  • 15.Grishman R, Westbrook D, Meyers A. NYU’s English ACE 2005 system description. ACE. 2005;.
  • 16. McClosky D, Surdeanu M, Manning CD. Event extraction as dependency parsing. In: ACL; 2011. [Google Scholar]
  • 17. Phi VT, Santoso J, Tran VH, Shindo H, Shimbo M, Matsumoto Y. Distant Supervision for Relation Extraction via Piecewise Attention and Bag-Level Contextual Inference. IEEE Access. 2019;7:103570–103582. 10.1109/ACCESS.2019.2932041 [DOI] [Google Scholar]
  • 18. Li Q, Ji H, Yu H, Li S. Constructing information networks using one single model. In: EMNLP; 2014. [Google Scholar]
  • 19.Wadden D, Wennberg U, Luan Y, Hajishirzi H. Entity, Relation, and Event Extraction with Contextualized Span Representations. In: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP); 2019. p. 5788–5793.
  • 20. Zhang J, Hong Y, Zhou W, Yao J, Zhang M. Interactive learning for joint event and relation extraction. International Journal of Machine Learning and Cybernetics. 2020;11(2):449–461. 10.1007/s13042-019-00985-8 [DOI] [Google Scholar]
  • 21. Nivre J. Algorithms for deterministic incremental dependency parsing Computational Linguistics. 2008;. [Google Scholar]
  • 22. Zhang Y, Clark S. Syntactic processing using the generalized perceptron and beam search. Computational linguistics. 2011;. 10.1162/coli_a_00037 [DOI] [Google Scholar]
  • 23. Dyer C, Ballesteros M, Ling W, Matthews A, Smith NA. Transition-Based Dependency Parsing with Stack Long Short-Term Memory. In: ACL; 2015. [Google Scholar]
  • 24. Wang Y, Che W, Guo J, Liu T. A neural transition-based approach for semantic dependency graph parsing. In: AAAI; 2018. [Google Scholar]
  • 25.Choi JD, Palmer M. Transition-based semantic role labeling using predicate argument clustering. In: Proceedings of the ACL 2011 Workshop on Relational Models of Semantics. Association for Computational Linguistics; 2011. p. 37–45.
  • 26.Lamers M, de Hoop H. Animacy information in human sentence processing: An incremental optimization of interpretation approach. In: International Workshop on Constraint Solving and Language Processing. Springer; 2004. p. 158–171.
  • 27. Foo S, Li H. Chinese word segmentation and its effect on information retrieval. Information processing & management. 2004;40(1):161–190. 10.1016/S0306-4573(02)00079-1 [DOI] [Google Scholar]
  • 28.Zhang Y, Yang J. Chinese NER Using Lattice LSTM. In: Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers); 2018. p. 1554–1564.
  • 29. Hochreiter S, Schmidhuber J. Long short-term memory. Neural computation. 1997;9(8):1735–1780. 10.1162/neco.1997.9.8.1735 [DOI] [PubMed] [Google Scholar]
  • 30.Shibata Y, Kida T, Fukamachi S, Takeda M, Shinohara A, Shinohara T, et al. Byte Pair encoding: A text compression scheme that accelerates pattern matching. Technical Report DOI-TR-161, Department of Informatics, Kyushu University; 1999.
  • 31.Li P, Zhu Q, Diao H, Zhou G. Joint modeling of trigger identification and event type determination in Chinese event extraction. In: Proceedings of COLING 2012; 2012. p. 1635–1652.
  • 32.Li P, Zhou G. Employing morphological structures and sememes for Chinese event extraction. In: Proceedings of COLING 2012; 2012. p. 1619–1634.
  • 33.Devlin J, Chang MW, Lee K, Toutanova K. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. arXiv preprint arXiv:181004805. 2018;.
  • 34. Gage P. A new algorithm for data compression. C Users Journal. 1994;12(2):23–38. [Google Scholar]
  • 35.Mikolov T, Chen K, Corrado G, Dean J. Efficient estimation of word representations in vector space. arXiv preprint arXiv:13013781. 2013;.
  • 36.Zhang M, Zhang Y, Fu G. Transition-based neural word segmentation. In: Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers); 2016. p. 421–431.
  • 37. Fredkin E. Trie memory. Communications of the ACM. 1960;3(9):490–499. 10.1145/367390.367400 [DOI] [Google Scholar]
  • 38. Srivastava N, Hinton G, Krizhevsky A, Sutskever I, Salakhutdinov R. Dropout: a simple way to prevent neural networks from overfitting. The Journal of Machine Learning Research. 2014;. [Google Scholar]
  • 39.Kingma DP, Ba J. Adam: A method for stochastic optimization. arXiv preprint arXiv:14126980. 2014;.
  • 40.Miwa M, Bansal M. End-to-End Relation Extraction using LSTMs on Sequences and Tree Structures. In: Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers); 2016. p. 1105–1116.
  • 41.Kovaleva O, Romanov A, Rogers A, Rumshisky A. Revealing the Dark Secrets of BERT. In: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP); 2019. p. 4356–4365.
  • 42. Nadeau D, Sekine S. A survey of named entity recognition and classification. Lingvisticae Investigationes. 2007;30(1):3–26. 10.1075/li.30.1.03nad [DOI] [Google Scholar]
  • 43. Zelenko D, Aone C, Richardella A. Kernel methods for relation extraction. Journal of machine learning research. 2003;3(Feb):1083–1106. [Google Scholar]
  • 44.Kumaran G, Allan J. Text classification and named entities for new event detection. In: Proceedings of the 27th annual international ACM SIGIR conference on Research and development in information retrieval; 2004. p. 297–304.
  • 45. Liu J, Chen Y, Liu K, Zhao J. Event detection via gated multilingual attention mechanism. In: AAAI; 2018. [Google Scholar]
  • 46. Roth D, Yih Wt. Global inference for entity and relation identification via a linear programming formulation Introduction to statistical relational learning. 2007; p. 553–580. [Google Scholar]
  • 47.Yang B, Cardie C. Joint inference for fine-grained opinion extraction. In: Proceedings of the 51st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers); 2013. p. 1640–1649.
  • 48. Li Q, Ji H, Huang L. Joint event extraction via structured prediction with global features. In: ACL; 2013. [Google Scholar]
  • 49. Nguyen TH, Cho K, Grishman R. Joint event extraction via recurrent neural networks. In: NAACL; 2016. [Google Scholar]
  • 50.Rei M. Semi-supervised Multitask Learning for Sequence Labeling. In: Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers); 2017. p. 2121–2130.
  • 51.Chen W, Zhang Y, Isahara H. Chinese named entity recognition with conditional random fields. In: Proceedings of the Fifth SIGHAN Workshop on Chinese Language Processing; 2006. p. 118–121.
  • 52.Peng N, Dredze M. Improving Named Entity Recognition for Chinese Social Media with Word Segmentation Representation Learning. In: Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers); 2016. p. 149–155.
  • 53. Chen Y, Zheng DQ, Zhao TJ. Chinese relation extraction based on deep belief nets. Ruanjian Xuebao/Journal of Software. 2012;23(10):2572–2585. [Google Scholar]
  • 54. Zeng Y, Yang H, Feng Y, Wang Z, Zhao D. A convolution BiLSTM neural network model for Chinese event extraction In: Natural Language Understanding and Intelligent Applications. Springer; 2016. p. 275–287. [Google Scholar]
  • 55.Zhang L, Moldovan D. Chinese relation classification using long short term memory networks. In: Proceedings of the Eleventh International Conference on Language Resources and Evaluation (LREC 2018); 2018.

Decision Letter 0

Jie Zhang

21 May 2020

PONE-D-20-12576

A Transition-based Neural Framework for Chinese Information Extraction

PLOS ONE

Dear Dr. Huang,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Please revise the paper by considering the reviewer's comments.

Please submit your revised manuscript by Jul 05 2020 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: http://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols

We look forward to receiving your revised manuscript.

Kind regards,

Jie Zhang

Academic Editor

PLOS ONE

Journal Requirements:

When submitting your revision, we need you to address these additional requirements.

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

2. We suggest you thoroughly copyedit your manuscript for language usage, spelling, and grammar. If you do not know anyone who can help you do this, you may wish to consider employing a professional scientific editing service.  

Whilst you may use any professional scientific editing service of your choice, PLOS has partnered with both American Journal Experts (AJE) and Editage to provide discounted services to PLOS authors. Both organizations have experience helping authors meet PLOS guidelines and can provide language editing, translation, manuscript formatting, and figure formatting to ensure your manuscript meets our submission guidelines. To take advantage of our partnership with AJE, visit the AJE website (http://learn.aje.com/plos/) for a 15% discount off AJE services. To take advantage of our partnership with Editage, visit the Editage website (www.editage.com) and enter referral code PLOSEDIT for a 15% discount off Editage services.  If the PLOS editorial team finds any language issues in text that either AJE or Editage has edited, the service provider will re-edit the text for free.

Upon resubmission, please provide the following:

  • The name of the colleague or the details of the professional service that edited your manuscript

  • A copy of your manuscript showing your changes by either highlighting them or using track changes (uploaded as a *supporting information* file)

  • A clean copy of the edited manuscript (uploaded as the new *manuscript* file)

3. Please ensure that the manuscript contains adequate English translations.

4. We note that you have indicated that data from this study are available upon request. PLOS only allows data to be available upon request if there are legal or ethical restrictions on sharing data publicly. For more information on unacceptable data access restrictions, please see http://journals.plos.org/plosone/s/data-availability#loc-unacceptable-data-access-restrictions.

In your revised cover letter, please address the following prompts:

a) If there are ethical or legal restrictions on sharing a de-identified data set, please explain them in detail (e.g., data contain potentially sensitive information, data are owned by a third-party organization, etc.) and who has imposed them (e.g., an ethics committee). Please also provide contact information for a data access committee, ethics committee, or other institutional body to which data requests may be sent.

b) If there are no restrictions, please upload the minimal anonymized data set necessary to replicate your study findings as either Supporting Information files or to a stable, public repository and provide us with the relevant URLs, DOIs, or accession numbers. For a list of acceptable repositories, please see http://journals.plos.org/plosone/s/data-availability#loc-recommended-repositories.

We will update your Data Availability statement on your behalf to reflect the information you provide.

5. PLOS requires an ORCID iD for the corresponding author in Editorial Manager on papers submitted after December 6th, 2016. Please ensure that you have an ORCID iD and that it is validated in Editorial Manager. To do this, go to ‘Update my Information’ (in the upper left-hand corner of the main menu), and click on the Fetch/Validate link next to the ORCID field. This will take you to the ORCID site and allow you to create a new iD or authenticate a pre-existing iD in Editorial Manager. Please see the following video for instructions on linking an ORCID iD to your Editorial Manager account: https://www.youtube.com/watch?v=_xcclfuvtxQ

6. Please include a copy of Table 6 which you refer to in your text on page 9.

7. We note you have included a table to which you do not refer in the text of your manuscript. Please ensure that you refer to Table 3 in your text; if accepted, production will need this reference to link the reader to the Table.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Partly

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: No

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: Summary: This paper proposes a joint extraction approach for Chinese information. The main idea is to combine lattice LSTM and transition system to build an end-to-end learning network to complete the joint extraction of entities, relations and events. The authors demonstrate improvements over the state of the art on ACE 2005 Chinese datasets.

The paper is thorough in its presentation of the approach and evaluation, and is generally well written and readable. However, it does requires further editing to address numerous grammatical errors, especially with regard to the use of articles.

Although datasets cannot be publicly shared, authors can provide an online replication package that includes some test samples, allowing reviewers and readers to validate the analyses and data filtering that was performed.

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2020 Jul 15;15(7):e0235796. doi: 10.1371/journal.pone.0235796.r002

Author response to Decision Letter 0


19 Jun 2020

Reviewers' comments:

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: Summary: This paper proposes a joint extraction approach for Chinese information. The main idea is to combine lattice LSTM and transition system to build an end-to-end learning network to complete the joint extraction of entities, relations and events. The authors demonstrate improvements over the state of the art on ACE 2005 Chinese datasets.

The paper is thorough in its presentation of the approach and evaluation, and is generally well written and readable. However, it does requires further editing to address numerous grammatical errors, especially with regard to the use of articles.

Although datasets cannot be publicly shared, authors can provide an online replication package that includes some test samples, allowing reviewers and readers to validate the analyses and data filtering that was performed.

---Response: Thank you for your comments, grammatical errors are addressed, test samples and replication package are available at https://github.com/zjcalva/chinese_information_extraction, which is added in the paper(Section Experiments)

Attachment

Submitted filename: Response to Reviewers.txt

Decision Letter 1

Jie Zhang

23 Jun 2020

A Transition-based Neural Framework for Chinese Information Extraction

PONE-D-20-12576R1

Dear Dr. Huang,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Jie Zhang

Academic Editor

PLOS ONE

Additional Editor Comments (optional):

Reviewers' comments:

Acceptance letter

Jie Zhang

6 Jul 2020

PONE-D-20-12576R1

A Transition-based Neural Framework for Chinese Information Extraction

Dear Dr. Huang:

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department.

If your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org.

If we can help with anything else, please email us at plosone@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Dr. Jie Zhang

Academic Editor

PLOS ONE

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    Attachment

    Submitted filename: Response to Reviewers.txt

    Data Availability Statement

    ACE2005 data used in our paper is owned by Linguistic Data Consortium, The Trustees of the University of Pennsylvania(https://www.ldc.upenn.edu/). Please refer to https://catalog.ldc.upenn.edu/LDC2006T06 for the data details. To facilitate following researches, we provide our code and sample data at https://github.com/zjcalva/chinese_information_extraction, which is added in the paper (Section Experiments). e have provide sample dataset (about 3% of full dataset) in this URL. To fully reproduce our study, readers need to replace "test_sample.txt" with full dataset obtained from LDC. The authors confirm that they had no additional access privilege, and that they accessed the LDC data in the same manner described above.


    Articles from PLoS ONE are provided here courtesy of PLOS

    RESOURCES