Skip to main content
Bioinformatics logoLink to Bioinformatics
. 2019 Jun 14;36(1):264–271. doi: 10.1093/bioinformatics/btz490

CoCoScore: context-aware co-occurrence scoring for text mining applications using distant supervision

Alexander Junge 1,, Lars Juhl Jensen 1,
Editor: Jonathan Wren
PMCID: PMC6956794  PMID: 31199464

Abstract

Motivation

Information extraction by mining the scientific literature is key to uncovering relations between biomedical entities. Most existing approaches based on natural language processing extract relations from single sentence-level co-mentions, ignoring co-occurrence statistics over the whole corpus. Existing approaches counting entity co-occurrences ignore the textual context of each co-occurrence.

Results

We propose a novel corpus-wide co-occurrence scoring approach to relation extraction that takes the textual context of each co-mention into account. Our method, called CoCoScore, scores the certainty of stating an association for each sentence that co-mentions two entities. CoCoScore is trained using distant supervision based on a gold-standard set of associations between entities of interest. Instead of requiring a manually annotated training corpus, co-mentions are labeled as positives/negatives according to their presence/absence in the gold standard. We show that CoCoScore outperforms previous approaches in identifying human disease–gene and tissue–gene associations as well as in identifying physical and functional protein–protein associations in different species. CoCoScore is a versatile text mining tool to uncover pairwise associations via co-occurrence mining, within and beyond biomedical applications.

Availability and implementation

CoCoScore is available at: https://github.com/JungeAlexander/cocoscore.

Supplementary information

Supplementary data are available at Bioinformatics online.

1 Introduction

Text mining of the scholarly literature for the purpose of information extraction is a fruitful approach to keep abreast of recent research findings. The first step in information extraction is named entity recognition (NER) (Jurafsky and Martin, 2008). Biomedical NER aims to identify relevant entities, such as genes, chemicals or diseases, in text. Entities of interest can either be predefined in a dictionary or predicted using a machine learning model. NER is followed by a normalization step mapping the entities to a fixed set of identifiers, such as HGNC gene symbols (Yates et al., 2017) or Disease Ontology terms (Kibbe et al., 2015). General approaches such as LINNAEUS (Gerner et al., 2010), Tagger (Pafilis et al., 2013), taggerOne (Leaman and Lu, 2016) or OGER (Basaldella et al., 2017) recognize diverse biomedical entities in text, while specialized tools recognize mentions of genetic variants (Allot et al., 2018) or chemicals (Jessop et al., 2011).

It is an active area of research to aggregate literature mentions of individual entities to extract higher-level information, such as pairwise biomedical relations, from the literature. Approaches to extract pairwise relations from a corpus of scientific articles, e.g. downloaded from PubMed, typically follow one of three main paradigms. Firstly, pattern-based approaches define a fixed set of regular expressions or linguistic patterns to match single phrases stating relations of interest. Pattern-based approaches typically achieve high precision but low recall in practice and require a labor-intensive manual construction of matching rules. Examples for this class of approaches are textpresso (Muller et al., 2004) or pattern-based approaches, as developed by Saric et al. (2004), used in STRING (Szklarczyk et al., 2017) and STITCH (Szklarczyk et al., 2016). Secondly, unsupervised counting approaches count how often two entities appear together and aggregate these counts over the whole corpus in a co-occurrence statistic. A major shortcoming of simple counting-based co-occurrence scoring approaches to find pairwise relations is that the context of each co-occurrence is ignored, which can lead to low precision. For instance, sentences explicitly stating the absence of an association or describing findings unrelated to a relation are counted, too. Furthermore, counting-based co-occurrence scoring approaches do not allow to differentiate between different kinds of associations, such as physical protein–protein interactions and transcription factor–target associations. The major strengths of counting approaches are that they typically achieve relatively high recall and require no annotated training data or manually crafted match patterns. Examples of this class of approaches are the text mining evidence contained in STRING and DISEASES (Pletscher-Frankild et al., 2015) as well as DisGeNet (Pinero et al., 2017). Thirdly, supervised machine learning approaches require a labeled training dataset of associations and train a model to recognize relations of interest. Machine learning approaches are prone to overfit to the often small, manually annotated training datasets resulting in brittle models that do not generalize well to other datasets. For example Rios et al., (2018) showed that generalization between datasets of protein–protein and drug–drug interactions is only achieved when additional techniques such as the use of adversarial neural networks for domain adaption are employed. Examples for machine learning-based approaches to relation extraction are BeFree (Bravo et al., 2015) and LocText (Cejuela et al., 2018). Bundschus et al. (2008) make use of conditional random fields.

Distant supervision, sometimes called weak supervision, can be used to generate datasets with a large number of samples with some amount of noise in the labels (Craven and Kumlien, 1999; Lamurias et al., 2017; Poon et al., 2015; Ravikumar et al., 2012). Distant supervision for relation extraction only requires access to a knowledge base of well-described associations as well as an unlabeled set of entity co-occurrences. Labels for the dataset of co-occurrences are then inferred based on the presence or absence of the co-mentioned entities in the knowledge base. Note that a manually annotated text corpus is not required when using distant supervision.

In this work, we describe a novel approach, CoCoScore, that combines an unsupervised counting approach with a machine learning approach based on distant supervision. This allows CoCoScore to train a machine learning model to score sentence-level co-mentions without requiring an expert-curated dataset of phrases describing associations. The model is based on fastText (Joulin et al., 2016) and relies on word embeddings that represent words as dense vectors. CoCoScore finally aggregates all sentence-level scores in a given corpus in a final context-aware co-occurrence score for each entity pair. We apply CoCoScore to different biomedical relation extraction tasks: tissue–gene, disease–gene, physical protein–protein interactions and functional protein–protein associations in H.sapiens, D.melanogaster and S.cerevisiae. CoCoScore consistently outperforms a baseline model that uses constant sentence scores, following previously proposed approaches. We show then that the performance of CoCoScore further benefits from an unsupervised pretraining of the underlying word embeddings. By querying CoCoScore with manually constructed sentences, we show that keywords indicating protein–protein interactions and, to a certain extent, negations and modality are reflected in the sentence scores. A Python implementation of CoCoScore is available for download. The software package is geared towards reusability across many text mining tasks by only requiring a list of co-mentions for scoring without relying on a particular NER approach.

2 Materials and methods

2.1 Corpus

The corpus used for text mining consists of PubMed abstracts as well as both open access and author’s manuscript full text articles available from PMC in BioC XML format (Comeau et al., 2018; Doğan et al., 2014). All abstracts were last updated on June 9th, 2018 and all full text articles were last updated on April 17, 2018. We removed full text articles that were not classified as English-language articles by fastText (Joulin et al., 2016) using a pretrained language identification model for 176 languages downloaded from https://fasttext.cc/docs/en/language-identification.html. We furthermore removed full text articles that could not be mapped to a PubMed ID and those that mention more than 200 entities of any type included in our dictionary of biomedical entities such as proteins, chemicals, diseases, species or tissues. The final corpus consists of 28 546 040 articles of which 2 106 542 are available as full text and the remainder as abstracts.

2.2 Datasets and distant supervision

We use tagger v1.1 to recognize named entities in the corpus using a dictionary-based approach. Tagger can be downloaded from https://bitbucket.org/larsjuhljensen/tagger/. The dictionaries used for named entity recognition, training and test datasets as well as pretrained word embeddings and fastText scoring models described below can be downloaded from: https://doi.org/10.6084/m9.figshare.7198280.v1. The named entity recognition step is followed by a normalization step to a common naming scheme. All gene/protein identifiers were mapped to identifiers of corresponding proteins used in STRING v10.5 (Szklarczyk et al., 2017). The normalization of disease and tissue identifiers is described below. We used placeholder tokens in all datasets to replace tissue, gene, protein and disease names found by tagger. This blanking of entity names is important to learn a co-occurrence scoring model independent of the identity of the entities mentioned. Finally, we retain sentences that co-mention at least two biomedical entities of interest, depending on the given dataset.

The assignments of binary class labels to the sentences in each dataset follows a distant supervision approach to obtain a weak labeling. Given a sentence co-mentioning two entities of interest, the sentence is assigned a positive class label (1) if the entity pair is found in a given gold standard set of pairwise associations. If the two entities appear in the gold standard individually but not in association, the sentence is assigned a negative class label (0). The gold standard is specific to each dataset and described in the following sections. Table 1 lists information about the final datasets. The final CoCoScore scoring scheme can then be defined as a sum over all articles in the corpus (see Section 2.3).

Table 1.

Entity, pair and co-mention counts as well as percentage of positive instances in all datasets

Dataset type Organism Entity count Pair count Pos. pairs Gold pairs Gold pairs cov. Co-mentions Pos. co-mentions
Disease–gene H.sapiens 698 diseases, 1972 genes 51 786 4.9% 2726 93.7% 1 182 951 45.2%
Tissue–gene H.sapiens 21 tissues, 14066 genes 174 916 13.2% 31 387 74.0% 15 706 365 39.4%
Functional PPA H.sapiens 4695 proteins 1 032 063 14.1% 361 744 40.4% 16 390 304 58.2%
Functional PPA D.melanogaster 1792 proteins 36 524 27.4% 88 604 11.3% 358 141 64.9%
Functional PPA S.cerevisiae 1567 proteins 47 005 32.6% 59 111 25.9% 301 582 65.4%
Binding PPI H.sapiens 6053 proteins 1 236 751 2.9% 76 299 47.4% 15 611 378 21.3%
Binding PPI D.melanogaster 2168 proteins 60 378 12.7% 32 541 23.6% 463 822 42.6%
Binding PPI S.cerevisiae 1612 proteins 35 786 17.2% 13 500 45.6% 240 112 59.0%

Note: ‘Gold pairs’ refers to the total number of pairs found in the gold standard. ‘Gold pairs cov.’ is the percentage of gold-standard pairs co-mentioned in at least one sentence in the dataset.

2.2.1. Disease–gene associations

We followed the approach in the DISEASES database (Pletscher-Frankild et al., 2015) and obtained an expert-curated gold standard of disease–gene associations from Genetics Home Reference (GHR) (Fomous et al., 2006) (downloaded on May 7, 2017) by parsing disease-associated genes from json-formatted disease entries in GHR. We also retained entity co-mentions in the literature that involved a Disease Ontology (DO) (Kibbe et al., 2015) child term of a disease found in the gold standard. This propagation upwards the DO hierarchy yields a larger dataset of disease–gene associations while not compromising quality. For instance, any article linking the gene APP to Alzheimer’s disease implies that APP has a more general role in neurodegenerative disease. Disease names and aliases were mapped to DO identifiers.

2.2.2. Tissue–gene associations

We followed the approach in the TISSUES database (Palasca et al., 2018) and downloaded manually curated tissue–gene associations from UniProtKB (SIB Members, 2016). We restricted the tissue–gene association dataset to 21 major tissues, following the benchmarking scheme of the TISSUES database, and employed ontology propagation upwards the BRENDA Tissue Ontology (BTO) (Gremse et al., 2011), similar to the previously described DO propagation for disease mentions. Tissue names were normalized to BTO identifiers.

2.2.3. Functional protein–protein associations

We obtained gold standard protein–protein associations (PPA) for H.sapiens, D.melanogaster and S.cerevisiae following the approach for benchmarking associations in STRING (Szklarczyk et al., 2017): Proteins found in at least one KEGG pathway map (Kanehisa et al., 2017) were considered positives since they are functionally associated in the given pathway. We also supplemented the original KEGG maps with artificial maps created by joining proteins from maps that share common metabolites.

2.2.4. Physical protein–protein interactions

We obtained gold standard physical protein–protein interactions (PPI) for H.sapiens, D.melanogaster and S.cerevisiae by obtaining interactions classified as ‘binding’ from STRING v10.5 and retained only the highest scoring interactions, with a score >0.9, as the gold standard. Binding interactions with a score 0.9, were added to a grey list. Co-mentions of grey-listed protein pairs were ignored and counted as neither positives nor negatives when creating the gold standard via distant supervision. While the resulting PPI datasets only contain protein pairs that physically bind to each other, the PPA datasets also encompass other functional associations as defined by membership in the same pathway.

2.3 Context-aware co-occurrence scoring

The context-aware co-occurrence scoring approach implemented in CoCoScore consists of two components. Firstly, a sentence-level classification model is trained to predict context-aware co-mention scores. Secondly, a scoring scheme combines sentence-level scores into a co-occurrence score taking evidence gathered over the whole corpus into account.

2.3.1. Unsupervised pretraining of word embeddings

Word embeddings represent each unique word in the corpus by a vector. We use a skipgram word embedding model that learns word vectors such that the vector representation of a word can be used to predict the words appearing in its context. This objective allows to represent words with similar syntax and semantics by similar vectors, as measured in terms of their inner product. Further details on the skipgram model and its training process can be found in Mikolov et al. (2013) and Bojanowski et al. (2016).

We pretrained word embeddings using fastText v1.0 (Joulin et al., 2016) on the whole corpus, not just on sentences in the dataset that co-mention entities of interest, which improves their generalization to downstream machine learning tasks. This step can be viewed as an instance of transfer learning where information is brought from a general task, the pretraining of word embeddings, to a specific task, the classification of sentences co-mentioning biomedical entities.

2.3.2. Training a sentence classification model

Our sentence-level classification model was implemented using fastText v.1.0 in supervised classification mode. Given a sentence, the pretrained vector representations of each word in the sentence are averaged. A logistic regression classifier M, implemented in fastText, then predicts a binary class label since each sentence is labeled as either positive or negative after distant supervision. The sentence classification model M returns a score between 0 and 1. We interpret this score as the probability that the sentence belongs to the positive class, i.e. that it states an associations.

The dimensionality of word embeddings was set to 300; we performed 50 epochs of stochastic gradient descent with learning rate of 0.005 to train the model. We used unigram as well as bigram embeddings in the logistic regression classifier M to partially capture local word order. Instead of performing a computationally expensive cross-validation for hyperparameter tuning, we set these fastText hyperparameters by manually optimizing the performance on a subset of the disease–gene association training data. Remaining hyperparameters were set to their defaults in the fastText v1.0 release.

2.3.3. Co-occurrence scoring

The final CoCoScore co-occurrence scores for a pair of entities aggregates the scores computed by sentence model M over all documents in the dataset. Given a corpus C and an entity pair (i, j), the co-occurrence count C(i, j) for the pair is

C(i,j)=kCsk(i,j),

where

sk(i,j)={maxuTk(i,j)ru(i,j)if i and j are co-mentioned in k0else.

Here, Tk(i,j) is the set of sentences co-mentioning i and j in document k and ru(i,j) is the sentence-level score returned by M for sentence u. The co-occurrence counts C(i, j) are converted to co-occurrence scores S(i, j) as follows:

S(i,j)=C(i,j)α(C(i,j)C(·,·)C(i,·)C(·,j))1α. (1)

C(i,·) and C(·,j) are the sums of all co-occurrence counts involving entity i and j, respectively, C(·,·) sums the co-occurrences of all entity pairs. The hyperparameter α trades off the influence of C(i, j) counts and the observed-over-expected ratio captured in the second term of Equation 1. α was set to 0.65 after cross-validation (see Section 3.1).

Figure 1 outlines the complete context-aware co-occurrence scoring approach, illustrating both C(i, j) and M.

Fig. 1.

Fig. 1.

Context-aware scoring of co-occurrences. (A) N sentences in the corpus co-mention the gene LRRK2 (i) and Parkinson’s disease (j). Context-aware sentence-level scores r(i, j) are summed to produce the final co-occurrence count C(i, j). (B) The score r1(i,j) is computed by blanking gene and disease names, mapping all remaining words to their word vectors and scoring the resulting sentence vector using a machine learning model. As described in Section 2.3.2, this model is based on logistic regression and trained via distant supervision

2.3.4 Baseline scoring scheme

We next defined a baseline model to compare CoCoScore to. Contrary to the context-aware model implemented in CoCoScore, the baseline model does not take context into account but scores all co-mentions equally. Given a corpus C and entity pair (i, j), the baseline co-occurrence count C˜(i,j) is defined as:

C˜(i,j)=kCs˜k(i,j),

where

s˜k(i,j)={1if i and j are co-mentioned in a sentence in k0else.

Note that the central difference between CoCoScore and this baseline is that s˜k(i,j) in the baseline uses a constant sentence score of 1 while sk(i,j) in CoCoScore relies on sentence scores learned by the model M. As before, the final co-occurrence scores S(i,j)˜ are computed from C(i,j)˜:

S˜(i,j)=C˜(i,j)α(C˜(i,j)C˜(·,·)C˜(i,·)C˜(·,j))1α (2)

For the datasets of sentence-level co-mentions used in this work, this baseline model is equivalent to the co-occurrence scoring model used in, e.g. STRING (Franceschini et al., 2013; Szklarczyk et al., 2017), STITCH (Szklarczyk et al., 2016), TISSUES (Santos et al., 2015; Palasca et al., 2018) and DISEASES (Pletscher-Frankild et al., 2015).

2.3.5. Performance evaluation

The area under the precision–recall curve (AUPRC) achieved by a given scoring model was computed by first sorting all entity pairs according to their co-occurrence scores in decreasing order and calculating precision as well as recall at each score threshold as follows:

precision=TPTP+FP

and

recall=TPTP+FN=TPP,

where TP is the number of true positives, FP is the number false of positives, FN is the number of false negatives and P is the number of positives. The AUPRC is the area under the precision–recall curve.

A random classifier has AUPRC equal to the fraction of positives in the dataset and a perfect classifier has an AUPRC of 1. Precision–recall curves are better suited than receiver operating characteristic curves (ROC) for imbalanced datasets since the latter give inflated performance estimates (Lever et al., 2016; Lichtnwalter and Chawla, 2012). However, for comparison, we also state model performance in terms of area under ROC (AUROC) in the Supplementary Material of this article.

All AUPRC performance measures reported below were adjusted to a fixed percentage of 10% positive samples in the dataset. This adjustment makes AUPRC values comparable between datasets with different fractions of positives. We picked a positive percentage of 10% since this seems to be a realistic prior given our datasets (Table 1). Adjusting the AUPRC to a fixed class distribution was performed as follows: Let a be the target fraction of positives in the dataset (0.1 in this work) and b be the observed fraction of positives in the dataset. To adjust the AUPRC, we replace precision with the following adjusted measure:

precisionadjusted=(a/b)·TP(a/b)·TP+((1a)/(1b))·FP

The adjusted AUPRC is then the area under the precisionadjusted–recall curve.

2.3.6. CV and train/test splitting

For each dataset, we reserved all co-mentions involving 20% of the entity pairs as the test set which is only used for the final model evaluation step. This pair-level splitting ensures independence of training and test datasets since all co-mentions of a specific pair of entities are found exclusively in the training or exclusively in the test set. Each training dataset consists of co-mentions of the remaining 80% of entity pairs. In the training dataset, we randomly sampled a maximum of 100 sentence-level co-mentions per pair to ensure that the sentence-level model M does not overly fits to pairs that appear very often in the literature. To ensure a realistic performance evaluation, no such sampling was done for the test dataset. 3-fold cross-validation (CV) on the training set was used to tune the hyperparameter α. For computational reasons, we randomly sampled 10% of interactions in the three biggest datasets (functional PPA H.sapiens, binding PPI H.sapiens, tissue–gene associations) during CV. This reduced the number of associations in the downsampled dataset to approximately the number of associations in the remaining datasets. 1000 bootstrap samples on the test dataset were drawn to compute performance standard deviations and to test statistical significance using Student’s t-test.

3 Results and discussion

3.1 Sentence scores of higher importance in CoCoScore than in baseline model

Before analyzing the performance on the test set, we tuned the weighting exponent hyperparameter α for both CoCoScore and the baseline model (see Section 2.3) via cross-validation (CV). α determines how much weight is put on the co-occurrence counts compared to the observed-over-expected ratio. The CoCoScore model achieved optimal CV performance for α0.65 and the baseline model for α0.55 for most datasets. Supplementary Figure S1 depicts CV performance depending on α for both models. We consider the tissue–gene dataset, where CV results in a considerably down-weighted observed-over-expected ratio term, an outlier due to the poor performance of both models on this dataset. The optimal α for CoCoScore was larger than the optimal α for the baseline model. This means that the best performing CoCoScore model put more weight on the co-occurrence counts than on the observed-over-expected-ratio, compared to the baseline model (Equations 1 and 2). We hypothesize that this is because CoCoScore down-weights uninformative sentences, compared to informative ones, making the sentence-level scores more reliable. Furthermore, the CoCoScore model outperformed the baseline on all datasets for the respective optimal α ranges as well as two alternative CoCoScore implementations that use the sentence scoring model M as a filter for removing low scoring sentences prior to score aggregation (see Supplementary Fig. S2). Below, we use α=0.65 for CoCoScore and α=0.55 for the baseline to compute test dataset performance.

3.2 CoCoScore outperforms baseline model in identifying disease–gene and tissue–gene associations

Table 2 lists AUPRC performance for both CoCoScore and the baseline model on the tissue–gene and disease–gene association datasets. Supplementary Table S1 depicts the performance in terms of the area under the receiver operating characteristic curve (AUROC). CoCoScore outperformed the baseline model on both dataset. Both approaches achieved considerably better performance on the disease–gene than on the tissue–gene association dataset. We thus manually inspected the 10 highest-scoring associations in the tissue–gene association dataset. Five of these tissue–gene pairs were counted as false positives, as defined by the gold standard derived from UniProtKB (Section 2.2.2). However, each of these pairs had more than 900 sentence-level co-mentions in articles and multiple sentences clearly stating the expression of the respective gene in the respective tissue. We concluded that these five associations are likely true positives that are missing in the gold standard rather than false positives. The seemingly poor performance on the tissue–gene association dataset can in part be explained by the incompleteness of the gold standard. At the same time, CoCoScore appears to be able to learn informative text patterns leading to an improved performance.

Table 2.

Adjusted area under the precision-recall curve (AUPRC) for CoCoScore and baseline model on tissue–gene and disease–gene association datasets generated via distant supervision

Method Disease–gene Tissue–gene
CoCoScore 0.86 0.19
Baseline 0.80 0.17

3.3 Physical protein–protein interactions are better identified than functional protein–protein associations

Figure 2 depicts performance on functional protein–protein associations (PPAs) and physical protein–protein interactions (PPIs) across H.sapiens, D.melanogaster and S.cerevisiae for both CoCoScore and the baseline. Supplementary Figure S3 depicts the performance in terms of AUROC. CoCoScore outperformed the baseline and yielded similar improvements on all functional PPA and physical PPI datasets. While both models performed better on the binding PPI datasets than on the functional PPA datasets, we did not observe a clear trend in performance differences between organisms. CoCoScore achieves best adjusted AUPRC of 0.67 for binding PPI in H.sapiens and adjusted AUPRC of 0.57 in D.melanogaster and of 0.58 in S.cerevisiae. On the other hand, CoCoScore achieves best adjusted AUPRC of 0.50 for functional PPA in both D.melanogaster and S.cerevisiae and an adjusted AUPRC of 0.44 for H.sapiens. Overall, CoCoScore outperformed the baseline model on all six protein–protein association datasets surveyed.

Fig. 2.

Fig. 2.

Performance on functional protein–protein associations and physical protein–protein interactions across H.sapiens, D.melanogaster and S.cerevisiae for both CoCoScore (blue) and the baseline model (red). Performance is depicted as adjusted area under the precision-recall curve (AUPRC) (Color version of this figure is available at Bioinformatics online.)

3.4 Pretrained word embeddings improve performance on most datasets

The default CoCoScore sentence classification model relies on word embeddings that were pretrained in an unsupervised manner on all articles in the corpus. To assess the impact of this pretraining step on CoCoScore’s performance, we compared the usage of pretrained word embeddings to the usage of embeddings that are learned at training time. In the latter scenario, the fastText-based sentence scoring model not only optimizes the weights of the logistic regression classifier but also the vectors representing words, starting from randomly initialized vectors. Since the latter approach never accesses the complete corpus, word embeddings are only trained on sentences co-mentioning entities that are found in the respective training dataset.

Figure 3 depicts adjusted AUPRC performance with and without pretrained word embeddings. The CoCoScore performance in Figure 3 is the same as shown in Figure 2 and Table 2. Supplementary Figure S4 depicts performance in terms of AUROC.

Fig. 3.

Fig. 3.

Performance using pretrained and not pretrained word embeddings. Not pretrained embeddings are learned at training time. The performance of CoCoScore with (blue) and without (green) pretrained word embeddings is shown for functional PPA and binding PPI datasets (A) as well as disease–gene and tissue–gene associations (B). Performance is depicted as adjusted area under the precision-recall curve (AUPRC). All performance standard deviations after bootstrap resampling were <0.006. The difference between models using pretrained and learned embeddings was significant for all datasets except disease–gene associations and binding PPIs in D.melanogaster at a significance level of 0.001 (Color version of this figure is available at Bioinformatics online.)

The CoCoScore model using pretrained word embeddings in most cases outperformed non-pretrained word embeddings. We observe that the pretraining step was more fruitful for datasets with poor performance. The small increase in performance for some datasets could be due to the considerable size of the distantly supervised dataset (the smallest dataset contains 240k sentence co-mentions) which are large enough to train adequate word embeddings without pretraining. However, we still recommend using CoCoScore with pretrained embeddings for best performance.

3.5 Making sense of CoCoScore’s sentence scoring model by manually querying the model

The 300 dimensions of the word embeddings are not easily interpretable making it hard to understand which features drive sentence score predictions for a model trained on a given dataset. We thus used an indirect approach to interpret the sentence-level scoring model learned by CoCoScore by querying the model trained to recognize binding PPIs in S.cerevisiae with hand-crafted example sentences.

We observed that the model returned high scores for sentences containing keywords linked to physical interactions, such as the words ‘complex’ or ‘subunit’, but did not pick up modality or uncertainty in sentences very well, once a keyword was present. For instance, the sentence ‘_ and _ form a complex.’ received a score of 0.99 while the sentences ‘_ and _ do not form a complex.’ and ‘_ and _ might form a complex.’ received a score of 0.98 and 0.99, respectively. Here, ‘_’ is a generic token used to blank protein names.

On the other hand, the model seemed to recognize negations and modality in sentences that contained the verb phrase ‘bind to’. The sentence ‘_ always binds to _.’ received a score of 0.72, ‘_ binds to _.’ received a score of 0.44, ‘_ possibly binds to _.’ received a score of 0.37, ‘_ does not bind to _.’ received a score of 0.34, and ‘_ never binds to _.’ received a score of 0.24. Based on this exploratory analysis, we conclude that the CoCoScore sentence scoring model for S.cerevisiae binding PPIs seems to rely on keywords and is able to detect modality and negations in certain situations.

3.6 Limitations and future work

While CoCoScore implements a novel context-aware co-occurrence scoring approach that improves upon a baseline model for all our test datasets, we see several limitations and directions for future research. Relations extracted by CoCoScore currently lack directionality. Many biomedical relations, such as protein phosphorylation, are directional that is not trivial to infer if, for instance, one protein kinase phosphorylates another, as commonly seen in signal transduction pathways. To address this shortcoming, CoCoScore’s distant supervision approach could be combined with pattern-based approaches to infer directionality. Alternatively, the word embeddings for the words in a sentence could not be collapsed into a single sentence vector but kept as a sequence of vectors fed into a sequence model such as a recurrent neural network.

We also plan to investigate the transferability of pretrained sentence scoring models between relation extraction tasks. For instance, a unified model could potentially be trained that recognizes not one specific type of relations, such as disease–gene associations, but also other relations, such as protein–protein interactions. Keywords and modality driving sentence scores (Section 3.5) should, to some extent, be transferrable between relation extraction tasks. Similarly, pretrained scoring models trained on one dataset could be combined with supervised learning performed on a second, expert-labeled dataset. This would enable the simultaneous use of large, distantly supervised dataset as well as small, accurately labeled dataset to boost performance. Magge et al. (2018) use a similar approach to identify geographic locations in sequence database entries.

Lastly, CoCoScore could be extended to score co-mentions beyond sentence-level by, for example, introducing a term in the scoring model that depends on the distance between entities co-mentioned outside a sentence.

4 Conclusion

Our newly developed approach, CoCoScore, performs pairwise co-occurrence scoring over a big corpus by combining an unsupervised counting scheme with a distantly supervised sentence scoring model based on pretrained word embeddings. This scoring model is in some situations able to recognize keywords, negations and modality in sentences. Our approach performs better than a baseline scoring scheme inspired by previously proposed approaches on all eight benchmark datasets used in this study, covering disease–gene, tissue–gene, physical protein–protein interactions and functional protein–protein associations. The performance improvements achieved by CoCoScore are driven by the underlying sentence scoring model.

CoCoScore is a versatile tool to aid biomedical relation extraction via text mining that is applicable to many applications beyond those presented in this paper. Our open source implementation only requires sentences co-mentioning entities as input and is available under a permissive license together with pretrained word embedding as well as the sentence scoring models trained in this work. This eases the integration of CoCoScore into existing text mining workflows for biomedical relation extraction.

Supplementary Material

btz490_Supplementary_Material

Acknowledgements

We would like to thank Rūdolfs Bērziņš for helpful discussions of this works, David Westergaard for feedback on the manuscript as well as the team of the Danish National Supercomputer for Life Sciences (Computerome) for HPC support.

Funding

This work was supported by the Novo Nordisk Foundation (NNF14CC0001) and the National Institutes of Health (NIH) Illuminating the Druggable Genome Knowledge Management Center (U54 CA189205 and U24 224370).

Conflict of interest: LJ Jensen is among the founders of a non-publicly traded company, Intomics A/S and owns between 10 and 15% in stocks through a holding company. Intomics A/S is a contract research organization specialized in deriving core biological insight from big data. He is also member of the scientific advisory board and have received an income of <$5000 USD in the past year for this position.

References

  1. Allot A. et al. (2018) LitVar: a semantic search engine for linking genomic variant data in PubMed and PMC. Nucleic Acids Res., 46, W530–W536. [DOI] [PMC free article] [PubMed] [Google Scholar]
  2. Basaldella M. et al. (2017) Entity recognition in the biomedical domain using a hybrid approach. J. Biomed. Seman., 8, 51.. [DOI] [PMC free article] [PubMed] [Google Scholar]
  3. Bojanowski P. et al. (2016) Enriching word vectors with subword information. arXiv, abs/1607.04606.
  4. Bravo A. et al. (2015) Extraction of relations between genes and diseases from text and large-scale data analysis: implications for translational research. BMC Bioinformatics, 16, 55. [DOI] [PMC free article] [PubMed] [Google Scholar]
  5. Bundschus M. et al. (2008) Extraction of semantic biomedical relations from text using conditional random fields. BMC Bioinformatics, 9, 207.. [DOI] [PMC free article] [PubMed] [Google Scholar]
  6. Cejuela J.M. et al. (2018) LocText: relation extraction of protein localizations to assist database curation. BMC Bioinformatics, 19, 15.. [DOI] [PMC free article] [PubMed] [Google Scholar]
  7. Comeau D.C. et al. (2018) PMC text mining subset in bioc: 2.3 million full text articles and growing. arXiv, abs/1804.05957. [DOI] [PMC free article] [PubMed]
  8. Craven M., Kumlien J. (1999) Constructing biological knowledge bases by extracting information from text sources. In: Proceedings of the 7th International Conference on Intelligent Systems for Molecular Biology, pp. 77–86. [PubMed]
  9. Doğan R.I. et al. (2014). BioCand simplified use of the PMC open access dataset for biomedical text mining. In: Proceedings of the 2014 Workshop on Biomedical Text Mining, Language Resources And Evaluation Conference.
  10. Fomous C. et al. (2006) ‘Genetics home reference’: helping patients understand the role of genetics in health and disease. Commun. Genet., 9, 274–278. [DOI] [PubMed] [Google Scholar]
  11. Franceschini A. et al. (2013) STRING v9.1: protein–protein interaction networks, with increased coverage and integration. Nucleic Acids Res., 41, D808–815. [DOI] [PMC free article] [PubMed] [Google Scholar]
  12. Gerner M. et al. (2010) LINNAEUS: a species name identification system for biomedical literature. BMC Bioinformatics, 11, 85.. [DOI] [PMC free article] [PubMed] [Google Scholar]
  13. Gremse M. et al. (2011) The BRENDA Tissue Ontology (BTO): the first all-integrating ontology of all organisms for enzyme sources. Nucleic Acids Res., 39, D507–513. [DOI] [PMC free article] [PubMed] [Google Scholar]
  14. Jessop D.M. et al. (2011) OSCAR4: a flexible architecture for chemical text-mining. J. Cheminform., 3, 41.. [DOI] [PMC free article] [PubMed] [Google Scholar]
  15. Joulin A. et al. (2016) Bag of tricks for efficient text classification. arXiv, abs/1607.01759.
  16. Jurafsky D., Martin J.H. (2008) Speech and Language Processing, Chapter 22 Information Extraction, 2nd edn.Prentice Hall, Upper Saddle River, New Jersey, pp. 900–951. [Google Scholar]
  17. Kanehisa M. et al. (2017) KEGG: new perspectives on genomes, pathways, diseases and drugs. Nucleic Acids Res., 45, D353–D361. [DOI] [PMC free article] [PubMed] [Google Scholar]
  18. Kibbe W.A. et al. (2015) Disease Ontology 2015 update: an expanded and updated database of human diseases for linking biomedical knowledge through disease data. Nucleic Acids Res., 43, D1071–1078. [DOI] [PMC free article] [PubMed] [Google Scholar]
  19. Lamurias A. et al. (2017) Extracting microRNA–gene relations from biomedical literature using distant supervision. PLoS One, 12, e0171929.. [DOI] [PMC free article] [PubMed] [Google Scholar]
  20. Leaman R., Lu Z. (2016) TaggerOne: joint named entity recognition and normalization with semi-Markov Models. Bioinformatics, 32, 2839–2846. [DOI] [PMC free article] [PubMed] [Google Scholar]
  21. Lever J. et al. (2016) Points of significance: classification evaluation. Nat. Methods, 13, 603–604. [Google Scholar]
  22. Lichtnwalter R., Chawla N.V. (2012) Link prediction: Fair and effective evaluation. In: 2012 IEEE/ACM International Conference on Advances in Social Networks Analysis and Mining, pp. 376–383.
  23. Magge A. et al. (2018) Deep neural networks and distant supervision for geographic location mention extraction. Bioinformatics, 34, i565–i573. [DOI] [PMC free article] [PubMed] [Google Scholar]
  24. Mikolov T. et al. (2013) Distributed representations of words and phrases and their compositionality. arXiv, abs/1310.4546.
  25. Muller H.M. et al. (2004) Textpresso: an ontology-based information retrieval and extraction system for biological literature. PLoS Biol., 2, e309.. [DOI] [PMC free article] [PubMed] [Google Scholar]
  26. Pafilis E. et al. (2013) The SPECIES and ORGANISMS resources for fast and accurate identification of taxonomic names in text. PLoS One, 8, e65390.. [DOI] [PMC free article] [PubMed] [Google Scholar]
  27. Palasca O. et al. (2018) TISSUES 2.0: an integrative web resource on mammalian tissue expression. Database (Oxford), 2018. [DOI] [PMC free article] [PubMed] [Google Scholar]
  28. Pinero J. et al. (2017) DisGeNET: a comprehensive platform integrating information on human disease-associated genes and variants. Nucleic Acids Res., 45, D833–D839. [DOI] [PMC free article] [PubMed] [Google Scholar]
  29. Pletscher-Frankild S. et al. (2015) DISEASES: text mining and data integration of disease–gene associations. Methods, 74, 83–89. [DOI] [PubMed] [Google Scholar]
  30. Poon H. et al. (2015) Distant supervision for cancer pathway extraction from text. In: Proceedings of the Pacific Symposium on Biocomputing, Kohala Coast, Hawaii. pp. 120–131. [PubMed] [Google Scholar]
  31. Ravikumar K. et al. (2012) Literature mining of protein-residue associations with graph rules learned through distant supervision. J. Biomed. Seman., 3, S2. [DOI] [PMC free article] [PubMed] [Google Scholar]
  32. Rios A. et al. (2018) Generalizing biomedical relation classification with neural adversarial domain adaptation. Bioinformatics, 34, 2973–81. [DOI] [PMC free article] [PubMed] [Google Scholar]
  33. Santos A. et al. (2015) Comprehensive comparison of large-scale tissue expression datasets. PeerJ, 3, e1054.. [DOI] [PMC free article] [PubMed] [Google Scholar]
  34. Saric J. et al. (2004) Extracting regulatory gene expression networks from Pubmed. In: Proceedings of the 42nd Annual Meeting of the Association for Computational Linguistics (ACL-04).
  35. SIB Members. (2016) The SIB Swiss Institute of Bioinformatics’ resources: focus on curated databases. Nucleic Acids Res., 44, 27–37. [DOI] [PMC free article] [PubMed] [Google Scholar]
  36. Szklarczyk D. et al. (2016) STITCH 5: augmenting protein–chemical interaction networks with tissue and affinity data. Nucleic Acids Res., 44, D380–384. [DOI] [PMC free article] [PubMed] [Google Scholar]
  37. Szklarczyk D. et al. (2017) The STRING database in 2017: quality-controlled protein–protein association networks, made broadly accessible. Nucleic Acids Res., 45, D362–D368. [DOI] [PMC free article] [PubMed] [Google Scholar]
  38. Yates B. et al. (2017) Genenames.org: the HGNC and VGNC resources in 2017. Nucleic Acids Res., 45, D619–D625. [DOI] [PMC free article] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

btz490_Supplementary_Material

Articles from Bioinformatics are provided here courtesy of Oxford University Press

RESOURCES