Skip to main content
Bioinformatics logoLink to Bioinformatics
. 2021 Jul 12;37(Suppl 1):i468–i476. doi: 10.1093/bioinformatics/btab331

Utilizing image and caption information for biomedical document classification

Pengyuan Li 1,, Xiangying Jiang 2,3, Gongbo Zhang 4,5, Juan Trelles Trabucco 6, Daniela Raciti 7, Cynthia Smith 8, Martin Ringwald 9, G Elisabeta Marai 10, Cecilia Arighi 11, Hagit Shatkay 12,
PMCID: PMC8346654  PMID: 34252939

Abstract

Motivation

Biomedical research findings are typically disseminated through publications. To simplify access to domain-specific knowledge while supporting the research community, several biomedical databases devote significant effort to manual curation of the literature—a labor intensive process. The first step toward biocuration requires identifying articles relevant to the specific area on which the database focuses. Thus, automatically identifying publications relevant to a specific topic within a large volume of publications is an important task toward expediting the biocuration process and, in turn, biomedical research. Current methods focus on textual contents, typically extracted from the title-and-abstract. Notably, images and captions are often used in publications to convey pivotal evidence about processes, experiments and results.

Results

We present a new document classification scheme, using both image and caption information, in addition to titles-and-abstracts. To use the image information, we introduce a new image representation, namely Figure-word, based on class labels of subfigures. We use word embeddings for representing captions and titles-and-abstracts. To utilize all three types of information, we introduce two information integration methods. The first combines Figure-words and textual features obtained from captions and titles-and-abstracts into a single larger vector for document representation; the second employs a meta-classification scheme. Our experiments and results demonstrate the usefulness of the newly proposed Figure-words for representing images. Moreover, the results showcase the value of Figure-words, captions and titles-and-abstracts in providing complementary information for document classification; these three sources of information when combined, lead to an overall improved classification performance.

Availability and implementation

Source code and the list of PMIDs of the publications in our datasets are available upon request.

1 Introduction

Biomedical research findings are typically reported via publications. To simplify access to domain-specific knowledge, while supporting the research community, several biomedical databases [e.g. UniProt (Bateman et al., 2021), BioGRID (Chatr-Aryamontri et al., 2017), Wormbase (Harris et al., 2020) and MGI (Blake et al., 2021)] invest significant effort in expert curation of the literature. The first step in the biocuration process is to identify articles that are relevant to a specific area on which the biomedical databases focus. For example, biocurators at the Jackson Laboratory’s Gene Expression Database (GXD) identify publications relevant to gene expression during mouse development (Finger et al., 2017). Manually selecting biomedical publications in such focus areas is often too labor-intensive and slow for effectively detecting all and only the relevant articles within a large volume of published literature. As such, automatically identifying publications relevant to a specific topic is an important task toward expediting biocuration and, in turn, biomedical research.

The vast majority of current methods for categorization of biomedical documents focus on textual contents which are typically extracted from the title and the abstract of the publication. Several supervised learning methods, including Support Vector Machines (SVMs) (Garcia et al., 2015), Decision Trees (Almeida et al., 2014) and Neural Networks (Burns et al., 2019; Fergadis et al., 2018), have been applied and studied to build document classifiers. Burns et al. (2019) investigated the application of several word embedding methods using different neural network configurations for identifying scientific literature containing information about molecular interaction. Rule-based methods have also been proposed for document classification (Hu et al., 2005; Karystianis et al., 2017). For instance, to identify epidemiological publications, Karystianis et al. (2017) developed a set of rules based on syntactical patterns observed from the training documents. Notably, current methods utilize only textual information, while important research processes and experimental results are often reported via images and their captions in publications.

Figures and captions convey fundamental, essential information in biomedical documents. As such, there is a growing interest in storing, browsing and in utilizing images and their respective captions as a source of knowledge. In particular, biomedical databases are beginning to store and to display images as evidence for a variety of processes and for experimental results (Finger et al., 2017; Liechti et al., 2017). Notably, most current biomedical publications are stored as Portable Document Format (PDF). An essential step toward making use of images is the extraction of figures and captions from the PDF files of publications. Several systems have been developed for identifying and extracting figures and captions from scientific documents (Clark et al., 2016; Li et al., 2019).

Another obstacle toward utilizing biomedical images is the abundance of compound figures comprising multiple panels (see e.g. Fig. 1), where each panel often conveys a distinct information type obtained via one of several possible diverse modalities. For instance, both graphs and gel images may appear side-by-side as panels in a single figure providing evidence for similar or for distinct findings. In order to utilize the information from individual subfigures, it is essential to segment compound images into their constituent panels. Identifying compound figures and their constituent panels is a topic of much research (Chhatkuli et al., 2013; Santosh et al., 2015), including our own (Li et al., 2018).

Fig. 1.

Fig. 1.

An example of our pipeline for figures, subfigures and captions extraction. The original PDF document (Kobayashi et al., 2012) is shown on the left. Figures (dashed red boxes) and captions are first extracted from the document using PDFigCapX. Figures then be further separated into subfigures (solid blue boxes) using FigSplit

Image captions have been shown effective for document classification in several studies (Burns et al., 2019; Jiang et al., 2017, 2020; Regev et al., 2002). For instance, Burns et al. (2019) compared classification performance under different information sources, when identifying publications containing molecular interaction information, relevant to the IntAct Molecular Interaction database (Kerrien et al., 2012). Their experiments showed that a classifier utilizing figure captions outperformed classifiers using information from either the title-and-abstract, MeSH terms, body text or figure description from the body text. Our group is one of the first to use image content information for biomedical document classification (Ma et al., 2015; Shatkay et al., 2006). Shatkay et al. (2006) first proposed to use the class label of figures (such as: line chart, gel electrophoresis and fluorescence microscopy) as image features to identify publications that are relevant to the Gene Ontology annotation task performed by the Mouse Genome Informatics at the Jackson Laboratory. A more recent work from our group integrates information extracted from figures using Optical Character Recognition with text information for identifying documents that are relevant to cis-regulatory modules (Ma et al., 2015). However, none of the current methods use image captions and image contents together. Thus, we aim to integrate information from both image contents and their respective captions, in addition to titles-and-abstracts, toward improving biomedical document classification.

Here we introduce a new scheme that utilizes information from images, captions and title-and-abstracts toward improved biomedical document classification. To do this, we first extract figures, subfigures/panels and captions from the documents. In order to represent figures within biomedical documents, we propose a new image representation, namely, Figure-word that encodes the combination of different types of panels within a figure. An image taxonomy is also introduced and used to train a classifier for categorizing the extracted panels. For handling text-contents, we employ word embeddings (Moen and Ananiadou, 2013), for both caption-based and title-and-abstract-based document representations. To utilize all three types of information sources (images, captions and titles-and-abstracts), two information integration methods are introduced. The first combines Figure-words and features obtained from captions and title-and-abstract into a single larger vector for document representation; while the second employs a meta- classification scheme.

The rest of the paper presents the details of our method, and demonstrates its effectiveness through a series of experiments. Section 2 describes the complete framework of our method; Section 3 presents experiments and results, assessing its performance; Section 4 discusses and analyzes the results, while Section 5 concludes and outlines directions for future work.

2 Methods

Our goal is to identify biomedical documents that are relevant to a specific domain by utilizing images and captions along with titles-and-abstracts. To do that, we first extract figures and their captions from the PDF files of biomedical documents, by employing the parsing tool that we have developed—and is now publicly available—PDFigCapX (https://www.eecis.udel.edu/∼compbio/PDFigCapX) (Li et al., 2019). As many of the extracted figures are compound images comprising multiple panels, we also separate such figures into their constituent panels, using our previously developed FigSplit (https://www.eecis.udel.edu/∼compbio/FigSplit) system (Li et al., 2018) for compound image separation.

To represent images within biomedical documents, we first introduce an image taxonomy comprising 12 categories, which serves as framework for classifying biomedical figures. Next, we train an image classifier to categorize the extracted panels. We introduce a new image representation, namely, Figure-word, which encodes the combination of different types of panels in a figure, and use it to generate an image-based representation of each document, dIMG. Word embeddings, which convert a word to a numerical vector of a fixed number of dimensions have been prevalently used for text representation (Mikolov et al., 2013). As such, we use word embeddings pre-trained over a corpus of biomedical articles to generate, for each document d, its caption-based representation, dCAP, as well as its title-and-abstract-based representation, dTA. We introduce two information integration methods to utilize the information from images and captions, in addition to titles-and-abstracts. The first method concatenates the representations dIMG,dCAP and dTA into a single larger vector for representing each document, d. The second is a meta-classification approach, combining the output of base classifiers that were trained separately over images, captions and titles-and-abstracts to train the final document classifier.

2.1 Extracting figures, subfigures and captions from biomedical documents

To utilize image information, we first extract images and their corresponding captions from the PDF file of biomedical publications. Extracting figures and captions is not a simple task due to the complex and diverse layout of biomedical publications and the variations in figure structure, texture and contents. To extract images and their captions from biomedical publications, which are primarily stored as PDF files, we use PDFigCapX (Li et al., 2019). Unlike other methods that extract figures by handling raw encoded contents of PDF documents, PDFigCapX begins by separating text from graphical contents, utilizing layout information to detect and disambiguate figures and captions. Files containing the figures and their associated captions are produced as output.

The vast majority of the extracted figures are compound images consisting of multiple panels. In order to utilize image information from each individual panel, we use our FigSplit tool (Li et al., 2018), segmenting compound images into their constituent panels. Unlike other methods that segment images using gaps between panels, FigSplit identifies panels based on Connected Component Analysis. It also overcomes the common issues of over- and under-segmentation by evaluating and self-correcting candidate segmentations that are likely to be inaccurate.

Both systems, PDFigCapX and FigSplit, were tested on existing and on newly assembled datasets, demonstrating robustness and significant improvement compared to other state-of-the-art methods (Li et al., 2018, 2019). Figure 1 shows an example of our pipeline for extracting figures, subfigures and captions from biomedical publications. The input to our pipeline is the original PDF document shown on the left. By using PDFigCapX, figure and caption pairs are extracted. The extracted images are shown in red dashed boxes. By applying FigSplit, compound images are split into their constituent panels—each shown within a solid blue box on the right.

2.2 Image-based document representation

Figures in biomedical publications are typically used to show the process and results of experiments. Different types of images are used to report certain types of experiments. For example, gel images are typically used in pull-down assays (Orchard et al., 2012). Class labels of figures have been shown useful for document representation in biomedical document classification in our previous work (Shatkay et al., 2006). As discussed in Section 2.1, the majority of figures within biomedical publications are compound images. Building upon our previous idea, we introduce here a new method to represent figures within documents based on class labels of their constituent panels here.

While several image taxonomies were proposed for classifying biomedical images (De Herrera et al., 2016; Lopez et al., 2013; Shatkay et al., 2006), as no standard exists, we extend the image taxonomy previously proposed by our group, through collaboration with GXD (Finger et al., 2017), Protein Information Resource (Wu et al., 2003) and WormBase (Harris et al., 2020), as shown in Figure 2. At the top level, images are classified into Graphics, Molecular Structure, Experimental and Other images. At the second level, Graphics are classified into Histogram, Line Chart and Other Diagram. Molecular Structure images are classified into Macromolecule Sequence and 3D Structure images. Experimental images are further classified into Fluorescence Microscopy, Light Microscopy, Whole Mount, Gel and Plate images. We also note that figure legends or margins are sometimes over-separated from their original images by the compound image separation process, thus forming individual panels. We refer to such panels formed by over-segmentation of compound images as separation residuals. As these residuals do not belong to any of the informative taxonomy’s classes, we augment our taxonomy with a separation residual class.

Fig. 2.

Fig. 2.

The image taxonomy used for panel classification

To automatically assign class label to individual panels, we build an image classifier. A pre-trained Convolutional Neural Network, VGG16 (Andrearczyk and Müller, 2018; Simonyan and Zisserman, 2015), is used for image classification. To train the classifier, we use the annotated image dataset that was introduced by Lopez et al. (2013) based on the Molecular INTeraction database dataset (Licata et al., 2012). The image dataset consists of 34 876 pre-labeled panels; its statistics are shown in Table 1. In addition, a set of 500 labeled whole mount images were provided by GXD. Trained and tested via a 5-fold cross validation, the classifier demonstrates 87.89% accuracy.

Table 1.

Distribution of image types included in our experiments, based on the image dataset introduced by Lopez et al. (2013)

Histogram Line chart Other diagram Macromolecule sequence 3D structure Fluorescence microscopy Gel/blot Plate Light microscopy Other Separation residual
No. of panels 4270 2664 3536 499 1424 5714 14865 508 1156 130 110

Once the class label of each panel is obtained, we represent each figure as an 11-dimensional binary vector <c1,c2,,ci,,c11>, where ci is 1 if a panel from class i is present in the figure and 0 otherwise. For instance, if the figure comprises only histograms and fluorescence microscopy panels (panels of type1 and type6 respectively), its corresponding vector is: <1, 0, 0, 0, 0, 1, 0, 0, 0, 0, 0>. We refer to each such vector as a Figure-word. Figure 3 shows the process for converting figures into their corresponding Figure-words. As the number of classes in our image-taxonomy is 11, the total number of possible Figure-words in our vocabulary is 211 (2048). A document d, in turn, is represented as a vector dIMG=<I1,I2,,Ii,,In>, where n=211 and Ii (1i2048) is 1 if the ith Figure-word appears in the document d, 0 otherwise.

Fig. 3.

Fig. 3.

The process for converting figures into their corresponding Figure-words. The set of figures extracted from the biomedical documents is shown on the left. The corresponding Figure-words along with their vector representations indicating the types of comprising panels, are shown on the right. The images on the left are taken from (Li et al., 2008, Fig. 4; Qiu and Dhe-Paganon, 2011, Fig. 6; Sugioka et al., 2014, Fig.3; Leung et al., 2012, Fig. 9; Heinonen et al., 2015, Fig. 4; Dai et al., 2012, Fig. 3)

2.3 Caption-based document representation

Captions associated with figures provide another important source of information for biomedical document classification. In order to make use of captions, we employ a standard preprocessing procedure that includes named-entity recognition (NER), stemming and stop-words removal as we have done in our earlier work (Jiang et al., 2017, 2020). For NER, we first identify all gene, disease, chemical, species, mutation and cell-line concepts using PubTator, which is widely used for annotations of biomedical concepts (Wei et al., 2019). We then substitute each of the identified concepts by its respective generic terms ‘gene’, ‘disease’, ‘chemical’, ‘species’, ‘mutation’ or ‘cell-line’. We also stem words using the Porter stemmer and remove standard stop words (Porter, 1980; Canese and Weis, 2013). The concatenated captions are used as the caption-based document representation.

Word embeddings map words to vectors of a fixed dimension so that words appearing in similar contexts are mapped to similar vectors. Such a vectorized representation has been widely used for text classification (Moen and Ananiadou, 2013), and more recently for biomedical named entity recognition and biomedical relation extraction (Lee et al., 2020). A word embedding model (Moen and Ananiadou, 2013) has been pre-trained specifically on a biomedical corpus, which consists of PubMed titles-and-abstracts and PubMed Central full text articles by employing the word2vec tool (Mikolov et al., 2013). We use such embeddings to represent the concatenated captions. Each word, wi, within the concatenated captions is converted to a word embedding vector wi whose dimensionality is 200. The document d is then represented as a 200-dimentional vector dCAP, calculated as dCAP=1n(w1+w2++wn), where n is the total number of distinct words in the concatenated captions, and wi is the embedding vector of the ith distinct word.

2.4 Title-and-abstract-based document representation

The title and the abstract of articles are the text components most often used for biomedical document classification. To represent a document based on those, we first obtain the title-and-abstract of each publication. Similar to the steps described in Section 2.3, we employ a standard preprocessing procedure that includes named-entity recognition, stemming and stop-words removal to each title-and-abstract. The same word embeddings described in Section 2.3 are employed to convert each word wi in the preprocessed text to a word embedding vector wi. The document d is then represented as a 200-dimensional vector, denoted as dTA, by calculating the mean of embedding vectors that are associated with words in the preprocessed text.

2.5 Information integration for document classification

So far, we have introduced document representations based on images (dIMG), captions (dCAP) and title-and-abstracts (dTA). Next, we present two schemes for integrating the information stemming from these three sources.

(1) Integration via concatenated vectors

Under this scheme, to represent a document d, we simply concatenate the vectors dIMG,dCAP and dTA into a single vector dALL, thus utilizing the information obtained from images, captions and titles-and-abstracts. Recall that the value of an entry in the dIMG vector is either 1 or 0 which indicates whether or not a Figure-word appears in a document, while the value of an entry in dCAP or dTA is obtained by calculating the mean of embedding vectors converted from words in a caption or a title-and-abstract. There is no specific limit on the range of embedding vectors. Therefore, the values of such entries are at different scales from that of entries in dIMG. As such, we standardize each feature within dALL by rescaling the features such that they have a mean of 0 and a standard deviation of 1. For classifying the documents, we conducted experiments with several classification schemes, including Random Forests, Naïve Bayes (not shown here) and SVMs. As SVMs have been commonly used for both image and text classification, and have shown the best performance in this context (Holzinger et al., 2014; Simpson et al., 2015), we use SVM as the model for classifying the final resulting document-vectors, and denote this SVM classifier CombV.

(2) Integration via meta-classification

Another approach we propose toward integrating the multiple types of information is to employ a meta-classification scheme, which combines results obtained from multiple base classifiers into a single classification output. To do that, we first separately train three base classifiers CIMG,CCAP and CTA using the representations of images (dIMG), captions (dCAP) and titles-and-abstracts (dTA). By applying a base classifier to a document d, we obtain the class label L and the probability P of document d to be assigned to the relevant class. Each document d is then represented as a 6-dimensional vector dCombC=<LIMG,PIMG,LCAP,PCAP,LTA,PTA>. This representation is then used for training another classifier, referred to as meta-classifier, denoted as CombC, which assigns the final class label to each document. Similar to the concatenation-based integration, we use SVMs both as base classifiers and as the ultimate classifier in the meta-classification.

3 Experiments and results

To evaluate our method we conduct two sets of experiments. The first aims to compare the classification performance obtained when using only a single type of information to represent documents versus the performance when employing a representation that combines all three types of information. The classifiers using the representations of Figure-words, captions and titles-and-abstracts, are denoted as CIMG,CCAP and CTA, respectively.

In the second set of experiments, we compare the performance of our system when utilizing a representation that combines images, captions and titles-and-abstracts to the performance attained by three state-of-the-art systems. The first system to which we compare is a random forest-based method (RFCAP) developed by Jiang et al. (2017) for identifying publications that are relevant to GXD. This classifier uses features extracted from the title-and-abstract and from caption text. The second is a convolutional neural network triage system (CNNBiLSTM) presented by Burns et al. (2019) for identifying publications containing information about molecular interactions. The CNNBiLSTM classifier uses captions only. The third is a hierarchical recurrent neural network classification system (HRNN) developed by Fergadis et al. (2018) for identifying publications containing information about protein-protein interactions affected by genetic mutations; and uses title-and-abstract only. We compare all three systems using the code provided by their respective authors. For comparison, we run five complete rounds of 5-fold cross validation with 5 different 5-way partitions of the dataset. All experiments are conducted on a DELL machine that uses an Intel Core i7-6700 processor, an Nvidia GTX 1070 GPU, 8 GB of RAM and 256 GB of SSD.

3.1 Datasets and evaluation

In our experiments, we use two datasets for which we have the ground-truth class-labels. The first dataset, denoted GXD2000, is a subset of the dataset used by Jiang et al. (2017), who is also the developer of RFCAP. The original dataset is a collection of 58 362 publications (provided as PDF), curated by the Jackson Lab’s GXD throughout the years 2012–2016. As a first test of our method, we selected at random 1000 relevant and 1000 irrelevant documents from these publications, while retaining the same distribution of publication-years as in the larger GXD dataset. In order to use figures and captions, we first apply PDFigCapX to the GXD2000 dataset. 8939 figures and 8594 captions are extracted from the relevant publications, while 8414 figures and 8042 captions are extracted from the irrelevant publications. We note that the number of figures extracted exceeds that of the captions, as some pages display figures (or parts of figures) without associated captions. FigSplit is then applied to separate compound figures into their constituent panels, resulting in 60 194 individual panels extracted from the 8939 figures associated with relevant publications and 41 015 panels obtained from the 8414 figures associated with the irrelevant publications.

The second dataset used in our experiments, denoted DSP, was introduced by Burns et al. (2019) for testing their system CNNBiLSTM. It comprises 537 publications relevant to molecular interactions and 451 irrelevant ones spanning the year range 1996-2017. Only publications for which PDF files are available are used in our experiments. As such, out of the 537 relevant publications only 534 are used, while out of the 451 irrelevant ones only 448 are retained, as their PDF files were available for download online. We then apply PDFigCapX and FigSplit to identify and extract figures, captions and constituent panels of extracted figures. From the 534 relevant publications, 3975 figures, 3912 captions and 21 421 panels are extracted, while 2928 figures, 2832 captions and 14 224 panels are extracted from the 448 irrelevant ones. Table 2 shows the statistics for these two datasets.

Table 2.

The number of figures, captions and panels identified and extracted from publications in the datasets used in our experiments

Datasets Classes No. of docs No. of figures No. of captions No. of panels
GXD2000 Relevant 1000 8939 8594 60 194
Irrelevant 1000 8414 8042 41 015
DSP Relevant 534 3975 3912 21 421
Irrelevant 448 2928 2832 14 224

The total time for PDFigCapX to process the GXD2000 dataset of 2000 publications is about 5.9 h (10.60 s per document, wall clock) where the average document contains 8.7 figures, 8.3 captions and is 7.2MB in size. It takes about 4.3 h (0.83 s per image, wall-clock) for FigSplit to process all extracted figures where on average 50.6 panels are extracted from each publication within the GXD2000 dataset. Over the DSP dataset, PDFigCapX takes about 2.4 h (8.69 s per document, wall-clock) to process all 982 publications where the average document contains 7.0 figures, 6.9 captions and the average file size is 2.8MB. FigSplit takes about 1.6 h (0.78 s per image, wall-clock) to process all extracted figures where on average 36.3 panels are extracted from each publication in the DSP dataset.

To evaluate the document classification performance, we use standard measures, Precision, Recall and F-score defined as:

graphic file with name btab331um1.jpg

3.2 Results

Table 3 presents the classification performance attained when using only a single type of information to represent documents, along with the performance when employing a representation that combines all three types of information in our first set of experiments. The second to fourth columns in the table show results from experiments over the GXD2000 dataset. Our classifier CIMG, utilizing the Figure-words alone, attained 80.5% precision, 77.0% recall and 78.7% F-score. By using information from the caption alone, our classifier CCAP attained 88.6% in precision, 87.1% in recall and 87.8% in F-score. A similar performance (87.5% precision, 87.7% recall and 87.6% F-score) is attained by the classifier CTA when titles-and-abstracts are used to represent documents. Notably, a significant improvement is attained by using the representation that combines information from images, captions and titles-and-abstracts. By employing the meta-classification scheme, our classifier CombC attained 88.7% precision, 89.9% recall and 89.3% F-score. Our classifier CombV attained the highest performance of 89.4% precision, 91.0% recall and 90.2% F-score when the concatenated vectors are used for document representation. The performance attained by using the integrated information is statistically significantly higher than the performance attained based on the Figure-words, the captions or the titles-and-abstracts alone (P<0.01, two-sample t-test for all measures).

Table 3.

Classification performance attained by using information from images (CIMG), captions (CCAP), the title-and-abstract (CTA), concatenated vectors from all three types (CombV), and by using the meta-classifier (CombC). The highest values attained are shown in boldface. Standard deviations are shown in parentheses.

GXD2000
DSP
Classifiers Precision Recall F-score Precision Recall F-score
CIMG 0.805 (.021) 0.770 (.026) 0.787 (.021) 0.679 (.018) 0.768 (.026) 0.731 (.026)
CCAP 0.886 (.027) 0.871 (.032) 0.878 (.021) 0.804 (.024) 0.809 (.034) 0.806 (.021)
CTA 0.875 (.021) 0.877 (.015) 0.876 (.013) 0.790 (.023) 0.807 (.023) 0.798 (.015)
CombC 0.887 (.019) 0.899 (.025) 0.893 (.008) 0.822 (.032) 0.826 (.044) 0.823 (.020)
CombV 0.894 (.019) 0.910 (.017) 0.902 (.008) 0.831 (.014) 0.834 (.031) 0.832 (.019)

The three rightmost columns in Table 3 present the results attained over the DSP dataset. When the Figure-words are used for document representation, our classifier CIMG attained 69.7% precision, 76.8% recall and 73.1% F-score. The classifier CCAP attained 80.4% precision, 80.9% recall and 80.6% F-score. A similar performance (79.0% precision, 80.7% recall and 79.8% F-score) is attained by CTA when titles-and-abstracts are used for document representation. Again, a significant improvement is attained by using the integrated information. We attained 82.2% precision, 82.6% recall and 82.3% F-score when the meta-classification scheme is applied. The highest performance of 83.1% precision, 83.4% recall and 83.2% F-score is attained when the concatenated vectors are used to represent documents. The performance attained by classifiers that integrate information from images, captions and titles-and-abstracts is statistically significantly higher than the performance attained by classifiers that are based on single information source (P<0.01, two-sample t-test for all measures). Our results demonstrate that our information integration schemes indeed improve biomedical document classification.

Table 4 compares the performance of our classifier CombV to that attained by the three other state-of-the-art systems, RFCAP, CNNBiLSTM and HRNN. Over the GXD2000 dataset, RFCAP attained 82.9% recall, 87.8% F-score and the highest precision of 93.4%. CNNBiLSTM achieved 87.6% precision, 85.0% recall and 86.2% F-score, while HRNN attained 85.5% in precision, 87.5% in recall and 86.4% in F-score. While the precision (89.4%) attained by our classifier is slightly lower than that reached by RFCAP, our classifier CombV attained the highest recall of 91.0% and the highest F-score of 90.2% over the GXD2000 dataset. Notably, recall is often viewed as more important than precision for biomedical document curation (Fang et al., 2012; Müller et al., 2004). Moreover, the differences between the results obtained by our system and those attained by RFCAP, CNNBiLSTM and HRNN are statistically significant (P<0.001, two-sample t-test). The three rightmost columns in Table 4 presents the results attained over the DSP dataset. RFCAP achieved 79.8% precision, 80.9% recall and 79.8% F-score, while HRNN attained 75.4% precision, 81.8% recall and an F-score of 78.3%. The CNNBiLSTM system, whose author introduced the DSP dataset itself, reached 82.0% precision, 79.6% recall and 80.9% F-score. Our classifier attained the highest performance of 83.1% precision, 83.4% recall and 83.2% F-score over the DSP dataset. Moreover, the differences between the results obtained by our classifier and those attained by other state-of-the-art systems are statistically significant (P<0.001, two-sample t-test).

Table 4.

Classification performance Comparison with other state-of-the-art systems. The highest values attained are shown in boldface. Standard deviations are shown in parentheses.

GXD2000
DSP
Classifiers Precision Recall F-score Precision Recall F-score
RFCAP 0.934 (.017) 0.829 (.037) 0.878 (.018) 0.798 (.068) 0.809 (.057) 0.798 (.030)
CNNBiLSTM 0.876 (.028) 0.850 (.031) 0.862 (.013) 0.820 (.023) 0.796 (.030) 0.809 (.018)
HRNN 0.856 (.044) 0.875 (.033) 0.864 (.010) 0.754 (.032) 0.818 (.054) 0.783 (.017)
CombV 0.894 (.019) 0.910 (.017) 0.902 (.008) 0.831 (.014) 0.834 (.031) 0.832 (.019)

4 Discussion

Notably, Figure-words provide important information for document classification. The image-based classifier, CIMG, which uses the newly proposed Figure-words alone for document representation attained 80.5% precision, 77.0% recall and 78.7% F-score over the GXD2000 dataset, while attaining 69.7% precision, 76.8% recall and 73.1% F-score when applied to the DSP dataset (Table 3). Moreover, Figure-words provide information distinct from that captured by captions or by titles-and-abstracts. Of the GXD2000 dataset, 71 relevant publications (7.1% of the relevant publications) were correctly identified by CIMG, but incorrectly classified by CCAP, while 71 publications were correctly identified by CIMG, but incorrectly classified by CTA. Of the DSP dataset, 59 relevant publications (11.0% of the relevant data) were correctly identified by CIMG, but incorrectly classified by CCAP, while 56 publications (10.5% of the relevant data) were correctly identified by CIMG, but incorrectly classified by CTA.

Another noteworthy point is that captions provide distinct information from that provided by titles-and-abstracts for document classification. As indicated in Section 3.2, the performances attained using a classifier based on captions or titles-and-abstracts alone (CCAP and CTA, respectively) are similar over both the GXD2000 and the DSP datasets. However, the relevant publications identified by classifiers CCAP and CTA are quite different. Of the GXD2000 dataset, 60 relevant publications (6.0% of the relevant ones) were correctly identified by CCAP, but incorrectly classified by CTA, while 66 distinct publications were correctly identified by CTA, but incorrectly classified by CCAP. Of the DSP dataset, 38 relevant publications (7.1% of the relevant data) were correctly identified by CCAP, but incorrectly classified by CTA, while 37 distinct publications were correctly identified by CTA, but incorrectly classified by CCAP. By their very nature, titles-and-abstracts form a high-level summary of an entire study, while captions present details of experimental processes and results. This difference is reflected in the vocabulary of titles-and-abstracts versus that of captions. For instance, words such as anterior, WT, dorsal, embryo, green, lateral and mount, are commonly found in captions of publications relevant to GXD when describing gene expression experiments in mouse embryos. As such, captions provide information that is distinct from that provided through titles-and-abstracts, thus supporting more effective classification.

Figure 4 illustrates the respective classification results attained by the classifiers CIMG, CCAP and CTA. Of the GXD2000 dataset, 18 relevant publications are identified only by CCAP, while 24 relevant ones are identified only by CTA. Notably, 29 relevant publications can only be identified by CIMG using Figure-words for document representation. Of the DSP dataset, classifier CIMG identified 32 relevant publications that are distinct from those identified by CCAP and CTA. These findings strongly suggest that the three data sources, namely, Figure-words, captions and titles-and-abstracts provide distinct and complementary information for document classification.

Fig. 4.

Fig. 4.

Comparison of classification results between the classifiers CIMG,CCAP and CTA. The solid circle corresponds to the relevant publications that are correctly identified by classifier CIMG. The relevant publications correctly classified by CCAP are indicated as a dashed circle, while those correctly identified by CTA are shown as a dotted circle. The region marked by horizontal stripes indicates the relevant publications identified by classifier CIMG only. The region marked by grid pattern corresponds to the relevant publications classified by CCAP only, while the region shown in solid gray indicates those identified by CTA only. (a) The comparison over the GXD2000 dataset. (b) The comparison over the DSP dataset

To better understand the contribution of Figure-words to improved classification, we identify the most distinguishing Figure-words by ranking them based on the Z-score Test (Myers et al., 1993), as we have done before for identifying distinguishing text terms (Jiang et al., 2017, 2020). Table 5a shows the top-5 scoring Figure-words, along with their occurrence frequency in the relevant and in the irrelevant publications of the GXD2000 dataset. There is a significant difference between the Figure-word distribution in relevant publications and their distribution in irrelevant ones. For instance, there are 1339 images consisting of fluorescence alone in relevant publications of the GXD2000, while only 437 such images in the irrelevant publications. Similarly, Table 5b shows that the distinguishing Figure-words identified with respect to the DSP dataset, demonstrate a clear difference in Figure-word distribution between relevant publications and irrelevant ones. Therefore, we believe that our newly proposed Figure-words have much potential for improving biomedical document classification.

Table 5.

Top scoring Figure-words that contribute to the document classification task. The left most column in each table shows the Figure-words and their corresponding vectors indicating the types of comprising panels. The other two columns show the occurrence frequencies of corresponding figures in the relevant and in the irrelevant dataset, respectively. (a) Top Figure-words identified over the GXD2000 dataset. (b) Top Figure-words identified over the DSP dataset.

graphic file with name btab331ilf1.jpg

Figure 5 shows examples of Figure-words. Notably, the top scoring Figure-words also correspond to images that are typically found in distinct biomedical experiments. For instance, fluorescence microscopy images often appear in publications relevant to the GXD as this is a common imaging technique for visualizing gene expression. As curators at the GXD focus on mouse embryo studies, Figure-words containing embryonic whole-mount images are also indicative of documents that are likely relevant to GXD. Similarly, Co-inmmunoprecipitation, and Pull Down experiments are commonly used in studies relevant to molecular interactions, thus Figure-words corresponding to the gel/blot images are important for document classification over the DSP dataset (Fig. 5b). As such, our newly proposed Figure-words compactly account for and convey certain types of biomedical experiments. Experimental evidence is important for identifying relevant biomedical documents (Burns et al., 2018; Han et al., 2006), thus our Figure-words can contribute much informative evidence to the document classification task.

Fig. 5.

Fig. 5.

Examples of images for top scoring Figure-words. The leftmost column shows Figure-words along with their vector representations and their respective combination of panel types. The rest of the columns show examples of corresponding images. (a) Image examples from the GXD2000 dataset. Original images are taken from (Vogt et al., 2012, Fig. 1; Grimsley-Myers et al., 2012, Fig. 7; Yamaguchi et al., 2014, Fig. 5; Quattrocolo & Maccaferri, 2014, Fig. 1; Liu et al., 2012, Fig. 6; Rooijen et al., 2012, Fig. S3). (b) Image examples from the DSP dataset. Original images are taken from (Shinohara et al., 2005, Fig. 3; Cheng et al., 2016, Fig. 4; Mysling et al., 2016, Fig. 1; Yoshida et al., 2014, Fig. 3; Graef et al., 2009, Fig. 6; Li et al., 2008, Fig. 4)

As discussed above, Figure-words, captions and titles-and-abstracts provide complementary information for document classification. A significant improvement is obtained by using the representation comprising all three sources. In our experiments, we attained statistically significantly improved performance by employing the meta-classification scheme CombC as well as by employing the classifier, CombV, where concatenated vectors are used for document representation, as compared to classification based on the title-and-abstracts, the captions and the Figure-words alone.

In the second set of experiments when our system is compared against three state-of-the-art systems, over the GXD2000 dataset, two classifiers (CNNBiLSTM, HRNN) utilizing only a type of information source attained similar performance. The RFCAP classifier which uses features extracted from combined title-and-abstract and caption text, indeed outperforms the systems that use just a single type of information. Our method, which integrates the information from Figure-words, captions and titles-and-abstracts attained the highest recall and F-score over the GXD2000 dataset. When applied to the DSP dataset, our method attained the highest score across all measures. These results demonstrate that Figure-words provide information distinct from that provided by titles-and-abstracts and by captions for supporting classification, and also prove the effectiveness of the integration methods that we introduced.

While our method indeed improves classification performance, there is still room for improvement, especially for the image taxonomy. In the work reported here, we utilized the image taxonomy consisting of 11 categories, as there is no unique standard image taxonomy for categorizing biomedical research images yet. A more comprehensive taxonomy has the potential to support a more informative Figure-words vocabulary and as such improve the overall document classification results. In our future work, we plan to expand and refine the image taxonomy we employ for categorizing biomedical research images. We are already in the process of applying our classification scheme to a larger dataset, namely the COVID-19 open research dataset comprising more than 50 000 articles (Chen et al., 2020; Wang et al., 2020), and plan to further apply it to the complete GXD dataset of more than 58 000 publications used by Jiang et al. (2017).

As our newly proposed Figure-words correspond to distinct images used in certain types of biomedical experiments, our image representation method can help biocurators identify images according to their experiment types. For example, a biocurator may want to identify images used in yeast two-hybrid experiments based on the images used to describe such experiments. We will also investigate the usage of our image representation for other tasks, such as biomedical image classification.

5 Conclusion

We presented a new scheme for identifying biomedical documents that are relevant to a certain domain, by using information derived from both images and captions, as well as from titles-and-abstracts. To do so, we first employed a pipeline for processing biomedical documents, comprising two parts: PDFigCapX that extracts figures with their captions from documents, and FigSplit for splitting the extracted compound figures into constituent panels, to biomedical documents.

A new image representation, Figure-word that encodes the combination of different types of panels is proposed for representing figures within documents. For captions and titles-and-abstracts, word embeddings are employed to represent documents as vectors. To utilize both image and caption information, in addition to titles-and-abstracts, we introduced two information integration methods. The first concatenates Figure-words and features obtained from captions and titles-and-abstracts into a single larger vector for document representation; the second employs a meta-classification scheme. Our experiments demonstrate the effectiveness of the newly proposed Figure-words for representing images. Moreover, classification performance is improved through the integration of information from all three sources, namely, images, captions and titles-and-abstracts.

As part of future work, we plan to build a more comprehensive taxonomy for refining image classification and improving the document classification performance. It is noteworthy that the newly proposed Figure-words correspond to certain distinct types of images used in reporting biomedical experiments. We will investigate the potential usage of Figure-words for other tasks, such as biomedical image classification.

Funding

This work was partially supported by National Institutes of Health (NIH)/National Library of Medicine (NLM) awards [R56LM011354A and R01LM012527]; NIH/National Institute of Child Health and Human Development (NICHD) award [P41 HD062499 to M.R.].

Conflict of Interest: none declared.

Contributor Information

Pengyuan Li, Department of Computer and Information Sciences, University of Delaware, Newark, DE 19716, USA.

Xiangying Jiang, Department of Computer and Information Sciences, University of Delaware, Newark, DE 19716, USA; Amazon, Seattle, WA 98109, USA.

Gongbo Zhang, Department of Computer and Information Sciences, University of Delaware, Newark, DE 19716, USA; Google, Mountain View, CA 94043, USA.

Juan Trelles Trabucco, Department of Computer Science, The University of Illinois at Chicago, Chicago, IL 60612, USA.

Daniela Raciti, Division of Biology and Biological Engineering, California Institute of Technology, Pasadena, CA 91125, USA.

Cynthia Smith, The Jackson Laboratory, Bar Harbor, ME 04609, USA.

Martin Ringwald, The Jackson Laboratory, Bar Harbor, ME 04609, USA.

G Elisabeta Marai, Department of Computer Science, The University of Illinois at Chicago, Chicago, IL 60612, USA.

Cecilia Arighi, Department of Computer and Information Sciences, University of Delaware, Newark, DE 19716, USA.

Hagit Shatkay, Department of Computer and Information Sciences, University of Delaware, Newark, DE 19716, USA.

References

  1. Almeida H.  et al. (2014) Machine learning for biomedical literature triage. PLoS One, 9, e115892. [DOI] [PMC free article] [PubMed] [Google Scholar]
  2. Andrearczyk V., Müller H. (2018) Deep multimodal classification of image types in biomedical journal figures. In: Proc. of CLEF, pp. 3–14. Avignon, France. [Google Scholar]
  3. Bateman A.  et al. (2021) UniProt: the universal protein knowledgebase in 2021. Nucleic Acids Res., 49, D480–D489. [DOI] [PMC free article] [PubMed] [Google Scholar]
  4. Blake J.A.  et al. ; The Mouse Genome Database Group. (2021) Mouse Genome Database (MGD): knowledgebase for mouse-human comparative biology. Nucleic Acids Res., 49, D981–D987. [DOI] [PMC free article] [PubMed] [Google Scholar]
  5. Burns G.  et al. (2018) Towards evidence extraction: analysis of scientific figures from studies of molecular interactions. In Proc. of ISWC, pp. 95–102. Monterey, CA, USA. [Google Scholar]
  6. Burns G.A.  et al. (2019) Building deep learning models for evidence classification from the open access biomedical literature. Database, 2019, baz034. [DOI] [PMC free article] [PubMed] [Google Scholar]
  7. Canese K., Weis S. (2013) PubMed: the bibliographic database. In: The NCBI Handbook [Internet]. 2nd edn. National Center for Biotechnology Information. Bethesda, MD, USA. [Google Scholar]
  8. Chatr-Aryamontri A.  et al. (2017) The BioGRID interaction database: 2017 update. Nucleic Acids Res., 45, D369–D379. [DOI] [PMC free article] [PubMed] [Google Scholar]
  9. Chen Q.  et al. (2020) LitCovid: an open database of COVID-19 literature. Nucleic Acids Res., 49, D1534–D1540. [DOI] [PMC free article] [PubMed] [Google Scholar]
  10. Cheng L.F.  et al. (2016) Incorporation of GM-CSF or CD40L enhances the immunogenicity of Hantaan virus-like particles. Front. Cell. Infect. Microbiol., 6, 185. [DOI] [PMC free article] [PubMed] [Google Scholar]
  11. Chhatkuli A.  et al. (2013) Separating compound figures in journal articles to allow for subfigure classification. Proc. SPIE Med. Imaging, 8674, 86740J. [Google Scholar]
  12. Clark C.  et al. (2016) Pdffigures 2.0: mining figures from research papers. In Proc. of IEEE/ACM JCDL, pp. 143–152. Newark, NJ, USA. [Google Scholar]
  13. Dai B.  et al. (2012) A comparative transcriptome analysis identifying FGF23 regulated genes in the kidney of a mouse CKD model. PLoS One, 7, e44161. [DOI] [PMC free article] [PubMed] [Google Scholar]
  14. De Herrera A.G.S.  et al. (2016) Overview of the ImageCLEF 2016 medical tasks. In: Working Notes of ImageCLEF, pp. 219–232. Evora, Portugal. [Google Scholar]
  15. Fang R.  et al. (2012) Automatic categorization of diverse experimental information in the bioscience literature. BMC Bioinformatics, 13, 16. [DOI] [PMC free article] [PubMed] [Google Scholar]
  16. Fergadis A.  et al. (2018) Hierarchical bi-directional attention-based RNNs for supporting document classification on protein–protein interactions affected by genetic mutations. Database, 2018, bay076. [DOI] [PMC free article] [PubMed] [Google Scholar]
  17. Finger J.H.  et al. (2017) The mouse gene expression database (GXD): 2017 update. Nucleic Acids Res., 45, D730–D736. [DOI] [PMC free article] [PubMed] [Google Scholar]
  18. Garcia M.A.M.  et al. (2015) Biomedical literature classification using encyclopedic knowledge: a Wikipedia-based bag-of-concepts approach. PeerJ, 3, e1279. [DOI] [PMC free article] [PubMed] [Google Scholar]
  19. Graef T.  et al. (2009) KIR2DS4 is a product of gene conversion with KIR3DL2 that introduced specificity for HLA-A 11 while diminishing avidity for HLA-C. J. Exp. Med., 206, 2557–2572. [DOI] [PMC free article] [PubMed] [Google Scholar]
  20. Grimsley-Myers C.M.  et al. (2012) Redundant functions of RAC GTPases in inner ear morphogenesis. Dev. Biol., 362, 172–186. [DOI] [PMC free article] [PubMed] [Google Scholar]
  21. Harris T.W.  et al. (2020) WormBase: a modern model organism information resource. Nucleic Acids Res., 48, D762–D767. [DOI] [PMC free article] [PubMed] [Google Scholar]
  22. Han B.  et al. (2006) Substring selection for biomedical document classification. Bioinformatics, 22, 2136–2142. [DOI] [PubMed] [Google Scholar]
  23. Heinonen H.  et al. (2015) Identification of several potential chromatin binding sites of HOXB7 and its downstream target genes in breast cancer. Int. J. Cancer, 137, 2374–2383. [DOI] [PMC free article] [PubMed] [Google Scholar]
  24. Hu Z.Z.  et al. (2005) Literature mining and database annotation of protein phosphorylation using a rule-based system. Bioinformatics, 21, 2759–2765. [DOI] [PubMed] [Google Scholar]
  25. Holzinger A.  et al. (2014) Biomedical text mining: state-of-the-art, open problems and future challenges. In Interactive Knowledge Discovery & Data Mining in Biomedical Informatics, Ch. 16. Springer, Berlin, Heidelberg, 271–300. [Google Scholar]
  26. Jiang X.  et al. (2017) Effective biomedical document classification for identifying publications relevant to the mouse Gene Expression Database. Database, 2017, bax017. [DOI] [PMC free article] [PubMed] [Google Scholar]
  27. Jiang X.  et al. (2020) Integrating image caption information into biomedical document classification in support of biocuration. Database, 2020, baaa024. [DOI] [PMC free article] [PubMed] [Google Scholar]
  28. Karystianis G.  et al. (2017) Evaluation of a rule-based method for epidemiological document classification towards the automation of systematic reviews. J. Biomed. Inf., 70, 27–34. [DOI] [PubMed] [Google Scholar]
  29. Kerrien S.  et al. (2012) The IntAct molecular interaction database in 2012. Nucleic Acids Res., 40, D841–D846. [DOI] [PMC free article] [PubMed] [Google Scholar]
  30. Kobayashi C.  et al. (2012) GMCSF-independent CD1a expression in epidermal Langerhans cells: evidence from human CD1A genome-transgenic mice. J. Invest. Dermatol., 132(1), 241–244. [DOI] [PMC free article] [PubMed] [Google Scholar]
  31. Lee J.  et al. (2020) BioBERT: a pre-trained biomedical language representation model for biomedical text mining. Bioinformatics, 36, 1234–1240. [DOI] [PMC free article] [PubMed] [Google Scholar]
  32. Leung L.  et al. (2012) Apolipoprotein E4 causes age-and sex-dependent impairments of hilar GABAergic interneurons and learning and memory deficits in mice. PLoS One, 7, e53569. [DOI] [PMC free article] [PubMed] [Google Scholar]
  33. Li P.  et al. (2018) Compound image segmentation of published biomedical figures. Bioinformatics, 34, 1192–2299. [DOI] [PMC free article] [PubMed] [Google Scholar]
  34. Li P.  et al. (2019) Figure and caption extraction from biomedical documents. Bioinformatics, 35, 4381–4388. [DOI] [PMC free article] [PubMed] [Google Scholar]
  35. Li Y.  et al. (2008) Crystal structure of human XLF/Cernunnos reveals unexpected differences from XRCC4 with implications for NHEJ. EMBO J., 27, 290–300. [DOI] [PMC free article] [PubMed] [Google Scholar]
  36. Licata L.  et al. (2012) MINT, the molecular interaction database: 2012 update. Nucleic Acids Res., 40, D857–D861. [DOI] [PMC free article] [PubMed] [Google Scholar]
  37. Liechti R.  et al. (2017) SourceData: a semantic platform for curating and searching figures. Nat. Methods, 14, 1021–1022. [DOI] [PubMed] [Google Scholar]
  38. Liu Y.  et al. (2012) Sexually dimorphic BDNF signaling directs sensory innervation of the mammary gland. Science, 338, 1357–1360. [DOI] [PMC free article] [PubMed] [Google Scholar]
  39. Lopez L.D.  et al. (2013) An image-text approach for extracting experimental evidence of protein-protein interactions in the biomedical literature. In Proc. of ACM BCB, pp. 412–418. Washington, DC, USA. [Google Scholar]
  40. Ma K.  et al. (2015) Utilizing image-based features in biomedical document classification. In Proc. of IEEE ICIP, pp. 4451–4455. Quebec city, Quebec, Canada. [Google Scholar]
  41. Mikolov T.  et al. (2013) Distributed representations of words and phrases and their compositionality. In Proc. of NIPS, pp. 3111–3119. Lake Tahoe, NV, USA. [Google Scholar]
  42. Moen S.P.F.G.H., Ananiadou T.S.S. (2013) Distributional semantics resources for biomedical text processing. In Proc. of LBM, pp. 39–44. Tokyo, Japan. [Google Scholar]
  43. Müller H.M.  et al. (2004) Textpresso: an ontology-based information retrieval and extraction system for biological literature. PLoS Biol., 2, e309. [DOI] [PMC free article] [PubMed] [Google Scholar]
  44. Myers S.L.  et al. (1993) Probability and Statistics for Engineers and Scientists. Macmillan, New York. [Google Scholar]
  45. Mysling S.  et al. (2016) The acidic domain of the endothelial membrane protein GPIHBP1 stabilizes lipoprotein lipase activity by preventing unfolding of its catalytic domain. Elife, 5, e12095. [DOI] [PMC free article] [PubMed] [Google Scholar]
  46. Orchard S.  et al. (2012) Protein interaction data curation: the International Molecular Exchange (IMEx) consortium. Nat. Methods, 9, 345–350. [DOI] [PMC free article] [PubMed] [Google Scholar]
  47. Porter M.F. (1980) An algorithm for suffix stripping. Program, 14, 130–137. [Google Scholar]
  48. Qiu L., Dhe-Paganon S. (2011) Oligomeric structure of the MALT1 tandem Ig-like domains. PLoS One, 6, e23220. [DOI] [PMC free article] [PubMed] [Google Scholar]
  49. Quattrocolo G., Maccaferri G. (2014) Optogenetic activation of Cajal-Retzius cells reveals their glutamatergic output and a novel feedforward circuit in the developing mouse hippocampus. J. Neurosci., 34, 13018–13032. [DOI] [PMC free article] [PubMed] [Google Scholar]
  50. Regev Y.  et al. (2002) Rule-based extraction of experimental evidence in the biomedical domain: the KDD Cup 2002 (task 1). ACM SIGKDD Explorations Newslett., 4, 90–92. [Google Scholar]
  51. Rooijen C.  et al. (2012) Evolutionarily conserved requirement of Cdx for post-occipital tissue emergence. Development, 139, 2576–2583. [DOI] [PubMed] [Google Scholar]
  52. Santosh K.C.  et al. (2015) Stitched multipanel biomedical figure separation. In Proc. of IEEE CBMS, pp. 54–59. Sao Carlos, Brazil. [Google Scholar]
  53. Shatkay H.  et al. (2006) Integrating image data into biomedical text categorization. Bioinformatics, 22, e446–e453. [DOI] [PubMed] [Google Scholar]
  54. Shinohara H.  et al. (2005) PKCβ regulates BCR-mediated IKK activation by facilitating the interaction between TAK1 and CARMA1. J. Exp. Med., 202, 1423–1431. [DOI] [PMC free article] [PubMed] [Google Scholar]
  55. Simonyan K., Zisserman A. (2015) Very deep convolutional networks for large-scale image recognition. In Proc. of ICLR, pp. 1409–1556. San Diego, CA, USA. [Google Scholar]
  56. Simpson M.S.  et al. (2015) Literature-based biomedical image classification and retrieval. Comput. Med. Imag. Graph., 39, 3–13. [DOI] [PubMed] [Google Scholar]
  57. Sugioka K.  et al. (2014) The roles of urokinase-type plasminogen activator in leukocyte infiltration and inflammatory responses in mice corneas treated with lipopolysaccharide. Investig. Ophthalmol. Vis. Sci., 55, 5338–5350. [DOI] [PubMed] [Google Scholar]
  58. Vogt E.J.  et al. (2012) Importance of the pluripotency factor LIN28 in the mammalian nucleolus during early embryonic development. Development, 139, 4514–4523. [DOI] [PMC free article] [PubMed] [Google Scholar]
  59. Wang L.  et al. (2020) CORD-19: the Covid-19 open research dataset. In Proc. of NLP COVID-19 workshop at ACL, pp. 1–12. Seattle, WA, USA. [Google Scholar]
  60. Wei C.H.  et al. (2019) PubTator central: automated concept annotation for biomedical full text articles. Nucleic Acids Res., 47, W587–W593. [DOI] [PMC free article] [PubMed] [Google Scholar]
  61. Wu C.H.  et al. (2003) The protein information resource. Nucleic Acids Res., 31, 345–347. [DOI] [PMC free article] [PubMed] [Google Scholar]
  62. Yamaguchi T.  et al. (2014) Skn-1a/Pou2f3 is required for the generation of Trpm5-expressing microvillous cells in the mouse main olfactory epithelium. BMC Neurosci., 15, 1–10. [DOI] [PMC free article] [PubMed] [Google Scholar]
  63. Yoshida S.  et al. (2014) Molecular cloning of rat and porcine retina-derived POU domain factor 1 (POU6F2) from a pituitary cDNA library. J. Reprod. Dev., 60, 288–294. [DOI] [PMC free article] [PubMed] [Google Scholar]

Articles from Bioinformatics are provided here courtesy of Oxford University Press

RESOURCES