Abstract
Multi-modal neuroimage retrieval has greatly facilitated the efficiency and accuracy of decision making in clinical practice by providing physicians with previous cases (with visually similar neuroimages) and corresponding treatment records. However, existing methods for image retrieval usually fail when applied directly to multi-modal neuroimage databases, since neuroimages generally have smaller inter-class variation and larger inter-modal discrepancy compared to natural images. To this end, we propose a deep Bayesian hash learning framework, called CenterHash, which can map multi-modal data into a shared Hamming space and learn discriminative hash codes from imbalanced multi-modal neuroimages. The key idea to tackle the small inter-class variation and large inter-modal discrepancy is to learn a common center representation for similar neuroimages from different modalities and encourage hash codes to be explicitly close to their corresponding center representations. Specifically, we measure the similarity between hash codes and their corresponding center representations and treat it as a center prior in the proposed Bayesian learning framework. A weighted contrastive likelihood loss function is also developed to facilitate hash learning from imbalanced neuroimage pairs. Comprehensive empirical evidence shows that our method can generate effective hash codes and yield state-of-the-art performance in cross-modal retrieval on three multi-modal neuroimage datasets.
Keywords: Deep Bayesian Hashing, Retrieval, Multi-Modal Neuroimage, MRI, PET
I. Introduction
Neuroimage analysis has made a profound contribution to modern clinical analysis [1]–[3], image-guided surgery [4]–[6], and automated diagnostic studies [7]–[13]. Currently, various digital imaging techniques have been developed to generating heterogeneous visual representations of brain tissues, such as structural magnetic resonance imaging (sMRI) [14]–[16], positron emission tomography (PET) [17], and computed tomography (CT) [18]. However, interpreting neuroimages is a complicated task, which usually requires extensive professional knowledge. In practice, it is important to provide physicians with previous cases (with visually similar neuroimages) and their corresponding treatment records to facilitate case-based reasoning and evidence-based medicine in clinical decision making [19], [20]. Hence, multi-modal neuroimage retrieval [21]–[23], which can return similar cases from heterogeneous neuroimage databases, has attracted increasing interest in the field.
In this paper, we focus on hashing-based multi-modal neuroimage retrieval, which makes a good balance between the retrieval quality and computation cost. Generally, cross-modal hashing aims to map heterogeneous inputs into binary codes in a common Hamming space. Based on the availability of supervisory signals, existing methods can be roughly categorized into two groups: (1) unsupervised hashing [24] that learns hash functions based on original data structures and distributions; and (2) supervised hashing [25] that learns hash functions by exploiting original data and their semantic labels. Recently, many unsupervised and supervised cross-modal hashing methods have been developed for natural image retrieval [26]–[29]. Unfortunately, existing approaches usually achieve sub-optimal results when applied directly to multi-modal neuroimage databases for the following reasons. On the one hand, neuroimages generally contain complex tissue textures and anatomical structures compared to natural images. And subtle lesions in the local brain region can significantly affect the diagnostic results with high confidence [8], implying that neuroimages may show small inter-class variation. On the other hand, even for the same subject, different neuroimaging techniques can generate different visual representations (e.g., a pair of sMRI and PET scans from the same subject), which introduces large inter-modal discrepancy. It’s highly desired to develop advanced hashing techniques to tackle the problems of small inter-class variation and large inter-modal difference for efficient retrieval of multi-modal neuroimages.
To this end, we propose a novel deep cross-modal neuroimage hashing method (see Fig. 1), dubbed CenterHash, which aims to map neuroimages from different modalities into a common Hamming space and well preserve cross-modal semantic similarities. Specifically, we devise a deep multi-modal Bayesian hash learning framework, which can simultaneously learn deep representations and binary hash codes for multimodal neuroimages. By assuming that optimal binary codes of subjects from the same class (even with different modalities) should be close to a common representation, we propose a center prior for the learned binary codes, which serves as the key component to address the challenges of small inter-class variation and large inter-modal discrepancy. We also propose a weighted contrastive likelihood loss function to enable effective hash code learning from imbalanced neuroimage pairs. Extensive experiments testify that CenterHash yields state-of-the-art performance on three multi-modal neuroimage datasets.
Fig. 1.
Proposed deep cross-modal hashing method (called CenterHash) for multi-modal neuroimage retrieval, including three key components: (1) two modality-specific networks (i.e., Backbone 3DCNNs) to project sMRI and PET to binary hash codes; (2) a CenterNet to learn center representations for each category in the Hamming space, with a multi-layer perceptron (MLP) architecture; (3) a weighted contrastive likelihood loss and a center prior, both derived from a Bayesian learning framework. Different colors denote different modalities, and different markers (except circles) denote different categories. Blue circles denote to-be-learned center representations. FCH: Fully-Connected Hash layer.
II. Related Work
We now briefly review the most relevant studies, including continuous representation and binary representation (via hash codes) based methods for multi/cross-modal image retrieval. For comprehensive surveys, we refer readers to [30] and [31].
A. Continuous Representation based Methods
Conventional approaches typically rely on continuous image representations for medical image retrieval [21]–[23], [32]. For instance, Cao et al. [21] extended the probabilistic Latent Semantic Analysis (pLSA) model to integrate the visual and textural features of medical images and developed a deep Boltzmann machine-based multi-modal learning framework to derive the missing modality. Vikram et al. [22] proposed a Latent Dirichlet Allocation-based (LDA) technique to encode the visual features and explored early fusion and late fusion to combine visual features with textual features. Gao et al. [23] proposed a multi-graph learning-based method, which consists of two stages: query category prediction and ranking. Even though these methods have achieved good results, they usually suffer from the curse of feature dimension and are not suitable for modern neuroimage search systems.
B. Binary Representation based Methods
To improve the efficiency of image retrieval, various fast multi-modal image retrieval methods have been proposed based on binary representation (e.g., hashing codes). According to whether supervisory signals are involved in the learning phase, existing multi-modal hashing methods can be divided into two groups: (1) unsupervised approaches, and (2) supervised methods. In the following, we summarize some representative methods from these two groups, respectively.
Unsupervised approaches usually learn hash functions from the original feature space to Hamming space by exploiting the data structure, topological information, and data distributions. Semantic-topic multi-modal hashing (STMH) [33] explains multi-modal data by a composition of semantic topics, and uses binary codes to indicate the existence of the corresponding topic. Alternating co-quantization (ACQ) [34] learns hash functions by simultaneously minimizing binary quantizers for each modality and preserving data similarities across different modalities. Unsupervised generative adversarial cross-modal hashing (UGACH) [28] proposes a correlation graph to capture the manifold structure and exploits generative adversarial networks (GANs) to match the generated data pairs and pairs from the correlation graph. Unsupervised deep cross-modal hashing (UDCMH) [29] solves the hashing problem by integrating deep learning and matrix factorization.
Supervised cross-modal hashing methods [25], [35] can explore the semantic information to enhance data correlation across different modalities, thus usually achieving superior performance compared to their unsupervised counterparts. The cross-view hashing (CVH) method [25] tries to optimize the similarity-weighted cumulative Hamming distance between different pairwise samples to learn the hash functions. Semantic correlation maximization (SCM) [36] seamlessly integrates semantic labels into the hashing learning procedure and can use all the supervised information with linear-time complexity. Semantics preserving hashing (SEPH) [27] approximates the semantic affinities in Hamming space and uses the obtained hash codes as supervised signals to learn hash functions via kernel logistic regression. Though these methods have progressed the field, they typically employ hand-crafted image features, which may degrade the retrieval performance.
Recently, deep learning has revolutionized computer vision, machine learning, and many other related areas. Deep learning-based cross-modal hashing methods [26], [37]–[39] have also been proposed for image retrieval. Deep cross-modal hashing (DCMH) [37] exploits pairwise labels across different modalities to preserve the semantic relationship in Hamming space. Pairwise relationship guided deep hashing (PGDH) [38] considers both intra-modal and inter-modal similarities and designs a regularization term to maximize the representation ability of the hash codes. Ranking-based deep cross-modal hashing (RDCMH) [40] learns hash functions by utilizing a multi-level ranking semantic structure from multi-labels. Self-supervised adversarial hashing (SSAH) [39] adopts adversarial learning to maximize the representation correlations between different modalities and designs a self-supervised semantic network to discover semantic information from multi-label annotations. Cross-modal deep variational hashing (CMDVH) [41] designs cross-modal fusion networks and modality-specific networks and utilizes variational learning to match the learned hash codes. Cross-modal Hamming hashing (CMHH) [26] utilizes the exponential distribution to push relevant data pairs to have further small Hamming distances. Note that these approaches usually focus on natural images, without considering the unique property of neuroimages (i.e., small inter-class variance and large inter-modal discrepancy). Thus, they may result in sub-optimal performance when applied directly to neuroimaging data.
III. Methodology
For multi-modal neuroimage search, the modality of query subjects could be different from that of subjects in the database. Here, we assume that is the given multi-modal neuroimage dataset, where and are medical image collections from two different modalities (e.g., sMRI and PET). Given o as the number of categories, contains n examples and is the label set, where is the corresponding label for xi, contains m examples and is the label set, where is the corresponding label for yi. and are binary values that represent whether xi and yi belong to the d-th category. The goal of multi-modal hashing is to learn modality-specific hash functions: hx(x) : x → {−1, +1}k and hy(y) : y → {−1, +1}k, which can encode original samples xi and yi into compact k-bit hash code and in a common Hamming space such that the original semantic similarity can be maximally preserved, where and indicate the d-th element of and respectively.
To address the challenges of small inter-class variation and large inter-modal discrepancy and enable learning from imbalanced neuroimaging data, this paper presents a deep Bayesian hash learning method (called CenterHash) for multi-modal neuroimage search, with the architecture shown in Fig. 1. The proposed CenterHash employs multi-modal neuroimages (i.e., sMRI and PET) and their labels as input and can simultaneously learn neuroimage representations and binary hash codes through an end-to-end pipeline. Three components are included: (1) two modality-specific networks (i.e., Backbone 3DCNNs) to project input images to binary hash codes; (2) the proposed CenterNet to learn center representations in Hamming space; (3) a weighted contrastive likelihood loss and a novel center prior, both derived from a Bayesian learning framework. Details can be found in the following.
A. Framework
In the customary settings for supervised hashing [37], [38], pairwise labels are usually used as supervised signals. Here, we construct a similar pairwise sample set , with each element denoting a pair of sMRI and PET scans from the same category. Also, we construct a dissimilar pairwise sample set , with each element denoting a pair of sMRI and PET scans from different categories. Accordingly, the pairwise similarity label can be defined as
| (1) |
For effective multi-modal neuroimage hashing, we want that semantically similar neuroimage pairs can have similar hash codes in Hamming space with a high probability, and vice versa. To effectively estimate this probability, we implement a deep Bayesian learning framework. Specifically, given the pairwise similarity labels S = {sij}, the maximum a posterior estimation of the learned binary hash codes for sMRI and for PET can be defined as
| (2) |
where p(Bx) and p(By) are prior distributions for the learned hash codes, p(S|Bx,By) is the weighted contrastive likelihood function defined as
| (3) |
where wij is the corresponding weight that is used to deal with the data imbalance problem by re-weighting each data pair (xi, yj) according to the importance of each category. We set wij as
| (4) |
where δ = |S1|/|S0| denotes the data imbalance rate with S1 = {sij ∈ S : sij = 1} and S0 = {sij ∈ S : sij = 0} represents the similar and dissimilar pairwise label sets respectively. | · | is used to indicate the cardinality of a set. is the conditional probability of pairwise label sij given the hash codes and , which can be naturally defined by the Bernoulli distribution
| (5) |
where σ(x) = 1/(1 + e−αx) is the adaptive sigmoid function with α to control the bandwidth, and indicates the similarity of hash codes and in Hamming space. Since we want to minimize the Hamming distance between similar neuroimage pairs and maximize the Hamming distance between dissimilar neuroimage pairs, we can naturally set as the negative Hamming distance:
| (6) |
where is the Hamming distance function. However, calculating requires discrete operations and is non-differentiable, directly optimizing Eq. (2) with back propagation is not feasible. As indicated in [42], for a pair of hash codes and , there exists a nice relationship between their inner product 〈·,·i〉 and Hamming distance distH(·,·):
| (7) |
Hence, we can adopt the inner product to quantify the pairwise similarity of hash codes in Hamming space by setting . Similar to binary logistic regression, we encourage that, if the hash codes and are similar, the conditional probability should be large, implying that the multi-modal image pair xi and yj should be semantically similar. Otherwise, they should be semantically dissimilar. Therefore, Eq. (5) can be regarded as a reasonable extension of the binary logistic regression to the pairwise Hamming classification scheme.
B. Formulation of CenterHash
With the framework proposed in Section III-A, we can easily instantiate a specific cross-modal hash model with any valid prior distribution for the learned hash codes. Many state-of-the-art hash methods neglect the prior distribution or only assume a uniform prior, and learn hash codes by optimizing the likelihood function in Eq. (5). However, directly applying these methods may fail to obtain optimal hash codes due to the unique property of multi-modal neuroimages (i.e., small inter-class variation and large inter-modal discrepancy). Since hash codes for scans from the same class are expected to be similar to each other, which is consistent with clustering, where similar data points are expected to be in the same group, inspired by k-means clustering [43], here we assume that multi-modal neuroimages in the same category share a common center representation in the Hamming space. We design a novel center prior distribution for the learned hash codes as follows
| (8) |
where and are the binary center representations in Hamming space for the class and the class , respectively. Such class-specific center presentations are shared across different modalities. This prior distribution assumes that the optimal muti-modal hash codes should distribute closely to their corresponding class centers, which is consistent with the requirement that objects with same semantic labels should have similar hash codes. By maximizing the center prior in Eq. (8), one can encourage to-be-learned hash codes (of each class) from different modalities to be close to their corresponding shared class centers, thus explicitly and simultaneously enhancing the intra-class compactness and minimizing the inter-modal discrepancy.
For Eq. (8), we need to specify the definition of the center representation cl. The most straightforward way is to set cl as the mean value of hash codes of two modalities (i.e., and ) for the class l. However, the mean operation will destroy the binary property and will also make cl suffer from the instability caused by mini-batch training. We alternatively propose a neural network (named CenterNet) to automatically learn the binary center presentation for each category as
| (9) |
where hl represents the mapping function of CenterNet, which maps the label l to the binary center representation. In this way, we can learn the multi-modal hash codes and the shared class center representations simultaneously.
By taking Eqs. (5)–(6) into the maximum a posterior estimation in Eq. (2), we obtain the overall optimization objective for the proposed CenterHash as
| (10) |
where L is the weighted contrastive likelihood loss, T is the center prior loss, and Θ is the network parameters. Specifically, L can be formulated as
| (11) |
Similarly, the center prior loss can be formulated as
| (12) |
where the hash codes and are learned from the corresponding modality-specific networks (i.e., Backbone 3DCNNs in Fig. 1), and the center representations and are learned from the CenterNet. Detailed configurations for all these networks will be elaborated in Subsection III-D.
By optimizing the overall objective in Eq. (8), we can jointly learn two hash functions hx(·) and hy(·) and one class-specific center representation learning function hl(·). Given a query neuroimage, we can obtain its hash codes by first forward propagating it from the corresponding modality-specific hash function and then thresholding it via the sign function sgn(·), where sgn(hi) = 1 if hi > 0, otherwise sgn(hi) = −1. Even though the proposed CenterHash is designed to retrieve neuroimages with two modalities (i.e., sMRI and PET), one can easily extend it to applications with more modalities.
C. Connection to Classification
In this subsection, we reveal that there exists a close relationship between the proposed prior distribution and classification. To enhance the discriminability of the learned hash codes, many studies assume that the hash codes should be optimal for a jointly learned classifier and usually model the relationship between hash codes (e.g., ) and semantic labels (e.g., ) with a linear classifier fx(·) defined as
| (13) |
where is the parameter for the classifier fx(·), and is the weight vector for the d-th class. The sigmoid cross-entropy loss is usually adopted to optimize the classifier
| (14) |
From Eq. (14), we can observe that for all d ∈ [1,2, · · · , o] and i ∈ [1,2, · · · , n], if , will be maximized; and it will be minimized, otherwise. This means that optimizing the classification loss in Eq. (14) will encourage all hash codes to be close to the corresponding class weight vectors and be away from weight vectors from other classes.
Considering the prior distribution in Section III-B, the class weight vectors act similar to the class centers in Eq. (8). However, the class weight vectors contain continuous values and may have different scales, while the class centers are all binary hash codes. To better compare these two vectors, we set , where cd is a binary vector for the d-th class and is a diagonal transformation matrix with all non-negative diagonal entries (i.e., ). By substituting into Eq. (14), we can get
| (15) |
From Eq. (15), we can see that optimizing the loss function actually minimizes or maximizes . Note that , compared with the inner product , re-weights the i-th hash dimension with for the d-th class. Since larger weights indicate more important hash dimensions, learning with Eq. (15) will make the learned hash codes focus on some specific hash dimensions with large weights. In addition, since the weight vectors for different classes are different, the hash dimension weights may be inconsistent across different classes, which can greatly hinder the accurate Hamming distance calculation and degrade the search performance. Furthermore, different scales of the weight vectors for different classes will make the learned hash codes focus on some specific classes and hinder the model to generalize to other classes. Thus learning with the linear classifier in Eq. (14) may be not optimal for hash code learning. To this end, our method explicitly learns the binary class centers and directly minimizes the distances between hash codes and their class centers, thus can generate more consistent hash codes with better generalization ability.
D. Implementation
The architecture of modality-specific networks for sMRI and PET (i.e., and ) are similar to the 3DCNN model designed in [16], with the last fully-connected (FC) layer replaced by a new FC layer with k units as the hash layer. Besides, tanh(·) is used as the activation function to relaxed the binary codes. In the Supplementary Materials, we show that optimizing both of the loss functions L and T with the relaxed hash codes obtained from tanh(·) will decrease the loss values with exact binary codes in each iteration. For our CenterNet, which is designed to learn class-specific center representations, we use a multi-layer perceptron (MLP) that consists of two FC layers, denoted as “FC-1” and “FC-2”. Besides, ReLU is used as the activation function for FC-1, while the tanh(·) function is employed for FC-2. The detailed configuration of CenterNet is shown in Table I. The learning algorithm, optimization procedure, and the computation complexity analysis can be found in the Supplementary Materials. In the experiments, we randomly sample data points from the training set to form a mini-batch. Then, for each iteration, the similar and dissimilar data pairs are constructed based on all data points in the mini-batch.
TABLE I.
THE CONFIGURATION OF CENTERNET. FC: FULLY-CONNECTED.
| Layer | Number of Units |
|---|---|
| FC-1 | (label dimension o + length of hash code k)/2 |
| FC-2 | length of hash code k |
IV. Experiments
A. Datasets
We evaluate our method on three popular benchmark datasets: (1) Alzheimer’s Disease Neuroimaging Initiative (ADNI1), (2) ANDI2 [44], and (3) Australian Imaging, Biomarkers and Lifestyle dataset (AIBL) [45]. In the following, we give a more detailed introduction for each dataset.
ADNI1 contains 821 subjects with 1.5T T1-weighted sMRI scans, and only 397 subjects among them have PET images. Each subject was annotated by a category-level label, i.e., Alzheimer’s disease (AD), normal control (NC), or mild cognitive impairment (MCI). Such labels were determined based on the standard clinical criteria, including mini-mental state examination scores and clinical dementia rating. Among subjects with sMRI scans in ADNI1, there are 229 NC, 393 MCI, and 199 AD subjects. For PET data in ADNI1, there are 100 NC, 93 AD, and 204 MCI subjects.
ADNI2 contains 636 subjects with 3T T1-weighted sMRI scans, and 309 subjects have PET images. With the same clinical criteria for ADNI1, these images were divided into three categories (i.e., AD, NC, and MCI). For sMRI data, there have 200 NC, 277 MCI, and 159 AD subjects. For PET data, there have 94 NC, 149 MCI, and 66 AD subjects.
AIBL consists of 612 subjects with 3T T1-weighted sMRI scans and 560 subjects with PET scans. Similar to ADNI1 and ADNI2, these images are grouped into AD, NC, and MCI, respectively. For sMRI modality, there are 447 NC, 94 MCI, and 71 AD subjects. For PET modality, there are 407 NC, 91 MCI, and 62 AD subjects.
For all the three datasets, we randomly select 10% images from each class to form the test set, and the remaining images are used as the retrieval set. To optimize the proposed method, we further randomly select 90% images from the retrieval set as the training set and treat the others as the validation set. Following [16], we pre-process all sMRI and PET scans using a standard pipeline, including anterior commissure (AC)-posterior commissure (PC) correction, intensity correction [46], skull stripping [47], and cerebellum removing. The affine registration is also performed to align each PET image to its corresponding sMRI scan.
B. Competing Methods and Evaluation Metrics
The proposed CenterHash method is first compared with five state-of-the-art cross-modal hashing methods based on traditional machine learning techniques, including ACQ [34], STMH [33], CVH [25], SEPH [27], and SCM [36] and then compared with three recently proposed deep learning methods for cross-modal hashing, including PGDH [38], DCMH [37], and CMHH [26]. The codes of the first seven competing methods are kindly provided by the authors. For CMHH, we implement the method by ourselves. Five traditional methods (i.e., ACQ, STMH, CVH, SCM, and SEPH) are implemented with MATLAB, while four deep learning methods (i.e., DCMH, PGDH, CMHH, and our CenterHash) are implemented with Keras [48] or Tensorflow [49]. Among eight competing approaches, two methods (i.e., ACQ and STMH) are unsupervised, while the remaining ones are supervised.
For traditional methods, we extract volumes of the grey matter from 90 regions-of-interest (ROIs) as features for representing sMR and PET images, and use these ROI features as input. For deep learning methods, we use raw images as inputs. All parameters from the modality-specific 3DCNNs and CenterNet in our CenterHash are randomly initialized. To accelerate the model training, we first pre-train each modality-specific 3DCNN by simply optimizing a single-modal version of Eq. 5. Then, we combine the modality-specific 3DCNNs and CenterNet together to simultaneously optimize the hash codes for different modalities and the center representations. For a fair comparison, DCMH and PGDH use the same backbone network architectures and pre-training strategy as our method. Adam is used to optimize the network parameters of four deep learning methods (i.e., DCMH, PGDH, CMHH, and CenterHash), where the initial learning rate is set as 0.01 and the mini-batch size is fixed as 5. Similar to [50], [51], we select the hyper-parameter α via cross-validation.
Four evaluation metrics are used in this paper, including mean of average precision (MAP), recall@K, topN-precision, and precision-recall. The first three metrics are based on Hamming space ranking, which rank returned data points based on their Hamming distances to the queries. The precision-recall metric is based on hash lookup, as it first builds a hash lookup table and returns data points within a pre-defined Hamming radius. Given a query and a list of Q ranked retrieval examples, the average precision (AP) for the query is defined as
| (16) |
where R denotes the number of ground-truth relevant points in the database, and P(q) denotes the precision value for the top q retrieved points. δ(q) = 1 when the q-th retrieval point is relevant with the query, otherwise δ(q) = 0. In our experiments, Q is set as 100 for all the three datasets. The MAP is defined as the average APs for all queries. Recall@K is defined as the percentage of ground-truth neighbors from the top K returned instances among all the semantically relevant points. TopN-precision indicates the average ratio of semantically relevant instances among the top N returned points for all given queries based on Hamming distance. In this paper, we set both K and N as 100. Precision-recall reveals the retrieval precision with different recall values and is considered as a good indicator of overall search performance.
C. Results
1). Results of Hamming Ranking:
We first present MAP results for CenterHash and all other baseline methods on the three datasets using different lengths of hash code (i.e., k) to provide a global evaluation. Then we report the topN-precision and recall@K curves with k = 64.
The MAP results achieved by all methods on the ADNI1, ADNI2, and AIBL datasets are reported in Table II. Here, ”M → P” represents the case where queries are sMRI scans and the database contains PET images, and ”P → M” represents the case where queries are PET images and the database has sMRI scans. From the results in Table II, one can observe an interesting finding. That is, the MAP values for ”M→P” are usually higher than those for ”P→M”, especially for ADNI1 and ADNI2. This may be due to the limited number of PET images. From Subsection IV-A, we can see that the numbers of sMRI scans on three datasets are usually larger than that of PET scans. Since deep models usually need a large number of training data to reduce over-fitting, the limited PET scans may degrade the generalization ability of the hash functions for PET modality. Therefore, the generated hash codes for MRI queries may be better than those for PET queries, which may induce the performance gap between ”M→P” and ”P→M” tasks. From the results, one can also obtain that CenterHash usually outperforms other competing methods by large margins. For instance, compared to SCM (the state-of-the-art traditional method), CenterHash can obtain absolute increase of 7.253%/3.105% and 13.74%/4.085% 5.473%/2.915% in terms of the average MAP in two cross-modal retrieval tasks ”M → P” and ”P → M” on ADNI1, ADNI2, and AIBL, respectively. Also, CenterHash outperforms DCMH (the state-of-the-art deep learning method), by margins of 2.928%/2.822%, 5.625%/4.395%, and 2.225%/3.323% in terms of the average MAP on three datasets, respectively. These results validate the superiority of CenterHash in cross-modal neuroimage retrieval.
TABLE II.
COMPARISON WITH BASELINES IN TERMS OF MAP (WITH BEST RESULTS SHOWN IN BOLDFACE).
| Task | Method | ADNI1 |
ADNI2 |
AIBL |
|||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
| k = 16 | k = 32 | k = 64 | k = 128 | k = 16 | k = 32 | k = 64 | k = 128 | k = 16 | k = 32 | k = 64 | k = 128 | ||
| M → P | ACQ | 0.4163 | 0.3713 | 0.4015 | 0.4623 | 0.3883 | 0.4341 | 0.4386 | 0.4096 | 0.5498 | 0.5535 | 0.5563 | 0.5583 |
| STMH | 0.4246 | 0.4093 | 0.4087 | 0.3864 | 0.3778 | 0.4449 | 0.4511 | 0.4225 | 0.5870 | 0.6087 | 0.5995 | 0.5969 | |
| CVH | 0.3933 | 0.3842 | 0.4163 | 0.3752 | 0.4023 | 0.3896 | 0.4238 | 0.4003 | 0.5694 | 0.5710 | 0.5622 | 0.5654 | |
| SEPH | 0.4378 | 0.4416 | 0.4063 | 0.4320 | 0.4225 | 0.4610 | 0.4097 | 0.4670 | 0.6231 | 0.5944 | 0.5385 | 0.5660 | |
| SCM | 0.5037 | 0.4713 | 0.5109 | 0.5174 | 0.4368 | 0.4944 | 0.4452 | 0.5009 | 0.5735 | 0.5919 | 0.5892 | 0.5807 | |
| CMHH | 0.4738 | 0.4323 | 0.4502 | 0.4144 | 0.4520 | 0.4348 | 0.4576 | 0.4223 | 0.5932 | 0.5741 | 0.5794 | 0.5962 | |
| PGDH | 0.5124 | 0.5181 | 0.5276 | 0.5225 | 0.6275 | 0.5735 | 0.5861 | 0.5543 | 0.6054 | 0.6094 | 0.5928 | 0.6025 | |
| DCMH | 0.5562 | 0.5529 | 0.5366 | 0.5306 | 0.5819 | 0.5712 | 0.5914 | 0.5573 | 0.6160 | 0.6268 | 0.6261 | 0.5963 | |
| CenterHash | 0.5855 | 0.5598 | 0.5663 | 0.5818 | 0.6463 | 0.6164 | 0.6363 | 0.6278 | 0.6430 | 0.6361 | 0.6380 | 0.6371 | |
| P → M | ACQ | 0.3985 | 0.4242 | 0.3725 | 0.3706 | 0.3923 | 0.4023 | 0.4107 | 0.4466 | 0.5544 | 0.5548 | 0.5503 | 0.5503 |
| STMH | 0.4246 | 0.4093 | 0.4087 | 0.3864 | 0.4281 | 0.3923 | 0.3779 | 0.4586 | 0.5825 | 0.5907 | 0.5731 | 0.5821 | |
| CVH | 0.3913 | 0.3992 | 0.4079 | 0.4138 | 0.4099 | 0.3615 | 0.3768 | 0.4106 | 0.6075 | 0.5880 | 0.5881 | 0.5699 | |
| SEPH | 0.4585 | 0.3962 | 0.4587 | 0.3688 | 0.3807 | 0.3803 | 0.4223 | 0.4472 | 0.5555 | 0.5651 | 0.5775 | 0.5482 | |
| sCM | 0.4806 | 0.4658 | 0.4599 | 0.4659 | 0.4370 | 0.4853 | 0.4280 | 0.4635 | 0.5798 | 0.5892 | 0.5992 | 0.6106 | |
| CMHH | 0.4017 | 0.4098 | 0.4072 | 0.4008 | 0.4070 | 0.3961 | 0.4301 | 0.4121 | 0.5888 | 0.5850 | 0.5740 | 0.5964 | |
| PGDH | 0.5097 | 0.4565 | 0.4422 | 0.4592 | 0.4478 | 0.4585 | 0.4254 | 0.4408 | 0.5848 | 0.5935 | 0.5949 | 0.5996 | |
| DCMH | 0.5248 | 0.4601 | 0.4518 | 0.4468 | 0.4490 | 0.4521 | 0.4556 | 0.4449 | 0.5923 | 0.5824 | 0.5760 | 0.6118 | |
| CenterHash | 0.5312 | 0.4837 | 0.4989 | 0.4826 | 0.4835 | 0.4970 | 0.5131 | 0.4838 | 0.6251 | 0.6105 | 0.6303 | 0.6295 | |
The recall@K curves with k = 64 achieved by all methods on the three datasets for both ”M → P” and ”P → M” tasks are shown in Fig. 2 and Fig. 3. From these results, one can obtain that CenterHash can usually achieve better recall performance than other baseline methods on both cross-modal retrieval tasks, which is consistent with the MAP evaluation.
Fig. 2.
Recall@K curves for ADNI1, ADNI2, and AIBL with sMRI query images and PET database images.
Fig. 3.
Recall@K curves for ADNI1, ADNI2, and AIBL with PET query images and sMRI database images.
The topN-precision curves with k = 64 achieved by different methods on the three datasets for both ”M → P” and ”P → M” tasks are shown in Fig. 4 and Fig. 5. These figures suggest that our CenterHash generally achieves higher precisions, which is consistent with the recall@K and MAP evaluations. In the task of medical image retrieval, users usually focus on the top returned results, and thus, it’s essential to provide users with top returned instances that are highly relevant to the query. From the results, one can see that CenterHash outperforms the other methods by a large margin when the number of returned instances is small.
Fig. 4.
TopN-precision curves for ADNI1, ADNI2, and AIBL with sMRI query images and PET database images.
Fig. 5.
TopN-precision curves for ADNI1, ADNI2, and AIBL with PET query images and sMRI database images.
2). Results of Hash Lookup:
Given a query object, the precision and recall values for the returned items within any Hamming radius can be computed. By investigating these values with every Hamming radius from 0 to k, we can draw the precision-recall curves. Fig. 6 and Fig. 7 report the precision-recall results of different methods on ADNI1, ADNI2, and AIBL for both ”M → P” and ”P → M” tasks with k = 64. From these figures, one can observe that CenterHash consistently achieves the best performance. Especially, CenterHash outperforms other methods by a large margin at low recall levels, which is desirable in practical search systems. From Table II and Figs. 2–7, one can see that CenterHash generally obtains superior performance in terms of both Hamming ranking metrics (i.e., MAP, recall@K, and topN-precision) and hash lookup metric (i.e., Precision-recall), which demonstrates that, compared with other baseline methods, the proposed CenterHash can learn more discriminative hash codes and enable more effective neuroimage search on both cross-modal retrieval tasks.
Fig. 6.
Precision-recall curves for ADNI1, ADNI2, and AIBL with sMRI query images and PET database images.
Fig. 7.
Precision-recall curves for ADNI1, ADNI2, and AIBL with PET query images and sMRI database images.
3). Results of Small Hamming Radius:
To compare the search performance when the returned examples are in a small Hamming radius, we set Q in Eq. (16) as the number of k/2 and report the corresponding MAP results in Fig. 8. From Fig. 8, one can see that CenterHash can achieve the best MAP results, further demonstrating that CenterHash can successfully learn hash codes with high quality.
Fig. 8.
MAP results for ADNI1, ADNI2, and AIBL with returned examples in a small Hamming radius.
D. Cross-data Generalization
In clinical applications, collecting sufficient medical images for model training may be challenging. Therefore, a retrieval system with strong generalization capabilities across different datasets is highly desirable. Here, to test the generalization ability of CenterHash, we select ADNI1 and ADNI2 datasets and evaluate its search performance in two scenarios: (1) cross-data training, where we train the model on ADNI1 (ADNI2) and test it on ADNI2 (ADNI1). The queries and database images are both from ADNI2 (ADNI1); (2) cross-data query, where we train the model on ADNI1 (ADNI2) and test its retrieval performance with queries from ADNI2 (ADNI1) and database images from ADNI1 (ADNI2). For both scenarios, two state-of-the-art methods (i.e., SCM and DCMH) are selected for comparison.
Cross-data Training.
We first train the models on ADNI1 and test them on ADNI2, with MAP results reported in Fig. 9(a). Similarly, we also train the models on ADNI2 and test them on ADNI1, with results shown in Fig. 9(b). These figures show that our method clearly outperforms SCM and DCMH in both tasks. From Figs. 9(a)–9(b) and Table II, we can observe that CenterHash outperforms or achieves comparable performance with other methods, even though the training and testing images are from different datasets.
Fig. 9.
Results of models using cross-data training strategy. (a) Models trained on ADNI1, and tested on ADNI2 (with query and database images from ADNI2). (b) Models trained on ADNI2, and tested on ADNI1 (with query and database images from ADNI1).
Cross-data Query.
We first train models on ADNI1. Given query images from ADNI2, we want to retrieval similar subjects in ADNI1. Fig. 10(a) reports the MAP results. Fig. 10(b) shows the results of models trained on ADNI2, where query images are from ADNI1. We can again see that CenterHash consistently achieves the best performance. The results in Figs. 9–10 together demonstrate that our method has strong generalization ability across different datasets.
Fig. 10.
Results of models using cross-data query strategy. (a) Models trained on ADNI1, and tested on ADNI2 (with query from ADNI2 and database images from ADNI1). (b) Models trained on ADNI2, and tested on ADNI1 (with query from ADNI1 and database images from ADNI2).
E. Parameter Sensitivity
We now study the influence of the hyper-parameter α that controls the bandwidth of the adaptive sigmoid function in Eq. (5). Specifically, we first set the hash code length k as 16 and then compute MAP values by varying α between 0.2 and 9.0. The map results achieved by our CenterHash method in the ”M→P” and ”P→M” tasks on the three datasets are reported in Fig. 11. From this figure, we can see that the MAP values first increase with the increasing of α and then maintain relatively high values in a large range of 1.0 ≤ α ≤ 5.0. Thus, in practice, we can empirically select α from [1.0, 5.0]. In other experiments in the paper, we fix α as 1.0. The influence of the hash code length can be found in the Supplementary Materials.
Fig. 11.
MAP results on the three datasets with different α.
F. Ablation Study
To investigate the role of different components in our framework, we investigate three variants of CenterHash, including (1) CenterHash-G that is trained on data from two categories (i.e., AD and NC) and tested on data from three categories (i.e., AD, NC, and MCI) to test the generalization ability of CenterHash from binary retrieval problems to three-class retrieval problems; (2) CenterHash-W that is trained with a standard contrastive likelihood loss [37], [38] (i.e., wij = 1 in Eq. (3). This variant is used to demonstrate the role of the weight wij that is designed to deal with imbalanced training pairs; (3) CenterHash-C without using the proposed center prior, which learns hash codes by directly optimizing Eq. (3). This variant is adopted to study the role of the center prior. Note that CenterHash, CenterHash-W and CenterHash-C are trained and tested on data from all three categories (i.e., AD, NC, and MCI). And CenterHash-G is trained on AD and NC subjects and tested on data from those three categories, respectively. The experimental results of CenterHash and its variants are reported in Table III.
TABLE III.
MAP RESULTS OF CENTERHASH AND ITS VARIANTS (WITH BEST RESULTS SHOWN IN BOLDFACE).
| Task | Method | ADNI1 |
ADNI2 |
AIBL |
|||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
| k = 16 | k = 32 | k = 64 | k = 128 | k = 16 | k = 32 | k = 64 | k = 128 | k = 16 | k = 32 | k = 64 | k = 128 | ||
| M → P | CenterHash-G | 0.4098 | 0.3848 | 0.3857 | 0.3902 | 0.4371 | 0.4092 | 0.4142 | 0.4046 | 0.5896 | 0.5872 | 0.5794 | 0.5713 |
| CenterHash-W | 0.5519 | 0.5433 | 0.5311 | 0.5359 | 0.5933 | 0.5719 | 0.6012 | 0.5760 | 0.6239 | 0.6105 | 0.6043 | 0.5977 | |
| CenterHash-C | 0.5624 | 0.5292 | 0.5347 | 0.5568 | 0.6191 | 0.5844 | 0.5980 | 0.6039 | 0.6185 | 0.6029 | 0.5994 | 0.6136 | |
| CenterHash | 0.5855 | 0.5598 | 0.5663 | 0.5818 | 0.6463 | 0.6164 | 0.6363 | 0.6278 | 0.6430 | 0.6361 | 0.6380 | 0.6371 | |
| P → M | CenterHash-G | 0.3864 | 0.3828 | 0.3621 | 0.3700 | 0.4236 | 0.3972 | 0.3735 | 0.3944 | 0.5898 | 0.5781 | 0.5850 | 0.5771 |
| CenterHash-W | 0.5076 | 0.4614 | 0.4537 | 0.4546 | 0.4575 | 0.4775 | 0.4694 | 0.4642 | 0.5849 | 0.5875 | 0.5931 | 0.5863 | |
| CenterHash-C | 0.5103 | 0.4600 | 0.4517 | 0.4623 | 0.4453 | 0.4608 | 0.4731 | 0.4517 | 0.6038 | 0.5914 | 0.5843 | 0.5889 | |
| CenterHash | 0.5312 | 0.4837 | 0.4989 | 0.4826 | 0.4835 | 0.4970 | 0.5131 | 0.4838 | 0.6251 | 0.6105 | 0.6303 | 0.6295 | |
From Table III, we can see that CenterHash outperforms CenterHash-G in most cases. The underlying reason is that CenterHash-G, which is only trained on two categories, cannot capture the data distribution well, resulting in poor generalizability. This also demonstrates the necessity of training models on all three categories. Besides, the performance of CenterHash-G on AIBL is generally better than those on ADNI1 and ADNI2. This may be because AIBL contains less MCI images (compared to ADNI1 and ADNI2), so only using AD and NC images for model training has less impact on AIBL. Moreover, the results also show that CenterHash can outperform CenterHash-W by substantially large margins of 3.280%/2.978%, 4.610%/2.720% and 2.945%/3.590% in terms of the average MAP for cross-modal retrieval tasks ”M → P” and ”P → M” on ADNI1, ADNI2, and AIBL, respectively. The possible reason is that CenterHash-W (using the standard contrastive likelihood loss) neglects the class imbalance problem, and may degrade the retrieval performance. In contrast, CenterHash with the proposed weighted contrastive likelihood can re-weight training pairs according to class importance, providing a flexible solution for the data imbalance problem. Furthermore, we can also obtain that with the proposed center prior, CenterHash outperforms CenterHash-C by substantial margins of 2.758%/2.803%, 3.035%/3.663%, and 2.995%/3.175% in terms of the average MAP in the tasks ”M → P” and ”P → M” on ADNI1, ADNI2, and AIBL, respectively. This implies that encouraging hash codes from different modalities to be close to the common class centers (as we do in CenterHash) can potentially enhance the inter-class difference and mitigate the inter-modal discrepancy, thus boosting the search performance.
V. Conclusion
This paper presents CenterHash for multi-modal neuroimage retrieval. Specifically, the proposed CenterHash formulates the hash code learning problem in a Bayesian learning framework, where a center prior is proposed to explicitly encourage hash codes of the same class from different modalities to be close. Besides, a weighted contrastive likelihood loss function is also developed to address the imbalanced data problem. Comprehensive empirical evidence suggests that our CenterHash yields state-of-the-art cross-modal retrieval performance on three multi-modal neuroimage datasets. As the future work, it is interesting to incorporate the relevance feedback [52] into our framework to further improve the search performance of information retrieval systems.
Supplementary Material
Acknowledgments
This work was partly supported by NIH grants (Nos. AG041721, AG053867).
Contributor Information
Erkun Yang, Department of Radiology and BRIC, University of North Carolina at Chapel Hill, Chapel Hill, NC 27599, USA.
Mingxia Liu, Department of Radiology and BRIC, University of North Carolina at Chapel Hill, Chapel Hill, NC 27599, USA.
Dongren Yao, Department of Radiology and BRIC, University of North Carolina at Chapel Hill, Chapel Hill, NC 27599, USA; Brainnetome Center and National Laboratory of Pattern Recognition, Institute of Automation, University of Chinese Academy of Sciences, Beijing, China.
Bing Cao, Department of Radiology and BRIC, University of North Carolina at Chapel Hill, Chapel Hill, NC 27599, USA; State Key Laboratory of Integrated Services Networks, Xidian University, Xi’an, China.
Chunfeng Lian, Department of Radiology and BRIC, University of North Carolina at Chapel Hill, Chapel Hill, NC 27599, USA.
Pew-Thian Yap, Department of Radiology and BRIC, University of North Carolina at Chapel Hill, Chapel Hill, NC 27599, USA.
Dinggang Shen, Department of Radiology and BRIC, University of North Carolina at Chapel Hill, Chapel Hill, NC 27599, USA; Department of Artificial Intelligence, Korea University, Seoul 02841, Republic of Korea.
References
- [1].Tajbakhsh N et al. , “Convolutional neural networks for medical image analysis: Full training or fine tuning?” IEEE Trans. Med. Imag, vol. 35, no. 5, pp. 1299–1312, 2016. [DOI] [PubMed] [Google Scholar]
- [2].Cheng C-H and Liu W-X, “Identifying degenerative brain disease using rough set classifier based on wavelet packet method,” Journal of Clinical Medicine, vol. 7, no. 6, p. 124, 2018. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [3].Graham RN, Perriss R, and Scarsbrook AF, “Dicom demystified: A review of digital file formats and their use in radiological practice,” Clinical Radiology, vol. 60, no. 11, pp. 1133–1140, 2005. [DOI] [PubMed] [Google Scholar]
- [4].Grimson WEL, Kikinis R, Jolesz FA, and Black P, “Image-guided surgery,” Scientific American, vol. 280, no. 6, pp. 54–61, 1999. [DOI] [PubMed] [Google Scholar]
- [5].Gioux S, Choi HS, and Frangioni JV, “Image-guided surgery using invisible near-infrared light: Fundamentals of clinical translation,” Molecular Imaging, vol. 9, no. 5, pp. 7290–2010, 2010. [PMC free article] [PubMed] [Google Scholar]
- [6].Mohamed A, Zacharaki EI, Shen D, and Davatzikos C, “Deformable registration of brain tumor images via a statistical model of tumor-induced deformation,” Medical Image Analysis, vol. 10, no. 5, pp. 752–763, 2006. [DOI] [PubMed] [Google Scholar]
- [7].Liu M, Zhang D, and Shen D, “Relationship induced multi-template learning for diagnosis of Alzheimer’s disease and mild cognitive impairment,” IEEE Trans. Med. Imag, vol. 35, no. 6, pp. 1463–1474, 2016. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [8].Owais M, Arsalan M, Choi J, and Park KR, “Effective diagnosis and treatment through content-based medical image retrieval (CBMIR) by using artificial intelligence,” Journal of Clinical Medicine, vol. 8, no. 4, p. 462, 2019. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [9].Purcaru MAP, Repanovici A, and Nedeloiu T, “Non-invasive assessment method using thoracic-abdominal profile image acquisition and mathematical modeling with Bezier curves,” Journal of Clinical Medicine, vol. 8, no. 1, p. 65, 2019. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [10].Avni U, Greenspan H, Konen E, Sharon M, and Goldberger J, “X-ray categorization and retrieval on the organ and pathology level, using patch-based visual words,” IEEE Trans. Med. Imag, vol. 30, no. 3, pp. 733–746, 2010. [DOI] [PubMed] [Google Scholar]
- [11].Liu M, Zhang J, Adeli E, and Shen D, “Joint classification and regression via deep multi-task multi-channel learning for Alzheimer’s disease diagnosis,” IEEE Trans. Biomed. Eng, vol. 66, no. 5, pp. 1195–1206, 2018. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [12].Fan Y, Gur RE, Gur RC, Wu X, Shen D, Calkins ME, and Davatzikos C, “Unaffected family members and schizophrenia patients share brain structure patterns: a high-dimensional pattern classification study,” Biological psychiatry, vol. 63, no. 1, pp. 118–124, 2008. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [13].Fan Y, Rao H, Hurt H, Giannetta J, Korczykowski M, Shera D, Avants BB, Gee JC, Wang J, and Shen D, “Multivariate examination of brain abnormality using both structural and functional MRI,” NeuroImage, vol. 36, no. 4, pp. 1189–1199, 2007. [DOI] [PubMed] [Google Scholar]
- [14].Slichter CP, Principles of magnetic resonance. Springer Science & Business Media, 2013, vol. 1. [Google Scholar]
- [15].Cheng B, Liu M, Shen D, Li Z, and Zhang D, “Multi-domain transfer learning for early diagnosis of Alzheimer’s disease,” Neuroinformatics, vol. 15, no. 2, pp. 115–132, 2017. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [16].Lian C, Liu M, Zhang J, and Shen D, “Hierarchical fully convolutional network for joint atrophy localization and Alzheimer’s disease diagnosis using structural MRI,” IEEE Trans. Pattern Anal. Mach. Intell, vol. 42, no. 2, pp. 880–893, 2020. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [17].Bailey DL, Maisey MN, Townsend DW, and Valk PE, Positron emission tomography. Springer, 2005. [Google Scholar]
- [18].Hsieh J, Computed tomography: principles, design, artifacts, and recent advances. SPIE press, 2003, vol. 114. [Google Scholar]
- [19].Holt A, Bichindaritz I, Schmidt R, and Perner P, “Medical applications in case-based reasoning,” The Knowledge Engineering Review, vol. 20, no. 3, pp. 289–292, 2005. [Google Scholar]
- [20].Liu M, Zhang J, Adeli E, and Shen D, “Landmark-based deep multi-instance learning for brain disease diagnosis,” Medical Image Analysis, vol. 43, pp. 157–168, 2018. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [21].Cao Y et al. , “Medical image retrieval: a multimodal approach,” Cancer Informatics, vol. 13, pp. CIN–S14053, 2014. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [22].Vikram M, Anantharaman A, and BS S, “An approach for multimodal medical image retrieval using latent dirichlet allocation,” in COMAD, 2019, pp. 44–51. [Google Scholar]
- [23].Gao Y, Adeli-M E, Kim M, Giannakopoulos P, Haller S, and Shen D, “Medical image retrieval using multi-graph learning for MCI diagnostic assistance,” in MICCAI. Springer, 2015, pp. 86–93. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [24].Song J, Yang Y, Yang Y, Huang Z, and Shen H, “Inter-media hashing for large-scale retrieval from heterogeneous data sources,” in SIGMOD, 2013, pp. 785–796. [Google Scholar]
- [25].Kumar S and Udupa R, “Learning hash functions for cross-view similarity search,” in IJCAI, 2011. [Google Scholar]
- [26].Cao Y, Liu B, Long M, and Wang J, “Cross-modal hamming hashing,” in ECCV, 2018, pp. 202–218. [Google Scholar]
- [27].Lin Z, Ding G, Hu M, and Wang J, “Semantics-preserving hashing for cross-view retrieval,” in CVPR, 2015, pp. 3864–3872. [Google Scholar]
- [28].Zhang J, Peng Y, and Yuan M, “Unsupervised generative adversarial cross-modal hashing,” in AAAI, 2018, pp. 539–546. [Google Scholar]
- [29].Wu G, Lin Z, Han J, Liu L, Ding G, Zhang B, and Shen J, “Unsupervised deep hashing via binary latent factor models for large-scale cross-modal retrieval.” in IJCAI, 2018, pp. 2854–2860. [Google Scholar]
- [30].Li Z, Zhang X, Müller H, and Zhang S, “Large-scale retrieval for medical image analytics: A comprehensive review,” Medical Image Analysis, vol. 43, pp. 66–84, 2018. [DOI] [PubMed] [Google Scholar]
- [31].Wang J, Shen H, Song J, and Ji J, “Hashing for similarity search: A survey,” arXiv preprint arXiv:1408.2927, 2014. [Google Scholar]
- [32].Toth CD, O’Rourke J, and Goodman JE, Handbook of discrete and computational geometry. CRC press, 2004. [Google Scholar]
- [33].Wang D, Gao X, Wang X, and He L, “Semantic topic multimodal hashing for cross-media retrieval,” in IJCAI, 2015, pp. 3890–3896. [Google Scholar]
- [34].Irie G, Arai H, and Taniguchi Y, “Alternating co-quantization for cross-modal hashing,” in ICCV, 2015, pp. 1886–1894. [Google Scholar]
- [35].Bronstein MM, Bronstein AM, Michel F, and Paragios N, “Data fusion through cross-modality metric learning using similarity-sensitive hashing,” in CVPR, 2010, pp. 3594–3601. [Google Scholar]
- [36].Zhang D and Li W-J, “Large-scale supervised multimodal hashing with semantic correlation maximization,” in AAAI, 2014. [Google Scholar]
- [37].Jiang Q and Li W, “Deep cross-modal hashing,” in CVPR, 2017, pp. 3232–3240. [Google Scholar]
- [38].Yang E, Deng C, Liu W, Liu X, Tao D, and Gao X, “Pairwise relationship guided deep hashing for cross-modal retrieval.” in AAAI, 2017, pp. 1618–1625. [Google Scholar]
- [39].Li C, Deng C, Li N, Liu W, Gao X, and Tao D, “Self-supervised adversarial hashing networks for cross-modal retrieval,” in CVPR, 2018, pp. 4242–4251. [Google Scholar]
- [40].Liu X, Yu G, Domeniconi C, Wang J, Ren Y, and Guo M, “Ranking-based deep cross-modal hashing,” in AAAI, 2019, pp. 4400–4407. [Google Scholar]
- [41].Erin Liong V, Lu J, Tan Y-P, and Zhou J, “Cross-modal deep variational hashing,” in ICCV, 2017, pp. 4077–4085. [Google Scholar]
- [42].Liu W, Wang J, Ji R, Jiang Y, and Chang S-F, “Supervised hashing with kernels,” in CVPR, 2012, pp. 2074–2081. [Google Scholar]
- [43].Lloyd S, “Least squares quantization in PCM,” IEEE Trans. Inf. Theory, vol. 28, no. 2, pp. 129–137, 1982. [Google Scholar]
- [44].Jack CR Jr et al. , “The Alzheimer’s disease neuroimaging initiative (ADNI): MRI methods,” Journal of Magnetic Resonance Imaging: An Official Journal of the International Society for Magnetic Resonance in Medicine, vol. 27, no. 4, pp. 685–691, 2008. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [45].Ellis KA et al. , “The australian imaging, biomarkers and lifestyle (AIBL) study of aging: methodology and baseline characteristics of 1112 individuals recruited for a longitudinal study of alzheimer’s disease,” International psychogeriatrics, vol. 21, no. 4, pp. 672–687, 2009. [DOI] [PubMed] [Google Scholar]
- [46].Sled JG, Zijdenbos AP, and Evans AC, “A nonparametric method for automatic correction of intensity nonuniformity in MRI data,” IEEE Trans. Med. Imag, vol. 17, no. 1, pp. 87–97, 1998. [DOI] [PubMed] [Google Scholar]
- [47].Wang Y, Nie J, Yap P-T, Shi F, Guo L, and Shen D, “Robust deformable-surface-based skull-stripping for large-scale studies,” in MICCAI. Springer, 2011, pp. 635–642. [DOI] [PubMed] [Google Scholar]
- [48].Chollet F et al. , “Keras,” https://keras.io, 2015.
- [49].Abadi M et al. , “Tensorflow: A system for large-scale machine learning.” in OSDI, vol. 16, 2016, pp. 265–283. [Google Scholar]
- [50].Cao Y, Long M, Liu B, Wang J, and KLiss M, “Deep cauchy hashing for hamming space retrieval,” in CVPR, 2018, pp. 1229–1237. [Google Scholar]
- [51].Cao Z, Long M, Wang J, and Yu PS, “Hashnet: Deep learning to hash by continuation,” in ICCV, 2017, pp. 5608–5617. [Google Scholar]
- [52].Rui Y, Huang TS, and Mehrotra S, “Content-based image retrieval with relevance feedback in mars,” in ICIP, 1997, pp. 815–818. [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.











