Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2021 Oct 1.
Published in final edited form as: IEEE Trans Med Imaging. 2020 Apr 14;39(10):3125–3136. doi: 10.1109/TMI.2020.2987796

A Novel Attribute-based Symmetric Multiple Instance Learning for Histopathological Image Analysis

Trung Vu 1,#, Phung Lai 2,#, Raviv Raich 3, Anh Pham 4, Xiaoli Z Fern 5, UK Arvind Rao 6
PMCID: PMC7561004  NIHMSID: NIHMS1633881  PMID: 32305904

Abstract

Histopathological image analysis is a challenging task due to a diverse histology feature set as well as due to the presence of large non-informative regions in whole slide images. In this paper, we propose a multiple-instance learning (MIL) method for image-level classification as well as for annotating relevant regions in the image. In MIL, a common assumption is that negative bags contain only negative instances while positive bags contain one or more positive instances. This asymmetric assumption may be inappropriate for some application scenarios where negative bags also contain representative negative instances. We introduce a novel symmetric MIL framework associating each instance in a bag with an attribute which can be either negative, positive, or irrelevant. We extend the notion of relevance by introducing control over the number of relevant instances. We develop a probabilistic graphical model that incorporates the aforementioned paradigm and a corresponding computationally efficient inference for learning the model parameters and obtaining an instance level attribute-learning classifier. The effectiveness of the proposed method is evaluated on available histopathology datasets with promising results.

Keywords: Histopathological image analysis, Multiple instance learning, Symmetric setting, Attribute learning, Cardinality constraints, Dynamic programming

I. Introduction

Histopathological image analysis is a critical task in cancer diagnosis. Generally, this process is performed by pathologists who are capable of identifying problem-specific cues in a digital image, or a whole slide image (WSI), in order to classify it into one of the disease categories. In recent years, there have been an increasing interest in the application of automatic histopathological image analysis using machine learning algorithms [1], [2]. The advantages of this approach include (i) reducing variability in human interpretations and hence improving classification accuracy, (ii) eliminating a significant amount of trivial cases to ease the burden on pathologists, and (iii) providing quantitative image analysis in the context of one specific disease.

Most conventional approaches to automatic histopathological image analysis fall into the category of fully supervised learning, where training labels are available for the WSI and all of its patches (small blocks extracted from the image). These methods often rely on feature extraction techniques that are customized for a variety of problems, namely texture features [3], [4], spatial features [5], [6], graph-based features [7], [8], and morphological features [9], [10]. Those features can then be used by various classification algorithms such as random forest, support vector machines (SVM), and convolutional neural networks (CNN). Recently, automatic feature discovery framework has also been proposed by Vu et al. [11]. Their discriminative feature-oriented dictionary learning (DFDL) method was shown to outperform many competing methods, particularly in low training scenarios. Nevertheless, one major disadvantage of the fully-supervised approach is the labeling cost. Since each WSI typically comprises hundreds of patches, it requires a large amount of labor to create even a small number of training data. Moreover, labeling a histopathology image at the region-level could be a challenging task with inherent uncertainty, even for experts in the field. To address these issues, many researchers have been studying weakly-supervised learning that focuses on coarse-grain annotations, i.e., only WSI labels are given.

Multiple Instance Learning (MIL) is a framework for weakly supervised learning (limited supervision) that relies on a training set of bags of instances labeled at the bag level only. In our scenario, each WSI (bag) contains a collection of tissue segments (instances), but the annotation of cancer-type is only available at the image/bag level. The goal is to develop a classifier to predict both bag level and instance level labels. Various MIL-based approaches have been applied successfully in biochemistry [12], [13], image classification and segmentation [14]-[16], text categorization [17], [18], object recognition, tracking and localization [19]-[21], behavior coding [22], anomaly detection [23], and co-saliency detection [24]. In histopathological image classification, only a limited number of MIL approaches have been studied in literature. In [25], Dundar et al. introduced a multiple instance learning approach (MILSVM) based on the implementation of the large margin principle with different loss functions defined for positive and negative samples. Later on, Xu et al. [26] adopted the clustering concept into MIL to propose an integrated framework of segmentation, clustering, and classification named multiple clustered instance learning (MCIL). The authors also extended their work by taking into consideration the contextual prior in the MIL training stage to reduce the intrinsic ambiguity. Most recently, a comparison of general MIL-based methods on histopathological image classification, namely, mi-SVM and MI-SVM [27], miGraph and MIGraph [28] has also been reported in [29]. All of the aforementioned methods, nonetheless, deal with predicting the presence or absence of cancer, and are based on an asymmetric assumption that all instances in a negative bag are negative while each positive bag contains at least one positive instance. This commonly-used MIL assumption may be not suitable for other MIL setting such as predicting the cancer type based on histopathology images. In this case, only a fraction of the tissue segments can be useful towards recognizing the cancer type of each WSI, and such MIL setting is symmetric in that both positive and negative bags contain stereotypical instances featured for each cancer type along with irrelevant ones.

In this paper, we consider the binary classification problem of cancer types based on histopathology images. Our contribution in this paper is as follows. First, we introduce a novel symmetric MIL where both negative and positive bags contain relevant and irrelevant instances. Second, we propose a probabilistic graphical model, named Attribute-based Symmetric Multiple Instance Learning (AbSMIL), that incorporates cardinality constraints on the relevant instances in each bag to leverage possible prior knowledge and develop a forward-backward dynamic programming algorithm for learning model parameters. To facilitate efficient inference, the online learning version of our algorithm is also presented. Finally, the advantages of the proposed framework are demonstrated by experiments on instance annotation and bag level label prediction using classical multi-instance image recognition datasets as well as histopathology datasets.

II. Related work

In MIL setting, it is important to make assumptions regarding the relationship between the instances within a bag and the class label of the bag. Most of the MIL algorithms follow the standard assumption that each positive bag contains at least one positive instance and negative bags contain only negative instances. In one of the early works, Maron and Lozano-Pérez [30] introduced the concept point that is close to one instance in each positive bag and far from all instances in negative bags. The diverse density (DD) framework [31], [32] is then developed based on the idea of finding the best candidate concept. Later on, Andrew et al. [27] extended SVM approach to mi-SVM and MI-SVM by finding a separating hyperplane such that at least one instance in every positive bag is located on one side and all instances in negative bags are located on the other side of the hyperplane. In a different approach, Zhou et al. [28] proposed MIGraph and miGraph methods that map every bag to a graph and explicitly model the relationships among the instances within a bag. A number of single-instance algorithms have also been adapted to a multiple-instance context such as MIL-Boost [33], KI-SVM [34], Latent SVM [35], MI-CRF [36]. They all maintain the classical asymmetric assumption in their algorithms.

For some MIL applications where the class of a bag is defined by instances belonging to more than one concept, the standard assumption may be viewed as too strict. Therefore, researchers have recently shown a general interest in other more loose assumptions such as the collective assumption [37], [38]. In this paper, we consider the problem of predicting cancer types in histopathology images and propose a symmetric assumption that treats classes equally (see Fig. 1). In a positive bag, there is at least one relevant instance that demonstrates certain attributes associated with the positive class (e.g., stripes going from bottom-left to top-right). Similarly, a negative bag also contains at least one relevant instance that demonstrates certain negative attributes (e.g., stripes going from bottom-right to top-left). Finally, both bags contain some irrelevant instances whose attributes do not contribute to the difference between the two classes (e.g., dotted and wave). The goal is to learn the positive, negative, and irrelevant attributes and use them to address various classification tasks within this framework.

Fig. 1.

Fig. 1.

The setting of the proposed attribute-based symmetric MIL framework. The goal is to classify positive and negative bags and to figure out distinctly positive and negative relevant attributes (stripes going from bottom-left to top-right or bottom-right to top-left, respectively) and irrelevant attributes (dotted and wave) in each bag category. The proposed model helps control the relevant instance proportions in the bags.

Our assumption is motivated by multi-instance multi-label learning (MIML), a generalization of single-instance binary classifiers to the multiple-instance case. In MIML setting, each instance is associated with a latent instance label and each bag is the union of its instance labels [39]. With the presence of novel class instances [40], the bag label only involves the known instance classes and does not provide information about the presence or absence of the novel class. This consideration is similar to our symmetric MIL setting where only a small portion of the relevant instances are labeled and irrelevant instances are often ignored by pathologists. However, a simple reduction of the model for MIML would fail in the cancer classification problem as the bag labels in MIL are binary but not a subset of the class labels. Recently, You et al. [41] introduced cardinality constraints to the MIML setting and demonstrated that optimizing the control over the maximum number of instances per bag can significantly improve the performance of the model. Motivated by the result, we propose using cardinality constraints to limit the number of relevant attributes in each bag. Given that there are hundreds of instances per bag, cardinality constraints can help control the model complexity. While the probabilistic machinery for implementing cardinality constraints is similar to the approach in You et al. [41], the graphical model and inference methods differ. The modeling difference: instance-level labels include unknown attribute/cluster labels (see Fig. 1), whereas in the MIML setting in [41] instance-level labels are taken directly from bag-level labels (e.g., the bag-level label {2, 5} implies that relevant instances must have labels of either 2 or 5). The inference difference: (i) to promote the cluster diversity, we introduce entropy regularization to the original log-likelihood objective, (ii) to reduce the computational burden of large histopathology images, we apply a stochastic gradient descent approach. Both of which are not included in [41].

III. Problem formulation and proposed model

In this section, we formulate the problem of cancer type classification and describe our proposed AbSMIL approach.

A. Problem Formulation

We consider a collection of B bags and their labels, denoted by {Xb,Yb}b=1B. The bag level label Yb ∈ {0, 1} represents each of the two cancer types. The bth bag contains nb instances Xb={xbi}i=1nb where xbiRd is a feature vector for the ith instance. Our goal is to predict the bag label based on the set of feature vectors for its instances. Moreover, we would like to learn a robust model that are capable of explaining the labeling decision. To that end, we consider the following attribute-based assumption on the data.

Relevant/irrelevant instances and attributes assumption.

We assume that each instance in a bag can be either relevant or irrelevant. Relevant instances provide useful information towards a specific class. Further, there may be more than one types of relevant instances for the same class, which we capture as instance attributes (clusters). In Fig. 1, for example, the positive (or negative) class always has two relevant attributes: 1 and 2 (or 3 and 4). Formally, we assume the attribute zbi, corresponding to the ith instance in the bth bag, belongs to the set {0,1,,C}, where 0 is reserved for irrelevant instances and the non-zero integers represent C attributes for relevant instances. For the bth bag, we denote the set of hidden attributes for all instances by zb = [zb1, zb2, … , znb]T, and the binary vector indicating the presence/absence of relevant attributes by yb=[yb1,yb2,,ybC]T{0,1}C.

No mixed-class attributes assumption.

Let us call attributes that provide sufficient information for predicting the positive bag as positive attributes (PAs) and similarly, for negative bags as negative attributes (NAs). Since the goal is to find distinct attributes that discriminate the two classes, we assume there is no shared relevant attribute between positive bags and negative bags. To be more specific, we consider the first half of the attribute set C+={1,,C2} as PAs and the second half C={C2+1,,C} as NAs (C is even in our assumption). Although we focus on a balanced number of PAs and NAs, the proposed model is not limited to this symmetry when extending to other settings.

Relevance cardinality constraints.

In many applications, the domain knowledge may provide practical information regarding the number of relevant instances. Such heuristics can be exploited with constraints on the maximum number of relevant instances per bag, i.e.,

i=1nbIzbi0nmax,forb=1,,B,

where Iσ denotes the indicator function taking the value 1 if σ is true and 0 otherwise. Unlike the standard MIL assumption, the relevance cardinality constraints require both positive and negative bags to have a bounded number of relevant instances. In our discriminative model, nmax is a tuning parameter that can be optimized to improve the classification performance. Although those make our model more amenable to cancer type recognition, we note that they are not equivalent to sparsity-promoting prior in generative models such as Laplacian [42] or spike-and-slab [43].

Potential extension to multinomial classification.

As the no mixed-class attributes assumption and relevance cardinality constraint apply for every bag, we can easily extend our model to the problem of multinomial MIL classification by considering more categories of attributes. For three-cancer-type classification, by way of illustration, we can consider three groups of labels {1,,C3}, {C3+1,,2C3} and {2C3+1,,C}.

B. Attribute-Based Symmetric MIL Model

The graphical representation of the proposed model is illustrated in Fig. 2. We assume that instances are independent given all the feature vectors in the bag. We follow the discriminative approach in [39] to model the relationship between the attribute of an instance zbi and its feature vector xbi by a multinomial logistic regression function

Pbic(w)=P(zbi=cxbi,w)=ewcTxbic=0CewcTxbi, (1)

where wcRd is the weight for the cth attribute and w=[w0,w1,w2,,wC]. For the cth attribute in the bth bag, a presence or absence indicator ybc is computed following the standard assumption [44]

ybc=1i=1nbIzbic,forc=1,,C.

Fig. 2.

Fig. 2.

Graphical model for the proposed AbSMIL model. Observed variables are shaded.

To encode the cardinality constraint i=1nbIzbi0nmax for the bth bag, we introduce an binary observation variable

Tb=(Ii=1nbIzbi01)(Ii=1nbIzbi0nmax).

By letting Tb = 1, we enforce the constraint on the number of relevant instances per bag during training stage. The bag level label Yb is computed based on the presence of positive and negative attributes

Ybyb={0,for(c=1C2{ybc=0})(c=C2+1C{ybc=1}),1,for(c=1C2{ybc=1})(c=C2+1C{ybc=0}),2,otherwise.}

For the completeness, the model may allow Yb = 2; however, the considered dataset contains only positive bags Yb = 1 and negative bags Yb = 0. Thus, we can ignore the case Yb = 2 in our derivation. To summarize, our model includes the observation {Yb,Tb,Xb}b=1B, unknown classifier parameter w, hidden variables {yb,zb}b=1B and a tuning parameter nmax.

IV. Inference

This section provides details of our graphical model, including the derivation of the incomplete log-likelihood, the expectation maximization (EM) approach to estimate the model parameters and the prediction of both instance and bag level labels. To facilitate efficient inference, we further present the dynamic programming method for the expectation step in combination with online learning for the maximization step.

A. Regularized Maximum Likelihood

Since the bags are independent, the normalized negative incomplete log-likelihood is given by

Lincm(w)=1Bb=1B(logP(Yb,TbXb,w)+logP(Xb))

where we assume that P(Xb) is a constant w.r.t. w. To ensure the attributes are distinctly different, we introduce entropy regularization that quantifies the cluster diversity (see [45], [46]):

Hb(w)=i=1nb(c=0CPbic(w)logPbic(w)), (2)

and a quadratic penalty to control the complexity of the model and avoid over-fitting. Now let us denote Lb(w)=logP(Yb,TbXb,w), the objective in our approach can be formulated as

L(w)=1Bb=1B(Lb(w)+λeHb(w))+λqw22, (3)

where λq, λe are parameters of the quadratic regularizer and the entropy minimizer, respectively. Following the principle of maximum likelihood estimation (MLE), our goal is to minimize L(w). However, this optimization problem is challenging as the probability P(Yb, TbXb, w) in the objective function is not trivial to compute. For instance, to obtain P(Yb = 1, TbXb, w), we marginalize the joint probability model of all the model variables over the hidden variables:

P(Yb=1,TbXb,w)=zb,ybP(Yb=1,Tb,zb,ybXb,w).

Using the graphical model in Fig. 2, we can expand the joint probability as

P(Yb=1,Tb,zb,ybXb,w)=P(Yb=1yb)P(ybzb)P(Tbzb)P(zbXb,w) (4)

Finally, by substituting (4) into the marginalization, we obtain

P(Yb=1,TbXb,w)=zb,ybP(Yb=1yb)P(ybzb)P(Tbzb)P(zbXb,w)=zbIi=1nbIzbi01Ii=1nbIzbi0nmaxi=1nbP(zbixbi,w)

where the summation in the intermediate step is over zb{0,1,,C}nb and yb{0,1}C, while the latter summation is over zb{0,1,,C2}nb. Notice that this summation includes a total of approximately k=1nmax(nbk)(C2)k terms, and hence, is computationally expensive or even intractable when nmax is large. Therefore, we consider an EM approach as the alternative approach to efficiently minimizing L(w).

B. Estimation of Model Parameters

Let us begin by identifying the negative complete log-likelihood as

Lcm(w)=1Bb=1BlogP(Yb,Tb,yb,zbXb,w)+K,

where K corresponds to the term 1Bb=1B log P(Xb) independent of the parameter vector w. Substituting the model dependence structure from (4) into Lcm and absorbing terms that do not depend on w into the constant yields

Lcm(w)=1Bb=1Bi=1nb(c=0CIzbi=cwcTxbi+log(c=0CewcTxbi))

up to a constant. The EM algorithm seeks to find the minimum of L(w) by iteratively applying the following two steps:

E-step.

The surrogate function is obtained by taking the expectation with respect to the current conditional distribution of the latent data {y, z} given the observed data {Y, T, X} and the current estimates of the parameters w(k)

Ey,zY,T,X,w(k)[Lcm(w)]=1Bb=1BJb(w,w(k))

where Jb(w,w(k)) is defined as

i=1nb(c=0CPbicpost(w(k))wcTxbi+log(c=0CewcTxbi)) (5)

and Pbicpost(w(k))=P(zbi=cYb,Tb,xbi,w(k)) denotes the posterior probability. Thus, our surrogate function with regularization is given by

Q(w,w(k))=1Bb=1B(Jb(w,w(k))+λeHb(w)+λqw22). (6)

In order to compute Q(w,w(k)), it is necessary to compute the posterior probability Pbicpost(w(k)). Using the conditional rule, this probability can be then determined by

Pbicpost(w(k))=P(zbi=c,Yb,Tbxbi,w(k))t=0CP(zbi=t,Yb,Tbxbi,w(k)). (7)

We provide a detailed calculation of (7) in Section IV-C.

M-step.

Since we consider the negative log-likelihood as the objective, this step is essentially to minimize the surrogate Q by solving

w(k+1)=argminwQ(w,w(k)).

Generally, minimizing Q(w,w) is a non-trivial optimization problem. Alternatively, we use the Generalized EM approach to facilitate a descent approach by taking steps along the gradient of Q(w,w). By Fisher’s identity, this gradient coincides with the gradient of the objective function

L(w)ww=w=Q(w,w)ww=w=1Bb=1BQb(w,w)ww=w.

The gradient Qb(w,w)wc can be computed as follows

Qb(w,w)wc=i=1nb(Pbic(w)Pbicpost(w))xbi+λei=1nbPbic(w)(t=0CPbit(w)(wtwc)Txbi)xbi+λqwc. (8)

The full gradient involves enumerating all the bags. To reduce the computation per iteration, we further propose a single-bag-based stochastic gradient descent approach in the flavor of the Pegasos algorithm [47], [48]. At the kth iteration, a random bag bk is chosen and a single-bag-based gradient is computed as follows

w(k+1)=Πτ(w(k)ηkQbk(w,w(k))ww=w(k)),

where ηk=1kλq, τ=2(λe+1)nbλqlog(C+1), and Πτ(v)=min{1,τv}v. Detailed derivation of the stochastic gradient update is provided in Appendix A.

C. Proposed Dynamic Programming for E-step

The brute-force calculation of P(zbi = c, Yb, Tbxbi, w(k)) in (7) requires marginalization over all other instance attributes (i.e., zbj for j = 1, … , nb and ji), which is exponential in the number of instances per bag (O(Cnb1)). Traditional approaches to address such intractable problems often resort to approximation techniques such as approximate inference [49], variational approximation [50], and black-box alpha [51]. In a recent line of work [39], [40], [52], Pham et al. proposed a dynamic programming approach for exact and efficient computation of the posterior. Motivated by the result, we follow a similar idea of converting the V-structure to the chain structure for efficient inference. Specifically, we define a new latent variable Nbi as the “signed” number of relevant attributes in the first i instances of the bth bag. The representation of Nbi is given by the finite state machine in Fig. 3(a). Each instance in a bag is represented by one of the input symbols {0, +, −} (corresponding to an irrelevant instance, a relevant instance in the positive class, or a relevant instance in the negative class, respectively). The number of relevant instances is represented by the set of states {0, ±1, ±2,…, ±nmax}, where the sign indicates whether instances are belong to the positive class or the negative class. For the completeness, we introduce the error state {x} for the case there is a mix of positive and negative attributes in the bag. Practically, this state will not be reached due to our no mixed-class attributes assumption. Thanks to the introduction of Nbi, the posterior probability can be calculated efficiently using the forward and backward message passing on the chain structure (see Fig. 3(b)-(e)). To further simplify our forward-backward message passing derivation, let us denote

Pbi0=Pbi0(w),Pbi+=cC+Pbic(w),Pbi=cCPbic(w) (9)

where the parameter w is omitted for simplicity. Below is the details of our dynamic programming algorithm for computing the posterior.

Fig. 3.

Fig. 3.

(a) The number of positively- (negatively-) labeled instances in the first i instances, Nbi, as a finite state machine. (b) A reformulation of Fig. 2 as a chain on Nbi2, and (c)-(e) recursive calculation of forward, backward messages, and posterior probability.

Step 1. Forward message passing.

This step computes the forward messages defined as

αbi(l)P(Nbi=lXb,w)forl=0,±1,,±nmax.

The first message is initialized by

αb1(l)=Il=0Pb10+Il=1Pb1++Il=1Pb1, (10)

The update equation for subsequent messages is given by

αbi(l)=Pbi0αb(i1)(l)+Il>0Pbi+αb(i1)(l1)+Il<0Pbiαb(i1)(l+1), (11)

for i = 2,…, nb.

Step 2. Backward message passing.

This step computes the backward messages defined as

βbi(l)P(Yb,TbNbi=l,Xb,w).

The first backward message is initialized by

βbnb(l)=IYb=1I0<lnmax+IYb=0Inmaxl<0, (12)

The update equation for subsequent messages is given by

βbi(l)=Pb(i+1)0βi+1(l)+I0l<nmaxPb(i+1)+βi+1(l+1)+I0l>nmaxPb(i+1)βi+1(l1), (13)

for i = nb – 1, … , 1.

Step 3. Joint probability calculation.

This step computes the joint probability defined as

Pbicjoint(w)P(zbi=c,Yb,Tbxbi,w),forc=0,1,,C,

First, initialize Pb1cjoint(w) by

(Ic=0β1(0)+IcC+β1(1)+IcCβ1(1))Pb1c(w) (14)

Next, for i = 2, … , nb, perform the update for Pbicjoint(w) by

(Ic=0l=nmax+nmaxβi(l)αi1(l)+IcC+l=0nmax1βi(l+1)αi1(l))+(IcCl=nmax+10βi(l1)αi1(l))Pbic(w). (15)

The detailed derivation for the forward messages, backward messages, and joint probability calculation are given in Appendix B, C, and D, respectively. We summarize the proposed AbSMIL approach in Algorithm 1.

D. Prediction

After learning the weight vector w, the instance level label for new test data can be predicted as follows:

z^bi=argmax0cCP(zbi=cxbi,w).

Note that this prediction is made without knowing the bag label. To this end, the bag level label can be predicted as

Y^b=argmaxm{0,1}P(Yb=m,Tb=1Xb,w)
Algorithm 1 Attribute-based Symmetric Multiple Instance Learning (AbSMIL)
1:Input:Training data{Xb,Yb}b=1B,cardinality constraintnmax,positive constantsλqandλe,initial weightw(0)2:Output:{w(k)}3:k=04:repeat5:Select a random bagb6:E-step:7:Compute prior probabilityPbic(w(k))using (1)8:Compute prior probabilityPbi0,Pbi+andPbiusing (9)9:Computeforwardmessageαbi(l)fori=1,,nbandl=0,±1,,±nmaxusing (10) and (11)10:Computebackwardmessageβbi(l)fori=nb,,1andl=0,±1,,±nmaxusing (12) and (13)11:ComputejointprobabilityPbicjoint(w(k))fori=1,,nbandc=0,1,,Cusing (14) and (15)12:ComputeposteriorprobabilityPbicpost(w(k))fori=1,,nbandc=0,1,,Cusing(7)13:M-step:14:τ=2(λe+1)nbλqlog(C+1)15:ComputeQb(w,w(k))wcforc=0,1,,Cusing(8)16:w(k+1)=Πτ(w(k)1kλqQb(w,w(k))w(k))17:k=k+118:untilstopping criteria is met

where P(Yb = m, Tb = 1 ∣ Xb, w) is given by

{l=1nmaxαbnb(l)form=1,l=nmax1αbnb(l)form=0.}

E. Complexity Analysis

To compute the posterior probability Pbicpost(w), we need to obtain the forward and backward messages over all instances of the bag (nb) and all possible numbers of relevant instances (2nmax + 1). Given the label Yb of the bag, the overall complexity of the E-step is O(nbnmax). Our proposed dynamic programming approach offers an efficient computation that is linear with the number of instances per bag nb when the number of relevant instances is constrained to be small. On the other hand, the M-step requires O(nbCd) to compute each single bag-based stochastic gradient. Thus, the total complexity per iteration is O(nb(Cd+nmax)). When each instance is a high-dimensional vector, i.e., dnmax, the M-step dominates other steps and the overall complexity per iteration of our algorithm is O(Cnbd). In terms of memory, the dominant factor stems from forward and backward messages. In order to store all possible messages, the space complexity per bag is O(nbnmax), which is often smaller than that of the instances per bag (O(dnb)) in practice.

Non-linear extension.

If there is a large number of attributes, a kernel extension can be used as an alternative to the linear model. By introducing non-linear kernel functions, data is transformed to capture the clustering nature of multiple attributes without assigning additional clusters. In practice, C=2 is often sufficient for binary classification. To implement a radial basis function (RBF) kernel, one can follow the approach of [53], replacing x with

ϕ(x)=1k[cos(g1Tx),sin(g1Tx),,cos(gkTx),sin(gkTx)]T (16)

where g1,,gki.i.d.N(0d,σg2Id). We demonstrate the effectiveness of the non-linear extension through the experiment in Section V-D.

V. Experiments

In this section, we evaluate the performance of the proposed algorithm (AbSMIL) in terms of runtime and classification accuracy. We compare its performance with state-of-the-art MIL frameworks on a number of datasets, including 4 different histopathology datasets.

A. Datasets

We consider 2 sets of experiments to evaluate the runtime (see V-C) and classification performance (see V-D). Firstly, in order to evaluate the runtime performance of AbSMIL, we create a synthetic MIL-based dataset from the MNIST dataset as follows. We select all the images of digits 0, 1, 2 and manually label the right/left-tilted 0s and 1s as the positive/negative relevant instances, and 2s as the irrelevant instances. Then, for each positive (negative) bag, we randomly select nrel relevant instances from the images of right-tilted (left-tilted) digits and nbnrel irrelevant instances from the images of 2. The total number of bags is B = 200, with a balanced number between positive and negative bags. The goal is to learn the orientation (left-tilted versus right-tilted) of 0 and 1 while the orientation of 2 is ignored.

Secondly, in order to evaluate the classification performance of AbSMIL, we use seven benchmark datasets in a wide range of applicability. (i) The first group includes three datasets popularly used in studies of MIL: Tiger, Fox, and Elephant datasets (see [12], [27], [28], [54], [55]). There are 200 bags in which 100 positive bags associated with the target animal images and 100 negative bags associated with other kinds of animal. For each of these datasets, 140 images are used for training and 60 images for test. The maximum number of instance per bag is 13. More details of these datasets can be found in [27]. (ii) The second group contains histopathology images of mammalian organs, provided by the Animal Diagnostics Lab (ADL) at Pennsylvania State University. Each of the three datasets (kidney, lung and spleen) contains 300 images of size 4000×3000 from either inflammatory or healthy tissues. A healthy tissue image largely consists of healthy patches while an inflammatory tissues have a dominant portion of diseased patches. 250 samples are used for training and the remaining ones are used for testing. There are 130 instances per bag in each dataset. More details of ADL datasets can be found in [11]. (iii) The last dataset, referred as the TCGA dataset [56], contains WSIs of brain cancer from The Cancer Genome Atlas (TCGA), provided by the National Institute of Health. This dataset contains 96 histopathology samples of two types of glioma: 48 samples for astrocytoma and 48 samples for oligodendroglioma. In each WSI (as one bag), cancerous regions occupy only a small portion of various shapes and color shading, and this portion is usually surrounded by benign cells, making the TCGA dataset inherently harder than the ADL datasets [11]. Noticeably, the problem of classifying cancer types in the TCGA dataset fits well the symmetric setting of AbSMIL. To obtain instances from each WSI, we first remove some redundant parts such as glass or folding areas, then randomly select a set of 100×100×3 patches in the image (with potential overlaps). The instances are obtained by featurizing these patches using 84 features including histogram of oriented gradients, histogram of gray images and SFTA-texture features [57]. The maximum number of instances per bag is 1258. A split of 76 training versus 20 test images is considered in our experiment. In all of these datasets, a balanced number of positive/negative bags is used in both training and testing stages.

B. Baselines

In our experiment, we compare the proposed method with a variety of popular approaches, including mi-SVM [27], MIL-Boost [33], miGraph [28], MCIL [26], DFDL [11], ORLR [39] and MIML-NC [40]. The first four methods are MIL-based approaches that utilize the standard asymmetric assumption. In particular, mi-SVM assumes that there is at least one pattern from every positive bag in the positive halfspace, while all patterns belonging to negative bags are in the negative halfspace. In [28], miGraph implicitly constructs a graph that model each bag and the relations among the instances within the bag. Both of these methods were previously used on the Tiger, Fox, Elephant datasets. In histopathological image classification, MCIL and DFDL are the state-of-the-art methods. While MCIL is designed for MIL setting and can be used directly in our experiment, DFDL learns a dictionary bases using manually extracted regions in the WSIs (i.e., annotating at instance level). In order to adapt DFDL to the MIL setting, we resort to assuming that all instances in positive bags are positive and all instances in negative bags are negative (similar to [11]). MIL-Boost can be seen as a special case of MCIL where there is one cluster in positive bags. The last two methods, ORLR and MIML-NC, are generally designed for the multiple instance multiple label learning (MIML) setting. As discussed in Section II, MIML setting is similar to our symmetric MIL assumption where both positive and negative bag contains instances from multiple clusters. However, while the MIML-based models are designed for a more general setting where the bag label is the union of all instance labels, the proposed AbSMIL method is designed for the specific setting of HIC where bag labels are binary based on the relevant instances.

C. Runtime Evaluation

In the following, we present the experiment to evaluate and compare the runtime performance of AbSMIL with the aforementioned algorithms.

Setting.

In the first set of simulations, we test the impact of the cardinality constraint nmax on the computational performance of AbSMIL by varying nb through the set of values {500, 1000, 2000, 5000} and nmax through the set of values {2, 5, 10, 20, 50, 100, 200, 500, 1000, 2000, 5000} (such that nmaxnb). Since among the aforementioned values we include fairly large values of nb and nmax, we consider the following to reduce the runtime. First, the original 784 = 28 × 28 image vector is trimmed down to a 30-dimensional vector by selecting only the first d = 30 elements. Next, the number of attributes is set to C=2 and the parameters λq and λe are fixed to 10−6 and 10−2, respectively. We also adjust the number of epochs inversely proportional to nmax so that each setting takes approximately the same amount of time. Finally, we report the average running time per epoch for each setting. In the second set of simulations, we test the impact of the number of attributes C on the computational performance of AbSMIL by varying nmax through the set of values {50, 100, 200, 500} and C through the set of values {2, 4, 10, 20, 50, 100, 200, 500, 1000, 2000, 5000}. The same setting for C, d, λq and λe from the first simulation is used in this simulation and the average running time per epoch is then reported. To reduce the runtime variance, each setting is repeated 10 times on the same computer.

Results and Analysis.

We plot our results from the simulations in Fig. 4. The left plot shows the average running time of AbSMIL as a function of nmax. It can be seen that the running time exhibits two different modes with respect to the change in nmax. Recall that our overall complexity per bag is O(nb(Cd+nmax)). When nmax is small, the dominant term in the sum is Cd (here, Cd=60) and we observe a flat region at the beginning of the four curves for different values of nb. When nmax becomes larger, linear behavior (i.e., runtime ∝ nmax) is observed as nmax gains dominance over Cd. Note that in practice, the implementation of AbSMIL for the special case when nmax = nb is less costly than reported and this case is equivalent to AbSMIL with no constraint. In the right plot, a similar behavior is observed: the running time remains stable when C is small and a linearly increasing runtime is observed when C is large. Note that for different values of nmax, the running time converges since the Cd becomes dominant, thereby making the different values of nmax negligible. Similar to the number of weak classifiers T in MCIL, the number of attributes C in our algorithm also contributes as a linear term in the computational complexity. We also notice variation in the running time due to the differences among the computational cluster nodes used in this experiment.

Fig. 4.

Fig. 4.

Running time of AbSMIL as a function of the cardinality constraint nmax (a) and the number of attributes C (b) in log-log scale. Each curve corresponds to a different setting in terms of nb (a) and in terms of nmax (b). The runtime values in each curve are calculated by averaging the runtime across 10 different runs (indicated by the markers). In (a), the dash-dotted lines and the dashed lines are added to demonstrate the asymptotic behavior of analytical complexity O(nb(Cd+nmax)) when nmax is small and when nmax is large, respectively.

D. Real-world Datasets

This subsection presents the experiment to evaluate and compare the accuracy performance of AbSMIL with the baseline algorithms on different real-world datasets.

Settings.

Since instance labels are unavailable for the aforementioned real-world datasets, all the results are evaluated using bag level prediction with 10-fold cross validation [61]. The value of nmax is selected in the set {5%, 10%, 20%, 30%, 40%, 50%, 100%} of nb in Tiger, Fox, and Elephant datasets; {5%, 10%, 20%, 50%, 75%} of nb in Kidney, Lung, Spleen datasets; and {1%, 2%, 5%, 10%, 20%} of nb in TCGA dataset. Note that xbi’s are assumed to be normalized such that the mean of each entry over the data is zero and the variance of each entry is 1. In AbSMIL, λq and λe are searched in a grid of {10−6, 10−5, 10−4, 10−3, 10−2}. For each setting, AbSMIL is initialized ten times, and the model that yields the lowest training negative log-likelihood is chosen to report the performance. We also report the results for two variants of the proposed method: AbSMIL - No constraint and AbSMIL - Kernel. AbSMIL - No constraint is essentially AbSMIL without using cardinality constraint. AbSMIL - Kernel is mentioned in (16), and its hyperparameter σg2 is selected in {0.999, 0.998, 0.995, 0.99, 0.98, 0.95, 0.9, 0.8, 0.5}. For mi-SVM, we change the loss constant C ∈ {10−2, 10−1, 1, 10, 102} and choose a linear kernel function. In MIL-Boost and MCIL, we search over the generalized mean (GM), the log-sum-exponential (LSE) softmax function with parameter named r ∈ {15, 20, 25} and the number of weak classifiers named T ∈ {150, 200, 250}. In miGraph, we tune the number of classes C ∈ {50, 100, 150, 200, 500}, RBF kernel γ ∈ {1, 5, 10, 15, 20, 25, 50} and threshold used in computing the weight of each instance ϵ ∈ {0.1, 0.2, 0.3, 0.4, 0.5}. ORLR and MIML-NC are tuning-free methods. For DFDL, we tune the regularization parameter ρ ∈ {10−5, 10−4, 10−3, 10−2, 10−1}, sparsity level parameter λ ∈ {10−5, 10−4,10−3, 10−2, 10−1}, and the number of dictionary bases k ∈ {100, 200, 500}. These tuning values are extracted from the corresponding aforementioned publications.

Results and Analysis.

Fig. 5 demonstrates the bag level accuracy of the aforementioned methods on different real-world datasets. Overall, the proposed AbSMIL method shows a competitive performance with other state-of-the-art methods for various datasets in our experiment. Additionally, we observe a major improvement by adding the cardinality constraint to the AbSMIL model. Let us discuss the detail of each group of datasets below.

Fig. 5.

Fig. 5.

Bag level accuracy various algorithms in seven real-world datasets. The proposed AbSMIL method shows a competitive performance with other state-of-the-art methods for various datasets in our experiment. Especially for the case of TCGA dataset, the outstanding result of AbSMIL indicates that the symmetric MIL assumption and the cardinality constraint are more suitable to such settings. The results with EM-DD [31], MI Kernel [58], MILES [59], and MICA [60] are observed from [60] and [55].

Tiger, Fox, and Elephant datasets:

Overall, Fig. 5 shows that AbSMIL obtains the highest accuracy on Elephant dataset, while AbSMIL - Kernel outperforms other methods on Fox and Tiger datasets. As pointed out in [27], due to the limited accuracy of the image segmentation, the relatively small number of region descriptors, and the small training set size, Fox dataset yields a harder classification problem than the other two datasets. Regarding result of ‘AbSMIL-No constraint’, it can be seen that removing the cardinality constraint lessens the performance of AbSMIL significantly (e.g., roughly 7%).

Kidney, Lung, and Spleen datasets:

The results reported for AbSMIL are obtained by setting nmax at 20% of the total instances per bag in Kidney and Lung datasets and at 50% in Spleen dataset. Again, it can be seen that AbSMIL and AbSMIL - Kernel slightly outperform other methods in terms of bag level accuracy. A comparison of images of instances from the same attribute obtained by AbSMIL and DFDL on Kidney dataset are shown in Fig. 6. For C=2, the cancerous tissue recognized by AbSMIL appears as a mixture of vascular proliferation (red parts) and necrosis (areas with no nuclei - appeared as the purple dot in the image). On the other hand, for C=4, cancerous tissue splits into two distinctly different categories. In comparison, the performance of DFDL appears worse since in this MIL setting, DFDL assumes that all instances in positive bags are positive and all instances in negative bags are negative.

Fig. 6.

Fig. 6.

Groups of instances in the same class predicted by AbSMIL and DFDL. Cancer groups are in solid blue while normal ones in dotted green. Compared to two DFDL bases, AbSMIL exhibits distinctly different tissue types between the two categories with C=2 and can be clustered to distinguish phenomenon in each category with C=4.

TCGA dataset:

The result obtained by AbSMIL in this experiment is at nmax of nearly 5% the total number of instances in a bag. As in Fig. 5, AbSMIL and AbSMIL - Kernel significantly outperform the other frameworks. This improvement is remarkable compared to the previous experiment and matches with our intuition on the TCGA dataset: in each WSI, cancerous regions occupy only a small portion, as opposed to the ADL dataset where inflammatory tissues dominate the entire WSI. Thereby, the cardinality constraint offers an advantage to AbSMIL over other methods in effectively gathering information from patches.

VI. Conclusion

In this paper, we introduced Symmetric MIL, a novel setting for multiple-instance learning where both positive and negative bags contain relevant class-specific instances as well as irrelevant instances that do not contribute to differentiating the classes. We presented a probabilistic model for attribute-based Symmetric MIL that accommodates for the presence of numerous irrelevant instances in the data, and takes into account prior information about the sparsity of the relevant instances. We developed an efficient inference approach that is linear in the number of instances and is suitable for the online learning scenario, updating the model using one bag at a time. We evaluated our framework on the real-world datasets: Tiger, Fox, Elephant, Kidney, Lung, Spleen, and TCGA. We obtained competitive results on all datasets and in particular for TCGA where bags contain mainly irrelevant instances. The results validate the merit of the proposed symmetric MIL framework.

Acknowledgment

The authors would like to thank Dr. Souptik Barua at Rice University for his help with TCGA data organization.

This work is partially supported by the National Science Foundation grants CCF-1254218, DBI-1356792, and IIS-1055113, American Cancer Society Research Scholar Grant RSG-RSG-16-005-01, and National Institutes of Health 1R37CA21495501A1.

Appendix A Gradient Derivation

From (6), the gradient of Qb w.r.t. wc can be decomposed into

Qb(w,w)wc=Jb(w,w)wc+Hb(w)wc+λqwc. (17)

On the one hand, differentiating Jb from (5) yields

Jb(w,w)wc=i=1nbPbicpost(w)xbi+i=1nbPbic(w)xbi.

On the other hand, differentiating Hb from (2) yields

Hb(w)wc=wc(i=1nbt=0CPbit(w)logPbit(w))=i=1nbt=0C(1+logPbit(w))Pbit(w)(It=cPbic(w))xbi=i=1nbPbic(w)xbi(t=0CPbit(w)(logPbit(w)logPbic(w)))=i=1nbPbic(w)(t=0CPbit(w)(wtwc)Txbi)xbi=i=1nbPbic(w)(t=0CPbit(w)(wtTxbit=0CPbit(w)xbi)xbi.

Thus, substituting the results back into (17) yields

Qb(w,w)wc=i=1nb(Pbic(w)Pbicpost(w))xbi+λei=1nbPbic(w)(t=0CPbit(w)(wtwc)Txbi)xbi+λqwc.

The ball radius τ is then computed as follows. Let w=argminwQb(w,w). Then the following always holds

Qb(w,w)Qb(0,w)Jb(w,w)+λeHb(w)+λqw22i=1nblog(C+1)+λei=1nblog(C+1)λqw22(λe+1)nblog(C+1)w22(λe+1)nblog(C+1)λqτ=w2(λe+1)nbλqlog(C+1)

where (*) stems from the fact that both Jb(w,w) and Hb(w) are non-negative.

Appendix B Forward Message Derivation

Initialize αi(l) for i = 1 and l = −nmax, … ,n+max:

α1(l)=P(N1=lXb,w)=cP(N1=l,zb1=cxb1,w)=cP(N1=lzb1=c)Pb1c(w)=c(Il=0Ic=0+Il=1IcC++Il=1IcC)Pb1c(w)=Il=0Pb10+Il=1cC+Pb1c+Il=1cCPblc=Il=0Pb10+Il=1Pb1++Il=1Pb1.

Update αi(l) for i = 2, … , nb and l = nmax, … , +nmax:

αi(l)=P(Ni=lXb,w)=c,kP(Ni=l,Ni1=k,zbi=cXb,w)=c,kP(Ni=lNi1=k,zbi=c)P(Ni1=kXb,w)Pbic(w)=c,k(Ic=0Ik=l+Il1IcC+Ik=l1)+(Il1IcCIk=l+1)αi1(k)Pbic(w)=αi1(l)Pbi0+Il>0αi1(l1)Pbi++Il<0αi1(l+1)Pbi.

Appendix C Backward Message derivation

Initialize βi(l) for i = nb and l = −nmax, … , +nmax:

βnb(l)=P(Yb,Tb=1Nnb=l,Xb,w)=P(Tb=1Nnb=l)P(YbNnb=l)=IYb=1I0<lnmax+IYb=0Inmaxl<0.

Update βi(l) for i = nb – 1, … , 1, l = −nmax, … , +nmax:

βi(l)=P(Yb,Tb=lNi=l,Xb,w)=c,kP(Yb,Tb=1,Ni+1=k,zb(i+1)=cNi=l,Xb,w)=c,kP(Yb,Tb=1Ni+1=k,Xb,w)P(Ni+1=kNi=l,zb(i+1)=c)Pb(i+1)c(w)=c,kβi+1(k)(Ic=0Ik=l+I0l<nmaxIcC+Ik=l+1)+(I0l>nmaxIcCIk=l1)Pb(i+1)c(w)=βi+1(l)Pb(i+1)0(w)+I0l<nmaxβi+1(l+1)Pb(i+1)++I0l>nmaxβi+1(l1)Pb(i+1).

Appendix D Joint Probability Derivation

For valid Yb and Tb, the state machine will not travel through the invalid state x. So we can safely ignore the invalid state x in our derivation and only consider valid states 0, ±1, … , ±nmax. Initialize Pbicjoint(w) for i = 1:

Pb1cjoint(w)=P(zb1=c,Yb,Tb=1Xb,w)=kP(Yb,Tb=1,N1=k,zb1=cXb,w)=kP(Yb,Tb=1N1=k,Xb,w)P(N1=kzb1=c)Pb1c(w)=kβ1(k)(Ic=0Ik=0+IcC+Ik=1)(+IcCIk=1)Pb1c(w)=(Ic=0β1(0)+IcC+β1(1)+IcCβ1(1))Pb1c(w).

Update rule for i = 2,…, nb – 1:

Pbicjoint(w)=P(zbi=c,Yb,Tb=1Xb,w)=k,lP(Yb,Tb=1,Ni=k,Ni1=l,zbi=cXb,w)=k,lP(Yb,Tb=1Ni=k,Xb,w)P(Ni=kNi1=l,zbi=c)P(Ni1=lXb,w)Pbic(w)k,lβi(k)(Ic=0Ik=l+Il<nmaxIcC+Ik=l+1)(+Il>nmaxIcCIk=l1)αi1(l)Pbic(w)=(Ic=0lβi(l)αi1(l)+IcC+l=0nmax1βi(l+1)αi1(l)+IcCl=nmax+10βi(l1)αi1(l))Pbic(w).

Contributor Information

Trung Vu, School of EECS, Oregon State University, Corvallis, OR.

Phung Lai, School of EECS, Oregon State University, Corvallis, OR.

Raviv Raich, School of EECS, Oregon State University, Corvallis, OR.

Anh Pham, School of EECS, Oregon State University, Corvallis, OR.

Xiaoli Z. Fern, School of EECS, Oregon State University, Corvallis, OR.

UK Arvind Rao, Department of Computational Medicine and Bioinformatics, and Department of Radiation Oncology, The University of Michigan, Ann Arbor, MI 48109.

References

  • [1].Gurcan MN, Boucheron LE, Can A, Madabhushi A, Rajpoot NM, and Yener B, “Histopathological image analysis: A review,” IEEE Rev. Biomed. Eng, vol. 2, pp. 147–171, 2009. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [2].Komura D and Ishikawa S, “Machine learning methods for histopathological image analysis,” Comput. Struct. Biotechnol. J, vol. 16, pp. 34–42, 2018. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [3].Kong J, Sertel O, Shimada H, Boyer KL, Saltz JH, and Gurcan MN, “Computer-aided evaluation of neuroblastoma on whole-slide histology images: Classifying grade of neuroblastic differentiation,” Pattern Recognit., vol. 42, no. 6, pp. 1080–1092, 2009. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [4].Dundar MM, Badve S, Bilgin G, Raykar V, Jain R, Sertel O, and Gurcan MN, “Computerized classification of intraductal breast lesions using histopathological images,” IEEE Trans. Biomed. Eng, vol. 58, no. 7, pp. 1977–1984, 2011. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [5].Huang P-W and Lee C-H, “Automatic classification for pathological prostate images based on fractal analysis,” IEEE Trans. Med. Imag, vol. 28, no. 7, pp. 1037–1050, 2009. [DOI] [PubMed] [Google Scholar]
  • [6].Srinivas U, Mousavi HS, Monga V, Hattel A, and Jayarao B, “Simultaneous sparsity model for histopathological image representation and classification,” IEEE Trans. Med. Imag, vol. 33, no. 5, pp. 1163–1179, 2014. [DOI] [PubMed] [Google Scholar]
  • [7].Doyle S, Agner S, Madabhushi A, Feldman M, and Tomaszewski J, “Automated grading of breast cancer histopathology using spectral clustering with textural and architectural image features,” in Int. Symp. Biomed. Imag.: From Nano to Macro. IEEE, 2008, pp. 496–499. [Google Scholar]
  • [8].Ta V-T, Lézoray O, Elmoataz A, and Schüpp S, “Graph-based tools for microscopic cellular image segmentation,” Pattern Recognit., vol. 42, no. 6, pp. 1113–1125, 2009. [Google Scholar]
  • [9].Boucheron LE, “Object- and spatial-level quantitative analysis of multispectral histopathology images for detection and characterization of cancer,” Ph.D. dissertation, University of California, Santa Barbara, March 2008. [Online]. Available: https://vision.ece.ucsb.edu/abstract/518 [Google Scholar]
  • [10].Zana F and Klein J-C, “Segmentation of vessel-like patterns using mathematical morphology and curvature evaluation,” IEEE Trans. Image Process, vol. 10, no. 7, pp. 1010–1019, 2001. [DOI] [PubMed] [Google Scholar]
  • [11].Vu TH, Mousavi HS, Monga V, Rao G, and Rao UA, “Histopathological image classification using discriminative feature-oriented dictionary learning,” IEEE Trans. Med. Imag, vol. 35, no. 3, pp. 738–751, 2016. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [12].Ray S and Craven M, “Supervised versus multiple instance learning: An empirical comparison,” in Proc. Int. Conf. Mach. Learn, 2005, pp. 697–704. [Google Scholar]
  • [13].Hajimirsadeghi H and Mori G, “Multiple instance real boosting with aggregation functions,” in Int. Conf. Pattern Recognit. IEEE, 2012, pp. 2706–2710. [Google Scholar]
  • [14].Wang X, Wang B, Bai X, Liu W, and Tu Z, “Max-margin multiple-instance dictionary learning,” in Proc. Int. Conf. Mach. Learn, 2013, pp. 846–854. [Google Scholar]
  • [15].Li D, Wang J, Zhao X, Liu Y, and Wang D, “Multiple kernel-based multi-instance learning algorithm for image classification,” J. Vis. Commun. and Image Represent, vol. 25, no. 5, pp. 1112–1117, 2014. [Google Scholar]
  • [16].Wu J, Zhao Y, Zhu J-Y, Luo S, and Tu Z, “MILCut: A sweeping line multiple instance learning paradigm for interactive image segmentation,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit, 2014, pp. 256–263. [Google Scholar]
  • [17].Settles B, Craven M, and Ray S, “Multiple-instance active learning,” in Adv. Neural Inf. Process. Syst, 2008, pp. 1289–1296. [Google Scholar]
  • [18].Alpaydın E, Cheplygina V, Loog M, and Tax DM, “Single-vs. multiple-instance classification,” Pattern Recognit, vol. 48, no. 9, pp. 2831–2838, 2015. [Google Scholar]
  • [19].Babenko B, Yang M-H, and Belongie S, “Robust object tracking with online multiple instance learning,” IEEE Trans. Pattern Anal. Mach. Intell, vol. 33, no. 8, pp. 1619–1632, 2011. [DOI] [PubMed] [Google Scholar]
  • [20].Kück H and de Freitas N, “Learning about individuals from group statistics,” in Conf. Uncertain. Artif. Intell. AUAI Press, 2005, p. 332339. [Google Scholar]
  • [21].Zhang K and Song H, “Real-time visual tracking via online weighted multiple instance learning,” Pattern Recognit., vol. 46, no. 1, pp. 397–411, 2013. [Google Scholar]
  • [22].Gibson J, Katsamanis A, Romero F, Xiao B, Georgiou P, and Narayanan S, “Multiple instance learning for behavioral coding,” IEEE Trans. Affect. Comput, vol. 8, no. 1, pp. 81–94, 2017. [Google Scholar]
  • [23].Quellec G, Lamard M, Cozic M, Coatrieux G, and Cazuguel G, “Multiple-instance learning for anomaly detection in digital mammography,” IEEE Trans. Med. Imag, vol. 35, no. 7, pp. 1604–1614, 2016. [DOI] [PubMed] [Google Scholar]
  • [24].Zhang D, Meng D, and Han J, “Co-saliency detection via a self-paced multiple-instance learning framework,” IEEE Trans. Pattern Anal. Mach. Intell, vol. 39, no. 5, pp. 865–878, 2017. [DOI] [PubMed] [Google Scholar]
  • [25].Dundar MM, Badve S, Raykar VC, Jain RK, Sertel O, and Gurcan MN, “A multiple instance learning approach toward optimal classification of pathology slides,” in Int. Conf. Pattern Recognit. IEEE, 2010, pp. 2732–2735. [Google Scholar]
  • [26].Xu Y, Zhu J-Y, Chang E, and Tu Z, “Multiple clustered instance learning for histopathology cancer image classification, segmentation and clustering,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit. IEEE, 2012, pp. 964–971. [Google Scholar]
  • [27].Andrews S, Tsochantaridis I, and Hofmann T, “Support vector machines for multiple-instance learning,” in Adv. Neural Inf. Process. Syst, 2003, pp. 577–584. [Google Scholar]
  • [28].Zhou Z-H, Sun Y-Y, and Li Y-F, “Multi-instance learning by treating instances as non-iid samples,” in Proc. Int. Conf. Mach. Learn, 2009, pp. 1249–1256. [Google Scholar]
  • [29].Kandemir M, Feuchtinger A, Walch A, and Hamprecht FA, “Digital Pathology: Multiple instance learning can detect Barrett’s cancer,” in Int. Symp. Biomed. Imag. IEEE, 2014, pp. 1348–1351. [Google Scholar]
  • [30].Maron O and Lozano-Pérez T, “A framework for multiple-instance learning,” in Adv. Neural Inf. Process. Syst, 1998, pp. 570–576. [Google Scholar]
  • [31].Zhang Q and Goldman SA, “EM-DD: An improved multiple-instance learning technique,” in Adv. Neural Inf. Process. Syst, 2002, pp. 1073–1080. [Google Scholar]
  • [32].Chen Y and Wang JZ, “Image categorization by learning and reasoning with regions,” J. Mach. Learn. Res, vol. 5, no. Aug, pp. 913–939, 2004. [Google Scholar]
  • [33].Zhang C, Platt JC, and Viola PA, “Multiple instance boosting for object detection,” in Adv. Neural Inf. Process. Syst, 2006, pp. 1417–1424. [Google Scholar]
  • [34].Li Y-F, Kwok JT, Tsang IW, and Zhou Z-H, “A convex method for locating regions of interest with multi-instance learning,” in Joint Eur. Conf. Mach. Learn. Knowl. Discov. Databases. Springer, 2009, pp. 15–30. [Google Scholar]
  • [35].Felzenszwalb PF, Girshick RB, McAllester D, and Ramanan D, “Object detection with discriminatively trained part-based models,” IEEE Trans. Pattern Anal. Mach. Intell, vol. 32, no. 9, pp. 1627–1645, 2010. [DOI] [PubMed] [Google Scholar]
  • [36].Deselaers T and Ferrari V, “A conditional random field for multiple-instance learning,” in Proc. Int. Conf. Mach. Learn, 2010, pp. 287–294. [Google Scholar]
  • [37].Foulds J and Frank E, “A review of multi-instance learning assumptions,” Knowl. Eng. Rev, vol. 25, no. 1, pp. 1–25, 2010. [Google Scholar]
  • [38].Carbonneau M-A, Cheplygina V, Granger E, and Gagnon G, “Multiple instance learning: A survey of problem characteristics and applications,” Pattern Recognit., vol. 77, pp. 329–353, 2018. [Google Scholar]
  • [39].Pham AT, Raich R, and Fern XZ, “Dynamic programming for instance annotation in multi-instance multi-label learning,” IEEE Trans. Pattern Anal. Mach. Intell, vol. 39, no. 12, pp. 2381–2394, 2017. [DOI] [PubMed] [Google Scholar]
  • [40].Pham AT, Raich R, Fern XZ, and Pérez Arriaga J, “Multi-instance multi-label learning in the presence of novel class instances,” in Proc. Int. Conf. Mach. Learn, 2015, pp. 2427–2435. [Google Scholar]
  • [41].You Z, Raich R, Fern XZ, and Kim J, “Weakly supervised dictionary learning,” IEEE Trans. Signal Process, vol. 66, no. 10, pp. 2527–2541, 2018. [Google Scholar]
  • [42].Tibshirani R, “Regression shrinkage and selection via the lasso,” J. Royal Stat. Soc.: Series B (Methodological), vol. 58, no. 1, pp. 267–288, 1996. [Google Scholar]
  • [43].Andersen MR, Winther O, and Hansen LK, “Bayesian inference for structured spike and slab priors,” in Adv. Neural Inf. Process. Syst, 2014, pp. 1745–1753. [Google Scholar]
  • [44].Weidmann N, Frank E, and Pfahringer B, “A two-level learning method for generalized multi-instance problems,” in Proc. Eur. Conf. Mach. Learn. Springer, 2003, pp. 468–479. [Google Scholar]
  • [45].Jaakkola T, Meila M, and Jebara T, “Maximum entropy discrimination,” in Adv. Neural Inf. Process. Syst, 1999, pp. 470–476. [Google Scholar]
  • [46].Hennig P and Schuler CJ, “Entropy search for information-efficient global optimization,” J. Mach. Learn. Res, vol. 13, no. Jun, pp. 1809–1837, 2012. [Google Scholar]
  • [47].Shalev-Shwartz S, Singer Y, Srebro N, and Cotter A, “Pegasos: Primal estimated sub-gradient solver for SVM,” Math. Program, vol. 127, no. 1, pp. 3–30, 2011. [Google Scholar]
  • [48].Cesa-Bianchi N, Shalev-Shwartz S, and Shamir O, “Efficient learning with partially observed attributes,” J. Mach. Learn. Res, vol. 12, no. Oct, pp. 2857–2878, 2011. [Google Scholar]
  • [49].Wang Z, Mülling K, Deisenroth MP, Ben HA , Vogt D, Schölkopf B, and Peters J, “Probabilistic movement modeling for intention inference in human-robot interaction,” Int. J. Rob. Res, vol. 32, no. 7, pp. 841–858, 2013. [Google Scholar]
  • [50].Futoma J, Sendak M, Cameron B, and Heller K, “Predicting disease progression with a model for multivariate longitudinal clinical data,” in Mach. Learn. Healthcare Conf, 2016, pp. 42–54. [Google Scholar]
  • [51].Hernández-Lobato JM, Li Y, Rowland M, Hernández-Lobato D, Bui T, and Turner R, “Black-box α-divergence minimization,” in Proc. Int. Conf. Mach. Learn, 2016, pp. 1511–1520. [Google Scholar]
  • [52].Pham AT, Raich R, and Fern XZ, “Efficient instance annotation in multi-instance learning,” in Proc. IEEE Workshop Stat. Signal Process. IEEE, 2014, pp. 137–140. [Google Scholar]
  • [53].Rahimi A and Recht B, “Random features for large-scale kernel machines,” in Adv. Neural Inf. Process. Syst, 2008, pp. 1177–1184. [Google Scholar]
  • [54].Dundar M, Krishnapuram B, Rao R, and Fung GM, “Multiple instance learning for computer aided diagnosis,” in Adv. Neural Inf. Process. Syst, 2007, pp. 425–432. [Google Scholar]
  • [55].Leistner C, Saffari A, and Bischof H, “MIForests: Multiple-instance learning with randomized trees,” in Proc. Eur. Conf. Comput. Vis. Springer, 2010, pp. 29–42. [Google Scholar]
  • [56].Tomczak K, Czerwińska P, and Wiznerowicz M, “The Cancer Genome Atlas (TCGA): an immeasurable source of knowledge,” Contemporary Oncology, vol. 19, no. 1A, p. A68, 2015. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [57].Costa AF, Humpire-Mamani G, and Traina AJM, “An efficient algorithm for fractal analysis of textures,” in SIBGRAPI Conference on Graphics, Patterns and Images. IEEE, 2012, pp. 39–46. [Google Scholar]
  • [58].Gärtner T, Flach PA, Kowalczyk A, and Smola AJ, “Multi-instance kernels,” in Proc. Int. Conf. Mach. Learn, 2002, pp. 179–186. [Google Scholar]
  • [59].Chen Y, Bi J, and Wang JZ, “MILES: Multiple-instance learning via embedded instance selection,” IEEE Trans. Pattern Anal. Mach. Intell, vol. 28, no. 12, pp. 1931–1947, 2006. [DOI] [PubMed] [Google Scholar]
  • [60].Mangasarian OL and Wild EW, “Multiple instance classification via successive linear programming,” J. Optim. Theory Appl, vol. 137, no. 3, pp. 555–568, 2008. [Google Scholar]
  • [61].Kohavi R, “A study of cross-validation and bootstrap for accuracy estimation and model selection,” in Proc. Int. Joint Conf. Artif. Intell, 1995, pp. 1137–1145. [Google Scholar]

RESOURCES