Skip to main content
Entropy logoLink to Entropy
. 2018 Mar 27;20(4):229. doi: 10.3390/e20040229

On the Coherence of Probabilistic Relational Formalisms

Glauber De Bona 1, Fabio G Cozman 1,*
PMCID: PMC7512744  PMID: 33265320

Abstract

There are several formalisms that enhance Bayesian networks by including relations amongst individuals as modeling primitives. For instance, Probabilistic Relational Models (PRMs) use diagrams and relational databases to represent repetitive Bayesian networks, while Relational Bayesian Networks (RBNs) employ first-order probability formulas with the same purpose. We examine the coherence checking problem for those formalisms; that is, the problem of guaranteeing that any grounding of a well-formed set of sentences does produce a valid Bayesian network. This is a novel version of de Finetti’s problem of coherence checking for probabilistic assessments. We show how to reduce the coherence checking problem in relational Bayesian networks to a validity problem in first-order logic augmented with a transitive closure operator and how to combine this logic-based approach with faster, but incomplete algorithms.

Keywords: relational Bayesian networks, probabilistic relational models, coherence checking

1. Introduction

Most statistical models are couched so as to guarantee that they specify a single probability measure. For instance, suppose we have N independent biased coins, so that heads has probability p for each one of them. Then, the probability of a particular configuration of all coins is exactly pn(1p)Nn, where n is the number of heads in the configuration. Using de Finetti’s terminology, we can say that the probabilistic assessments and independence assumptions are coherent as they are satisfied by a probability distribution [1]. The study of coherence and its consequences has influenced the foundations of probability and statistics, serving as a subjectivist basis for probability theory [2,3], as a broad prescription for statistical practice [4,5] and generally as a bedrock for decision-making and inference [6,7,8].

In this paper, we examine the coherence checking problem for probabilistic models that enhance Bayesian networks with relations and first-order formulas: more precisely, we introduce techniques that allow one to check whether a given relational Bayesian network, or a given probabilistic relational model is guaranteed to specify a probability distribution. Note that “standard” Bayesian networks are, given some intuitive assumptions, guaranteed to be coherent [9,10,11]. The challenge here is to handle models that enlarge Bayesian networks with significant elements of first-order logic; we do so by resorting to logical inference itself as much as possible. In the remainder of this section, we explain the motivation for this study and the basic terminology concerning it, and at the end of this section, we state our goals and our approach in more detail.

To recap, a Bayesian network consists of a directed acyclic graph, where each node is a random variable, and a joint probability distribution over those variables, such that the distribution and the graph satisfy a Markov condition: each random variable is independent of its non-descendants given its parents. (In a directed acyclic graph, node X is a parent of node Y if there is an edge from X to Y. The set of parents of node X is denoted Pa(X). Similarly, we define the children of a node, the descendants of a node, and so on.)

If all random variables X1,,Xn in a Bayesian network are categorical, then the Markov condition implies a factorization:

P(X1=x1,,Xn=xn)=i=1nP(Xi=xi|Pa(Xi)=πi), (1)

where πi is the projection of {X1=x1,,Xn=xn} on Pa(Xi).

Typically, one specifies a Bayesian network by writing down the random variables X1,,Xn, drawing the directed acyclic graph, and then settling on probability values P(Xi=xi|Pa(Xi)=πi), for each Xi, each xi and each πi. By following this methodological guideline, one obtains the promised coherence: a unique joint distribution is given by Expression (1).

The following example introduces some useful notation and terminology.

Example 1.

Consider two neighbors, Mary and Tina. The probability that a house is burglarized is 0.001 in their town. The alarm of a house rings with probability 0.9 given that the house is burglarized and with probability 0.01 given the house is not burglarized. Finally, if either alarm rings, the police are called. This little story, completed with some assumptions of independence, is conveyed by the Bayesian network in Figure 1, where burglary(x) means that the house of x (either Mary or Tina) is burglarized; similarly alarm(x) means that the alarm of x’s house rings; and finally calls just means that the police are called by someone.

In this paper, every random variable is binary with values zero and one, the former meaning “false” and the latter meaning “true”. Furthermore, we often write P(X), where X is a random variable, to mean the event {X=1}, and we often write P(¬X) to mean the event {X=0}.

Note also that we use, whenever appropriate, logical expressions with random variable names, such as alarm(Mary)burglary(Tina) to mean the disjunction of the proposition stating that alarm(Mary) is true and the proposition that burglary(Tina) is true. A random variable name has a dual use as a proposition name.

From the Bayesian network in Figure 1, we compute P(alarm(Mary))=0.9×0.001+0.01×0.999=0.01899 and P(calls)=0.01899+0.01899(0.01899)2.

Figure 1.

Figure 1

Bayesian network modeling the burglary-alarm-call scenario with Mary and Tina. In the probabilistic assessments (right), the logical variable x stands for Mary and for Tina.

Here are some interesting scenarios that enhance the previous example:

Example 2.

(Scenario 1) Consider that now we have three people, Mary, Tina and John, all neighbors. We can easily imagine an enlarged Bayesian network, with two added nodes related to John, and a modified definition where calls=alarm(Mary)alarm(Tina)alarm(John).

(Scenario 2) It is also straightforward to expand our Bayesian network to accommodate n individuals a1,a2,,an, all neighbors. We may even be interested in reasoning about calls without any commitments to a fixed n, where calls is a disjunction over all instances of alarm(x). For instance, we have that P(¬calls)=(10.01899)n; hence, the probability of a call to the police will be larger than half for a city with more than 36 inhabitants. No single Bayesian network allows this sort of “aggregate” inference.

(Scenario 3) Consider a slightly different situation with three people, where: Mary and Tina are neighbors Tina and John are neighbors, but Mary and John are not neighbors. Suppose also that each person may call the police, depending on neighboring alarms. This new situation is codified into the Bayesian network given in Figure 2.

(Scenario 4) Suppose we want to extend Scenario 3 to a town with n people. Without knowing which pairs are neighbors, there is no way we can predict in advance the structure of the resulting Bayesian network. However, we can reason about the possible networks: for instance, we know that each set of n people produces a valid Bayesian network, without any cycles amongst random variables.

Figure 2.

Figure 2

Bayesian network modeling Scenario 3 in Example 2. Probabilistic assessments are just as in Figure 1, except that, for each x, calls(x) is the disjunction of its corresponding parents.

There are many other scenarios where probabilistic modeling must handle repetitive patterns such as the ones described in the previous examples, for instance in social network analysis or in processing data in the semantic web [12,13,14]. The need to handle such “very structured” scenarios has led to varied formalisms that extend Bayesian networks with the help of predicates and quantifiers, relational databases, loops and even recursion [15]. Thus, instead of dealing with a random variable X at a time, we deal with parameterized random variables [16]. We write X(x) to refer to a parameterized random variable that yields a random variable for each fixed x in a domain; if we consider individuals a and b in a domain, we obtain random variables X(a) and X(b).

Plates offer a popular scheme to manipulate parameterized random variables [17]. A plate is a set of parameterized random variables that share a logical variable, meaning that they are indexed by elements of the same domain. A plate is usually drawn as a rectangle (associated with a domain) containing parameterized random variables. Figure 3 shows simple plate models for the burglary-alarm-call scenario described in Scenario 2 of Example 2.

Figure 3.

Figure 3

Plate models for Scenario 2 of Example 2; that is, for the burglary-alarm-call scenario where there is a single random variable calls. Left: A partial plate model (without the calls random variable), indicating that parameterized random variables burglary(x) and alarm(x) must be replicated for each person x; the domain consists of the set of persons as marked in the top of the plate. Note that each parameterized random variable must be associated with probabilistic assessments; in this case, the relevant ones from Figure 1. Right: A plate model that extends the one on the left by including the random variable calls.

Plates appeared with the BUGS package, to facilitate the specification of hierarchical models, and have been successful in applications [18]. One restriction of the original plate models in the BUGS package is that a parameterized random variable could not have children outside of its enclosing plate. However, in practice, many plate models violate this restriction. Figure 3 depicts a partial plate model that satisfies the restriction of the original BUGS package (left), and the plate model that violates it (right). Note that as long as the graph consisting of parameterized random variables is acyclic, we know that every Bayesian network generated from the plate model is indeed consistent.

Several other combinations of parameterized random variables and graph-theoretical representations have been proposed, often grouped under the loose term “Probabilistic Relational Model (PRM)” [10,19,20]. Using PRMs, one can associate parameterized random variables with domains, impose constraints on domains and even represent limited forms of recursion [19,21]. A detailed description of PRMs is given in Section 4; for now, it suffices to say that a PRM is specified as a set of “classes” (each class is a set of individuals), where each class is associated with a set of parameterized random variables and additionally by a relational database that gives the relations amongst individuals in classes. The plate model in Figure 3 (left) can be viewed as a diagrammatic representation for a minimalistic PRM, where we have a class Person containing parameterized random variables. Note that such a minimalistic PRM with a single class Person cannot encode Scenario 4 in Example 2, as in that scenario, we have pairs of interacting individuals.

Suppose that we want a PRM to represent Scenario 4 in Example 2. Now, the class Person must include parameterized random variables burglary, alarm and calls. The challenge is how to indicate which Persons are parents of a particular calls(x). To do so, one possibility is to introduce another class, say Neighborhood, where each element of Neighborhood refers to two elements of Person. In Section 4 we show how the resulting PRM can be specified textually; for now, we want to point out that finding a diagrammatic representing this PRM is not an obvious matter. Using the scheme suggested by Getoor et al. [19], we might draw the diagram in Figure 4. There, we have a class Person, a class Neighborhood and a “shadow” class Person that just indicates the presence of a second Person in any Neighborhood pair. Dealing with all possible PRMs indeed requires a very complex diagrammatic language, where conditional edges and recursion can be expressed [21].

Figure 4.

Figure 4

A Probabilistic Relational Model (PRM) for Scenario 4 in Example 2, using a diagrammatic scheme suggested by Getoor et al. [19]. A textual description of this PRM is presented in Section 4.

Instead of resorting to diagrams, one may instead focus just on textual languages to specify repetitive Bayesian networks. A very solid formalism that follows this strategy is Jaeger’s Relational Bayesian Networks (RBNs) [22,23]. In RBNs, relations within domains are specified using a first-order syntax as input, returning an output that can be seen as a typical Bayesian network. For instance, using syntax that will be explained later (Section 2), one can describe Scenario 4 in Example 2 with the following RBN:

  • burglary(x) = 0.001;

  • alarm(x) = 0.9 * burglary(x) + 0.01 * (1-burglary(x));

  • calls(x) = NoisyOR { alarm(y) | y; neighbor(x,y) };

One problem that surfaces when we want to use an expressive formalism, such as RBNs or PRMs, is whether a particular model is guaranteed to always produce consistent Bayesian networks. Consider a simple example [19].

Example 3.

Suppose we are modeling genetic relationships using the parameterized random variable gene(x), for any person x. Now, the genetic features of x depend on the genetic features of the mother and the father of x. That is, we want to encode:

If y and z are such that motherOf(y,x) and fatherOf(z,x) are true, then the probability of gene(x) depends on gene(y) and gene(z).

It we try to specify a PRM for this setting, we face a difficulty in that some instances of gene could depend on other instances of the same parameterized random variable. Indeed, consider drawing a diagram for this PRM, using the conventions suggested by Getoor et al. [19]. We would need a class Person, containing parameterized random variable gene, and two shadow classes, one for the father and one for the mother; a fragment of the diagram is depicted in Figure 5. If we could have a Person that appears as the father of his own father, we would have a cycle in the generated Bayesian network. Of course, we know that such a cycle can never be generated because neither the transitive closure of motherOf, nor of fatherOf can contain a cycle. However, just by looking at the diagram, without any background understanding of motherOf and fatherOf, we cannot determine whether coherence is guaranteed.

Figure 5.

Figure 5

The PRM for the genetic example, as proposed by Getoor et al. [19].

The possibility that RBNs and PRMs may lead to cyclic (thus inconsistent) Bayesian networks has been noticed before. Jaeger [23] suggested that checking whether an RBN always produces consistent Bayesian networks, for a given class of domains, should be solved by logical inference, being reducible to deciding the validity of a formula from first-order logic augmented with a transitive closure operator. This path has been explored by De Bona and Cozman [24], yielding theoretical results of very high computational complexity. On a different path, Getoor et al. [19] put forward an incomplete, but more intuitive way of ensuring coherence for their PRMs; in fact, assuming that some sets of input relations never form cycles, one can easily identify a few cases where coherence is guaranteed.

Thus, we have arrived at the problems of interest in this paper: Suppose we have an RBN or a PRM. Is it coherent in the sense that it can be always satisfied by a probability distribution? Is it coherent in the sense that it always produces a unique probability distribution? Such are the questions we address, by exploring a cross-pollination of ideas described in the previous paragraph. In doing so, we bring logical rigor to the problem of coherence of PRMs and present a practical alternative to identifying coherent RBNs.

After formally introducing relational Bayesian networks in Section 2, we review, in Section 3, how its coherence problem can be encoded in first-order logic by employing a transitive closure operator. Section 4 presents PRMs and the standard graph-based approach to their coherence checking. The logical methods developed for the coherence problem of RBNs are adapted to PRMs in Section 5. Conversely, in Section 6, we adapt the graph techniques presented for tackling the coherence of PRMs to the formalism of RBNs.

2. Relational Bayesian Networks

In this section, we briefly introduce the formalism of Relational Bayesian Networks (RBNs). We use the version of RBNs presented in [23], as that reference contains a thorough exposition of the topic.

Let S and R be disjoint sets of relation symbols, called the predefined relations and probabilistic relations, respectively. We assume that S contains the equality symbol =, to be interpreted in the usual way. Each predicate symbol is associated with a positive integer k, which is its arity. Given a finite domain D={d1,,dn}, if V is a set of relation symbols (as R or S), a V-structure D is an interpretation of the symbols in V into sets of tuples in D. Formally, a V-structure D maps each relation symbol vV with arity k into a subset of Dk. We denote by ModD(V) the set of all V-structures over a given finite domain D. Given a domain D, a vV with arity k and a tuple tDk, v(t) is said to be a ground V-atom. A V-structure D defines truth values for ground atoms: if v is mapped to a relation containing t, we say that v(t) is satisfied by D, which is denoted by Dv(t).

Employing the syntax of function-free first-order logic, we can construct formulas using a vocabulary of relations V, together with variables, quantifiers and Boolean connectives. We call these V-formulas, and their meaning is given by the first-order logic semantics, as usual, through the V-structures. We denote by φ(x1,,xm) a V-formula where x1,,xk are free variables, in the usual sense. If φ is a V-formula and D is a V-structure, Dφ denotes that φ is satisfied by D.

A random relational structure model for S and R is a partial function P that takes an S-structure D, over some finite domain D, and returns a probability distribution P(D):ModD(R)[0,1] over the R-structures on the same domain. As an R-structure can be seen as total assignment over the ground R-atoms, P(D) can be seen as a joint probability distribution over these ground atoms. An example of random relational structure model would be a function PS4 in Scenario 4 of Example 2 that receives an S-structure of neighbors and returns a joint probability distribution over ground atoms for burglary(·), alarm(·), calls(·). In that scenario, a given configuration D of neighbors, over a given domain D, implies a specific Bayesian network whose variables are the ground atoms for burglary(·), alarm(·), calls(·), which encodes a joint probability distribution, PS4(D), over these variables. If D is the configuration of neighbors from Scenario 4 of Example 2, PS4(D) would be captured by the Bayesian network in Figure 2.

Relational Bayesian networks provide a way to compactly represent random relational structure models. This is achieved by mapping each S-structure into a ground Bayesian network that encodes a probability distribution over R-structures. To begin, this ground Bayesian network has nodes representing r(t) (ground atoms), for each rR and tDk, where k is the arity of r. Thus, given the domain D of the input S-structure, the nodes in the corresponding Bayesian network are already determined. To define the arcs and parameters of the Bayesian network associated with an arbitrary S-structure, relational Bayesian networks employ their central notion of probability formula.

Probability formulas are syntactical constructions intended to link the probability of a ground atom r(t) to the probabilities of other ground atoms r(t), according to the S-structure. Once an R-structure and an S-structure are fixed, then for elements t1,,tk in the domain D, a probability formula F(t1,,tk) should evaluate to a number in [0,1].

The definition of probability formulas makes use of combination functions, which are functions from finite multi-sets over the interval [0,1] to numbers in the same interval. We use {|·|} to denote multi-sets. For instance, NoisyOR is a combination function such that, if c1,,cn[0,1], NoisyOR{|c1,,cn|}=1i=1n(1ci).

Definition 1.

Given disjoint sets S and R of relation symbols and a tuple x of k variables, F(x) is a (S,R)-probability formula if:

  • (constants) F(x)=c for a c[0,1];

  • (indicator functions) F(x)=r(x) for an rR with arity k;

  • (convex combinations) F(x)=F1(x)F2(x)+(1F1(x))F3(x), where F1(x),F2(x),F3(x) are probability formulas, or;

  • (combination functions) F(x)=comb{|F1(x,y),,Fm(x,y)|y;φ(x,y)|}, where comb is a combination function, F1(x,y),,Fm(x,y) are probability formulas, y is a tuple of variables and φ(x,y) is an S-formula.

Relational Bayesian networks associate a probability formula Fr(x) to each probabilistic relation rR, where x is a tuple of k variables, with k the arity of r:

Definition 2.

Given disjoint sets of relation symbols S and R, the predefined and probabilistic relations, a relational Bayesian network is a set Φ={Fr(x)rR}, where each Fr(x) is a (S,R)-probability formula.

To have an idea of how probability formulas work, consider a fixed S-structure DS over a domain D. Then, an R-structure DR over D entails a numeric value for each ground probability formula Fr(t), denoted by Fr(t)[DR], where t is tuple of elements in D. This is done inductively, by initially defining r(t)[DR]=1 if DRr(t), and r(t)[DR]=0 otherwise, for each r(t), for all rR. If Fr(x)=c, then Fr(t)[DR]=c, for any tuple t. The numeric value of Fr(t)[DR] for probability formulas that are convex combinations or combination function will require the evaluation of its subformulas Fi, which recursively end at the evaluation of ground atoms r(t) or constants c. As the set of ground atoms whose evaluation is needed to compute Fr(t)[DR] depends only on the S-structure DS, and not on DR, it is denoted by α(Fr(x),t,DS) and can be defined recursively:

  • α(c,t,DS)=;

  • α(r(x),t,DS)={r(t)};

  • α(F1(x)F2(x)+(1F1(x))F3(x),t,DS)=i=13α(Fi(x),t,DS);

  • α(comb{|F1(x,y),,Fm(x,y)|y;φ(x,y)|},t,DS) is given by:
    t s.t. DSφ(t,t)i=1mα(Fi(x,y),(t,t),DS).

    Here, (t,t) denotes the concatenation of the tuples t and t.

For a given S-structure DS, we can define a dependency relation between the nodes r(t) and r(t) in the Bayesian network via the probability formulas Fr and Fr by employing the corresponding α(·,·,·). Intuitively, α(Fr(x),t,DS) contains the ground atoms r(t) whose truth value in a structure DR determines the value of Fr(t), which is meant to be the probability of r(t). That is, α(Fr(x),t,DS) contains the parents of r(t).

Definition 3.

Relation ⪯, over ground R-atoms, is defined as follows:

r(t)r(t)iffr(t)α(Fr(x),t,DS).

When this relation is acyclic, a relational Bayesian network Φ={FrrR} defines, for a given S-structure DS over a finite domain D, a probability distribution over the R-structures DR over D via:

PDSΦ(DR)=rRt,DRr(t)Fr(t)[DR]t,DRr(t)(1Fr(t)[DR])

Example 4.

Scenario 4 of Example 2: We can define a relational Bayesian network that returns the corresponding Bayesian network for each number and configuration of neighbors. Let S={neighbor(·,·)} and R={burglary(·),alarm(·),calls(·)}. We assume that the relation neighbor is reflexive and symmetrical. For each relation in R, we associate a probability formula, forming the relational Bayesian network Φ:

  • Fburglary(x)=0.001; a constant;

  • Falarm(x)=0.9burglary(x)+0.01(1burglary(x)); a convex combination;

  • Fcall(x)= NoisyOR{|alarm(y)y;neighbor(x,y)|}; a combination function.

Note that, if F1(x) and F2(x) are probability formulas, then 1F1(x) and F1(x)F2(x) are convex combinations and, thus, probability formulas. As the inputs of the NoisyOR above are in {0,1}, the combination function actually works like a disjunction.

Given an S-structure DS over a domain D, Φ determines a joint probability distribution over the ground R-atoms, via a Bayesian network. If we take an S-structure DS over a domain D={d1,d2,d3} such that DSneighbor(d1,d2)neighbor(d2,d3), but DSneighbor(d1,d3), the resulting PDSΦ is the model for Scenario 3 in Example 2, whose Bayesian network is given in Figure 2.

3. The Coherence Problem for RBNs

It may happen for a relational Bayesian network Φ that some S-structures yield a cyclic dependency relation ⪯. When the relation ⪯ is cyclic for an S-structure, no probability distribution is defined over the R-structures. In such a case, we say Φ is incoherent for that S-structure. This notion can be generalized to a class of S-structures S, so that we say that Φ is coherent for S iff the resulting relation ⪯ is acyclic for each S-structure in S. To know whether a relational Bayesian network is coherent for a given class of S-structures is precisely one of the problems we are interested in this work.

In order to reason about the relation between a class of S-structures and the coherence of a relational Bayesian network Φ for it, we need to formally represent these concepts. To define a class of S-structures, note that they can be seen as first-order structures over which S-formulas are interpreted. That is, an S-formula defines the set of S-structures satisfying it. If φ is a closed S-formula (without free variables), we say that [[φ]] is the set of S-structures DS such that DSφ. We denote by θS an S-formula satisfied exactly by the S-structures in a given class S; that is, [[θS]]=S.

To encode the coherence of Φ, we need to encode the acyclicity of the dependency relation ⪯ resulting from an S-structure. Ideally, we would like to have a (first-order) S-formula, say ψΦ, that would be true only for S-structures yielding acyclic dependency relations ⪯. If that formula were available, a decision about the coherence of Φ for the class S would be reduced to a decision about the validity of the first-order formula θSψΦ: When the formula is valid, then every S-structure in the class S guarantees that the resulting dependency relation ⪯ for Φ is acyclic; hence, Φ is coherent for S; otherwise, there is an S-structure in S yielding a cyclic dependency relation ⪯ for Φ. Note that for S-formulas, only S-structures matter, and we could ignore any relation not in S. To be precise, if a first-order structure D falsifies θSψΦ, then there is an S-structure DS (formed by ignoring non-S relations) falsifying it.

Alas, to encode cycles in a graph, one needs to encode the notion of path, which is the transitive closure of a relation encoding arcs. It is a well-known fact that first-order logic cannot express transitivity. To circumvent that, we can add a (strict) transitive closure operator to the logic, arriving at the so-called transitive closure logics, as described for instance in [25].

This approach was first proposed by Jaeger [23], who assumed one could write down the S-formula ψΦ by employing a transitive closure operator. He conjectured that with some restrictions on the arity of the relations in S and R, one could hope to obtain a formula θSψΦ that is decidable. Nevertheless, no hint was provided as to how to construct such a formula, or as to its general shape. A major difficulty is that, if an S-structure D satisfying θS has domain D={d1,,dn}, the size of the resulting Bayesian network is typically greater than n, with one node per ground atom, so a cycle can also contain more nodes than n. There seems to be no direct way of employing the transitive closure operator to devise a formula ¬ψΦ that encodes cycles with more than n nodes and that is to be satisfied by some structures D over a domain with only n elements. In the next sections, we review a technique (introduced by the authors in [24]) to encode ψΦ for an augmented domain, through an auxiliary formula whose satisfying structures will represent both the S-structure and the resulting ground Bayesian network. Afterwards, we adapt the formula θS accordingly.

3.1. Encoding the Structure of the Ground Bayesian Network

Our idea to construct a formula ψΦ, for a given relational Bayesian network Φ, is first to find a first-order V-formula BΦ, for some vocabulary V containing S, that is satisfiable only by V-structures that encode both an S-structure DS and the structure of the ground Bayesian network resulting from it. These V-structures should contain, besides an S-structure DS, an element for each node in the ground Bayesian network and a relation capturing its arcs. Then, we can use a transitive closure operator to define the existence of paths (and cycles) via arcs, for enforcing acyclicity by negating the existence of a cycle.

Suppose we have two disjoint vocabularies S and R={r1,,rm} of predefined and probabilistic relations, respectively. We use a(v) to denote the arity of a relation v. Consider a relational Bayesian network Φ={Fr(x)rR}, where each Fr(x) is a (S,R)-probability formula. Let D be a V-structure satisfying BΦ. We want D to be defined over a bipartite domain D=DSDB, where DS is used to represent an S-structure DS and DB=DDS is the domain where the structure of the resulting ground Bayesian network is encoded. We overload names by including in V a unary predicate DS(·) that shall be true for all and only the elements in DS. The structure D shall represent the structure of the ground Bayesian network BΦ(DS), over the elements of DB, that is induced by the S-structure DS codified in DS. In order to accomplish that, D must have an element in DB for each ground atom over the domain DS. Furthermore, the V-structure D must interpret a relation, say Parent(·,·), over DB according to the arcs of the Bayesian network BΦ(DS).

Firstly, we need to define a vocabulary V that includes the predefined relations in S and contains the unary predicate DS (recall that the equality symbol (=) is included in S). Furthermore, V must contain a binary relation Parent to represent the arcs of the ground Bayesian network. As auxiliary relations for defining Parent, we will need a relation Depij, for each pair ri,rjR, whose arity is a(ri)+a(rj). For elements in DB to represent ground atoms r(t1,,tn), we use relations to associate elements in DB to relations r and to tuples t1,,tn. For each relation riR, we have a unary relation r¯iV, where r¯i(x) is intended to mean that the element xDB represents a ground atom of the form ri(·). As for the tuples, recall that each ti represents an element in the set DS over which the S-structure DS is codified. Hence, we insert in V binary relations ti for every 1imaxia(ri), such that ti(x,y) should be true iff the element xDB corresponds to a ground atom r(t1,,tk) where ti=y, for a yDS and some rR.

To save notation, we use Ri(x,y1,,yk) to denote r¯i(x)t1(x,y1)tk(x,yk) henceforth, meaning the element x in the domain represents the ground atom ri(y1,,yk), where a(ri)=k.

Now, we proceed to list, step-by-step, the set of conjuncts required in ψΦ, together with their meaning, for the V-structure D in [[ψΦ]] to hold the desired properties. To illustrate the construction, each set of conjuncts is followed by an example based on the RBN in Example 4, possibly given in an equivalent form for clarity.

We have to ensure that the elements in DB correspond exactly to the ground atoms in the ground Bayesian network BΦ(DS).

  • Each element in DB=DDS should correspond to a ground atom for some riR. Hence, we have the formula:
    x¬DS(x)i=1mr¯i(x). (2)
    x¬DS(x)burglary¯(x)alarm¯(x)calls¯(x).
  • No element may correspond to ground atoms for two different riR. Therefore, the formula below is introduced:
    x1i,jmij(¬r¯i(x)¬r¯j(x)). (3)
    x(¬burglary¯(x)¬alarm¯(x))(¬burglary¯(x)¬calls¯(x))(¬alarm¯(x)¬calls¯(x)).
  • Each element corresponding to a ground atom should correspond to exactly one tuple. To achieve that, let k=maxja(rj), and introduce the formula below:
    xyzj=1k(tj(x,y)tj(x,z)y=z). (4)
    xyz(t1(x,y)t1(x,z)y=z.
  • Each element corresponding to a ground atom for a riR should be linked a to tuple with arity a(ri). Thus, let k=maxja(rj), and introduce the formula below for each riR:
    xr¯i(x)(y1ya(ri)Ri(x,y1,,ya(ri))z¬ta(ri)+1(x,z)¬tk(x,z)). (5)
    xburglary¯(x)(yt1(x,y));xalarm¯(x)(yt1(x,y));xcalls¯(x)(yt1(x,y)).
  • Only elements in DB=DDS should correspond to ground atoms. This is enforced by the following formula, where k=maxia(ri):
    yDS(y)(i=1m¬r¯i(y)xj=1k¬tj(y,x)). (6)
    yDS(y)(¬burglary¯(y)¬alarm¯(y)¬calls¯(y)x¬t1(y,x)).
  • Each ground atom must be represented by at least one element (in DB=DDS). Therefore, for each riR, with a(ri)=k, we need a formula:
    y1ykDS(y1)DS(yk)xRi(x,y1,,yk). (7)
    yDS(y)(x1burglary¯(x1)t1(y,x1));sameforalarm¯andcalls¯.

    These formulas enforce that each ground atom r(t) is represented by an element x that is in DB, due to the formula (6).

  • No ground atom can be represented by two different elements. Hence, for each riR, with a(ri)=k, we introduce a formula:
    y1,ykxzRi(x,y1,,yk)Ri(z,y1,,yk)x=z. (8)
    yxzburglary¯(x)t1(y,x)burglary¯(z)t1(z,y)x=z;sameforalarm¯andcalls¯.

The conjunction of all formulas in (2)–(8) is satisfied only by structures D over the domain D=DSDB such that there is a bijection between DB and the set of all possible ground atoms {r(t) for some rR and tDSa(r)}. Now, we can put the arcs over these nodes to complete the structure of the ground Bayesian network BΦ(DS).

The binary relation Parent must hold only between elements in the domain D representing ground atoms r(t) and r(t) such that r(t)r(t). Recall that the dependency relation ⪯ is determined by the S-structure DS. While the ground atoms represented in DB, for a fixed R, are determined by the size of DS by itself, the relation Parent between them depends also on the S-formulas that hold for the S-structure DS. We want these S-structures to be specified by D over DS only, not over DB. To ensure this, we use the following group of formulas:

  • For all sS, consider the formula below, where a(s)=k:
    y1yks(y1,,yk)DS(y1)DS(yk). (9)
    y1y2neighbor(y1,y2)DS(y1)DS(y2).

The formula above forces that s(t), for any sS, can be true only for tuples tDSa(s).

For a known S-structure DS, it is straightforward to determine which ground atoms r(t) are the parents of r(t) in the ground Bayesian network BΦ(DS). One can use recursively the definition of the set of parents α(Fr(x),t,DS) given in Section 2. Nonetheless, with an unknown S-structure DS specified in D over DS, the situation is a bit trickier. The idea is to construct, for each pair ri(t) and rj(t), an S-formula Depij(t,t) that is true iff ri(t)rj(t) for the DS encoded in D. To define Depij(t,t), we employ auxiliary formulas CF(t)r(t), for a ground probability formula F(t) and a ground atom r(t), that will be an S-formula that is satisfied by D iff r(t)α(F(x),t,S). We define CF(t)r(t) recursively, starting from the base cases.

  • If F(t)=c, for a c[0,1], then CF(t)r(t)=;e.g.,CFburglary(t)alarm(t)=.

  • If F(t)=r(t), then CF(t)r(t)=(t=t) if r=r; and CF(t)r(t)= otherwise;

    e.g.,CFburglary(t)burglary(t)=(t=t) and CFburglary(t)calls(t)=.

Above, (t=t) is a short form for (t1=t1)(tk=tk), where k is the arity of t. These base cases are in line with the recursive definition of α(F(x),t,S) presented in Section 2. The third case is also straightforward:

  • If F(t)=F1(t)F2(t)+(1F1(t))F3(t), then CF(t)r(t)=i=13CFi(t)r(t).

    CFalarm(t)burglary(t)=CFburglary(t)burglary(t)C0.9burglary(t)C0.01burglary(t)=(t=t)

In other words, the computation of F(t)[DR] depends on r(t)[DR], for some DR, if the computation of some Fi(t)[DR], for 1i3, depends on r(t)[DR].

The more elaborated case happens when F(x) is a combination function, for which there is an S-formula involved. Recall that if F(x)=comb{|F1(x,y),,Fm(x,y)|y;φ(x,y)|}, then the parents of F(t) are given by t,DSφ(t,t)i=1mα(Fi(x,y),(t,t),DS). Thus, to recursively define CF(t)r(t), we need an S-formula that is satisfied by an S-structure DS iff:

r(t)t,DSφ(t,t)i=1mα(Fi(x,y),(t,t),DS).

The inner union is analogous to the definition of CF(t)r(t) for convex combinations. However, to cope with any t such that DSφ(t,t), we need an existential quantification:

  • If F(x)=comb{|F1(x,y),,Fm(x,y)|y;φ(x,y)|}, then we have that:
    CF(t)r(t)=tφ(t,t)i=1mCFi(t,t)r(t).
    CFcalls(t)alarm(t)=tneighbor(t,t)CFalarm(t)alarm(t)=tneighbor(t,t)(t=t)

Now, we can employ the formulas CF(t)r(t) to define the truth value of the ground relation Depij(t,t), that codifies when ri(t)rj(t).

  • For each pair ri,rjR, with a(ri)=k and a(rj)=k, we have the formula:
    x1xky1ykDepij(x1,,xk,y1,,yk)CFri(x1,,xk)rj(y1,,yk). (10)
    xyDepcallsalarm(x,y)zneighbor(x,z)(z=y);xyDepalarmburglary(x,y)(x=y).

In the formula above, CFri(x1,,xk)rj(y1,,yk) has free variables x1,,xk,y1,,yk and is built according to the four recursive rules that define CF(t)r(t), replacing the tuples t and t by x and y. We point out that such construction depends only on probability formulas in the relational Bayesian network Φ, and not on any S-structure. To build each CFri(x)rj(y), one just starts from the probability formula Fri(x) and follows the recursion rules until reaching the base cases, when CFri(x)rj(y) will be formed by subformulas like ,, S-formulas φ(·) and equalities (·=·), possibly quantified on variables appearing in φ.

The relation Parent(·,·) is defined now over elements that represent ground atoms ri(t) and rj(t) such that Depij(t,t), meaning that ri(t)rj(t). This can be achieved in two parts: ensuring that each ri(t)rj(t) implies Parent(t,t); and guaranteeing that Parent(t,t) is true only if ri(t)rj(t) for a pair of relations ri,rj.

  • For each pair ri,rjR, with a(ri)=k and a(rj)=k, let y and y denote y1,,yk and y1,,yk, respectively:
    xxy1yky1ykRi(x,y)Rj(x,y)Depij(y,y)Parent(x,x). (11)
    xxyycalls(x)t1(y,x)alarm(x)t1(y,x)Depcallsalarm(y,y)Parent(x,x).
  • Let k=maxja(rj) be the maximum arity in R, and let y and y denote the tuples y1,,ya(ri) and y1,,ya(rj), respectively:
    xxParent(x,x)y1yky1yk1i,jmRi(x,yri)Rj(x,yrj)Depij(yri,yrj). (12)

Definition 4.

Given disjoint sets of relations S and R and a relational Bayesian network Φ={FririR}, the formula BΦ is the conjunction of all formulas in (2)–(12).

For some fixed relational Bayesian networks Φ, the formula BΦ is satisfied only by V-structures D over a bipartite domain DSDB such that:

  • the relations in S are interpreted in DS, forming an S-structure DS;

  • there is a bijection b between the domain DB=DDS and set of all ground R-atoms formed by the tuples in DS;

  • each xDB is linked exactly to one riR, via the predicate r¯i(x), and exactly k=a(ri) elements in DS, via the relations t1(x,.),tk(x,.), and no ground atom is represented through these links twice;

  • the relation Parent(·,·) is interpreted as arcs in DB in such a way that DB,Parent form a directed graph that is the structure of the ground Bayesian network BΦ(DS).

3.2. Encoding Coherence via Acyclicity

The original formula ψΦ was intended to capture the coherence of the relational Bayesian network Φ. Our idea is to check the coherence by looking for cycles in the ground Bayesian network BΦ(DS) encoded in any V-structure satisfying BΦ. Hence, we replace ψΦ by an implication BΦψ, which is to be satisfied only by V-structures D such that, if D represents an S-structure DS and the resulting ground Bayesian network BΦ(DS), then BΦ(DS) is acyclic. Thus, ψ should avoid cycles of the relation Parent in the V-structures satisfying it.

There is a cycle with Parent-arcs in a V-structure D over a domain D iff there exists a xD such that there is a path of Parent-arcs from x to itself. Consequently, detecting Parent-cycles reduces to computing Parent-paths or Parent-reachability. We say y is Parent-reachable from x, in a V-structure D, if there are z0,,zkD such that x=z0, y=zk, and D1ikParent(zi1,zi). Thus, for each k, we can define reachability through k Parent-arcs: ParentPathk(x,y)=z0zk(z0=x)(zk=y)1ikParent(zi1,zi). Unfortunately, the size of the path (k) is unbounded a priori, as the domain D can be arbitrarily large. Therefore, there is no means in the first-order logic language to encode reachability, via arbitrarily large paths, with a finite number of formulas. In order to circumvent this situation, we can resort to a transitive closure logic.

Transitive closure logics enhance first-order logics with a transitive closure operator TC that we assume to be strict [25]. If φ(x,y) is a first-order formula, TC(φ)(x,y) means that y is φ-reachable from x, with a non-empty path. Accordingly, a V-structure D, over a domain D, satisfies TC(φ)(x,y) iff there is a kN and there are z0,,zkD such that x=z0, y=zk and D1ikφ(zi1,zi).

Employing the transitive closure operator, the existence of a Parent-path from a node x to itself (a cycle) is encoded directly by TC(Parent)(x,x); similarly, the absence of a Parent-cycle is enforced by ψ=x¬TC(Parent)(x,x).

At this point, the V-structures D over a domain D satisfying BΦψ have the following format:

  • either it encodes an S-structure in DSD (the part of the domain satisfying DS(·)) and the corresponding acyclic ground Bayesian network BΦ(DS) in DB=DDS.

  • or it is not the case that D encodes both an S-structure in DSD and the corresponding ground Bayesian network BΦ(DS) in DB=DDS;

Back to the coherence-checking problem, we need to decide, for a fixed relational Bayesian network Φ, whether or not a given class S of S-structures ensures the acyclicity of the resulting ground Bayesian network BΦ(DS). Recall that the class S must be defined via a (first-order) S-formula θS. As we are already employing the transitive closure operator in ψ, we can also allow its use in θS, which is useful to express S-structures without cycles, for instance.

To check the coherence of Φ for a class S, we cannot just check the validity of:

θS(BΦψ), (13)

because θS specifies S-structures over D, while BΦψ presupposes that the S-structure is given only over DS={dDDDS(d)}D. To see the kind of problem that might occur, think of the class S of all S-structures D where each dD is such that si(d) holds, for some unary predefined relation siS. Consider an S-structure DS (DθS), over a domain D. The formula BΦ cannot be satisfied by D, for DS(x) must hold for all xD, because of the formulas in (9), so no xD can represent ground formulas, due to the formulas in (6), contradicting the restrictions in (7) that require all ground atoms to be represented. Hence, this D satisfies θS without encoding the ground Bayesian network, thus falsifying BΦ and satisfying BΦψ, yielding the satisfaction of Formula (13). Consequently, Formula (13) is valid for this specific class S, no matter what the relational Bayesian network Φ looks like. Nonetheless, it is not hard to think of a Φ that is trivially incoherent for any class of S-structures, like Φ={Fr(x)=r(x)}, with S= and R={r}, where the probability formula associated with the relation rR is the indicator function r(x), yielding a cyclic dependency relation ⪯.

In order to address the aforementioned issue, we need to adapt θS, constructing θS to represent the class S in the extended, bipartite domain D=DSDB. The unary predicate DS(·) is what delimits the portion of D that is dedicated to define the S-structure. Actually, we can define DS as the set {xDDDS(x)}D. Therefore, we must construct a V-formula θS such that the V-structure D satisfies θS iff the S-structure DS, formed by DSD and the interpretation of the S relations, satisfies θS. That is, the S-formulas that hold in an S-structure DS must hold for the subset of a V-structure D defined over the part of its domain that satisfies DS(·). This can be performed by inserting guards in the quantifiers inside θS.

Definition 5.

Given a (closed) S-formula θS, θS is the formula resulting from applying the following substitutions to θS:

  • Replace each xφ(x) in θS by xDS(x)φ(x);

  • Replace each xφ(x) in θS by xDS(x)φ(x).

Finally, we can define the formula that encodes the coherence of a relational Bayesian network Φ for a class of S-structures S:

Definition 6.

For disjoint sets of relations S and R, a given relational Bayesian network Φ and a class of S-structures defined by θS, CΦ,S=θS(BΦψ).

Putting all those arguments together, we obtain the translation of the coherence-checking problem to the validity of a formula from the transitive closure logic:

Theorem 1

(De Bona and Cozman [24]). For disjoint sets of relations S and R, a given relational Bayesian network Φ and a class of S-structures S defined by θS, Φ is coherent for S iff CΦ,S is valid.

As first-order logic in general is already well-known to be undecidable, adding a transitive closure operator clearly does not make things easier. Nevertheless, decidability remains an open problem, even restricting the relations in R to be unary and assuming a decidable θS (even though there are some decidable fragments of first-order logic with transitive closure operators [25,26]). Similarly, a proof of general undecidability remains elusive.

3.3. A Weaker Form of Coherence

Jaeger introduced the coherence problem for RBNs as checking whether every input structure in a given class yields a probability distribution via an acyclic ground Bayesian network. Alternatively, we might define the coherence of an RBN as the existence of at least one input structure, out of a given class, resulting in an acyclic ground Bayesian network. This is closer to the satisfiability-like notion of coherence discussed by de Finetti and closer to work on probabilistic logic [27,28].

In this section, we show that, if one is interested in a logical encoding for this type of coherence for RBNs, the transitive closure operator can be dispensed with.

Suppose we have an RBN Φ and class S of input structures codified via a first-order formula θS and we want to decide whether Φ is coherent for some structure in S. This problem can be reduced to checking the satisfiability of a first-order formula, using the machinery introduced above, with the bipartite domain. This formula can be easily built as θSBΦψ. By construction, this formula is satisfiable iff there is a structure D over a bipartite domain D=DSDB where DS encodes an S-structure in S (DθS), DB encodes the corresponding ground Bayesian network (DBΦ) and the latter is acyclic (Dψ). Nonetheless, since now we are interested in satisfiability instead of validity, we can replace ψ by a formula ψ that does not employ the transitive closure operator.

The idea to construct ψ is to use a binary relation Parent(·,·) and to force it to extend, or to contain, the transitive closure of Parent(·,·). The formula ψ then also requires Parent(·,·) to be irreflexive. If there is such Parent(·,·), then Parent(·,·) must be acyclic. Conversely, if Parent(·,·) is acyclic, then Parent(·,·) can be interpreted as its transitive closure, being irreflexive. In other words, we want a structure to satisfy ψ iff it interprets a relation Parent(·,·) that both is irreflexive and extends the transitive closure of Parent(·,·).

In order to build ψ, the vocabulary V is augmented with the binary relation Parent. Now, we can define ψ as the conjunction of two parts:

  • xyzParent(x,y)Parent(x,y)Parent(x,y)Parent(y,z)Parent(x,z), forcing Parent to extend the transitive closure of Parent;

  • x¬Parent(x,x), requiring Parent to be irreflexive.

By construction, one can verify the following result:

Theorem 2.

For disjoint sets of relations S and R, a given relational Bayesian network Φ and a class of S-structures S defined by θS, Φ is coherent for some structure in S iff θSBΦψ is satisfiable.

The fact that θSBΦψ does not use the transitive closure operator makes its satisfiability decidable for any decidable fragment of first-order logic.

4. Probabilistic Relational Models

In this section, we introduce the machinery of PRMs by following the terminology by Getoor et al. [19], focusing on the simple case where uncertainty is restricted to descriptive attributes, which are assumed to be binary. We also review the coherence problem for PRMs and the proposed solutions in the literature. In the next section, we show how this coherence problem can also be tackled via logic, as the coherence of RBNs.

4.1. Syntax and Semantics of PRMs

To define a PRM, illustrated in Example 5, we need a relational model, with classes associated with descriptive attributes and reference slots that behave like foreign keys. Intuitively, each object in a class is described by the values of its descriptive attributes, and reference slots link different objects. Formally, a relational schema is described by a set of classes X={X1,,Xn}, each of which associated with a set of descriptive attributes A(Xi) and a set of reference slots R(Xi). We assume descriptive attributes take values in {0,1}. A reference slot ρ in a class X (denoted X.ρ) is a reference to an object of the class Range[ρ] (its range type) specified in the schema. The domain type of ρ, Dom[ρ], is X. We can view this reference slot ρ as a function fρ taking objects in Dom[ρ] and returning singletons of objects in Range[ρ]. That is, fρ(x)={y} is equivalent to x.ρ=y.

For any reference slot ρ, there is an inverse slot ρ1 such that Range[ρ1]=Dom[ρ] and Dom[ρ1]=Range[ρ]. The corresponding function, fρ1 takes an object x from the class Range[ρ] and returns the set of objects {yfρ(y)={x}} from the class Dom[ρ]. A sequence of slots (inverted or not) K=ρ1,,ρk is called a slot chain if Range[ρi]=Dom[ρi+1] for all i. The function corresponding to a slot chain K=ρ1,ρ2, fK, is a type of composition of the functions fρ1,fρ2, taking an object x from Range[ρ1] and returning a set objects {zy:yfρ1(x)zfρ2(y)} from Range[ρ2]. The corresponding function can be obtained by applying this type of composition two-by-two. We write yx.K when yfK(x).

An instance I of a relational schema populates the classes with objects, associating values with the descriptive attributes and reference slots. Formally, I is an interpretation specifying for each class XX: a set of objects I(X); a value A.x{0,1} for each descriptive attribute in AA(X) and each object xI(X); and an object x.ρI(Range[ρ]) for each reference slot ρR(X) and object xI(X). Note that, if x.ρ=y, fρ(x)={y}. We use Ix.A and Ix.ρ to denote the value of x.A and x.ρ in I.

Given a relational schema, a PRM defines a probability distribution over its instances. In the simplest form, on which we focus, objects and the relations between them are given as input, and there is uncertainty only over the descriptive attributes values. A relational skeleton σr is a partial specification of an instance that specifies a set of objects σr(Xi) for each class Xi in the schema besides the relation holding between these objects: σrx.ρ for each xσr(Xi) and ρR(Xi). A completion of a relational skeleton σr is an instance I such that, for each class XiX: I(Xi)=σr(Xi) and, for each xI(Xi) and ρR(Xi), Ix.ρ=σrx.ρ. We can see a PRM as a function taking relational skeletons and returning probability distributions over the completions of these partial instances, which can be seen as joint probability distributions for the random variables formed by the descriptive attributes of each object.

The format of a PRM resembles that of a Bayesian network: for each attribute X.A, we have a set of parents Pa(X.A) and the corresponding parameters P(X.APa(X.A)). The parent relation forms a direct graph, as usual, called the dependency graph; and the set of parameters define the conditional probability tables. The attributes in Pa(X.A) are called formal parents, as they will be instantiated for each object x in X according to the relational skeleton. There two types of formal parents: X.A can depend either on another attribute X.B of the same object or on an attribute X.K.B of other objects, where K is a slot chain.

In general, for an object x, x.K.B is a multiset {y.Byx.K}, whose size is defined by the relational skeleton. To compactly represent the conditional probability distribution when X.K.BPa(X.A), the notion of aggregation is used. The attribute x.A will depend on some aggregate function γ of this multiset, like its mean value, mode, maximum or minimum, and so on; that is, γ(X.K.B) will be a formal parent of X.A.

Definition 7.

A probabilistic Relational model Π for a relational schema R is defined as a pair ΠS,Πθ where:

  • ΠS defines, for each class XX and each descriptive attribute AA(X), a set of formal parents Pa(X.A)={U1,,Ul}, where each Ui has the form X.B or γ(X.K.B);

  • Πθ is the set of parameters defining legal Conditional Probability Distributions (CPDs) P(X.APa(X.A)) for each descriptive attribute AA(X) of each class XX.

The semantics of a PRM is given by the ground Bayesian network induced by a relational skeleton, where the descriptive attributes of each object are the random variables.

Definition 8.

A PRM Π=ΠS,Πθ and a relational skeleton σr define a ground Bayesian network where:

  • There is a node representing each attribute x.A, for all xσr(Xi), AA(Xi) and XiX;

  • For each XiX, each xσr(Xi) and each AA(Xi), there is a node representing γ(x.K.B) for each γ(Xi.K.B)Pa(Xi.A);

  • Each x.A depends on parents x.B, for formal parents X.BPa(X.A), and on parents γ(x.K.B), for formal parents γ(X.K.B)Pa(X.A), according to ΠS;

  • Each γ(x.K.B) depends on parents y.B with yx.K;

  • The CPD for P(x.APa(x.A)) is P(X.APa(X.A)), according to Πθ.

  • The CPD for P(γ(x.K.B)Pa(γ(x.K.B))) is computed through the aggregation function γ.

The joint probability distribution over the descriptive attributes can be factored as usual to compute the probability of a specific instance I that is a completion of the skeleton σr. If we delete each γ(x.K.B) from the ground Bayesian network, making its children depend directly on the nodes y.B with yx.K (defining a new parent relation Pa) and updating the CPDs accordingly, we can construct a simplified ground Bayesian network. The latter can be employed to factor the joint probability distribution over the descriptive attributes:

P(Iσr,Π)=xσrAA(x)P(Ix.AIPa(x.A))=Xixσr(Xi)AA(x)P(Ix.AIPa(x.A)).

Viewing Π as a function from skeletons to probability distributions over instances, we use Π(σr) to denote the probability distribution P(Iσr,Π) over the completions I of σr.

Example 5.

Recall again Scenario 4 in Example 2. We can define a PRM that returns the corresponding Bayesian network for each number and configuration of neighbors. In our relational schema, we have a class Person, whose set of descriptive attributes is A(Person)={burglary,alarm,calls}. Furthermore, to capture multiple neighbors, we also need a class Neighborhood, with two reference slots, R(Neighborhood)={neighbor1,neighbor2}, whose domain is Person. For instance, to denote that Alice and Bob are neighbors, we would have an object, say nAB, in the class Neighborhood, whose reference slots would be nAB.neighbor1=Alice and nAB.neighbor2=Bob.

We assume that the relation neighbor is reflexive (that is, for each Person x, there is always a Neighborhoodnx with nx.neighbor1=nx.neighbor2=x) and symmetrical (if xy.neighbor11.neighbor2, we also have yx.neighbor11.neighbor2).

For each descriptive attribute in our relational schema, we associate a set of formal parents and a conditional probability table, forming the following PRM Π to encode Scenario 4:

  • Pa(Person.burglary)=; P(Person.burglary)=0.001;

  • Pa(Person.alarm)={burglary}; P(Person.alarmburglary)=0.9 and P(Person.alarm¬burglary)=0.1;

  • Pa(Person.calls)={or(Person.neighbor11.neighbor2)};

    P(Person.callsor(Person.neighbor11.neighbor2)=c)=c, for c{0,1}.

Given a relational skeleton σr with persons and neighbors, Π determines a joint probability distribution over the the descriptive attributes, via a Bayesian network. Consider a skeleton σr with σr(Person)={x1,x2,x3} and n12,n23σr(Neighborhood), with σrnij.neighbor1=xi and σrnij.neighbor2=xj, for each nijσr(Neighborhood), but such that no nσr(Neighborhood) has n.neighbor1=x1 and n.neighbor2=x3. Then, the resulting probability distribution is the model of Scenario 3 in Example 2, whose Bayesian network is given in Figure 2.

4.2. Coherence via Colored Dependency Graphs

As with RBNs, for the model to be coherent, one needs to guarantee that the ground Bayesian network is acyclic. Getoor et al. [19] focused on guaranteeing that a PRM yields acyclic ground Bayesian networks for all possible relational skeletons. To achieve that, possible cycles are detected in the class dependency graph.

Definition 9.

Given a PRM Π, the class dependency graph GΠ is a directed graph with a node for each descriptive attribute X.A and the following arcs:

  • Type I arcs: X.B,X.A, where X.B is a formal parent of X.A;

  • Type II arcs: Y.B,X.A, where γ(X.K.B) is a formal parent of X.A and Y=Range[X.K].

When the class dependency graph is acyclic, so is the ground Bayesian network for any relational skeleton. Nevertheless, it may be the case that, even for cyclic class dependency graphs, any relational skeleton occurring in practice leads to a coherent model. In other words, there might be classes of skeletons for which the PRM is coherent. To easily recognize some of these classes, Getoor et al. [19] put forward an approach based on identifying slot chains that are acyclic in practice. A set of slot chains Kga={K1,,Km} is guaranteed acyclic if we are guaranteed that, for any possible relational skeleton σr, there is a partial ordering ⪯ over its objects such that, for each KiKga, xy for any pair x and yx.Ki (we use xy to denote xy and xy).

Definition 10.

Given a PRM Π and a set of guaranteed acyclic slot chains Kga, the colored class dependency graph GΠ is a directed graph with a node for each descriptive attribute X.A and the following arcs:

  • Yellow arcs: X.B,X.A, where X.B is a formal parent of X.A;

  • Green arcs: Y.B,X.A, where γ(X.K.B) is a formal parent of X.A, Y=Range[X.K] and KKga;

  • Red arcs: Y.B,X.A, where γ(X.K.B) is a formal parent of X.A, Y=Range[X.K] and KKga.

Intuitively, yellow cycles in the colored class dependency graph correspond to attributes of the same object, yielding a cycle in the ground Bayesian network. If we add some green arcs to such a cycle, then it is guaranteed that, departing from a node x.A in the ground Bayesian network, these arcs form a path to y.A, where xy, since ⪯ is transitive. Hence, x is different from y, and there is no cycle. If there is a red arc in a cycle, however, one may have a skeleton that produces a cycle.

A colored class dependency graph is stratified if every cycle contains at least one green arc and no red arc. Then:

Theorem 3

(Getoor et al. [19]). Given a PRM Π and a set of guaranteed acyclic slot chains Kga, if the colored class dependency graph GΠ is stratified, then the ground Bayesian network is acyclic for any possible relational skeleton.

In the result above and in the definition of guaranteed acyclic slot chains, “possible relational skeleton” refers to the class of skeletons that can occur in practice. The user must detect the guaranteed acyclic slot chains, taking advantage of his a priori knowledge on the possible skeletons in practice. For instance, consider a slot chain motherOf linking objects of the same class Person (Example 3). A genetic attribute, like Person.blueEyes, might depend on Person.motherOf.blueEyes. Mathematically, we can conceive of a skeleton with a cyclic relation motherOf, resulting in a red cycle in the colored class dependency graph. Nonetheless, being aware of the intended meaning of motherOf, we know that such skeletons are not possible in practice, so the cycle is green, and coherence is guaranteed.

Identifying guaranteed acyclic slot chains is by no means trivial. In fact, Getoor et al. [19] also define guaranteed acyclic (g.a.) reference slots and g.a. slot chains are defined as those formed only by g.a. reference slots. Still, these maneuvers miss the cases where two possible reference slots cannot be g.a. according to the same ⪯, but combine to form a g.a. slot chain. Getoor et al. [19] mention the possibility of assuming different partial orders to define different sets of g.a. slot chains: in that case, each ordering would correspond to a shade of green in the colored class dependency graph, and coherence would not be ensured if there were two shades of green in a cycle.

5. Logic-Based Approach to the Coherence of PRMs

The simplest approach to the coherence of PRMs, via the non-colored class dependency graph, is intrinsically incomplete, in the sense that some skeletons might yield a coherent ground Bayesian network even for cyclic graphs. The approach via colored class dependency graph allows some cyclic graphs (the stratified ones) to guarantee consistency for the class of all possible skeletons. However, this method depends on a pre-specified set of guaranteed acyclic slot chains, and the colored class dependency graph being stratified for this set is only a sufficient, not a necessary condition for coherence. Therefore, the colored class dependency graph method is incomplete, as well. Even using different sets of g.a. slot chains (corresponding to shades of green) to eventually capture all of them, it is still possible that a cycle with red arcs cannot entail incoherence in practice. Besides being incomplete, the graph-based method is not easily applicable to an arbitrary class of skeletons. Given a class of skeletons as input, the user would have to detect somehow which slot chains are guaranteed acyclic for that specific class; this can be considerably more difficult than ensuring acyclicity in the general case.

To address these issues, thus obtaining a general, complete method for checking the coherence of PRMs for a given class of skeletons, we can resort to the logic-based approach we introduced for the RBNs in previous sections. The goal of this section is to adapt the logic-basic techniques to PRMs.

PRMs can be viewed as RBNs, as conditional probability tables of the former can be embedded into combination functions of the latter. This translation is out of our scope though, and it suffices for our purposes to represent PRMs as random relational structures, taking S-structures to probability distributions on R-structures. While the S-vocabulary is used to specify classes of objects and relations between them (that is, the relational skeleton), the R-vocabulary expresses the descriptive attributes of the objects. Employing this logical encoding of PRMs, we can apply the approach from Section 3.1 to the coherence problem for PRMs.

To follow this strategy, we first show how a PRM can be seen as a random relational structure described by a logical language.

5.1. PRMs as Random Relational Structures

Consider a PRM Π=ΠS,Πθ over a relational schema described by a set of classes X={X1,Xn}, each associated with a set of descriptive attributes A(Xi) and a set of reference slots R(Xi). Given a skeleton σr, which is formed by objects and relations holding between them, the PRM Π yields a ground Bayesian network over the descriptive attributes of these objects, defining a probability distribution Π(σr) over the completions of σr. Hence, if the relational skeleton is given as a first-order S-structure over a set of objects and a set of unary relations R denotes their attributes, the PRM becomes a random relational structure.

We need to represent a skeleton σr as a first-order S-structure Σ. Objects in σr can be seen as the elements of the domain D of Σ. Note that PRMs are typed, with each object belonging to specific class XiX. Thus, we use unary relations X1,,Xn in the vocabulary S to denote the class of each object. Accordingly, for each xD, Xi(x) holds in Σ iff xσr(Xi). As each object belongs to exactly one class in the relational skeleton, the class of possible first-order structures is restricted to those where the relations X1,,Xn form a partition of the domain.

The first-order S-structure Σ must also encode the relations holding between the objects in the skeleton that are specified via the values of the reference slots. To capture these, we assume they have unique names and consider, for each reference slot Xi.ρR(Xi) with Range[ρ]=Xj, a binary relation Sρ. In Σ, Sρ(x,y) holds iff σrx.ρ=y. Naturally, Sρ(x,y) should imply Xi(x) and Xj(y). Now, Σ encodes, through the vocabulary S, all objects of a given class, as well as the relations between them specified in the reference slots. In other words, there is a computable function bS from relational skeletons σr to S-structures Σ=bS(σr). For bS to be a bijection, we make its codomain equal to its range.

The probabilistic vocabulary of the random relational structure corresponding to a PRM is formed by the descriptive attributes of every class in the relational schema. We assume that attributes in different classes have different names, as well, in order to define the vocabulary of unary relations R={AA(Xi)XiX}. If Aj is an attribute of Xi, x.Aj=1 (resp. x.Aj=0) in the PRM is mirrored by the ground R-atom Aj(x) being true (resp. false) in the random relational structure. Thus, as a completion I corresponds to a value assignment to descriptive attributes of objects x1,,xm from a relational skeleton σr, it also corresponds to an R-structure DI over a domain D={x1,,xm} in the following way: DIAi(xj) iff xj.Ai=1. Note that we assume that for DI to correspond to a completion I of σr, DIAi(xj) whenever Ai is not an attribute of the class XX such that xjσr(X). Let bR denote the function taking instances I and returning the corresponding R-structures DI=bR(I). As we cannot recover the skeleton σr from the R-structure DI=bR(I), bR is not a bijection. Nevertheless, fixing a skeleton σr, there is a unique I such that bR(I)=DI.

Now, we can define a random relational structure PΠ that corresponds to the PRM Π. For every relational skeleton σr over a domain D, let PΠ(bS(σr)):ModR(D)[0,1] be a probability distribution over R-structures such that PΠ(bS(σr))(DR)=Π(σr)(IR), if DR=bR(IR), for a completion IR of σr, and PΠ(bS)(DR)=0 otherwise.

5.2. Encoding the Ground Bayesian Network and its Acyclicity

The probability distribution PΠ(bS(σr)) can be represented by a ground Bayesian network BPΠ(bS(σr)), where nodes represent the ground R-atoms. The structure of this network is isomorphic to the simplified ground Bayesian network yielded by Π for the skeleton σr, if we ignore the isolated nodes representing the spurious Ai(xj)=0, when Ai is not an attribute of the class to which xj belongs. The coherence of Π(σr) depends on the acyclicity of the corresponding ground Bayesian network BΠ(σr), which is acyclic iff BPΠ(σr) is so. Therefore, we can encode the coherence of a PRM Π for a skeleton σr via the acyclicity of BPΠ(bs(σr)) by applying the techniques from Section 3.

We want to construct a formula that is satisfied only by those S-structures bS(σr) such that Π(σr) is coherent. Again, we consider an extended, bipartite domain D=DSDB, with bS(σr) encoded over DS and the structure of BPΠ(σr) encoded in DB. We want to build a formula BΠ that is satisfied by structures D over D=DSDB such that, if D encodes bS(σr) over DS, then D encodes the structure of BPΠ(bS(σr)) over DB. The nodes are encoded exactly as shown in Section 3.1.

To encode the arcs, we employ once more a relation Parent(·,·). Parent(y,y) must hold only if x,yDB denote ground R-atoms Ai(x) and Aj(x) such that x.AjPa(x.Ai) in the simplified ground Bayesian network, which is captured by the formula Depij(x,x), as in Section 3.1. The only difference here is that now Depij(x,x) can be defined directly. We use Depij(x,x) here to denote a formula recursively defined, not an atom over the binary relation Depij(·,·). For each pair Ai,AjR, we can simply look at ΠS to see the conditions on which x.Aj is a parent of x.Ai in the simplified ground Bayesian network (x.AjPa(x.Ai)), in which case Aj(x) will be a parent of Ai(x) in BPπ(bS(σr)). If X.AjPa(X.Ai), then Depij(x,x) should be true whenever x=x. If γ(X.K.Aj)Pa(X.Ai), for a slot chain K, then x.K.AjPa(x.Ai) and Depij(x,x) should be true whenever x is related to x via K=ρ1,,ρk. This is the case if:

y1,y2yk1Sρ1(x,y1)Sρ2(y1,y2)Sρk(yk1,x)

is true. If K=ρ, this formula is simply Sρ(x,x).

Note that it is possible that both X.Aj and γ(X.K.Aj) are formal parents of X.Ai, and there can even be different parents γ(X.K.Aj), for different K. Thus, we define Depij(x,x) algorithmically. Initially, make Depij(x,x)=. If X.AjPa(X.Ai) for some Aj, make Depij(x,x)=Depij(x,x)(x=x). Finally, for each γ(X.K.Aj) in Pa(X.Ai), for a slot chain K=ρ1,,ρk, make:

Depij(x,x)=Depij(x,x)y1,y2yk1Sρ1(x,y1)Sρ2(y1,y2)Sρk(yk1,x),

using fresh y1,,yk1.

Analogously to Section 3.1, we have a formula BΠ, for a fixed PRM Π, that is satisfied only by structures D over a bipartite domain DSDB where the Parent(·,·) relation over DB brings the structure of the ground Bayesian network BPΠ(Σ) corresponding to the S-structure Σ encoded in DS. Again, acyclicity can be captured via a transitive closure operator: ψΠ=x¬TC(Parent)(x,x). The PRM Π is coherent for a skeleton σr if, for every structure D over a bipartite domain DSDB encoding bS(σr) in DS, we have DBΠψΦ.

Consider now a class S of skeletons σr such that {bS(σr)σrS} is the set of S-structures satisfying a first-order formula θS. To check whether the PRM Π is coherent for the class S, we construct θS by inserting guards to the quantifiers, as explained in Definition 5. Finally, the PRM Π is coherent for a class S of relational skeletons iff θS(BΠψΠ) is valid.

We have thus succeeded in turning coherence checking for PRMs into a logical inference, by adapting techniques we developed for RBNs. In the next section, we travel, in a sense, the reverse route: we show how to adapt the existing graph-based techniques for coherence checking of PRMs to coherence checking of RBNs.

6. Graph-Based Approach to the Coherence of RBNs

The logic-based approach to the coherence problem for RBNs can be applied to an arbitrary class of input structures, as long as the class can be described by a first-order formula, possibly with a transitive closure operator. Given any class of input structures S, via the formula θS, we can verify the coherence of a RBN Φ via the validity of θS(BΦψ), as explained in Section 3. Furthermore, this method is complete, as Φ is coherent for S if and only if such a formula is valid. Nonetheless, completeness and flexibility regarding the input class come at a very high price, as deciding the validity of this first-order formula involving a transitive closure operator may be computationally hard, if at all decidable. Therefore, RBNs users can benefit from the ideas introduced by Getoor et al. [19] for the coherence of PRMs, using the (colored) dependency graphs. While Jaeger [23] proposes to investigate the coherence for a given class of models described by a logical formula, Getoor et al. [19] are interested in a single class of inputs: the skeletons that are possible in practice. With a priori knowledge, the RBN user perhaps can attest to the acyclicity of the resulting ground Bayesian network for all possible inputs.

Any arc r(t),r(t) in the output ground Bayesian network BΦ(DS), for an RBN Φ and input DS reflects that the probability formula Fr(x), when x=t, depends on r(t). Hence, possible arcs in this network can be anticipated by looking into the probability formulas Fr(x), for the probabilistic relations rR, in the definition of Φ. In other words, by inspecting the probability formula Fr(x), we can detect those rR for which an arc r(t),r(t) can possibly occur in the ground Bayesian network. Similarly to the class dependency graph for PRMs, we can construct a high-level dependency graph for RBNs that brings the possible arcs, and thus cycles, in the ground Bayesian network.

Definition 11.

Given an RBN Φ, the R-dependency graph GΦ is a directed graph with a node for each probabilistic relation rR and the following arcs:

  • Type I arcs: r,r, where r(x) occurs in Fr(x) outside the scope of a combination function;

  • Type II arcs: r,r, where r(y) occurs in Fr(x) inside the scope of a combination function.

Intuitively, a Type I arc r,r in the R-dependency graph of an RBN Φ means that, for any input structure DS over D and any tuple tDa(r), Fr(t) depends on r(t) in the ground Bayesian network BΦ(DS); formally, r(t)α(Fr(x),t,DS). For instance, if Fr1(x)=mean({|r2(y)y;S(x,y)|})(1r3(x)), then, given any S-structure, Fr1(t) depends on r3(t) for any t. Type II arcs capture dependencies that are contingent on the S-relations holding in the input structure. In other words, a Type II arc r,r means that Fr(t) will depend on r(t) if some S-formula φ holds in the input structure DS; DSφ. For instance, if Fr1(x)=(mean({|r2(y)y;S(x,y)|})(1r3(x)), r1(t) depends on r2(t) (for t,t in the domain D) in the output ground Bayesian network iff the input DS is such that DSS(t,t). If combination functions are nested, the corresponding S-formula might be fairly complicated. Nevertheless, the point here is simply noting that, given a Type II arc r,r, the conditions on which r(t) is actually a child of r(t) in the ground Bayesian network can be expressed with an S-formula parametrized by t,t, which will be denoted by φSr,r(t,t). Consequently, for t,tD, DSφSr,r(t,t) iff r(t)α(Fr(x),t,DS), i.e., r(t) depends on r(t) in BΦ(DS).

As each arc in the ground Bayesian network corresponds to an arc on the R-dependency graph, when the latter is acyclic, so will be the former, for any input structure. As it happens with class dependency graphs and PRMs, though, a cycle in the R-dependency graph does not entail a cycle in the ground Bayesian network if a Type II arc is involved. It might well be the case that the input structures DS found in practice do not cause cycles to occur. This can be captured via a colored version of the R-dependency graph.

In the same way that Type I arcs in the class dependency graph of a PRM relate to attributes of different objects, in the R-dependency graph of an RBN, these arcs encode the dependency between relations r,rR to be grounded with (possibly) different tuples. For a PRM, the ground Bayesian network can never reflect a cycle with green arcs, but no red one, in the class dependency graph, for a sequence of green arcs guarantees different objects, according to a partial ordering. Analogously, with domain knowledge, the user can identify Type II arcs in the R-dependency graph whose sequence will prevent cycles in the ground Bayesian network, via a partial ordering over the tuples.

For a vocabulary S of predefined relations, let TD={DaaN} denote the set of all tuples with elements of D. We say a set Aga={ri,ri1in} of Type II arcs is guaranteed acyclic if, for any possible input structure DS over D, there is a partial ordering ⪯ over TD such that, if DSφSr,r(t,t) for some t,tTD, then tt. Here, again, “possible” means “possible in practice”.

Definition 12.

Given the R-dependency graph of an RBN Φ and a set Aga of guaranteed acyclic type II arcs, the colored R-dependency graph GΠ is a directed graph with a node for each rR and the followin arcs:

  • Yellow arcs: Type I arcs in the R-dependency graph;

  • Green arcs: Type II arcs r,r in the R-dependency graph such that r,rAga;

  • Red arcs: The remaining (Type II) arcs in the R-dependency graph.

Again, yellow cycles in the colored R-dependency graph correspond to relations rR grounded with the same tuple t, yielding a cycle in the ground Bayesian network. If green arcs are added to a cycle, then it is guaranteed that, departing from a node r(t) in the ground Bayesian network, these arcs form a path to r(t), where tt for a partial ordering ⪯, and there is no cycle. Once more, red arcs in cycles may cause t=t, and coherence is not ensured. Calling stratified a R-dependency graph whose every cycle contains at least one green arc and no red arc, we have:

Theorem 4.

Given the R-dependency graph of an RBN Φ and a set Aga of guaranteed acyclic Type II arcs, if the colored class dependency graph GΦ is stratified, then the ground Bayesian network is acyclic for any possible input structure.

Of course, detecting guaranteed acyclic Type II arcs in R-dependency graphs of RBNs is even harder than, as a generalization of, detecting guaranteed acyclic slot chains in PRMs. In any case, if the involved relations r,rR are unary, one is in a position similar to finding acyclic slot chains, as the arguments of r,r can be seen as objects, and only a partial ordering over the elements of the domain (not tuples) is needed.

7. Conclusions

In this paper, we examined a new version of coherence checking, a central problem in the foundations of probability as conceived by de Finetti. The simplest formulation of coherence checking takes a set of events and their probabilities and asks whether there can be a probability measure over an appropriate sample space [1]. This sort of problem is akin to inference in propositional probabilistic logic [28]. Unsurprisingly, similar inference problems have been studied in connection with first-order probabilistic logic [27]. Our focus here is on coherence checking when one has events specified by first-order expressions, on top of which one has probability values and independence relations. Due to the hopeless complexity of handling coherence checking for any possible set of assessments and independence judgments, we focus on those specifications that enhance the popular language of Bayesian networks. In doing so, we address a coherence checking problem that was discussed in the pioneering work by Jaeger [23].

We have first examined the problem of checking the coherence of relational Bayesian networks for a given class of input structures. We used first-order logic to encode the output ground Bayesian network into a first-order structure, and we employed a transitive closure operator to express the acyclicity demanded by coherence, finally reducing the coherence checking problem to that of deciding the validity of a logical formula. We conjecture that Jaeger’s original proposal concerning the format of the formula encoding the consistency of a relational Bayesian network Φ for a class S cannot be followed as originally stated; as we have argued, the possible number of tuples built from a domain typically outnumbers its size, so that there is no straightforward way to encode the ground Bayesian network, whose nodes are ground atoms, into the input S-structure. Therefore, it is hard to think of a method that translates the acyclicity of the ground Bayesian network into a formula φΦ to be evaluated over an input structure in the class S (satisfying θS). Our contribution here is to present a logical scheme that bypasses such difficulties by employing a bipartite domain, encoding both the S-structure and the corresponding Bayesian network. We have also extended those results to PRMs, in fact mixing the existing graph-based techniques for coherence checking with our logic-based approach. Our results seem to be the most complete ones in the literature.

Future work includes searching for decidable instances of the formula encoding the consistency of a relational Bayesian network for a class of input structures and exploring new applications for the logic techniques herein developed.

Acknowledgments

GDB was supported by Fapesp, Grant 2016/25928-4. FGC was partially supported by CNPq, Grant 308433/2014-9. The work was supported by Fapesp, Grant 2016/18841-0.

Abbreviations

The following abbreviations are used in this manuscript:

CPD Conditional Probability Table
g.a. guaranteed acyclic
iff if and only if
PRM Probabilistic Relational Model
PSAT Probabilistic Satisfiability
RBN Relational Bayesian Network

Author Contributions

Both authors have contributed to the text and read and approved the final manuscript.

Conflicts of Interest

The authors declare no conflict of interest. The founding sponsors had no role in the design of the study; in the collection, analyses or interpretation of data; in the writing of the manuscript; nor in the decision to publish the results.

References

  • 1.De Finetti B. Theory of Probability. Volumes 1 and 2 Wiley; New York, NY, USA: 1974. [Google Scholar]
  • 2.Coletti G., Scozzafava R. Probabilistic Logic in a Coherent Setting. Kluwer; Dordrecht, The Netherlands: 2002. Trends in Logic, 15. [Google Scholar]
  • 3.Lad F. Operational Subjective Statistical Methods: A Mathematical, Philosophical, and Historical, and Introduction. John Wiley; New York, NY, USA: 1996. [Google Scholar]
  • 4.Berger J.O. In Defense of the Likelihood Principle: Axiomatics and Coherency. In: Bernardo J.M., DeGroot M.H., Lindley D.V., Smith A.F.M., editors. Bayesian Statistics 2. Elsevier Science; Amsterdam, The Netherlands: 1985. pp. 34–65. [Google Scholar]
  • 5.Regazzini E. De Finetti’s Coherence and Statistical Inference. Ann. Stat. 1987;15:845–864. doi: 10.1214/aos/1176350379. [DOI] [Google Scholar]
  • 6.Shimony A. Coherence and the Axioms of Confirmation. J. Symb. Logic. 1955;20:1–28. doi: 10.2307/2268039. [DOI] [Google Scholar]
  • 7.Skyrms B. Strict Coherence, Sigma Coherence, and the Metaphysics of Quantity. Philos. Stud. 1995;77:39–55. doi: 10.1007/BF00996310. [DOI] [Google Scholar]
  • 8.Savage L.J. The Foundations of Statistics. Dover Publications, Inc.; New York, NY, USA: 1972. [Google Scholar]
  • 9.Darwiche A. Modeling and Reasoning with Bayesian Networks. Cambridge University Press; Cambridge, UK: 2009. [Google Scholar]
  • 10.Koller D., Friedman N. Probabilistic Graphical Models: Principles and Techniques. MIT Press; Cambridge, MA, USA: 2009. [Google Scholar]
  • 11.Pearl J. Probabilistic Reasoning in Intelligent Systems: Networks of Plausible Inference. Morgan Kaufmann; San Mateo, CA, USA: 1988. [Google Scholar]
  • 12.Getoor L., Taskar B. Introduction to Statistical Relational Learning. MIT Press; Cambridge, MA, USA: 2007. [Google Scholar]
  • 13.De Raedt L. Logical and Relational Learning. Springer; Berlin, Heidelberg: 2008. [Google Scholar]
  • 14.Raedt L.D., Kersting K., Natarajan S., Poole D. Statistical Relational Artificial Intelligence: Logic, Probability, and Computation. Morgan & Claypool; San Rafael, CA, USA: 2016. [Google Scholar]
  • 15.Cozman F.G. Languages for Probabilistic Modeling over Structured Domains. Tech. Rep. 2018 submitted. [Google Scholar]
  • 16.Poole D. First-order probabilistic inference; Proceedings of the International Joint Conference on Artificial Intelligence (IJCAI); Acapulco, Mexico. 9–15 August 2003; pp. 985–991. [Google Scholar]
  • 17.Gilks W., Thomas A., Spiegelhalter D. A language and program for complex Bayesian modeling. Statistician. 1993;43:169–178. doi: 10.2307/2348941. [DOI] [Google Scholar]
  • 18.Lunn D., Spiegelhalter D., Thomas A., Best N. The BUGS project: Evolution, critique and future directions. Stat. Med. 2009;28:3049–3067. doi: 10.1002/sim.3680. [DOI] [PubMed] [Google Scholar]
  • 19.Getoor L., Friedman N., Koller D., Pfeffer A., Taskar B. Probabilistic relational models. In: Getoor L., Taskar B., editors. Introduction to Statistical Relational Learning. MIT Press; Cambridge, MA, USA: 2007. [Google Scholar]
  • 20.Koller D. Probabilistic relational models; Proceedings of the International Conference on Inductive Logic Programming; Bled, Solvenia. 24–27 June 1999; pp. 3–13. [Google Scholar]
  • 21.Heckerman D., Meek C., Koller D. Probabilistic Entity-Relationship Models, PRMs, and Plate Models. In: Getoor L., Taskar B., editors. Introduction to Statistical Relational Learning. MIT Press; Cambridge, MA, USA: 2007. pp. 201–238. [Google Scholar]
  • 22.Jaeger M. Relational Bayesian networks; Proceedings of the Thirteenth Conference on Uncertainty in Artificial Intelligence; Providence, RI, USA. 1–3 August 1997; pp. 266–273. [Google Scholar]
  • 23.Jaeger M. Relational Bayesian networks: A survey. Electron. Trans. Art. Intell. 2002;6:60. [Google Scholar]
  • 24.De Bona G., Cozman F.G. Encoding the Consistency of Relational Bayesian Networks. [(accessed on 23 March 2018)]; Available online: http://sites.poli.usp.br/p/fabio.cozman/Publications/Article/bona-cozman-eniac2017F.pdf.
  • 25.Alechina N., Immerman N. Reachability logic: An efficient fragment of transitive closure logic. Logic J. IGPL. 2000;8:325–337. doi: 10.1093/jigpal/8.3.325. [DOI] [Google Scholar]
  • 26.Ganzinger H., Meyer C., Veanes M. The two-variable guarded fragment with transitive relations; Proceedings of the 14th IEEE Symposium on Logic in Computer Science; Trento, Italy. 2–5 July 1999; pp. 24–34. [Google Scholar]
  • 27.Fagin R., Halpern J.Y., Megiddo N. A Logic for Reasoning about Probabilities. Inf. Comput. 1990;87:78–128. doi: 10.1016/0890-5401(90)90060-U. [DOI] [Google Scholar]
  • 28.Hansen P., Jaumard B. Probabilistic Satisfiability. École Polytechique de Montréal; Montreal, Canada: 1996. Technical Report G-96-31; Les Cahiers du GERAD. [Google Scholar]

Articles from Entropy are provided here courtesy of Multidisciplinary Digital Publishing Institute (MDPI)

RESOURCES