Skip to main content

This is a preprint.

It has not yet been peer reviewed by a journal.

The National Library of Medicine is running a pilot to include preprints that result from research funded by NIH in PMC and PubMed.

ArXiv logoLink to ArXiv
[Preprint]. 2025 Aug 6:arXiv:2310.07543v5. Originally published 2023 Oct 11. [Version 5]

Ordinal Characterization of Similarity Judgments

Jonathan D Victor *, Guillermo Aguilar , Suniyya A Waraich
PMCID: PMC10593068  PMID: 37873008

Abstract

Characterizing judgments of similarity within a perceptual or semantic domain, and making inferences about the underlying structure of this domain from these judgments, has an increasingly important role in cognitive and systems neuroscience. We present a new framework for this purpose that makes limited assumptions about how perceptual distances are converted into similarity judgments. The approach starts from a dataset of empirical judgments of relative similarities: the fraction of times that a subject chooses one of two comparison stimuli to be more similar to a reference stimulus. These empirical judgments provide Bayesian estimates of underling choice probabilities. From these estimates, we derive indices that characterize the set of judgments in three ways: compatibility with a symmetric dis-similarity, compatibility with an ultrametric space, and compatibility with an additive tree. Each of the indices is derived from rank-order relationships among the choice probabilities that, as we show, are necessary and sufficient for local consistency with the three respective characteristics. We illustrate this approach with simulations and example psychophysical datasets of dis-similarity judgments in several visual domains and provide code that implements the analyses at https://github.com/jvlab/simrank.

Keywords: perceptual spaces, maximum likelihood estimation, ultrametric space, additive trees, triads, multidimensional scaling

MSC2020 subject classifications: 91E30, 62P15, 92-08, 92-10, 51-08

1. Introduction

Characterization of the similarities between elements of a domain of sensory or semantic information is important for many reasons. First, these similarities, and the relationships between them, [8, 15, 34], reveal the cognitive structure of the domain. Similarities are functionally important as they are the substrate for learning, generalization, and categorization [15, 27, 30, 45]. At a mechanistic level, the quantification of similarities provides a way to test hypotheses concerning their neural substrates [18]. Thus, measuring of perceptual similarities, and using these judgments to make inferences about the geometry of the underlying perceptual spaces, plays an important role in cognitive and systems neuroscience.

The goal of this work is to present a novel approach that complements the standard strategies used for this purpose. The starting point for the present approach, in common with standard strategies, is a set of triadic similarity judgments: is stimulus x or stimulus y more similar to a reference stimulus r ? To make geometric inferences from such data, one standard approach is to make use of a variant of multidimensional scaling [6, 17, 22, 33, 37, 41], i.e., to associate the stimuli with points in a space, so that the distances between the points account for the perceptual similarities. Once these points are determined, inferences can be made about the dimensionality of the space, its curvature, and its topology. A second approach, topological data analysis, makes use of the distances directly, and then invokes graph-theoretic procedures [5, 11, 12, 31, 47] to infer these geometric features.

In applying these approaches to experimental data, one must deal with the fact that even if a forced-choice response is required, the response likely represents an underlying choice probability – and that this choice probability may depend on sensory noise, noise in how distances are mentally computed and transformed into dis-similarities, and noise in the decision process in which dis-similarities are compared. As a consequence, analysis of an experimental dataset requires, at least implicitly, substantial assumptions. Such assumptions are not always benign: a monotonic transformation of distances – which preserves binary similarity judgments – can alter the dimensionality and curvature of a multidimensional scaling model [19]. Topological data analysis via persistent homologies, which also only makes use of rank orders of distances, is invariant to a global monotonic transformation of distances, but makes other assumptions (for example, that this transformation is the same across the domain), and does not typically take into account a noise model.

With these considerations in mind, here we pursue an approach to make inferences from the choice probabilities themselves, as estimated from repeated triadic judgments. Our main assumption is that if, for any particular triad, comparison stimulus x is chosen more often than comparison stimulus y as closer to a reference stimulus r, then the distance between x and r is less than the distance between y and r. Note that we do not make any assumptions about how relative or absolute distances are transformed into choice probabilities within an individual triad, or whether this transformation is the same across the domain.

As we show, despite the relative paucity of assumptions, the approach nevertheless provides indices that characterize a set of similarity judgments in three useful ways. The first index quantifies the extent to which the similarity judgments are consistent with conditions necessary for symmetry – i.e., that in judging similarity, reference and comparison stimuli are treated in the same way. While at first glance one might expect that similarity judgments always reflect symmetric distances and most models (including those considered here) make this assumption, this need not be the case [34, 35] – so it is useful to have a way to determine whether an experimental dataset implies violations of symmetry. Applying our approach to test for symmetry is also the simplest of the three we consider.

The second index quantifies consistency with an ultrametric model, a geometry that is a formalization of strict hierarchical structure [29]. In an ultrametric model, elements of the space correspond to the leaf nodes on a rooted tree, and distance between two nodes is determined by the height of the first common ancestor. Such structure has been postulated for perceptual domains with semantic content [15, 27, 32, 36]. Consideration of ultrametric models is also motivated by evidence that representation of olfactory [47] and physical [46] space may have a hyperbolic geometry, as hyperbolic geometries provide natural embeddings of trees [25].

The third index quantifies consistency with an addtree (or additive tree) model, a generalization of the ultrametric model. In an addtree model, the distance between two points is determined by the length of the path on an acyclic graph. An addtree model allows for more flexible clustering and may be characteristic of semantic domains [15, 26]. Moreover, since a one-dimensional domain is a special case of an addtree [26], exclusion of addtree structure implies that a one-dimensional model cannot account for the rank order of similarity judgments – and therefore, cannot account for any model in which perceptual distances are monotonically related to these judgments. Addtree models are in a sense piecewise linear, and thus also may be appropriate models for olfactory perceptual spaces [44].

The organization of this paper is as follows. The first section sets out a formal framework of the approach and includes rigorous mathematical results. The second section moves from this framework to procedures that can be applied to experimental data, yielding indices that characterize consistency with symmetry, ultrametric structure and addtree structure. To deal with the fact that the choice probabilities are estimated quantities, we use a Bayesian approach, and details of its implementation differ for the three indices. In particular, integration over a local prior extends exactly to a global prior for the symmetry index and the ultrametric index, but is only approximate for addtree. We then apply the method to synthetic datasets. These examples demonstrate the ability of the indices to characterize similarity structure, that the characterization is largely insensitive to the Bayesian prior, and illustrate how interpretation can be augmented by analysis of surrogate datasets. The final results section applies the method to experimental data from three visual domains. The Discussion considers caveats, limitations, alternative strategies, implications for experimental design, and avenues for further development.

2. Theory

Overview, key terms, and preliminaries

Our goal is to develop indices that characterize a dataset of triadic similarity judgments, in a way that provides insight into the structure of the underlying perceptual space. Our central assumption is that, within a triadic judgment, the probability that a participant chooses one pair of stimuli as more similar than an alternative is monotonically related to the similarity. Typical datasets include large numbers of similarity judgments of overlapping triads, and the relationship between these judgments contains information about the underlying perceptual space. We show how this information can be accessed, without further making assumptions about the specifics of the monotonic relationship between choice probability and (dis)-similarity, whether it is constant throughout the space, or the decision process itself.

We define a triad, denoted (r;x,y), to be an ordered triple of stimuli, consisting of a reference stimulus r and two comparison stimuli, x and y, all drawn from a set S of all stimuli (Figure 1A). A triad forms the basic unit of data collection in an experiment: the participant is asked to decide, in a forced-choice response, which of the two comparison stimuli is more similar to the reference. The mirror (Figure 1B) of the triad (r;x,y) is defined as the triad (r;y,x). Several groupings of triads will be useful below. A triplet (Figure 1C) is defined as a set of the triads {(z;x,y),(x;y,z),(y;z,x)} that can be formed from three stimuli, with each stimulus serving as the reference for one of the triads. A tripod (Figure 1D) is defined as a set of triads {(w;x,y),(w;y,z),(w;z,x)}. A tent (Figure 1E) is defined as the union of a triplet (the base of the tent) and a tripod whose comparison stimuli are the stimuli in the triplet and whose common reference stimulus is the vertex of the tent, and is denoted {w;x,y,z}, where w is the vertex and the base is the triplet {(z;x,y),(x;y,z),(y;x,z)}.

Figure 1:

Figure 1:

Panel A. A triad (r;x,y) consists of a reference stimulus r and two comparison stimuli, x and y; the first comparison stimulus is indicated by the open arrow and the second by the closed arrow. Panel B. In the mirror triad (r;y,x), the roles of the two comparison stimuli are reversed (note the arrowheads near the reference stimulus r). Panel C. A triplet is a set of three triads {(z;x,y),(x;y,z),(y;z,x)} that can be formed from three stimuli, with each stimulus serving as a reference for the other two. Panel D. A tripod (shown here in “top view”) is a set of three triads constructed with a single stimulus (here, w) serving as a reference for the other three stimuli taken in pairs. Panel E. A tent is a union of a triplet and a tripod.

We define the choice probability for the triad (r;x,y), denoted R(r;x,y), to be the probability that a participant judges x as more similar to r than y is to r. We assume that the two comparison stimuli in a triad are treated equivalently, i.e., that the response to a triad can also be considered to be the alternative response to its mirror. That is,

R(r;x,y)+R(r;y,x)=1. (2.1)

We view the choice probability R(r;x,y) as an unknown to be estimated from an experiment in which the triad (r;x,y), among others, is presented; this is discussed in detail in the Implementation section below. We do not require that the experiment explore each triad or its mirror (in an experiment with M stimuli, there are MM12=12M(M1)(M2) pairs of triads and their mirrors), though clearly greater coverage of the triads, and more repeats of each, will lead to better estimates of the choice probabilities. We also assume that an experimenter chooses, randomly, whether to present a triad or its mirror.

We assume that a triadic judgment is the result of a two-step process: first, estimation of the dis-similarity between the reference and each of the comparison stimuli, and second, comparison of these dis-similarities. We denote the dis-similarity of a comparison stimulus z to a reference stimulus r by D(r,z). Our central assumption is that a participant is more likely to judge that x is more similar to r than y is to r if, and only if, D(r,x)<D(r,y). That is,

R(r;x,y)>12D(r,x)<D(r,y). (2.2)

Because (2.1) holds for every triad and its mirror, an immediate consequence is that a choice probability of exactly 12 only occurs when dis-similarities are exactly equal: combining (2.1) with (2.2) yields

R(r;x,y)=12D(r,x)=D(r,y). (2.3)

We also assume that dis-similarities are non-negative, and that a dis-similarity of zero only occurs for stimuli that are identical:

D(x,y)0andD(x,y)=0x=y. (2.4)

Note, however, that we do not assume that dis-similarities are symmetric, i.e., that D(x,y)=D(y,x). Rather, we will test whether a dataset is compatible with this constraint – essentially, whether the reference and comparison stimuli are treated equivalently. We use the term “symmetric distance” for a function d(x,y) that satisfies (2.4) and also d(x,y)=d(y,x). Note that a symmetric distance need not satisfy the triangle inequality, and therefore need not be a metric (see Appendix A).

Eqs. (2.2) and (2.3) formalize our focus on rank order of dis-similarities. That is, rather than assume (or attempt to infer) a quantitative relationship between the choice probability and perceived dis- similarities, we only make use of the sign of comparisons – for example, that an alligator and toothpaste are more dis-similar than an alligator and a panda – but we do not attempt to infer the size of this difference, in absolute terms, relative to other dis-similarities not included in triads that have been presented, or relative to an internal noise.

The analyses below will ask whether the choice probabilities, and hence, the dis-similarities, suffice to rule out a model whose distances conform to a specific kind of geometric space. We consider two ways of formalizing this notion, each based on a set of stimuli, a set T of triads formed from them, and a function f on (reference, comparison) pairs within the triads. Given this set-up, we define a basic notion of compatibility: a set of dis-similarities D is pointwise compatible with f if, for all the triads (r;x,y)T, the rank-order of pairwise dis-similarities is the same as the rank-order of values assigned by f:

D(r,x)<D(r,y)f(r,x)<f(r,y). (2.5)

This contrasts with a stronger definition: the dis-similarities D is setwise compatible with f if for all (r;x,y)T, there is a monotonic-increasing function F for which

f(x,y)=F(D(x,y)). (2.6)

We focus primarily on pointwise compatibility, a focus that considers only the relationships between dis-similarities that jointly occur in triadic comparisons. Since setwise compatibility implies pointwise compatibility, a set of choice probabilities that rules out pointwise compatibility necessarily rules out setwise compatibility.

The present analysis, which applies most directly to a paradigm in which each trial is devoted to judgment of a single triad, is applicable to other paradigms in which individual trials yield judgments about more than one triad, provided that each judgment can be considered to be independent of the context in which it is made. For example, in the “odd one out” paradigm (also known as the “oddity task” [16]), three stimuli are presented and the participant is asked to choose which one is the outlier. Here, a selection of a stimulus xj out of a set xj,xk,xl can be interpreted as a judgment that Dxk,xj>Dxk,xl and also that Dxl,xj>Dxl,xk, and thus contributes to estimates of choice probabilities for two triads, xk;xj,xl and xl;xj,xk. The analysis is also applicable to paradigms in which they are to rank m comparison stimuli x1,,xm in order of similarity to a shared reference stimulus r [41]. The ranking obtained on each trial then contributes to estimation of choice probabilities for m2 triads r;xk,xl, one for each pair of comparison stimuli.

Choice probabilities and conditions for compatibility with distance-based models

The elements introduced above enable characterization of the choice probabilities in three ways: compatibility of dis-similarities with a symmetric distance, compatibility with an ultrametric model, and compatibility with an addtree model. The characterizations for symmetry and ultrametric structure are the most straightforward since these properties do not make use of additive structure.

Symmetry

We consider symmetry first, as it is a fundamental property of distances and it is also the simplest of the three characterizations. We focus on the choice probabilities among members of a triplet of stimuli, {x,y,z}, which, for brevity, we denote R1R(x;y,z), R2R(y;z,x), R3R(z;x,y). These values fully characterize the choice probabilities among the three stimuli, as the three other triads consisting of these stimuli are all mirrors of one of the Ri.

Proposition 2.1 (ordinal conditions for symmetry).

For any triplet composed of the stimuli {x,y,z}, the dis-similarities are pointwise compatible with a symmetric distance if, and only if, one of the following conditions hold. With nhalf as the number of Ri that are exactly equal to 12,

  1. nhalf{0,1} and the Ri12 include both positive and negative values or

  2. nhalf=3.

Proof.

To show that these conditions are required for pointwise compatibility with symmetry: If none of them hold, then either nhalf=2, or nhalf{0,1}, and the nonzero values of Ri12 are all of the same sign.

Case 1: nhalf=2. Without loss of generality (WLOG), say R1=R2 but R312. Then D(x,y)=D(x,z),D(y,z)=D(y,x). If D were pointwise compatible with a symmetric d, then

d(z,x)=d(x,z)=d(x,y)=d(y,x)=d(y,z)=d(z,y). (2.7)

Pointwise compatibility then requires that D(z,x)=D(z,y), which contradicts R312.

Case 2: nhalf{0,1}. There are at least two nonzero values of Ri12, and these nonzero values are all the same sign. WLOG assume that two of the Ri are >12, and R2 is one of these. Then, R112, R2>12, R312 requires that

D(x,y)D(x,z)D(z,x)D(z,y)D(y,z)<D(y,x). (2.8)

If D is pointwise compatible with a symmetric distance d, then (2.8) implies that

d(x,y)d(x,z)d(z,x)d(z,y)d(y,z)<d(y,x). (2.9)

Symmetry of d leads to a contradiction,

d(x,y)d(x,z)=d(z,x)d(z,y)=d(y,z)<d(y,x). (2.10)

To show that these conditions suffice for pointwise compatibility with a symmetric distance: nhalf=3 is trivial; all dis-similarities are equal. For nhalf{0,1}:, WLOG assume that R112 and R312 have opposite signs; R2, which compares D(y,x) with D(y,z) is unconstrained. Choose d(x,y)=d(y,x)=D(y,x), d(z,y)=d(y,z)=D(y,z); this guarantees that the rank-order of dis-similarities required by R2 is respected by d. If R1>12 and R3<12, we have D(x,y)<D(x,z), D(z,x)>D(z,y). Thus choosing d(x,z)=d(z,x)=max{d(x,y),d(z,y)}+k (where k>0) yields pointwise compatibility with D on {x,y,z}. If instead, R1<12 and R3>12, a similar argument applies: these require D(x,y)>D(x,z), D(z,x)<D(z,y). Set d(x,z)=max{D(x,z),D(z,x)} and put d(x,y)=d(y,x)=D(y,x)+k, d(z,y)=d(y,z)=D(y,z)+k, where k is large enough to ensure that d(x,y) and d(z,y) are both larger than d(x,z). □

Remark 2.2.

Appendix A shows that compatibility with a symmetric distance implies compatibility with a metric.

Proposition 2.1 can be rephrased as follows: a necessary and sufficient condition for a triad to be pointwise compatible with a symmetric distance is that either all choice probabilities are 12, or if not, at least one of the choice probabilities Ri is strictly greater than 12 and at least one of them is strictly less than 12. That is, the triplet of choice probabilities Ri is compatible with a symmetric distance if the three choice probabilities lie in a subset Ωsym of [0,1]3 consisting of the [0,1]3 cube from which two smaller cubes, 0,123Ri12 and 12,13Ri12, are removed:

Ωsym=[0,1]30,12312,13. (2.11)

This domain is diagrammed in Figure 2A.

Figure 2:

Figure 2:

Panel A. The domain Ωsym (eq. (2.11)) in which the choice probabilities in a triplet are compatible with a symmetric distance (grey volumes, left) and its complement in [0,1]3 (pink volumes, right). See Proposition 2.1 for further details. Panel B. The domain Ωumi (eq. (2.15), in which the choice probabilities in a triplet are compatible with an ultrametric distance. Ωumi is the union of three orthogonal components (gray), each of which is a square of side length 12 and its interior.

Note that the condition in Proposition 2.1 is not sufficient for setwise compatibility across all stimuli, or even pointwise compatibility for sets with more than 3 stimuli. The chain of inequalities of (2.8) or (2.9) is the simplest of a series of necessary conditions for setwise compatibility: more generally, if, for any n-cycle a1,a2,,an, the choice probabilities satisfy

Ra1;a2,an12Ran;a1,an112Ra2;a3,a112,at least one inequality strict, (2.12)

then pointwise compatibility in a1,a2,,an with a symmetric distance is impossible. For if (2.12) holds, with the final inequality strict, then (generalizing (2.10)), there would be a contradiction:

da1,a2da1,an=dan,a1dan,an1=dan1,an=da2,a3<da2,a1. (2.13)

Figure 3 illustrates the case of n=4, and shows that pointwise compatibility for comparisons among three stimuli does not imply pointwise compatibility for comparisons among four stimuli. In this diagram, eq. (2.12) means that traversing the perimeter of the configuration yields four choice probabilities that are not all exactly 12, but also not on both sides of 12. As this leads to the contradiction (2.13), these dis-similarities among the four points and four triads cannot be pointwise compatible with symmetric distances. This scenario can occur even though Proposition 2.1 holds for all triplets. To see this, set Da2,a4 and Da4,a2 to be larger than the any other dis-similarity. This guarantees that the choice probabilities for triplets a1;a2,a4,a2;a4,a1,a4;a1,a2 and a3;a4,a2,a4;a2,a3,a2;a3,a4 include values that are both <12 and >12, which satisfies the conditions of Proposition 2.1.

Figure 3:

Figure 3:

A configuration of four stimuli and four triads (solid arrows) corresponding to the comparisons of eq. (2.12) for n=4. Dis-similarities among the four stimuli may fail to be pointwise compatible with a symmetric distance even though the dis-similarities among the two triplets (upper and lower triangles, and their associated solid and dashed arrows) are compatible with a symmetric distance. For further details, see text following eq. (2.12).

Ultrametric

The motivation for considering compatibility with ultrametric distances begins with the observation that pointwise compatibility with a symmetric dis-similarity guarantees pointwise compatibility with a metric-space structure (Appendix A). It is therefore natural to ask whether the dis-similarities have further properties associated with specific kinds of metric spaces. Ultrametric spaces [29] are one important such kind, as they abstract the notion of a hierarchical organization – and have therefore been proposed as models for perceptual representations [15, 27, 32, 36]. Points in an ultrametric space correspond to the terminal nodes of a tree, and the distance between two points corresponds to the height of their first common ancestor. Formally, a distance d is said to satisfy the ultrametric inequality if, for any three points x, y, and z,

d(x,y)max(d(x,z),d(y,z)), (2.14)

a condition that implies the triangle inequality (Appendix A). Eq. (2.14) states that for the three pairwise distances among three points, none can be strictly greater than the other two, i.e., at least two of the distances must be equal and the third cannot be longer.

Proposition 2.3 (ordinal conditions for ultrametric).

For a triplet composed of the stimuli {x,y,z}, the dis-similarities are pointwise compatible with an ultrametric distance if and only if the dis-similarities are symmetric and at least one of the following three hold:

R112,R212,R3=12R1=12,R212,R312R112,R2=12,R312 (2.15)

where, as before, R1R(x;y,z), R2R(y;z,x), R3R(z;x,y). This domain, denoted Ωumi is illustrated in Figure 2B.

Proof.

To show that (2.15) and symmetry implies pointwise compatibility with an ultrametric: WLOG, assume the first set holds. Then

R112D(x,y)D(x,z),R212D(y,z)D(y,x),R3=12D(z,x)=D(z,y). (2.16)

Using symmetry, this implies

D(x,y)D(x,z)=D(y,z). (2.17)

Consider the transformation d=G(D) where

G(D)=0,D=01+DD+1,D>0 (2.18)

As this is strictly monotonic-increasing, it demonstrates (2.14). Moreover, G satisfies the triangle inequality (see Appendix A, eq. (A.6)). Thus G is a transformation that demonstrates setwise compatibility with an ultrametric distance on the triad, and, a fortiori, pointwise compatibility.

Conversely, if d is an ultrametric distance and D is pointwise compatible with d, then (2.14), which holds for d, must also hold for D. Therefore, of the three values D(x,y), D(y,z), D(z,x), two must be equal and the third can be no larger than the others. WLOG, assume D(y,z)=D(x,z) and D(x,y) is no larger. Then the first alternative of (2.15) holds. □

Remark 2.4.

If the conditions of Proposition 2.3 hold for all triads composed of stimuli in S, the dis-similarities are setwise compatible with an ultrametric distance on S. This follows immediately from the proof of Proposition 2.3, as setwise compatibility with an ultrametric distance on each triplet is demonstrated via a transformation (2.18) that is the same for all triplets.

Addtree

Like the ultrametric model, the additive similarity tree model [26] is a metric space model that places constraints on the properties of the distance, but these constraints are less-restrictive than the constraints of the ultrametric model (Appendix B) and therefore may have greater suitability as a model for perceptual representations. In this model, here referred to as “addtree”, the distance between two points is determined by a graph that has a tree structure, in which each link has a specified nonzero weight. The distance between two points is given by the total weight of the path that connects the points. Because of the requirement that the graph is a tree structure, there are no loops – and this places constraints on the inter-relationships of the distances.

To determine the extent to which the dis-similarities implied by a set of triadic judgments are compatible with the distances in an addtree model, the starting point is a necessary and sufficient condition for distances in a metric space to be compatible with an addtree structure [4, 7, 26]. This condition, known as the “four-point condition,” is that given any four points u, v, w, and x,

none of the three quantitiesd(u,v)+d(w,x)d(u,w)+d(v,x)d(u,x)+d(v,w)is strictly greater than the other two. (2.19)

Put another way, of the three pairwise sums in eq. (2.19), two must be equal, and the third can be no larger. Appendix B shows that this condition is weaker than the ultrametric inequality and stronger than the triangle inequality, and that a one-dimensional arrangement of points is always compatible with an addtree model.

Since the four-point condition is based on adding distances, we cannot apply it directly to dis-similarities – as distances are linked to dis-similarity via an unknown monotonic function. However, there are conditions on the dis-similarities that are necessary for the four-point inequality to hold.

Proposition 2.5 (necessary ordinal conditions for addtree).

If, in a tent {z;a,b,c}, the inequalities

R(z;c,b)12,R(z;c,a)12,R(a;b,c)12,R(b;a,c)12. (2.20)

hold, along with strict inequalities

R(z;c,b)<12orR(a;b,c)<12andR(z;c,a)<12orR(b;a,c)<12 (2.21)

then the dis-similarities in the tent {z;a,b,c} are not pointwise compatible with an addtree distance.

Proof.

Via eqs (2.2) and (2.3), these conditions may be rewritten as inequalities among dis-similarities. The inequalities (2.20) become

D(z,c)D(z,b)andD(z,c)D(z,a)andD(a,b)D(c,a)andD(a,b)D(b,c). (2.22)

The inequalities (2.21) become

(D(z,c)>D(z,b)orD(a,b)>D(a,c))and(D(z,c)>D(z,a)orD(a,b)>D(b,c)). (2.23)

Either alternative of the first portion of eq. (2.23) leads to

D(z,c)+D(a,b)>D(z,b)+D(c,a). (2.24)

This follows from eqs. (2.22) and (2.23) via a term-by-term comparison of the two sides of eq. (2.24): (2.22) implies that each term on the left of (2.24) is no smaller than the corresponding term on the right, and (2.23) implies that at least one of these inequalities is strict. Further, note that conditions (2.22) are unchanged by swapping a with b, and the two portions of (2.23) are interchanged by this swap. Similarly, either alternative of the second portion of (2.23) leads to

D(z,c)+D(a,b)>D(z,a)+D(c,b). (2.25)

If the dis-similarities D were pointwise-compatible with an addtree distance d, then (2.24) and (2.25) would imply (via (2.5)) that

d(z,c)+d(a,b)>d(z,b)+d(c,a)andd(z,c)+d(a,b)>d(z,a)+d(c,b), (2.26)

in contradiction to (2.19). □

It is helpful to think of Proposition 2.5 geometrically. If all dis-similarities are unequal, its conditions state that, for any tent, the largest dis-similarity in a tripod cannot be opposite the largest dis-similarity in the base. If some dis-similarities are equal, then its conditions state that there is strict inequality for comparisons between one of the oppositely-paired dis-similarities and the other two oppositely-paired dis-similarities, either in the tripod or in the base (Figure 1C-E).

Proposition C.1 (Appendix C) is a partial converse to Proposition 2.5: if the dis-similarities among four points are all unequal and the conjunction (2.20) is false, we construct a monotonic transformation that demonstrates setwise (and therefore pointwise) compatibility between the dis-similarities and an addtree distance. But note that even though the “4-point” condition (2.19) on distances suffices to ensure a global addtree model, Proposition C.1 (Appendix C) stops short of showing that the monotonic transformations needed to transform dis-similarities to distances for each quadruple of points can be made in a globally-consistent way – though we do not have examples to the contrary.

3. Implementation

In this section, we move from the results described above to procedures that can be applied to experimental data. Specifically, we develop indices, computable from a set of triadic dis-similarity judgments, that express the likelihood that these judgments are compatible with an underlying symmetric distance (the index Isym), an ultrametric distance (the index Iumi), and an addtree distance (the index Iadd). Isym is the simplest and illustrates the basic strategy; Iumi and Iadd each build on this strategy in different ways.

Common to all three indices is the hurdle that the choice probabilities R(r;x,y) are unknown, and must be estimated from experimental data. We denote the number of such trials in which the triad (r;x,y) is presented by N(r;x,y), and the number of trials in which the participant judges x as more similar to r than y is to r by C(r;x,y). This provides a naïve estimate of the choice probability:

Robs(r;x,y)=C(r;x,y)N(r;x,y). (3.1)

To ensure that the estimated choice probabilities obey (2.1), we consider any presentation of a triad (r;x,y) to also be a presentation of its mirror, so that any response contributes to C(r;x,y) (if x is judged as more similar to r than y is to r) or to C(r;y,x) (if otherwise). With this convention,

C(r;x,y)+C(r;y,x)=N(r;x,y)=N(r;y,x) (3.2)

and

Robs(r;x,y)+Robs(r;y,x)=1. (3.3)

However, Robs(r;x,y) is only an estimate of the choice probability R(r;x,y). Thus, rather than determine whether the naïve estimates Robs(r;x,y) satisfy the conditions of Propositions 2.1, 2.3, 2.5 and C.1, we take a Bayesian approach: given the observed data, what is the likelihood that the underlying choice probabilities R(r;x,y) are consistent with the requisite inequalities?

The Bayesian approach requires a prior for the distribution of the set of choice probabilities. For the assessment of compatibility with symmetry between reference and comparison stimuli, the prior assumes that the choice probabilities are independently drawn from a specified univariate distribution (see below). For the assessment of compatibility with an ultrametric or an addtree structure, we then modify the prior by eliminating combinations of choice probabilities that are inconsistent with symmetry.

We implement this strategy using beta functions as a family of priors for choice probabilities. Since we have assumed that an experimenter randomly chooses whether to present a triad or its mirror, (2.1) means that appropriate priors for the distribution of choice probabilities, p(R), should satisfy p(R)=p(1R). (This “symmetry” refers to interchangeability of the two comparison stimuli x and y in the triad (r;x,y); not to interchange between reference and comparison, which is our focus.). Thus, it suffices to consider the family of priors

pa(R)1B(a,a)Ra1(1R)a1, (3.4)

where B(a,a) is the symmetric specialization of the beta function B(a,b), defined in the standard fashion by

B(a,b)01ua1(1u)b1du=Γ(a)Γ(b)Γ(a+b). (3.5)

As the parameter a varies over the positive reals, the shape of the prior (3.4) changes from heavily weighted near the extremes of R=0 and R=1 (a near zero), to heavily weighted near R=1/2(a1), thus capturing scenarios ranging from those in which most judgments are near certainty, to those in which most judgments are close to equivocal. In between, a=1 corresponding to the scenario in which choice probabilities are evenly distributed in [0,1].

While this choice of priors is fundamentally a heuristic one, it has both theoretical motivations and practical advantages. The beta distribution is the univariate case of the Dirichlet distribution, a distribution that has theoretical justification as a prior for multivariate probabilities [9] – so the beta distribution is a natural choice for a decision model in which an internal multivariate state is collapsed to a binary choice. In addition to encompassing a range of qualitatively different shapes, the family includes important special cases. For a=1, it is flat. For a=1/2, (3.4) is the Jeffreys prior for probabilities in the [0,1] interval, i.e., the “uninformative” prior in an information-theoretic sense [14]. At a=0, the R-dependence of (3.4), 1R(1R), is the (improper) Haldane prior, the unique prior for which the expected value of R is equal to the naïve estimate CN [14]. At a practical level, this family of priors has shown its utility as a prior distribution for choice probabilities in the context of improving the estimation of psychometric functions [28]. Finally, the key integrals involving (3.4) are easy to compute (e.g., eq. (3.7), leading to eq. (3.8)), providing a computationally efficient means to select the parameter a that is most appropriate given the observed data. Note also that the prior (3.4) is symmetric about R=1/2, so that Bayesian inference of R from an Robs that satisfies (3.3) will necessarily satisfy (2.1).

We use a maximum likelihood approach to determine the parameter a (alternative choices for a will be considered in the example applications below). Specifically, we maximize the likelihood of the observed set of responses across the entire experiment, assuming that the individual responses for the i-th triad ri;xi,yi are independently drawn from a Bernoulli distribution with parameter Ri=Rri;xi,yi, and that each Ri is independently drawn from the distribution (3.4). That is, for a given Ri, the probability that the subject reports D(r,x)<D(r,y) in Ci of Ni presentations is

pCiRi,Ni=NiCiRiCi1RiNiCi. (3.6)

Integrating over the prior (3.4) for Ri yields the probability of observing Ci reports of D(r,x)<D(r,y) in Ni presentations, given the parameter a:

pCiNi,a=NiCi011B(a,a)Ra11Ria1RiCi1RiNiCidRi=NiCiBa+Ci,a+NiCiB(a,a) (3.7)

Making use of the independence of each triad yields the overall log-likelihood:

LL(a)=logipCiNi,a=logK+ilogBa+Ci,a+NiCiB(a,a), (3.8)

where K is a combinatorial factor independent of a, and the sum ranges over all triads. Maximizing (3.8) then determines the value of a for which independent draws of choice probabilities are most likely to yield the experimental data.

We then use this value of a to determine the posterior likelihood for choice probabilities within a triplet or a tent. That is, for a set R of choice probabilities Ri, the prior is

P(R)=i1B(a,a)Ria11Ria1. (3.9)

Via Bayes rule, this prior determines the posterior likelihood of a set of choice probabilities:

p(RC,N)=p(CR,N)P(R)p(CN). (3.10)

where C denotes the responses Ci to each of the triads, N denotes the number of times that each triad Ni was presented, and p(CR,N) denotes a product of the terms specified by eq. (3.6). p(CN), the a priori probability of C and N, is unknown, but as is standard in Bayesian analyses, it is eliminated when likelihood ratios are calculated.

Symmetry

We now use this machinery to estimate the probability that the choice probabilities underlying a set of observations are compatible with a symmetric distance. We first focus on the data within a single triplet, and then consider extension of the analysis to the entire dataset.

For the analysis within a single triplet (i.e., given the observations CT and NT for the triads in a triplet T), we compare the posterior likelihood for the corresponding choice probabilities RT for which the inequalities of Proposition 2.1 hold, to the likelihood that the observations result from choice probabilities within the entire space of choice probabilities. We denote the posterior likelihoods by

Lsym(T)=ΩsympRTCT,NTdRT, (3.11)

and

L(T)=ΩpRTCT,NTdRT, (3.12)

and their ratio by

LRsym(T)=Lsym(T)L(T), (3.13)

where Ω is the space in which all choice probabilities in the triplet T range independently over [0,1], and Ωsym is the subset of the space consistent with the conditions of Proposition 2.1, namely, the cube [0,1]3 from which 0,123 and 12,13 are removed (eq. (2.11), Figure 2A)).

Both quantities (3.11) and (3.12) can be expressed in terms of the prior via Bayes’ rule:

pRTCT,NT=pCTRT,NTPRTpCTNT=1pCT,NTiTNiCiRiCi1RiNiCipaRi, (3.14)

where the prior for each choice probability, paRi, is given by (3.4). The likelihood ratio (3.13) LRsym(T) is thus

LRsym(T)=ΩsymiTRiCi1RiNiCipaRidRTiT01RiCi1RiNiCipaRidRi (3.15)

Based on (2.11), the numerator of (3.15) is a combination of three terms,

ΩsymiTRiCi1RiNiCipaRidRT=[0,1]3iTRiCi1RiNiCipaRidRT12,13iTRiCi1RiNiCipaRidRT0,123iTRiCi1RiNiCipaRidRT, (3.16)

each of which can be written in terms of incomplete beta functions:

vwRC(1R)NCpa(R)dR=1B(a,a)vwRC+a1(1R)NC+a1dR=B(w;a+C,a+NC)B(v;a+C,a+NC)B(a,a) (3.17)

where

B(w;a,b)=0wua1(1u)b1du. (3.18)

Each factor of the denominator of (3.15) can also be expressed in terms of beta functions:

01RiCi1RiNiCipaRidRi=1B(a,a)01RiCi1RiNiCiRia11Ria1dRi=1B(a,a)01RiCi+a11RiNiCi+a1dRi=Ba+Ci,a+NiCiB(a,a) (3.19)

Combining (3.15) through (3.19) yields

LRsym(T)=1iT1B12;a+Ci,a+NiCiBa+Ci,a+NiCiiTB12;a+Ci,a+NiCiBa+Ci,a+NiCi. (3.20)

Many software packages (e.g., MATLAB) provide the normalized beta function

Bnorm(w;a,b)B(w;a,b)B(a,b)=1B(a,b)0wua1(1u)b1du, (3.21)

which simplifies (3.20) to

LRsym(T)=1iT1Bnorm12;a+Ci,a+NiCiiTBnorm12;a+Ci,a+NiCi. (3.22)

To extend this analysis to the entire set of observations C and N, we note that the conditions of Proposition 2.1 concern the triads within a single triplet, and that the triads within distinct triplets are non-overlapping. Thus, the posterior likelihood for choice probabilities consistent with Proposition 2.1 at all triplets, LsymC,N, is an integral over a product space Ωsym=TtripΩsym(T) with one component for each triplet. The integral over the component corresponding to the triplet T is given by (3.11). The unrestricted posterior likelihood LC,N is an integral over a product space Ω of choice probabilities for all triads. Since each triad is a member of exactly one triplet, this integral also factors into a product of terms corresponding to (3.12), one for each triplet. Consequently, the likelihood ratio for the entire dataset is given by

LRsym(C,N)=Lsym(C,N)L(C,N)=Ttrip1iT1Bnorm12;a+Ci,a+NiCiiTBnorm12;a+Ci,a+NiCi (3.23)

In view of (3.10), LRsym1LRsym (and the analogous likelihood ratios for the indices below) may be interpreted as a Bayes factor for the triplet: the ratio between the likelihood of the data C,N given that all of the choice probabilities are pointwise-compatible with symmetry, vs. the likelihood of the data given that the choice probabilities are drawn from the complementary portion of the prior.

While it is thus tempting to interpret (3.23) as a global measure of compatibility with symmetry, there is an important caveat: as mentioned above, pointwise compatibility at each triplet does not guarantee compatibility across the entire stimulus space S. To emphasize this point, we normalize LRsym by the number of triplets, yielding a quantity that can rigorously be interpreted an average measure of pointwise compatibility within triplets:

Isym(C,N)=1#(trip)logLRsym(C,N). (3.24)

Values of Isym(C,N) that are close to zero indicate that nearly all of the posterior distribution of choice probabilities lies within the product space Ωsym=TtripΩsym(T) (i.e., is pointwise compatible with a symmetric distance); progressively more negative values indicate that the posterior shifts into its complement ΩΩsym in which symmetry is necessarily violated.

A useful benchmark in interpreting Isym(C,N) is its value in the absence of any data (i.e., C=N=0). In that case, each of the normalized beta functions has a value of Bnorm12;a,a=12, so

Isym(0,0)=1#(trip)logTtrip11123123=log340.2877. (3.25)

Thus, values of Isym(C,N) greater than log34 are more compatible with symmetry than an index derived from choice probabilities drawn randomly from the prior.

Note also that deviations from this a priori value can only be driven by triplets in which there are observations for at least two of the triads. This follows from the fact that Bnorm12;a,a=12, so that if only one triad kT has a nonzero number of observations,

1iT1Bnorm(12;a+Ci,a+NiCi)iTBnorm(12;a+Ci,a+NiCi)=112121Bnorm(12;a+Ck,a+NkCk)1212Bnorm(12;a+Ck,a+NkCk)=34. (3.26)

This is a reassuring result: we should only be able to make inferences about the structure of the dis-similarity judgments if there is experimental data about more than one triad within a triplet. With data about only one triad, knowing the sign of the comparison is useless since this sign is arbitrarily determined by how the triad is labeled, i.e., (r;x,y) vs (r;y,x).

Ultrametric

We use a parallel strategy to apply the present approach to evaluate compatibility with an ultrametric distance. Necessary and sufficient criteria for compatibility with an ultrametric distance concern a single triplet (Proposition 2.3), as is the case for symmetry (Proposition 2.1). Thus, we first focus on a single triplet and then extend to the entire dataset.

An immediate difficulty arises: the conditions of Proposition 2.3 are only satisfied on a set of measure zero, since at least one of the Ri must be exactly equal to 12. So a Bayesian analysis based on a continuous prior (including the beta-function prior of (3.4)) will always lead to a likelihood ratio of zero, since a posterior derived from a continuous prior via Bayes rule and Bernoulli trials cannot have a discrete mass at 12.

It is nevertheless possible to capture the spirit of ultrametric behavior in a rigorous way, and at the same time, address a way in which beta-function prior may be unrealistic. To do this, we posit that there is a fraction of triads for which the underlying choice probability is exactly 12. For example, such triads could consist of stimuli for which there is no basis for comparison: is a toothbrush or a mountain more similar to an orange? But we don’t know, a priori, what fraction h of the triads have this property. To take this into account, we generalize the prior for choice probabilities to be a sum of two components: one component is the beta-function prior used above (3.4), normalized to 1h; the second component is a point mass at 12, normalized to h:

pa,h(R)(1h)pa(R)+hδ(R12)=1hB(a,a)Ra1(1R)a1+hδ(R12). (3.27)

For fixed h, the parameter a is determined by maximizing the likelihood of the observed responses (generalizing (3.8)):

LL(a;h)=logipCiNi,a;h=logK+ilog(1h)Ba+Ci,a+NiCiB(a,a)+2Nih, (3.28)

With this prior, we can then determine the likelihood ratio as a function of h, in the limit that h approaches zero. We anticipate (and will confirm below) that for small values of h, the likelihood ratio will be proportional to h, since the mass in the prior at 12 is proportional to h. This proportionality serves as an index of compatibility with the ultrametric property: how quickly does the likelihood ratio increase, if a small fraction of the underlying choice probabilities are exactly 12. An alternative approach (not taken here) is that if the experimental dataset suggests that a prior pa,h(R) with h>0 is a substantially better fit to the distribution of choice probabilities than pa,0(R), this prior can be used directly to calculate a likelihood ratio, and the best-fitting value of h then provides an additional descriptor of the dataset.

To implement this strategy for individual triads, we define

Lumi(T;h)=ΩumipRTCT,NTdRT, (3.29)

and consider the likelihood ratio,

LRumi(T;h)=Lumi(T;h)Lsym(T;h), (3.30)

where Ωumi is the space in which all choice probabilities in the triplet T satisfy the conditions (2.15) of Proposition 2.3 (Figure 2B). Note that since ultrametric behavior is only defined with respect to a symmetric distance, the denominator in the likelihood ratio is given by (3.11) (but using the more general prior (3.27)), which only considers combinations of choice probabilities in Ωsym.

In view of Proposition 2.3, Ωumi is a union of regions defined by combinations of the signs of Ri12, for i{1,2,3} (Figure 2B), and Proposition 2.1 implies that the same is true for Ωsym. Thus, numerator and denominator of (3.30) can be written

Lq(T;h)=[0,1]kVqsgnR112,,sgnRk12R1C11R1N1C1RkCk1RkNkCkpa,hR1pa,hRkdR1dRk (3.31)

where k=3, Ri are the choice probabilities in the triad T, Ci and Ni tally the responses, and Vqσ1,,σk, for q=umi or q=sym, is an indicator function over the configuration of signs, which defines either Ωumi or Ωsym. (We write (3.31) in a general form; k=3 suffices for analyzing ultrametric behavior and symmetry but the analysis of addtree behavior will require k=6.)

For the numerator of (3.30), Proposition 2.3 dictates that the nonzero values of Vumi are:

Vumi(+1,1,0)=1,corresponding toR1>12,R2<12,R3=12Vumi(0,+1,1)=1,corresponding toR1=12,R2>12,R3<12Vumi(1,0,+1)=1,corresponding toR1<12,R2=12,R3>12Vumi(0,0,0)=1,corresponding toR1=12,R2=12,R3=12. (3.32)

All other values of Vumi(σ) are zero, since either they don’t correspond to any of the conditions, or to exactly two of those conditions. The latter is impossible, as it would require two equalities and one strict inequality among the dis-similarities.

For the denominator of (3.30), we find the nonzero values of Vsym from Proposition 2.1:

nhalf=0:Vsym(±1,±1,1)=Vsym(±1,1,±1)=Vsym(1,±1,±1)=1nhalf=1:Vsym(±1,1,0)=Vsym(0,±1,1)=Vsym(1,0,±1)=1nhalf=3:Vsym(0,0,0)=1, (3.33)

where, as before, nhalf is the number of Ri that are exactly equal to 12.

To establish the behavior of the likelihood ratio (3.30) as h0, we use (3.27) to isolate the dependence of integrals (3.31) on h. This is a polynomial:

Lq(T;h)=σhZ(σ)(1h)kZ(σ)Vqσ1,,σkWσ1;C1,N1Wσk;Ck,Nk, (3.34)

where the sum is over all 3k assignments of the elements of σ=σ1,,σk to {1,0,+1}, Z(σ) is the number of entries in σ that are equal to zero (each such entry incurring a factor of h), and W(σ;C,N) is the integral of the prior (3.27), weighted by the experimental data, over one segment of the domain:

W(σ;C,N)=01/2RC(1R)NCpa(R)dR,σ=1RC(1R)NCR=12,σ=01/21RC(1R)NCpa(R)dR,σ=+1 (3.35)

These evaluate to

W(σ;C,N)=1B(a,a)B(12;a+C,a+NC),σ=112N,σ=01B(a,a)1B(12;a+C,a+NC),σ=+1 (3.36)

Consequently,

Lq(T;0)=Z(σ)=0Vqσ1,,σkWσ1;C1,N1Wσk;Ck,Nk, (3.37)
ddhLq(T;h)=σZ(σ)hZ(σ)1(1h)kZ(σ)(kZ(σ))hZ(σ)(1h)kZ(σ)1Vqσ1,,σkWσ1;C1,N1Wσk;Ck,Nk, (3.38)

and

ddhLq(T;h)h=0=Z(σ)=1Vqσ1,,σkWσ1;C1,N1Wσk;Ck,NkkZ(σ)=0Vqσ1,,σkWσ1;C1,N1Wσk;Ck,Nk. (3.39)

For the numerator of (3.30), the small- h behavior is proportional to (3.39), because Lumi(T;0)=0, as the nonzero values of Vumi(σ) all have Z(σ)1. The denominator of (3.30), Lsym(T;0) is nonzero, because Vsym(σ)=1 for six triplets of nonzero arguments (the cases nhalf=0 in eq. (3.33)). Thus, for small h, the likelihood ratio (3.30) is proportional to h. This proportionality indicates to what extent adding a small amount of mass to the prior at R=12 leads to triplets of choice probabilities that are compatible with the ultrametric property.

Since the triplets in each triad form non-overlapping sets (as was the case for the analysis of symmetry), we can combine the likelihood ratios for each triplet to form a likelihood ratio for the entire dataset:

LRumi(C,N;h)=TtripLRumi(T;h). (3.40)

The analysis of the limiting behavior of LRumi(T;h) then motivates an index of the extent to which a set of observations is compatible with an ultrametric distance:

Iumi(C,N)=1#triplimh0logLRumi(C,N;h)logh. (3.41)

This index is an average measure of pointwise compatibility of choice probabilities with an ultrametric distance across all triads. Moreover, in view of the remark following Proposition 2.3 and noting that the triads in each triplet are independent, it can also be considered as a measure of setwise compatibility across the entire stimulus space S. In applying (3.41) to data, the limiting behavior can be determined by setting h to a small nonzero value, e.g., 0.01 or 0.001, as we will show below.

As is the case for symmetry Isym (eq. (3.24)), a useful benchmark is the a priori value, Iumi(0,0). To calculate this, it suffices to consider a single triad T0 for which there are no observations:

Iumi(0,0)=limh0logLRumiT0;hlogh. (3.42)

The numerator of LRumiT0;h (eq. (3.30)), and its behavior for small h is given by ddhLumiT0;hh=0. This can be computed from (3.39), noting that (from eq. (3.36))

W(σ;0,0)=12,σ=±11,σ=0 (3.43)

and that there are three nonzero contributors to Vumi with Z(σ)=1 (eq. (3.32)). Thus, (3.39) yields

LumiT0;h=322h+Oh2. (3.44)

The denominator of LRumiT0;h is LsymT0;h (eq. (3.30)), which, as mentioned above, has a nonzero value at h=0. Using (3.43) and noting that there are six nonzero contributors to Vsym with Z(σ)=0 (the cases nhalf=0 in eq. (3.33)), (3.37) yields

LsymT0;0=623. (3.45)

Combining (3.30), (3.42), (3.44), and (3.45) yields the a priori value of the index:

Iumi(0,0)=limh0log34h+Oh234logh=0. (3.46)

In sum, the index Iumi(C,N) (eq. (3.41)) evaluates whether an experimental set of dis-similarity responses is compatible with an ultrametric model, and does so in a way that recognizes the intrinsic limitation that experimental data can never show that a choice probability is exactly 12. If the index is greater than 0, the observed data are more likely to be compatible with an ultrametric model than a set of unstructured choice probabilities; values less than 0 indicate progressively greater deviations from an ultrametric model.

Addtree

Formulation of the addtree index follows along similar lines, but with a focus on tents rather than triplets – corresponding to the necessary and sufficient conditions for pointwise compatibility (respectively Propositions 2.5 and C.1). For consistency with the machinery developed above for Iumi, we continue with the more general prior (3.27), though the specialization to h=0 (the beta-function prior (3.4)), is of primary interest. For each tent T, we consider the likelihood ratio

LRaddtree(T;h)=Laddtree(T;h)Lsymtent(T;h). (3.47)

Here,

Laddtree(T;h)=ΩaddtreepRTCT,NTdRT, (3.48)

where Ωaddtree is the space in which the six choice probabilities in the tent T are compatible with symmetry but falsify the conditions (2.20) and (2.21) of Proposition 2.5, and

Lsymtent(T;h)=ΩsymtentpRTCT,NTdRT, (3.49)

where Ωsymtent is the space in which the six choice probabilities in the tent T are merely compatible with symmetry.

These integrals have the same form as (3.31), so it suffices to specify the values of VqsgnR112,,sgnR612. For definiteness, given a tent T={z;a,b,c} with z at the vertex and {a,b,c} at the base, we specify the six choice probabilities needed to compute V as follows: for the tripod component, R1R(z;b,c),R2R(z;c,a),R3R(z;a,b); for the base, R4R(a;b,c),R5R(b;c,a),R6R(c;a,b). These choice probabilities or those of the mirror triads (but no other choice probabilities) enter into determining whether the conditions (2.20) and (2.21) are falsified for this tent: the choice probabilities R1,R2,R4, and R5 are explicit in (2.20) and (2.21) and all of the Ri are used equally as the base elements {a,b,c} are permuted. Since Vaddtree and Vsymtent has six arguments, each of which can take on any of three values {1,0,1}, there are 36=729 values to specify.

For Vaddtree, these values may be determined as follows. For the choice probabilities to be pointwise compatible with an addtree distance (i.e., for Vaddtree=1), the conditions (2.20) and (2.21) cannot hold for any of the permutations of {a,b,c}. Since these conditions are symmetric under interchange of a and b, it suffices to consider the cyclic permutations. So the region of [0,1]6 in which Vaddtree=1 is the intersection of the region that falsifies the conditions (2.20) and (2.21), which we denote Vaddtree[c], with the regions that falsify these conditions after cyclic permutation of a,b,c, which we denote Vaddtree[a] and Vaddtree[b]. Additionally, Vaddtree=0 for sets of choice probabilities that are incompatible with a symmetric distance. Thus,

Vaddtree=Vaddtree[a]Vaddtree[b]Vaddtree[c]Vsymtent. (3.50)

Vaddtree[c]=1 except when all of the inequalities (2.20) hold, or, as specified by (2.21), when D(z,c)>D(z,a) or D(a,b)>D(b,c) (but not both) is replaced by equality, and D(z,c)>D(z,b) or D(a,b)>D(c,a) (but not both) is replaced by equality. Thus, Vaddtree[c]=0 as follows:

Vaddtree[c]+τ1,τ2,σ3,τ4,+τ5,σ6=0forτ1,τ4andτ2,τ5{(1,1),(1,0),(0,1)};σ3andσ6{1,0,+1}. (3.51)

Here, the paired τi’s – not both of which can be zero – handle the allowed equalities specified by (2.21) and the σ’s handle the lack of a dependence on the third and sixth arguments. Vaddtree[a] and Vaddtree[b] are then determined by cyclic permutation:

Vaddtree[a]σ1,σ2,σ3,σ4,σ5,σ6=Vaddtree[c]σ2,σ3,σ1,σ5,σ6,σ4Vaddtree[b]σ1,σ2,σ3,σ4,σ5,σ6=Vaddtree[c]σ3,σ1,σ2,σ6,σ4,σ5 (3.52)

Vsymtent occurs both in the likelihood Laddtree(T;h) as a factor via (3.50) and alone in the likelihood Lsymtent(T;h). The choice probabilities of the three triads in the base depend on dis-similarities between the elements of the triplet {a,b,c}, so the choice probabilities compatible with symmetry correspond to Vsymσ4,σ5,σ6=1 (eq. (3.33)). The three triads in the tripod are comparisons between D(z,a),D(z,b), and D(z,c). While these are unconstrained by symmetry, they must be consistent with transitivity. That is, all of the inequalities:

D(z,a)<D(z,b)D(z,b)<D(z,c)D(z,c)<D(z,a) (3.53)

cannot hold, nor can it hold if up to two of the inequalities are non-strict, nor if all signs of comparison are inverted. This precisely matches the constraints on three dis-similarities required for compatibility with symmetry in Proposition 2.1, so it is captured by Vsymσ1,σ2,σ3. Thus,

Vsymtentσ1,σ2,σ3,σ4,σ5,σ6=Vsymσ1,σ2,σ3Vsymσ4,σ5,σ6 (3.54)

In sum, the likelihood ratio LRaddtree(T;h)=Laddree(T;h)Lsymtent(T;h) is determined by

Laddtree(T;h)=[0,1]6Vaddtreesgn(R112),,sgn(R612)R1C11R1N1C1R6C61R6N6C6pa,hR1pa,hR6dR1dR6 (3.55)

and

Lsymtent(T;h)=[0,1]6Vsymtentsgn(R112),,sgn(R612)R1C11R1N1C1R6C61R6N6C6pa,hR1pa,hR6dR1dR6 (3.56)

where Vaddtree and Vsymtent are given by eqs. (3.50) and (3.54).

As with the other indices, we can now average the likelihood ratios across all tents to form an index of compatibility with an addtree distance:

Iaddtree(C,N;h)=1#(tent)TtentlogLRaddtree(T;h), (3.57)

where the sum is over all tents T. By virtue of Propositions 2.5 and C.1, for h=0, this is the average log likelihood for pointwise compatibility with an addtree distance at each tent (h=0 is needed since Proposition C.1 only considers strict inequalities). Interpreted as a global measure of compatibility, the caveat mentioned for Isym applies here too, as setwise compatibility for the stimuli in every tent considered individually does not imply setwise compatibility on the entire stimulus space S. Here, though, there is an additional caveat: averaging the likelihood ratios in (3.57) is tantamount to assuming that each tent’s contribution to the log likelihood is independent. This is only an approximation since tents may have overlapping triads.

Finally, we calculate the benchmark value of Iaddtree based on the prior alone, for h=0. The two instances of Vsym in its denominator (eq. (3.54)) each contribute a factor of 34 (for each, six of 23 combinations of nonzero signs yield values of 1, as in the calculation of (3.25)). In the numerator, by direct enumeration, 24 of 26 nonzero sign combinations yield a value of Vaddtree=1. So

Iaddtree(0,0;0)=log24643434=log230.4055. (3.58)

4. Application to simulated datasets

Methods

To explore the utility of the indices Isym, Iumi, and Iaddtree we applied them to a range of simulated datasets. The main simulations considered a domain S with 15 stimuli, with distances assigned according to a range of geometries (Figure 4, top). In four of the geometries, stimuli were nodes in a four-level binary tree, and the geometries are distinguished by how the distances are calculated: TreeUM, an ultrametric space in which distances are given by height of the first common ancestor, TreeAdd, an addtree space in which distances are given by the graph distance, i.e., the number of links in the shortest path, TreeAddWt, an addtree space in which distances are given by a weighted graph distance, i.e., the sum of the lengths of the links in the shortest path, and TreeEucl, a non-addtree space in which distances are given by the Euclidean distance between the node as embedded in the plane. The fifth geometry, Line, is an addtree space in which the stimuli are arranged in a straight line and distances are given by the graph distance (which is equivalent to the Euclidean distance). The final geometry, Circle, is a non-addtree space in which the stimuli are arranged in a circle and distances are given by the graph distance. Within each geometry, distances were scaled to have a root-mean-squared value of 1.

Figure 4:

Figure 4:

Configurations used for generation of simulated datasets. Simulated dis-similarity judgments are determined by comparing the distances between two stimuli, followed by adding a Gaussian noise to simulate uncertainty. The distance types are: hierarchical, the height of the first common ancestor; graph, number of links in the shortest path; weighted graph: total length of links in the shortest path; and Euclidean, the Euclidean distance in the illustrated embedding. The final three columns indicate setwise compatibility based on the rank order of dis-similarities with a symmetric distance, an ultrametric distance, and an addtree distance. Distances within each configuration are normalized to have a root-mean-squared value of 1. The asterisk for the 150° configuration indicates that, while the dis-similarities are setwise compatible with an addtree distance, the distances themselves are not addtree.

For each geometry, we simulated an experiment in which all possible triads were presented an equal number of times N, for N{1,2,4,8,16,32}. The decision rule compared the distances and then added a random draw from a Gaussian of standard deviation σ for a range of values σ{0.0625,0.125,0.25,0.5,1,2}, to model uncertainty. (In typical experiments, N is in the range of 4–8, and a value of σ=0.25 yields a distribution of choice probabilities that is similar to those observed [41]).

These simulated responses were then used to compute the indices, using the beta-function prior (3.4) for Isym (eq. (3.24)) and Iaddtree (eq. (3.57) with h=0), and the modified prior (3.27) for Iumi, with h=0.001 to approximate the limit in eq. (3.42). For most calculations, the parameter a was determined by maximizing the log likelihood (eq. (3.8)) or its generalization for the modified prior (eq. (3.28)). In these cases, indices were only computed for N2. This is because for N=1, all empiric choice probabilities are 0 or 1 and the likelihood function is independent of a, so this parameter cannot be determined. Alternatively, to test sensitivity to the shape parameter a, we recalculated the indices for a range of fixed values for a (including the flat prior, a=0). To test sensitivity to the point mass parameter of the prior, we recalculated indices for a range of fixed values of h, and also with maximum likelihood fitted values for h, fitting a by maximum likelihood in all cases. When both a and h are determined by maximum likelihood, indices are only computed for N4.

Surrogates

To provide insight as to whether the indices are dominated by the structure of the data and not merely by the prior on choice probabilities, we take a heuristic approach based on parallel calculations with surrogate datasets. The logic is as follows. If the values of the indices are dominated by the prior, then surrogate datasets, drawn from the prior, would yield values of the indices similar to the values that were derived from the experimental data. To construct these surrogate datasets, we make use of the fact that the prior on choice probabilities is an independent one: the a priori likelihood of a set of choice probabilities is the product of the likelihoods assigned to each choice probability. Moreover, the prior for each choice probability R, either the beta-function prior (3.4) or its modification (3.27) that includes a discrete mass, is symmetric about an exchange of R for 1R. Thus, exchanging any of the observed choice probabilities R for 1R yields a set of choice probabilities which is a priori as likely as the experimental observations to result from the prior. The range of indices obtained from surrogates with random replacements of R for 1R is thus the range of indices expected when the relationships among the choice probabilities in the data are removed, but the choice probability distribution is retained. Conversely, values of the indices outside of this range indicate that the index values are dominated by relationships in the data that are not inherent in the prior.

We implemented this strategy with the “flip any” surrogate: the triads were independently selected with a probability of 12, and the responses to the chosen triads were replaced by their complement (so that C out of N responses that D(r,x)<D(r,y) were replaced by NC out of N responses that D(r,x)<D(r,y)). Note also that these surrogates match the original data in terms of whether “easy” R near 0 or 1) and “hard” R near 12) triads tend to co-occur in triplets, so this approach tests not only the effect of the distribution of choice probabilities, but also, the effect of the co-occurrence of easy and hard choice probabilities within triplets.

Since (as shown below) indices calculated from the data typically deviated substantially from the “flip any” surrogate, we then extended this logic to a second kind of surrogate, to test whether the index values could be accounted for by a more refined alteration of the choice probabilities. In this surrogate, called “flip all,” the choice probabilities have the same marginal distribution as the independent prior, but they are no longer independent. Instead, for the “flip all” surrogate, the choice probabilities within a triplet retain all of the pairwise correlation present in the data. To achieve this, we select triplets with a probability of 12, and, for each selected triplet, replace all of the responses of the three triads within the triplet by their complements. These surrogate sets of choice probabilities have the same likelihood as the data in the independent prior. They also match the pairwise correlations of choice probabilities within triplets, since the choice probabilities corresponding to two triads within a triplet are either both inverted, or both unchanged. However, this surrogate destroys any systematic relationship between the choice probabilities in different triplets, as well as any third-order correlations of the choice probabilities within triplets.

The range of indices obtained from the “flip all” surrogates is thus the range of indices expected when the pairwise correlations among the choice probabilities in a triplet are retained, as well as the choice probability distribution. For Iumi and Iaddtree, if the indices calculated from the data deviate from the range of values calculated from the surrogates, this indicates that the extent of ultrametricity or addtree structure in the data cannot be accounted for merely by pairwise correlations within triplets. (For Isym, the value calculated from the “flip all” surrogate necessarily matches the value calculated from the data, since the likelihood ratio (3.15) is unchanged.)

Since the indices are sums of values that are independently computed either from triplets or tents, the exact means and standard deviations of the surrogate indices could be computed efficiently by exhaustive sampling of each triplet or tent separately, rather than approximating them via a random sampling procedure.

Results

Identifying compatibility with underlying distances

The results for Isym (Figure 5) demonstrate that the approach efficiently distinguishes choice data that are compatible with symmetric distances from their “flip any” surrogates, which are not. For all geometries, as the number of observations for each triad N increases, Isym ascends from its a priori value (approximately −0.2877, eq. (3.25)), indicating compatibility. When judgments are reliable (σ0.5, first two columns of Figure 5 and first four columns of Supplementary Figure S1), Isym reaches its maximal value of 0 with only a few trials (N=2 or 4) per triad. For σ=1 (third column of Figure 5), the maximsal value is reached for N=8 or 16 trials per triad, and for highly unreliable judgments (σ=2, sixth column of Supplementary Figure S1), the maximal value is not reached even for N=32 judgments. The increasing values of Isym as calculated from the simulated data are paralleled by decreasing values of Isym as calculated from the “flip any” surrogate, since this manipulation destroys the necessary relationships within each triad (Isym calculated from the “flip all” surrogate is identical to the calculation from the original data because the requirements for symmetry in Proposition 2.1 are preserved when triads are replaced by their complements). These findings are all as expected and reassuring: the datasets are constructed from symmetric distances, and the ability to determine that the responses are compatible with symmetric distances increases with the number of observations N and decreases with judgment uncertainty σ.

Figure 5:

Figure 5:

Panel A. Behavior of Isym for the 15-point configurations, a range of trials per triad (N, abscissa), and a range of decision rules (σ, standard deviation of added noise). Solid symbols indicate values of Isym computed from simulated decisions. The dashed horizontal line indicates the a priori value for Isym. Hollow boxes indicate values of Isym computed from “flip any” surrogates, showing the ±1 standard deviation range. The thin vertical lines are to aid visualization, and do not represent ranges. The beta-function prior (3.4) is used, with values of the parameter a determined by maximum likelihood. Panel B expands the vertical scale. See Supplementary Figure S1 for additional values of σ and the values of the prior parameter a.

The results for Iumi (Figure 6) demonstrate that the approach is able to distinguish a dissimilarity structure that is compatible with ultrametric structure (TreeUM) from those that are not (all other geometries). Specifically, Iumi rises progressively from its a priori value of 0 for this geometry (first row of Figure 6), but falls for all of the others. Additionally, for the TreeUM geometry, analyses of both kinds of surrogates show a fall from the a priori value. The distinction between the TreeUM geometry and the surrogate datasets, or between the TreeUM geometry and the other geometries, is evident for N=4 trials per triad when judgments are reliable (σ0.5, first two columns of Figure 6 and first four columns of Supplementary Figure S2), but requires 8 or more trials per triad when they are unreliable (σ1, third column of Figure 6 and last two columns of Supplementary Figure S2).

Figure 6:

Figure 6:

Panel A. Behavior of Iumi for the 15-point configurations, a range of trials per triad (N, abscissa), and a range of decision rules (σ, standard deviation of added noise). Solid symbols indicate values of Iumi computed from simulated decisions. The dashed horizontal line indicates the a priori value for Iumi. Hollow boxes indicate values of Iumi computed from surrogates, showing the ±1 standard deviation range: wider boxes for “flip any” surrogates, narrower boxes for “flip all” surrogates. The thin vertical lines are to aid visualization, and do not represent ranges. The modified prior (3.27) is used with the value of the parameter a determined by maximum likelihood and h=0.001. Panel B expands the vertical scale. See Supplementary Figure S2 for additional values of σ and the values of the prior parameter a. Blue overlay indicates the simulated datasets that are incompatible with the ultrametric property.

The results for Iaddtree (Figure 7) demonstrate that the approach is able to distinguish dissimilarity data that are compatible with an addtree structure (TreeUM, TreeAdd, TreeAddWt, Line), from those that are not (TreeEucl, Circle). For the graph and weighted-graph geometries (TreeAdd, TreeAddWt) as well as the Line, Iaddtree ascends from its a priori (approximately −0.4055, eq. (3.58)) to a saturating value of 0, with similar dependence on the number (N) and reliability (σ) of judgments seen above. For the surrogates and for Circle, Iaddtree has the opposite behavior, descending from its a priori value as the amount of data increases. The behavior of Iaddtree for the other two non-addtree geometries also contrasts with addtree behavior, but also suggests some caveats. For the Euclidean distance on the tree (TreeEucl), values of Iaddtree below the a priori value – indicating incompatibility with an addtree distance – are only prominent when judgments are highly reliable σ0.25 and with a larger number of judgments N8. This is not surprising, since the Euclidean and weighted graph distances are often identical or similar.

Figure 7:

Figure 7:

Panel A. Behavior of Iaddtree for the 15-point configurations, a range of trials per triad (N, abscissa), and a range of decision rules (σ, standard deviation of added noise). Solid symbols indicate values of Iaddtree computed from simulated decisions. The dashed horizontal line indicates the a priori value for Iaddtree. Hollow boxes indicate values of Iaddtree computed from surrogates, showing the ±1 standard deviation range: wider boxes for “flip any” surrogates, narrower boxes for “flip all” surrogates. The thin vertical lines are to aid visualization, and do not represent ranges. The beta-function prior (3.4) is used, with the value of the parameter a determined by maximum likelihood. Panel B expands the vertical scale. See Supplementary Figure S3 for additional values of σ and the values of the prior parameter a. Blue overlay indicates the simulated datasets that are incompatible with the addtree property.

This point is explored in greater detail in Figure 8 and shows a limitation of Iaddtree. We consider configurations of a stimulus set consisting of 13 points, arranged along two rays with a common origin, as shown in Figure 4, bottom. (These specifics are motivated by the textures dataset below). The angle between the rays is set at 180°, 150°, 90°, and 30°, producing a series of simulations in which the distinction between the graph distance and the Euclidean distance is parametrically varied. For an angle of 180°, which is equivalent to a straight line of 13 points, the graph distance and the Euclidean distance are identical. The graph distance is independent of the angle, but the Euclidean distance departs progressively from the graph distance as the angle decreases. For an angle of 150°, the underlying choice probabilities for the Euclidean distance differs from that of the graph distance, but the rank orders of the distances are identical. All triadic choice probabilities based on a Euclidean distance are thus pointwise compatible with an addtree metric. In fact, they are setwise compatible: a monotonic transformation that rounds up each Euclidean distance to the nearest integer multiple of the nearest-neighbor distance yields the graph distance, which is addtree. In this scenario, the values of Iaddtree are nearly identical to those for the 180° configuration (they are not precisely identical, because the maximum likelihood beta-function prior for the two datasets are slightly different since the choice probabilities are not identical). For an angle of 90°, some of rank-orders of distances are different in the two geometries, and at this point, Iaddtree appears to show a slight difference in behavior for the Euclidean model at σ=0.25. However, this is more a manifestation of uncertainty in determining the choice probabilities, rather than evidence that the choice probabilities are incompatible with an addtree geometry – since Iaddtree never falls below the a priori value, and rises with increasing amounts of data N, or decreasing σ. In contrast, for an angle of 30°, where there are major differences between rank orders of graph and Euclidean distances, Iaddtree clearly identifies the incompatibility with an addtree geometry, as its value remains below 0.2 and, with sufficient data of sufficient quality (e.g., N=8 with σ=0.0625 in Figure 8, or N=8 with σ=0.125 in Supplementary Figure S4), it falls below the a priori value. In sum, these examples underscore the fact that Iaddtree is only sensitive to rank orders of dis-similarities, not to metrical structure, and show that modest deviations of rank orderings from compatibility with addtree distances may not be detectable.

Figure 8:

Figure 8:

Behavior of Iaddtree for the 13-point configurations, a range of trials per triad (abscissa), and a range of decision rules (σ, standard deviation of added noise). Graphical conventions as in Figure 7. See Supplementary Figure S4 for additional values of σ and the values of the prior parameter a. Blue overlay indicates the simulated datasets that are incompatible with the addtree property.

The second notable point in Figure 7 is that for the TreeUM, values of Iaddtree remain close to the a priori value across the entire range of N and σ, rather than rise, even though TreeUM has an ultrametric geometry. This might appear surprising, since the ultrametric geometry is a special case of an addtree geometry, and thus TreeUM is compatible with an addtree distance. The reason for this behavior is a severe mismatch between the best-fitting prior and the choice probability distribution. That is, the simulated data had some choice probabilities are exactly 12, while the analysis used the unmodified beta-function prior (eq. (3.4)), which does not have a point mass at 12. Consequently, the prior often forced the inferred choice probabilities to one side or the other of 12 with little evidence, and thus the pattern of a posteriori probabilities is spuriously close to the pattern of a priori probabilities. This scenario is unlikely to arise with real data (where underlying choice probabilities are not likely to be exactly 12). More importantly, this potential confound can be simply addressed by using the modified prior with a point mass at 12 (eq. (3.27)), as we show below.

Considerations for experimental design

An experimenter is often faced with the question of the best way to deploy a limited total number of observations: whether to sample a large number of triads with few repeats, or alternatively to concentrate on a small number of triads with proportionately more repeats. To provide some guidance on this point, Figure 9 replots the indices from Figure 5 to Figure 7 after dividing their departure from the a priori value by the number of trials in which each triad is sampled, N.

Figure 9:

Figure 9:

Normalized values of Isym (panel A), Iumi (panel B), and Iaddtree (panel C), for the 15-point configurations, a range of trials per triad (N, abscissa), and a range of decision rules (σ, standard deviation of added noise). Data from Figure 5, Figure 6, and Figure 7, replotted after normalizing the deviation from the a priori value by the number of trials per triad: I0+II0N, where I is the index from Figures 5-7, I0 is the a priori value, and N is the number of trials per triad. Blue overlay indicates the simulated datasets that are incompatible with the ultrametric property (B) or the addtree property (C). Other graphical conventions as in Figure 5, Figure 6, and Figure 7. See Supplementary Figure S5 for additional values of σ and results for the TreeAddWt and Line geometries.

The values in Figure 9 can be interpreted as the expected contribution to the log likelihood ratio on a per-observation basis. For the Isym and Iaddtree examples shown, the greatest departures occur for N=2. That is, once a set of triads have been sampled twice, it is more informative to observe additional triads, than to observe the same triads more extensively. This optimal number of repeated observations, however, may well be larger if the incompatibility with an addtree geometry is more subtle, as in Figure 8. It may therefore be advisable to carry out targeted simulations if there is a particular model to be excluded.

For Iumi (Figure 9B), the normalized upwards departure from a priori values peaks for N=4 or N=8 for the compatible geometry TreeUM, and the normalized downwards departure for the incompatible geometries does not appear to saturate. Thus, for a focus on Iumi, a more intensive sampling of a limited number of triads is expected to be more informative.

Dependence on the prior

Here we explore the influence of the prior for the distribution of choice probabilities. The prior plays a key role in this approach in that the likelihoods of the underlying choice probabilities are estimated by combining the observed responses with a prior in a Bayesian fashion. As we will show, while the prior is necessary to carry out this estimation, it typically has only a minor influence on the conclusions.

Up to this point, we used the prior specified in the Implementation section: the beta-function prior (3.4) for Isym and Iaddtree, in which the shape parameter a was determined by maximum-likelihood, and the modified prior (3.27) for Iumi, with a point mass of weight h=0.001 to approximate the limit of eq. (3.41). Here we consider instead specific choices of the shape parameter a (Figure 10) and alternative choices for the point mass h, including fitting h by maximum likelihood (Figure 11). Note that these choices provided distributions that cover a wide range of shapes: for h=0, these choices specify priors that range from U-shaped (a<1, typical of experimental data, where most judgments are clearcut) to an inverted Ua>1, and that h>0 adds a point mass at 12, leading to trimodal distributions for a<1. The flat prior corresponds to a=1 and h=0. We consider four representative geometries: TreeUM (ultrametric), TreeAdd, (addtree), TreeEucl (non-addtree but only modestly incompatible), and Circle (non-addtree but highly incompatible), and the decision rule specified by σ=0.25.

Figure 10:

Figure 10:

Dependence on the prior’s shape parameter, a, for Isym (panel A), Iumi (panel B), and Iaddtree (panel C), for the 15-point configurations and a range of trials per triad (N, abscissa). The decision rule σ=0.25 is used. Column 1: a is determined by maximum likelihood; other columns: a is assigned the value indicated over each column. The beta-function prior (3.4) is used for Isym and Iaddtree; the modified prior (3.27) is used for Iumi with h=0.001. Blue overlay indicates the simulated datasets that are incompatible with the ultrametric property (B) or the addtree property (C). Other graphical conventions as in Figure 5, Figure 6, and Figure 7. See Supplementary Figure S6 for additional fixed values of a and values of the prior parameter a determined by maximum likelihood.

Figure 11:

Figure 11:

Dependence on the prior’s point mass parameter, h, for Isym (panel A), Iumi (panel B), and Iaddtree (panel C), for the 15-point configurations and a range of trials per triad (N, abscissa). The decision rule σ=0.25 is used. Column 1: h is determined by maximum likelihood; other columns: h is assigned the value indicated over each column. The parameter a is determined by maximum likelihood in all cases. The modified prior (3.27) is used for all indices. Blue overlay indicates the simulated datasets that are incompatible with the ultrametric property (B) or the addtree property (C). Other graphical conventions as in Figure 5, Figure 6, and Figure 7. See Supplementary Figure S7 for additional fixed values of h and the values of the prior parameters a and h determined by maximum likelihood.

Figure 10 shows the behavior of the indices when a is set to a fixed value rather than determined by maximum likelihood. Supplementary Figure S6 shows their behavior for additional values of a and also gives the values of a determined by the maximum-likelihood fit. As expected, if a is set to a value that is far from the maximum likelihood value, the indices require more data to depart from their a priori value. This effect, however, is modest: for a flat prior a=1 and even for a poorly-chosen prior a=2, recovery of the behavior seen for the fitted prior is typically seen with N=4 or N=8 trials per triad.

Figure 11 shows a parallel analysis of alternative choices for the point mass h. Supplementary Figure S7 shows their behavior for additional values of h and also gives the values of a and h determined by the maximum-likelihood fit. For Isym and Iaddtree, the influence of this choice is again modest. The main finding is that the inability of Iaddtree to detect compatibility of the strictly ultrametric dataset TreeUM with addtree geometry when using a beta-function prior (Figure 7, top row) is remedied by allowing h>0 (Figure 11C, top row) as this prior allows for capturing the large fraction of underlying choice probabilities that are exactly 12. For Iumi, the dependence on the prior is somewhat complex. For the non-ultrametric geometries, there is a modest dependence on the prior: Iumi descends similarly from its a priori value with increasing amounts of data. But for TreeUM, there is a sharper ascent of Iumi when h is chosen to be small (0.001 or 0.01), than when h is close to its fitted value of ∼ 0.4. This is due to the fact that Iumi is defined by its limiting behavior when h approaches zero. This limiting behavior is accurately estimated with h=0.001 or 0.01, but not larger values.

5. Application to sample experimental datasets

Here we demonstrate the utility of the present approach via application to sample datasets from three psychophysical experiments, encompassing two methods for acquiring similarity judgments and spanning low- and high-level visual domains.

Methods

The first two experiments (“textures” and “faces”) make use of the method of Waraich et al. [41]: on each trial, participants rank the eight comparison stimuli c1,,c8, in order of similarity to a common reference r. These rank-orderings are then interpreted as a set of similarity judgments: ranking ci as more similar than cj to the reference r is interpreted as a triadic judgment that Dcj,r>Dci,r. Data are accumulated across all trials in which ci and cj are presented along with the reference r, leading to an estimate of Robsr;ci,cj. Stimulus sets consisted of 24 or 25 items (described in detail with Results below), and 10 sessions of 100 trials each are presented. On each trial, stimuli are randomly chosen to be the reference or the comparison stimuli. As there were 10 sessions of 100 self-paced trials each and each trial yielded 82=28 triadic judgments, each participant’s dataset contained 28,000 triadic judgments. This corresponds to an average of 28000252426.08 trials per triad for the 25-item sets and 28000242326.92 trials per triad for the 24-item sets.

For the “texture” and “faces” datasets (described in detail below), stimuli were generated in MATLAB, and were displayed and sequenced using open-source PsychoPy software on a 22-inch LCD screen (Dell P2210, resolution 1680×1050, color profile D65, mean luminance 52 cd/m2). The display was viewed binocularly from a distance of 57 cm. The visual angle of the stimulus array was 24 degrees; each stimulus (a texture patch or a face) subtended 4 degrees. Tallying of responses and multidimensional scaling as described in [41] were carried out via Python scripts. Computation of the indices and visualization was carried out in MATLAB using code that is publicly available at https://github.com/jvlab/simrank.

The third experiment (“brightness”) uses an odd-one-out paradigm. On each trial, three stimuli are presented, each consisting of a central disk drawn from one of eight luminances, and an annular surround. The surround was either of minimal or maximal luminance, and was perceived as black or white, respectively. The participant is asked to judge the brightness of the central disk, and to choose which of the three is the outlier. We interpret selection of a stimulus xj out of a triplet xj,xk,xl as a judgment that the pairwise dis-similarities involving this stimulus are larger than the dis-similarity of the two non-outliers, i.e., that Dxk,xj>Dxk,xl and also that Dxl,xj>Dxl,xk. Each trial thus contributes to estimates of choice probabilities for two triads, xk;xj,xl and xl;xj,xk, and these judgments are tallied across the experiment. Note though that, in contrast to the “textures” and “faces” datasets, here the specific triadic comparisons that enter into the tallies depend on the participant’s responses.

For the “brightness” dataset, stimuli were generated in Python 3.10 and the NumPy library. Stimuli were displayed on a calibrated 24-inch ViewPixx monitor (1920×1080 pixel resolution, mean luminance 70 cd/m2, Vpixx Technologies, Inc.), running custom Python libraries that handle high bit-depth grayscale images (https://github.com/computational-psychology/hrl). Monitor calibration was accomplished using a Minolta LS-100 photometer (Konica Minolta, Tokyo, Japan). The display was viewed binocularly from a distance of 76 cm. The visual angle of the display was 39 degrees; each stimulus subtended 5 degrees, with the central disk subtending 1.67 deg. The three stimuli were arranged in a triangular manner, 4 degrees equidistant from the center (Fig. 15A). There were 16 unique stimuli, consisting of all pairings of 8 values for the luminance of the center disk (14, 33, 55, 78, 104, 131, 163 and 197 cd/m2) and 2 values of luminance for the surrounding annulus (0.77 and 226 cd/m2). The 16 stimuli generated 163=560 possible triplet combinations, which were presented in randomized order and position, constituting one block. Each session consisted of two blocks, and each participant ran four sessions. In total, we collected 4480 trials per participant. As each trial gives information for two triadic judgments (as mentioned above), there were 8960 triadic judgments per participant. The 560 triplets contain 3 triads each, so there were an average of 89603.560=5.33 trials per triad.

Figure 15:

Figure 15:

Panel A. Stimuli for the brightness experiment. Each stimulus had a disk-and-annulus configuration, in which the disk had one of 8 luminances (columns) and either a black (upper row) or a white (lower row) surround annulus. The colored lines encircle three of the stimulus subsets used in Panel C. Panel B. A sample trial. Panel C. Indices Isym, Iumi, and Iaddtree for the brightness experiment calculated from all trials (black symbols), from trials with 8-element subsets with only one of the two kinds of surround (blue symbols), and from 8-element subsets with both surrounds (green symbols). Other graphical conventions as in Figure 13.

The texture and faces experiments were performed at Weill Cornell Medical College, in four participants (3F, 1M), ranging in age from 23 to 62. Participants MC and SAW (an author) were experienced observers and familiar with the “texture” stimulus set from previous studies; participants BL and ZK were naïve observers. All participated in the textures experiment; 2F (SAW and MC) participated in the faces experiment and neither had prior familiarity with those stimuli. The brightness experiment was performed at Technische Universität Berlin in three participants (1F, 2M), ranging in age from 31 to 39. Participant JP was a naïve observer; participants GA (an author) and JXV were experienced observers. All participants had normal or corrected-to-normal vision. They provided informed consent following institutional guidelines and the Declaration of Helsinki, according to a protocol that was approved by the relevant institution.

In addition to the calculations described above, we also calculated the indices Isym, Iumi, and Iaddtree for surrogate datasets, as detailed in the Simulated Datasets section. Briefly, the “flip any” surrogate was created by randomly selecting individual triads, and flipping the choice probabilities for the selected triads. The “flip all” surrogate was created by randomly selecting triplets and replacing Robs(r;x,y) by 1Robs(r;x,y) in the selected triplets.

Finally, we estimated the standard errors for the indices calculated from the original datasets via a jackknife on triplets (for Isym, Iumi) or tents (for Iaddtree). Maximum likelihood parameters a and h were not re-calculated for the jackknife subsets, as pilot analyses confirmed that removal of one triplet or tent made very little change in the maximum likelihood value.

Results

Textures

The textures experiment made use of the stimulus space described in [39], a 10-dimensional space of binary textures with well-characterized discrimination thresholds [38]. We chose a two-parameter component of this domain (Figure 12A) that allowed a focus on testing for compatibility for addtree structure. The two parameters chosen, β and β|, determine the nearest-neighbor correlations in the horizontal or vertical direction: the probability that a pair of adjacent checks have the same luminance (either both black or both white) is (1+β)/2, and the probability that a pair of adjacent checks have the opposite luminance (one black, the other white) is (1β)/2. Other than these constraints, the textures are maximum-entropy (see [39] for details). For these experiments, we chose values of β or β| from −0.9 to 0.9 in steps of 0.15. That is, six stimuli had positive values of β(0.15,0.30,0.45,0.60,0.75,0.90) with β=0, six had the corresponding negative values of β, six had positive values of β with β=0, six had negative values of β with β=0, and one had β=β=0. In the experiment, each stimulus example was unique – that is, a stimulus is specified by a particular β,β pair, but the texture example used on each trial was a different random sample from that texture.

Figure 12:

Figure 12:

Panel A. Stimuli used in the texture experiment. Each stimulus is an array of 16×16 black or white checks. For stimuli enclosed in dark blue, checks are correlated (or anticorrelated) along rows. Correlation strength is parameterized by β, where β>0 indicates positive correlation and β<0 indicates negative correlation. For stimuli enclosed in light blue, checks are correlated (or anticorrelated) along columns, similarly parameterized by β. The full stimulus set consists of 6 equally-spaced values positive and negative values of β and β, and an uncorrelated stimulus (center), where β=β=0. Panel B. Multidimensional scaling of similarity judgments for the stimuli in panel A for four participants. The data from each participant have been rotated into a consensus alignment via the Procrustes procedure (without rescaling). Lines connect stimuli along each of the rays in Panel A. One unit indicates one just-noticeable difference in an additive noise model [41].

The rationale for this stimulus set is that we anticipated that certain subsets of stimuli would be more compatible with the addtree model than others. The basis for these expectations is shown in Figure 12B, which presents non-metric multidimensional scaling of the similarity data. This analysis, carried out with the procedure detailed in [41], uses a maximum likelihood criterion to place the 25 stimulus points in a space, so that the Euclidean distances between them best account for the choice probabilities (assuming a uniform, additive decision noise). Consistently across participants, the points along each stimulus axis (β or β) map to a gradually curving trajectory. For this reason, we anticipate that comparison data from the stimuli on one of these trajectories (the 13 points with either β or β equal to zero, here called an “axis”) when analyzed in isolation, will be close to an addtree model. However, the two trajectories are not perpendicular: rays with same signs of β meet at an acute angle of 45° or less. That is, stimuli with strong positive correlations (β>0 compared to β>0) are seen as relatively similar to each other. This is anticipated to make the subset consisting of the 13 points with either β or β positive (a “vee”) incompatible with the addtree model, as the shortest perceptual path between two points at the end of the positive β or β rays is much shorter than a path that traverses each ray back to the origin. Similar reasoning indicates that the “vee” formed by the two negative rays should also be incompatible with an addtree model. Note, though, that this intuition assumes that the Euclidean distances in Figure 12B are an accurate account of the perceptual dis-similarities; the analysis via Iaddtree does not make this assumption.

Figure 13 shows the indices Isym, Iumi, and Iaddtree computed from the full datasets for each participant, and for the “axis” and “vee” subsets. As expected from the above analysis, the addtree index Iaddtree is substantially higher for the “axis” subsets than for the “vee” subsets – comparable to the difference between the values of Iaddtree for the simulated 180° and 30° datasets in Figure 8 – and the values for the “axis” and “vee” subsets straddle the value for the full dataset. Note that “axis” and “vee” subsets are equated in terms of the number of stimuli, and were collected simultaneously within a single experiment. This finding supports the efficacy of Iaddtree in determining compatibility with addtree distances: it is close to zero for data along an axis, which is anticipated to be compatible with an addtree distance, and decreases reproducibly for subsets that form an acute “vee”, for which the rank orders of similarity expected to be incompatible with an addtree distance. Note also that in all cases, it is higher than the a priori value, and substantially higher than values computed from surrogate datasets in which choice probabilities are randomly flipped. This latter point indicates (not surprisingly) that for the full dataset and the selected subsets, there are portions of the data that are more compatible with an addtree model than datasets with the same choice probabilities, but no relationship between the triads.

Figure 13:

Figure 13:

Indices Isym, Iumi, and Iaddtree for the texture experiment. Stimulus subsets are indicated by symbol color; participants by symbol shape. The vertical extent of the wide boxes indicate ±1 standard deviation for the “flip any” surrogates; the vertical extent of the narrow boxes indicate ±1 standard deviation for the “flip all” surrogates (not plotted for Iumi, as this index is unchanged by the “flip all” operation). The thin vertical lines are to aid visualization, and do not represent ranges. Standard errors for the experimental datasets are smaller than the symbol sizes. Null-hypothesis values of the indices are indicated by the horizontal dashed line.

For this dataset, values of Isym were quite close to zero (usually > −0.1), indicating that nearly all (>e0.1) of the posterior distribution of choice probabilities was compatible with a symmetric dis-similarity. Iumi, which measures compatibility with the ultrametric model, was typically −0.25 or less, substantially below the a priori value of zero. But interestingly, the highest values of Iumi were seen in the “vee” subsets, suggesting a partially hierarchical structure – e.g., that the two directions of correlation formed categories. As was the case for Iaddtree, all indices were higher than for the surrogates. For Isym, this is unsurprising, as randomly flipping choice probabilities would be unlikely to lead to a set of symmetric judgments. For Iumi, this finding indicates that, even though the ultrametric model is excluded, the data has islands of relatively greater compatibility with the ultrametric structure.

The above results were insensitive to the parameters a and h of the prior for the distribution of choice probabilities (eqs. (3.4) and (3.27)). The beta-function parameter a obtained by maximum likelihood (eq. (3.8)) ranged from 0.25 to 1.25 (with the lowest values for the full texture dataset), but very similar results to Figure 13 were obtained with a=0.5 for all datasets. For Iumi, the limit in (3.41) was estimated by setting h=0.001 but similar values were obtained for h=0.01. The findings for Isym and Iaddtree, here shown for h=0, were not substantially changed when h was determined by maximum likelihood. These values of h were typically quite small (median, 7 · 10−5).

Faces

The faces experiment used stimuli drawn from the public-domain library of faces, at https://faces.mpdl.mpg.de/imeji/, which contained color photographs of 171 individuals, stratified in three age ranges (“young”, “middle”, “old”). We randomly selected two males and two females from each age range, and for each individual in the faces database, used the two example photographs with neutral expressions, for a total of 24 unique images (2 genders × 3 age ranges × 2 individuals × 2 photographs of each).

The rationale for this choice of stimuli was that the above hierarchical organization might lead to a similarity structure close to ultrametric behavior. As shown in Figure 14, upper row, while this was not the case for analysis of the full dataset (Iumi<0 the a priori level), it was the case for the 8-stimulus subsets within each age bracket Iumi>0. Values of Iumi>0 were also seen in data from some observers, for some subsets subdivided by gender (restricted to two age ranges, to equate the number of stimuli), as shown in Figure 14, lower row. Values of Isym were again quite close to zero (usually > −0.1), indicating compatibility with a symmetric dis-similarity. Values of Iaddtree were similar to the a priori value, but much larger than for the surrogates. As was the case for the texture experiment, these results were insensitive to the parameters a and h of the prior for the distribution of choice probabilities. Here, values of the beta-function parameter a obtained by maximum likelihood ranged from approximately 0.1 to 0.5; results similar to those of Figure 14 were obtained with setting a=0.3 for all datasets. Also as was the case for the texture experiment, findings for Isym and Iaddtree, were not substantially changed when h was determined by maximum likelihood – even though the typical values of h were larger (median, 6 · 10−2), supporting the idea that some underlying choice probabilities were close to 0.5.

Figure 14:

Figure 14:

Indices Isym, Iumi, and Iaddtree for the faces experiment. Stimulus subsets are indicated by symbol color; participants by symbol shape. Upper row: full stimulus set (black symbols), and subsets partitioned by age. Lower row: full stimulus set (black symbols, repeated), and subsets partitioned by gender, with two age ranges each. Other graphical conventions as in Figure 13.

Brightness

The brightness experiment consisted of judgments of brightness dis-similarity for the set of disk-and-annulus stimuli as shown in Figure 15A. This disk-and-annulus configuration has been extensively used to study the effect of the context surround on the appearance of the inner disk [10, 40]. A light surround is expected to make the inner disk appear darker, and conversely, darker surround is expected to make the inner disk appear lighter. While it is generally assumed that this shift in appearance is along a one-dimensional brightness continuum, the evidence is ambiguous [23]. For example, Madigan and Brainard [21] found that one dimension suffices to explain brightness similarity ratings, while Logvinenko & Maloney [20] found that dis-similarity ratings under different illuminations required a 2-dimensional perceptual space.

This open question motivated the stimuli used in the present experiment: a gamut of 8 disk luminances, presented with either of 2 surround contexts (Figure 15A). Participants judged the brightness of the inner disk for triplets constructed from all possible combinations of disk luminance and surround (Figure 15B). If brightness is one dimensional, then dis-similarity judgments for the full set of stimuli should be described by a one-dimensional model, which is a special case of an addtree model (Appendix B). If, on the other hand, the surround produces differences in appearance that are not one dimensional, then the full set of judgments should be incompatible with an addtree distance. Under this hypothesis, restricting the judgments to stimuli with the same surround (the subsets encircled by the dark and light blue lines in Figure 15A) should recover a one-dimensional structure and compatibility with an addtree model, while restriction to a same-sized set but with two kinds of surrounds (green lines in Figure 15A) should preserve the incompatibility.

Figure 15C shows the results. For the full stimulus set (black symbols), Iaddtree is close to zero (> −0.17), and substantially higher than the a priori value, for all three participants. Even higher indices are found for the 8-element stimulus subsets of only black (Iaddtree>0.09) or only white Iaddtree>0.05 surrounds (blue symbols in Figure 15C). This is consistent with the notion that, when context (i.e., the surround) is held constant, dis-similarity judgments correspond to distances in a one-dimensional space. However, when Iaddtree was computed for 8-element subsets of the stimuli in which judgments were made across two surrounds (green symbols in Figure 15C), Iaddtree was lower, and varied substantially across participants. Data from GA always yielded the lowest value (Iaddtree was −0.26 to −0.17) and JP the highest value close to zero (Iaddtree was −0.08 to −0.03).

These findings show that when the surround context is constant, judgments are compatible with an addtree model, but there is inter-observer variability when judgments are made across two surround contexts. The variability is not surprising, as previous research has shown that individual idiosyncrasies can play a substantial role when disk-in-context stimuli are used to study brightness or color [24]. Our method seems to be capturing these inter-individual differences, but – as we are focusing on a demonstration of the analysis methods – we do not attempt to probe the basis for this difference here.

Similar to the texture and faces experiments, values of Isym are all close to zero for the brightness dataset, indicating compatibility with symmetric dis-similarity judgments. Ultrametric indices Iumi are below the a priori value for all cases, indicating incompatibility with an ultrametric model, as expected for a one-dimensional geometry.

Also as in the texture and faces experiments, results were robust to changes of analysis details. Values of the beta-function parameter a obtained by maximum likelihood ranged from approximately 0.07 to 0.22; results similar to those of Figure 15C were obtained with setting a=0.1 for all datasets. Findings for Isym and Iaddtree were not substantially changed when h was determined by maximum likelihood, yielding values of h with a median of 5 · 10−2, comparable to the faces dataset.

6. Discussion

The main goal of this paper is to advance a strategy for connecting similarity judgments of a collection of stimuli to inferences about the structure of the domain from which the stimuli are drawn. The starting point is an experimental dataset in which the judgments are assumed to be independently drawn binary choices distributed according to the underlying choice probabilities. We assume that for each triad (a reference stimulus and two comparison stimuli), the comparison stimulus that is more often judged to be more similar to the reference is at a shorter distance from it, but we do not assume, or attempt to infer, a quantitative relationship between choice probabilities and the distances. This approach also takes into account the possibility that each triad may have its own “pointwise” transformation that links choice probability and distance. While we recognize that judgments may be uncertain, we refrain from postulating a noise model or a decision model – or even that sensory or decision noise is uniform throughout the space.

Despite the relative paucity of assumptions, we show that it is possible to characterize dis-similarity judgments along three lines: compatibility with symmetry, compatibility with an ultrametric model, and compatibility with an additive tree model. These characteristics are functionally significant aspects of a domain’s organization. Symmetry (or its absence) has implications for the mechanism(s) by which comparisons are made [34, 35]. For symmetric similarity judgments, addtree models, but not ultrametric models, are consistent with the Tversky contrast model [26]. More broadly, semantic domains are anticipated to be consistent with a hierarchical model of similarity judgments (ultrametric or addtree), while domains of features are not [15, 27]. It is also worth noting that one-dimensional domains are a special case of the addtree model, so the present approach can address whether the apparent “curvature” in a one-dimensional perceptual space can be eliminated by alternative choices of the linkage between distance and decision – a limitation of the analysis in [3]. Furthermore, our method is sensitive enough to reveal inter-individual differences: for some participants data are compatible with the addtree model and for others, not (or less so) – consistent with other studies of the influence of context [24], and an interesting area for further investigation.

Comparisons to other methods

The present strategy, in which a main consideration is to keep assumptions about metrical distances to a minimum, is complementary to other ways of analyzing similarity judgments. A classical and commonly-used approach, non-metric multidimensional scaling [6, 33], explicitly postulates that the original data (here, the choice probabilities) reflect a monotonic transformation of a metric distance. The distance is taken to be the Euclidean distance, but distances in a hyperbolic or spherical geometry can also be used. An important related approach for one-dimensional models is maximum likelihood difference scaling [17, 22], which – via a decision model – takes into account the noisy nature of psychophysical judgments. This approach can also be extended to multidimensional models, but the need for a decision model remains [37, 41].

The spirit of our approach is similar in some ways to that of [36], who used analyses of the statistics of nearest neighbors (“centrality” and “reciprocity”) to assess models of perceptual spaces. Their study showed that tree-like models are preferred to low-dimensional metrical models for “cognitive” categories (furniture, musical instruments), but not for low-level domains (lights, shapes, tones). As in the current approach their analyses depended only on rank orders of distances, rather than metrical relationships. However, the starting point for their analysis is a complete set of pairwise distances with a definite rank order. In contrast, the present approach begins with a set of triadic judgments and is explicit that these serve only as estimates of an underlying choice probability.

Our approach also is related to that of topological data analysis (TDA) via persistent homologies [5, 11, 31, 47]. Like our approach, TDA avoids the need to postulate a specific relationship between dis-similarities and distances, as the Betti numbers are calculated from a sequence of graphs that depend only on the rank order of similarity judgments. However, construction of this sequence of graphs requires a globally uniform linkage between triadic judgments and relative distance, and also, that every pairwise distance is included in the measured triads. The characterizations yielded by TDA are also complementary: they focus on dimensionality and homology class, rather than the characterizations considered here.

Finally, we note another approach that directly seeks to identify features of ultrametric behavior in neural data. Treves [32] developed an information-theoretic measure of dis-similarity of neural responses to faces. The strategy for seeking evidence of ultrametric behavior was to examine the ratio, within each triplet, of the middle distance to the largest. This ratio, which would be 1 for an ultrametric, was found in that study to be closer to 1 than for expected by chance. Nonparametric generalizations of this approach may permit a relaxation of the assumed linkage between extensions the information-theoretic measure and dis-similarity, and even an evaluation of addtree models – but in contrast to our approach, it begins with a set of responses to each stimulus, rather than a sampling of triadic comparisons.

Experimental design

The analysis of simulated and experimental datasets carry implications for experimental design. From the point of view of demonstrating departure of the indices from their a priori value, the simulated datasets show that as few as 2 observations per triad suffice under some circumstances, and may even be optimal (Figure 9). However, this optimal number of samples per trial is likely to be dependent on how severely the properties tested by each index are violated and the actual geometry of the underlying space: for subtle departures (e.g., Figure 8), a larger number of observations (8 or more) may be needed. Note, however, that observations contribute to the indices only if the triads are in the same triplet or tent, a consideration that may have bearing on the way that triads are selected for presentation in an experiment.

The experimental datasets show that analyses of subsets of the full dataset can be revealing: in the texture dataset (Figure 13), the responses to the subset of stimuli contained in the “vees” are less compatible with an addtree distance than for subsets contained in the “axes”; in the faces dataset (Figure 14), data from images restricted to one gender and including the three age ranges were less compatible with an ultrametric distance than data restricted to one age range but including both genders; for the brightness dataset (Figure 15), subsets that included both luminance contexts were less compatible with an addtree distance than subsets that included only one luminance context. These observations suggest that the current approach can be incorporated into an adaptive design strategy, in which calculation of indices within subsets can be used to select stimuli or triads for more intensive study. This is a potential area for future work.

Limitations and caveats

Keeping assumptions to a minimum necessarily leads to certain limitations. The indices for symmetry and addtree structure reflect necessary, but not sufficient, conditions for global compatibility. Moreover, though the indices can be used to compare the behavior of the data with behavior of a model, they are not measures of goodness of fit: the indices merely measure to what extent the data act to concentrate the a priori distribution of choice probabilities into the subset of choice probabilities that have a particular characteristic. Critically, the extent of concentration will depend on the typical coverage of each triad: a greater number of trials of each triad provides better estimates of the underlying choice probabilities, and thus, can move the indices further from their a priori values. For this reason, the examples above (“axes” vs. “vees” in Figure 13; subdivision by age vs. gender in Figure 14, same context vs. different contexts in Figure 15C) focus on comparisons of indices between datasets with a similar number of stimuli, and a similar coverage of each.

Because the approach relies on rank-order comparisons and does not attempt to estimate perceptual distances in a metrical sense, it cannot distinguish an addtree geometry from a Euclidean one unless the geometries lead to differences in the rank orders of distances. Thus, while the incompatibility of a circular stimulus space with an addtree geometry is readily detected by Iaddtree (Figure 7, bottom row), the Euclidean distance in a two-ray stimulus space (Figure 8) is only distinguishable from an addtree geometry when the angle between the rays is acute.

A necessary component of putting this approach into practice is the need to infer underlying choice probabilities from a finite set of experimental observations, and this step – even if principled – entails caveats. Here, we use a Bayesian framework and a two-parameter family of priors: the beta-function prior of eq. (3.4), which suffices for Isym and Iaddtree, and its generalization that includes point mass, eq. (3.27), which is needed for Iumi. While this choice of prior has some theoretical justification, it is primarily based on heuristics and computational advantages. For the model scenarios considered here, simulations (Figure 10) show that within this family (which includes concave, convex, and flat distributions) the influence of the prior is mild, especially when there are several observations per triad. The need to consider priors with a point mass at 12 for Iumi arises because a strictly ultrametric geometry requires that some choice probabilities are exactly 12; had we only considered the beta-function prior, this has a posterior probability of zero, for any finite dataset. So, in order to formalize a useful sense in which the data have an ultrametric character, we ask how rapidly this probability moves away from zero as we allow the prior to include a fraction h of choice probabilities that are exactly 12. As we show in the simulations (Figure 6) and also in the analysis of the faces dataset (Figure 14), this is an effective strategy – but there may be other approaches more suitable in other specific circumstances.

The analysis of the synthetic ultrametric dataset TreeUM reveals one situation in which the choice of the prior is important. In this dataset, the underlying choice probabilities have a substantial point mass at 12. Excluding this possibility from the prior (Figure 7, top row) fails to reveal compatibility with addtree structure, while including it (Figure 11C, top row) demonstrates compatibility. With this in mind, it may be prudent to analyze experimental datasets both with assuming the beta-function prior (h=0) and allowing h to be set by maximum likelihood. In the experimental datasets presented here, this did not result in a substantial change.

Our approach also provides two strategies to help ensure that the indices are dominated by the data, rather than by the prior. First, each index has a definite a priori value, independent of the prior for the choice probabilities, provided only that it is symmetrical: log(3/4) (eq. (3.25)) for Isym, 0 (eq. (3.46)) for Iumi, and log(2/3) (eq. (3.58)) for Iaddtree. Second, the two kinds of surrogates (“flip any” and “flip all”) have the same a priori probability as the data, though they lack the inter-relationships among choice probabilities within triplets or tents. Computed values of the indices that do not deviate substantially from the a priori values, or do not deviate substantially from indices computed from the surrogates, should be viewed with caution, as they suggest that the indices are dominated by the prior, rather than by the data. The “flip all” surrogate provides a stronger test, since it retains not only the a priori choice probabilities, but also within-triplet pairwise correlations.

Variations, extensions and open issues

There are a number of variations and alternative analytic strategies that might be used to modify and enhance the present framework, as well as some open issues that would enable it to be further extended.

One such variation relates to how the locally-computed likelihood ratios are combined to form an overall characterization of the dataset. Here, we elected to simply average their logs, i.e., to compute an index that can be interpreted as an average measure of local compatibility. Instead, one could look at other aspects of the distribution of these ratios. For example, one could also transform the likelihood ratio computed for each triplet or triad into a Bayes factor, via BF=LR1LR (see comment following eq. (3.23)), and then combine the Bayes factors – yielding a Bayes factor that compares the hypotheses that all triplets are compatible, vs. no triplets are compatible.

One might also consider circumventing the difficulties related to underlying choice probabilities that are exactly 12 by allowing subjects to respond that the two comparison stimuli are equally dis-similar. At first glance, this sounds promising, but the problem is that, by allowing subjects to be explicitly indecisive, it (i) reduces the extent to which their responses will reveal subtle but systematic differences in perceived dis-similarity, and (ii) introduces a major confound, namely, that subjects may differ substantially in their criterion for using this alternative. The situation worsens if subjects change their criterion during the experiment, which introduces yet another uncontrolled confound. So although some psychophysicists have studied the usefulness of allowing an “uncertain” response (e.g. [42]), it is not recommended (e.g. [16, 43]), and consequently, it is rarely used in psychophysical studies. Accordingly, our approach and our sample datasets all used forced-choice paradigms without the possibility of an “uncertain” response.

There are also alternative strategies for surrogate generation. The “flip any” and “flip all” surrogates used here rely on replacing the observed responses for a triad by their complement. Alternatively, one could replace the data from each triad with the data from a randomly-selected triad. As is the case for the “flip any” and “flip all” surrogates, this permuted surrogate would have the same a priori probability as the data. Correlations within a triad be destroyed, and, in addition, judgments that were more certain (i.e,. observed choice probabilities near 0 or 1) would be randomly replaced by judgments that were less certain. Thus, a permutation test would sample the independent prior more evenly than the “flip” surrogates, but it would not determine whether the values of the indices could be due to co-occurrence of easy vs. hard judgments within triplets.

While the present approach aims at determining compatibility with an ultrametric or addtree distance, it stops short of attempting to determine the specific ultrametric or addtree structure. Existing methods for taking this step require a complete set of dis-similarity measures [1, 2, 13, 26], along with the assumption that the transformation from triadic choice probabilities into dis-similarities is uniform across the space. Choice probabilities provide constraints even without this assumption – for example certain relationships among the triadic judgments involving four points are sufficient for a pointwise addtree model – but it is unclear whether these constraints are sufficient for a global model, or how such a global model can be determined.

The central strategy of the present framework is to translate the consequences of metrical constraints on distances and their sums into constraints on rank orders of dis-similarities. Here, we have chosen to implement this strategy in its simplest form. In considering compatibility with a symmetric distance, we only considered triadic comparisons within a triplet – even though (see eq. (2.12)) there are further constraints on triadic comparisons within larger sets of triads. On larger stimulus sets, a complete analysis would also need to take into account the constraints of transitivity (eq. (3.53)). If these constraints are taken into account, the theoretical benefit that each of the local contributions to Isym is determined from a non-overlapping set of triads (eq. (3.24) would be lost. Relatedly, an alternative strategy for assessing addtree structure is to consider all 12 of the triadic comparisons among four stimuli together, rather than just the six-triad subsets that constitute a tent. It is unclear whether the additional complexity entailed by any of these considerations would translate into practical benefit.

We speculate that the overall framework of relating graphical distance models to rank-order comparisons can be extended to more complex graphs. Specifically, the observation that a metric that obeys the four-point condition can always be realized by the path metric on a weighted acyclic graph [4] suggests the possibility of a succession of further characterizations. By definition, acyclic graphs have no three-cycles. An isolated 3-cycle with nodes a1, a2, and a3 can always be removed by adding a node c, with distances dc,a1=da1,a2+da1,a3da2,a3/2, etc.; this quantity is guaranteed to be non-negative via the triangle inequality, and dc,a1+dc,a2=da1,a2. Thus, ruling out compatibility with an addtree model via Proposition 2.5 also implies that the dis-similarity structure is incompatible with distances on a weighted acyclic graph, or on a weighted graph with only isolated three-cycles. Consequently, a graph with two non-disjoint three-cycles or a four-cycle, is required. Perhaps more elaborate conditions analogous to those of Proposition 2.5 could then rule out compatibility with distances on graphs with some greater level of cycle structure or connectivity.

In this regard, it is interesting to note that the ultrametric condition and the four-point condition have a similar structure: both state that among three numbers (three single distances for the ultrametric, three pairwise sums for the four-point condition), the largest two must be identical. This intriguing similarity raises the further possibility of a sequence of analogous conditions, each specifying a progressively less-restrictive aspect of a set of dis-similarity judgments – such as compatibility with planar graphs (and, more generally, dimensionality), or other topological characterizations, such as statements about Betti numbers.

Supplementary Material

Supplement 1

Supplementary figures. Supplement to main text Figures 511

Acknowledgments.

This work was supported by NIH NEI EY7977 (JV), NSF: 2014217 (JV), and the Fred Plum Fellowship in Systems Neurology and Neuroscience (SAW). GA would like to thank Marianne Maertens for all her support. We thank Laurence T. Maloney for his many helpful discussions, especially concerning the addtree model.

Appendix A: Metric Spaces

This appendix demonstrates that pointwise or setwise compatibility of a set of choice probabilities with a symmetric dis-similarity D(x,y) implies pointwise or setwise compatibility with a metric-space structure.

A metric space is defined to be a set of points {x,y,z,}, along with a metric d(x,y) that associates each pair of points with a real number, and that satisfies three properties:

d(x,y)0andd(x,y)=0x=y, (A.1)
symmetry :d(x,y)=d(y,x), (A.2)

and

the triangle inequality :d(x,y)d(x,z)+d(y,z). (A.3)

As we have generically assumed (in eq. (2.4)) that dis-similarity D(x,y) satisfies (A.1) and we now add that it is symmetric (eq. (A.2)), we need to show that we can replace D(x,y) – which need not satisfy the triangle inequality – by a metric d(x,y) which both satisfies the triangle inequality and also accounts for the choice probabilities via

R(r;x,y)>12d(r,x)<d(r,y). (A.4)

It suffices to find a monotonic transformation of the dis-similarity

d(x,y)=G(D(x,y)). (A.5)

for which d(x,y) satisfies the triangle inequality. A suitable transformation for this purpose is

G(u)=0,u=01+u1+u,u>0 (A.6)

For distinct elements a and b, D(a,b)>0 and consequently d(a,b)=G(D(a,b)) is between 1 and 2. So the left-hand side of (A.3) is always < 2 and each term on the right-hand side is > 1, implying that the triangle inequality holds.

Appendix B: Addtree Models and the Four-Point Condition

This Appendix demonstrates some well-known and basic [4, 7, 26] relationships between the four-point condition, the triangle inequality, and the ultrametric inequality.

To show that the four-point condition (2.19) implies the triangle inequality, we set w=x. Then (2.19) becomes

None of the three quantitiesd(u,v)d(u,x)+d(v,x)d(u,x)+d(v,x)is strictly greater than the other two. (B.1)

That is, d(u,v)d(u,x)+d(v,x), which is the triangle inequality.

To show that the ultrametric inequality (2.14) implies the four-point condition, we first relabel the points u, v, w, and x so that d(u,v) is the smallest distance, and

d(u,v)d(u,w)d(u,x). (B.2)

Consider the triangle with vertices u, v, and w. Since d(u,v)d(u,w), the ultrametric inequality, which says that the two longest sides must have the same length, requires that the length of the third side, d(v,w), is equal to d(u,w):

d(u,v)d(u,w)d(v,w)=d(u,w). (B.3)

Similarly, applied to the triangle with vertices u, v, and x, the ultrametric inequality yields

d(u,v)d(u,x)d(v,x)=d(u,x). (B.4)

Combining these two yields

d(u,w)+d(v,x)=d(v,w)+d(u,x). (B.5)

Applied to the triangle with vertices u, w, and x, the ultrametric inequality yields,

d(u,w)d(u,x)d(w,x)=d(u,x). (B.6)

Combining this with the assumption that d(u,v) is the shortest distance, (B.2) yields

d(u,v)+d(w,x)=d(u,v)+d(u,x)d(v,w)+d(u,x). (B.7)

Together, (B.5) and (B.7) constitute the four-point condition.

Examples:

For four distinct points on a line and distances given by the ordinary Euclidean distance d(y,z)=|zy|, the addtree conditions hold. Taking u<v<w<x,

d(u,w)+d(v,x)=(wu)+(xv)=(wv)+(xu)=d(u,x)+d(v,w), (B.8)

while

d(u,v)+d(w,x)=(vu)+(xw)=(v+x)(u+w)<(w+x)(u+v)=(xu)+(wv)=d(u,x)+d(v,w). (B.9)

However, the ultrametric inequality does not in general hold, since d(u,v)<d(u,w)<d(u,x).

For the standard distance between four distinct points in general position in a plane, the four-point condition does not hold, since the three pairwise sums are typically distinct.

Appendix C: Setwise Sufficiency for the Four-Point Condition

Here, we prove a partial converse to Proposition 2.5: that falsification of the conjunction (2.22) suffices to ensure that the dis-similarities are setwise-compatible with an addtree model. More precisely:

Proposition C.1 (sufficient ordinal conditions for addtree).

If, for each relabeling of a set of four points {z,a,b,c}, (2.22) does not hold, and all of the dis-similarities are unequal, then there is a strictly monotonic-increasing transformation d(x,y)=F(D(x,y)) of the dis-similarities into distances, for which the four-point condition (2.19) holds.

Proof.

We choose F to be of the form

F(D)=D+k,DD0D,D<D0. (C.1)

so the demonstration rests on finding an appropriate choice of D0 and k. We note that we can ignore whether the quantity that results from the monotonic transformation satisfies

Figure 16:

Figure 16:

Three rank-orderings of the dis-similarities among four points consistent with falsification of the conjunction (2.22) used in the proof of Proposition C.1 (Appendix C). In all cases, D(z,c) is largest (heavy line) and D(z,a) is second-largest (intermediate line). The third-largest dis-similarity can be either D(z,b), D(a,c), or D(b,c) (thin solid lines). See Appendix C for details.

the triangle inequality, since (as is well-known) Appendix B shows that the four-point condition implies the triangle inequality.

We refer to the three quantities in the four-point condition (2.19) and the corresponding sums of pairs of dis-similarities as “pairsums.” For convenience, we reproduce (2.22) below, replacing the non strict inequalities by strict ones, since the dis-similarities are assumed to be unequal:

D(z,c)D(z,b)andD(z,c)D(z,a)andD(a,b)D(c,a)andD(a,b)D(b,c). (C.2)

WLOG, relabel the points so that D(z,c) is the greatest dis-similarity.

Case A. D(z,c)+D(a,b) is not the largest pairsum. Choose D0=D(z,c) in (C.1). Then d(z,c)+d(a,b)=F(D(z,c))+F(D(a,b))=D(z,c)+D(a,b)+k, but the other pairsums are unchanged by F since their components are all strictly less than D0. Setting k equal to the difference between D(z,c)+D(a,b) and the next-largest pairsum yields a transformation to distances that satisfies the four-point condition.

Case B. D(z,c) is the largest dis-similarity and D(z,c)+D(a,b) is the largest pairsum. Take D0>D(a,b) but smaller than the next-largest dis-similarity. Then again d(z,c)+d(a,b)=F(D(z,c))+F(D(a,b))=D(z,c)+D(a,b)+k. We show that applying F increases at least one of the other two pairsums by 2k. Then, choose k to be the difference between D(z,c)+D(a,b) and the next-largest pairsum. Since F cannot increase a pairsum by more than 2k, applying F results in equating the two largest pairsums.

To show that applying F increases at least one of the other two pairsums by at least 2k, we need to show that for at least one of the other two pairsums, both terms are greater than D(a,b). Since D(z,c) is the largest dis-similarity, the second-largest dis-similarity cannot be D(a,b), since then (2.22) would hold. So the second-largest dis-similarity must be between a point in {z,c} and a point in {a,b}. We then relabel the points (switching a and b if necessary, and switching z and c if necessary) so that the second-largest dis-similarity is D(z,a). In this relabeling, D(a,b) cannot be the third-largest dis-similarity, since then (C.2) would hold. Thus, D(z,c) is the largest dis-similarity, D(z,a) is the second-largest, and there are three possibilities for the third-largest dis-similarity: D(z,b), D(a,c), and D(b,c) (Figure 16).

Case B1. The third-largest dis-similarity is D(z,b): Either D(a,c)>D(a,b) or D(b,c)>D(a,b) since otherwise (C.2) would hold. Since D(z,b) and D(z,a) are both greater than D(a,b), both terms of at least one of D(z,b)+D(a,c) or D(z,a)+D(b,c) are greater than D(a,b).

Case B2. The third-largest dis-similarity is D(a,c):D(a,b) cannot be larger than both D(z,b) and D(b,c), because otherwise, a conjunction like (C.2) would hold for the tent {b;z,c,a}, as D(b,a)=D(a,b) would be the largest of the tripod at b, and D(z,c) would be the largest of the base. So at least one of D(z,b)>D(a,b) or D(b,c)>D(a,b). Thus, both terms of either pairsum D(z,b)+D(a,c), or D(z,a)+D(b,c) are greater than D(a,b).

Case B3. The third-largest dis-similarity is D(b,c): Both terms of D(z,a)+D(b,c) are larger than D(a,b).

References

  • [1].Abraham I., Bartal Y., and Neiman O., Embedding metrics into ultrametrics and graphs into spanning trees with constant average distortion, SIAM J. Computing 44 (2015), no. 1, 160–192, doi. [Google Scholar]
  • [2].Barthélemy J.-P. and Guénoche A., Trees and proximity representations, John Wiley and Sons, Chichester, England, 1991. [Google Scholar]
  • [3].Bujack R., Teti E., Miller J., Caffrey E., and Turton T. L., The non-riemannian nature of perceptual color space, Proc Natl Acad Sci U S A 119 (2022), no. 18, e2119753119, doi. [Google Scholar]
  • [4].Buneman P., A note on the metric properties of trees, Journal of Combinatorial Theory 17 (1974), 48–50. [Google Scholar]
  • [5].Dabaghian Y., Memoli F., Frank L., and Carlsson G., A topological paradigm for hippocampal spatial map formation using persistent homology, PLoS Comput Biol 8 (2012), no. 8, e1002581, doi. [Google Scholar]
  • [6].de Leeuw J. and Heiser W., Theory of multidimensional scaling, Handbook of Statistics (Krishnaiah P.R. and Kanal L.N., eds.), vol. 2, North-Holland, 1982, pp. 285–316. [Google Scholar]
  • [7].Dobson A., Unrooted trees for numerical taxonomy, J. Appl. Prob. 11 (1974), 32–42. [Google Scholar]
  • [8].Edelman S., Representation is the representation of similarities, Behav Brain Sci 21 (1998), no. 4, 449–67; discussion 467–98. [DOI] [PubMed] [Google Scholar]
  • [9].Ferguson T.S., A bayesian analysis of some nonparametric problems, Ann. Statist. 1 (1973), no. 2, 209–230, doi. [Google Scholar]
  • [10].Gilchrist A., Seeing black and white, Oxford University Press, New York, NY, 2006. [Google Scholar]
  • [11].Giusti C., Pastalkova E., Curto C., and Itskov V., Clique topology reveals intrinsic geometric structure in neural correlations, Proc Natl Acad Sci U S A 112 (2015), no. 44, 13455–60, doi. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [12].Guidolin A., Desroches M., Victor J. D., Purpura K. P., and Rodrigues S., Geometry of spiking patterns in early visual cortex: a topological data analytic approach, J R Soc Interface 19 (2022), no. 196, 20220677, doi. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [13].Hornik K., A clue for cluster ensembles, J. Statistical Software 14 (2005), no. 12, 1–25. [Google Scholar]
  • [14].Jeffreys H., An invariant form for the prior probability in estimation problems, Proc R Soc Lond A Math Phys Sci 186 (1946), no. 1007, 453–61, doi. [DOI] [PubMed] [Google Scholar]
  • [15].Kemp C. and Tenenbaum J. B., The discovery of structural form, Proc Natl Acad Sci U S A 105 (2008), no. 31, 10687–92, doi. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [16].Kingdom F. A.A. and Prins N., Psychophysics: A practical introduction, second edition ed., Academic Press, Cambridge, MA, 2016, doi. [Google Scholar]
  • [17].Knoblauch K. and Maloney L. T., Mlds: Maximum likelihood difference scaling in r, J. Statistical Software 25 (2008), no. 2, 1–26, doi. [Google Scholar]
  • [18].Kriegeskorte N. and Kievit R. A., Representational geometry: integrating cognition, computation, and the brain, Trends Cogn Sci 17 (2013), no. 8, 401–12, doi. [Google Scholar]
  • [19].Kruskal J. B. and Wish M., Multidimensional scaling, Sage, Beverly Hills, 1978. [Google Scholar]
  • [20].Logvinenko A. D. and Maloney L. T., The proximity structure of achromatic surface colors and the impossibility of asymmetric lightness matching, Percept Psychophys 68 (2006), no. 1, 76–83, doi. [DOI] [PubMed] [Google Scholar]
  • [21].Madigan S. C. and Brainard D. H., Scaling measurements of the effect of surface slant on perceived lightness, Iperception 5 (2014), no. 1, 53–72, doi. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [22].Maloney L. T. and Yang J. N., Maximum likelihood difference scaling, J Vis 3 (2003), no. 8, 573–85, doi. [DOI] [PubMed] [Google Scholar]
  • [23].Murray R.F., Lightness perception in complex scenes, Annual Review of Vision Science 7 (2021), 417–436, doi. [Google Scholar]
  • [24].Radonjic A., Cottaris N. P., and Brainard D. H., Color constancy supports cross-illumination color selection, J Vis 15 (2015), no. 6, 13, doi. [Google Scholar]
  • [25].Sala F., De Sa C., Gu A., and Re C., Representation tradeoffs for hyperbolic embeddings, Proceedings of the 35th International Conference on Machine Learning (Jennifer Dy and Andreas Krause, eds.), Proceedings of Machine Learning Research, vol. 80, PMLR, 10–15 Jul 2018, pp. 4460–4469, https://proceedings.mlr.press/v80/sala18a.html. [Google Scholar]
  • [26].Sattath S. and Tversky A., Additive similarity trees, Psychometrika 42 (1977), no. 3, 319–345. [Google Scholar]
  • [27].Saxe A. M., McClelland J. L., and Ganguli S., A mathematical theory of semantic development in deep neural networks, Proc Natl Acad Sci U S A 116 (2019), no. 23, 11537–11546, doi. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [28].Schutt H. H., Harmeling S., Macke J. H., and Wichmann F. A., Painfree and accurate bayesian estimation of psychometric functions for (potentially) overdispersed data, Vision Res 122 (2016), 105–123, doi. [DOI] [PubMed] [Google Scholar]
  • [29].Semmes S., An introduction to the geometry of ultrametric spaces, 2007, https://arxiv.org/abs/0711.0709. [Google Scholar]
  • [30].Shepard R. N., Stimulus and response generalization: tests of a model relating generalization to distance in psychological space, J Exp Psychol 55 (1958), no. 6, 509–23, doi. [DOI] [PubMed] [Google Scholar]
  • [31].Singh G., Memoli F., Ishkhanov T., Sapiro G., Carlsson G., and Ringach D. L., Topological analysis of population activity in visual cortex, J Vis 8 (2008), no. 8, 11 1–18, doi. [Google Scholar]
  • [32].Treves A., On the perceptual structure of face space, Biosystems 40 (1997), no. 1–2, 189–96, doi. [DOI] [PubMed] [Google Scholar]
  • [33].Tsogo L., Masson M. H., and Bardot A., Multidimensional scaling methods for many-object sets: A review, Multivariate Behav Res 35 (2000), no. 3, 307–19, doi. [DOI] [PubMed] [Google Scholar]
  • [34].Tversky A., Features of similarity, Psychol Rev 84 (1977), no. 4, 327–352. [Google Scholar]
  • [35].Tversky A. and Gati I., Similarity, separability, and the triangle inequality, Psychol Rev 89 (1982), no. 2, 123–54, http://www.ncbi.nlm.nih.gov/entrez/query.fcgi?cmd=Retrieve&db=PubMed&dopt=Citation&list_uids=7089125. [PubMed] [Google Scholar]
  • [36].Tversky A. and Hutchinson J.W., Nearest neighbor analysis of psychological spaces, Psychological Review 93 (1986), no. 1, 3–22. [Google Scholar]
  • [37].Victor J. D., Rizvi S. M., and Conte M. M., Two representations of a high-dimensional perceptual space, Vision Res 137 (2017), 1–23, doi. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [38].Victor J. D., Thengone D. J., Rizvi S. M., and Conte M. M., A perceptual space of local image statistics, Vision Res 117 (2015), 117–35, doi. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [39].Victor J.D. and Conte M.M., Local image statistics: maximum-entropy constructions and perceptual salience, J Opt Soc Am A Opt Image Sci Vis 29 (2012), no. 7, 1313–45, doi. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [40].Wallach H., Brightness constancy and the nature of achromatic colors, J Exp Psychol 38 (1948), no. 3, 310–24, doi. [DOI] [PubMed] [Google Scholar]
  • [41].Waraich S. A. and Victor J. D., A psychophysics paradigm for the collection and analysis of similarity judgments, JoVE (2022), no. 181, e63461, doi. [Google Scholar]
  • [42].Watson C.S., Kellogg S.C., Kanawishi D.T., and Lucas P.A., The uncertain response in detection-oriented psychophysics, Journal of Experimental Psychology 99 (1973), no. 2, 180–185, doi. [Google Scholar]
  • [43].Wickens T.D., Elementary signal detection theory, first edition ed., Oxford University Press, 2001. [Google Scholar]
  • [44].Wilson C. D., Serrano G. O., Koulakov A. A., and Rinberg D., A primacy code for odor identity, Nat Commun 8 (2017), no. 1, 1477, doi. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [45].Zaidi Q., Victor J., McDermott J., Geffen M., Bensmaia S., and Cleland T. A., Perceptual spaces: mathematical structures to neural mechanisms, J Neurosci 33 (2013), no. 45, 17597–602, doi. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [46].Zhang H., Rich P. D., Lee A. K., and Sharpee T. O., Hippocampal spatial representations exhibit a hyperbolic geometry that expands with experience, Nat Neurosci 26 (2023), no. 1, 131–139, doi. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [47].Zhou Y., Smith B. H., and Sharpee T. O., Hyperbolic geometry of the olfactory space, Sci Adv 4 (2018), no. 8, eaaq1458, doi. [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Supplement 1

Articles from ArXiv are provided here courtesy of arXiv

RESOURCES