Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2014 Mar 27.
Published in final edited form as: J Opt Soc Am A Opt Image Sci Vis. 2012 Oct 1;29(10):2204–2216. doi: 10.1364/JOSAA.29.002204

Asymptotic ideal observers and surrogate figures of merit for signal detection with list-mode data

Eric Clarkson 1
PMCID: PMC3967985  NIHMSID: NIHMS563840  PMID: 23201670

Abstract

The asymptotic form for the likelihood ratio is derived for list-mode data generated by an imaging system viewing a possible signal in a randomly generated background. This calculation provides an approximation to the likelihood ratio that is valid in the limit of large number of list entries, i.e., a large number of photons. These results are then used to derive surrogate figures of merit, quantities that are correlated with ideal-observer performance on detection tasks, as measured by the area under the receiver operating characteristic curve, but are easier to compute. A key component of these derivations is the determination of asymptotic forms for the Fisher information for the signal amplitude in the limit of a large number of counts or a long exposure time. This quantity is useful in its own right as a figure of merit (FOM) for the task of estimating the signal amplitude. The use of the Fisher information in detection tasks is based on the fact that it provides an approximation for ideal-observer detectability when the signal is weak. For both the fixed-count and fixed-time cases, four surrogate figures of merit are derived. Two are based on maximum likelihood reconstructions; one uses the characteristic functional of the random backgrounds. The fourth surrogate FOM is identical in the two cases and involves an integral over attribute space for each of a randomly generated sequence of backgrounds.

1. INTRODUCTION

For signal-detection problems, there is an optimal observer known as the ideal observer. Optimality in this case can be defined in several ways, one of which is the area under the receiver operating characteristic (ROC) curve. For a given statistical model of the data in the signal-absent and signal-present populations, the ideal observer maximizes the area under the ROC curve, also known as the area under the curve (AUC). The AUC is a figure of merit (FOM) for detection tasks that can be directly related to observer performance on two-alternative forced choice tests, where the observer is presented with two data vectors and must decide which one was drawn from the signal-present population. The ideal-observer AUC therefore can be used as a FOM quantifying the usefulness of the data produced by an imaging system for the particular detection task at hand.

The ideal observer is a mathematical observer that makes a decision for a given data vector by computing a likelihood ratio from the data and comparing it to a threshold [1]. Another quantity, the ideal-observer detectability, is monotonically related to the ideal-observer AUC and therefore provides an equivalent FOM for signal detection. The detectability for the ideal observer can be difficult to compute when there is randomness in the background of the signal as well as noise in the imaging system itself. It is therefore desirable to have a surrogate FOM (SFOM), a quantity that is easier to compute than the ideal-observer detectability but that correlates with it over a wide range of parameters specifying the background model, the signal model, and the imaging system. We have developed some SFOMs for imaging systems with binned data [2], and in this work we want to apply the same methods to systems that produce list-mode data [3,4]. A key component of this method is the Fisher information matrix (FIM).

The FIM is usually thought of as an important component of statistical estimation theory. The FIM is defined for a parameterized family of probability distribution functions (PDFs), where the parameter is a finite dimensional vector. In estimation theory, the Cramer–Rao bound is derived from the FIM and provides a lower bound on the variances of unbiased estimators of the components of the vector parameter [5]. One problem with the Cramer–Rao lower bound is that the minimum variance is only achieved for exponential families of PDFs, a very restricted class. In imaging, if we want to include object variability as well as system noise, then, as we will see below, the PDFs are mixture distributions, which do not form exponential families. Another problem with the Cramer–Rao bound is that it depends on the true parameter value, which is unknown and often a random quantity itself. A second property of the FIM is that the inverse of this matrix is asymptotically equal to the covariance matrix for the maximum likelihood estimator of the vector parameter [5]. Other asymptotic results relate the FIM to ideal-observer detection performance [6]. The difficulty with using these results in the imaging context is that we do not typically operate in an asymptotic regime, where you have many independent samples from a given member of the family of PDFs. In the imaging context we usually only have one image to work with and need to estimate a vector parameter and/or detect a signal without generating more images from the given object.

In previous publications [2,7] we have shown that there is a connection between the FIM and the ideal-observer detectability. The context in which this connection arises is the problem of detecting a small change in the vector parameter of a parameterized family of PDFs for the data. For example, if a signal is weak, then we may introduce a signal amplitude parameter and reduce the signal-detection problem to the detection of a small change in this parameter from zero. If we plot the detectability of the ideal observer versus signal amplitude, then the slope of this curve at the origin is the Fisher information for the amplitude parameter. One useful aspect of this relationship is that there are no special requirements on the form of the family of PDFs. This is also not an asymptotic result; it applies to the problem of detecting a signal from one data vector generated by the imaging system. In the sections below, we will use the FIM approximation to ideal-observer detectability to derive SFOMs for list-mode data. The essential strategy is to compute the FIM for the signal amplitude and then find approximations to this quantity that are valid when the number of list entries, the number of detected photons, is a large number. We will consider the fixed-counts case first, and then move on to the fixed-time case.

2. FIXED COUNTS

We do fixed counts first because it is easier in some ways and results from this case can be modified for use in the fixed-time case. We assume therefore that the number of counts is fixed ahead of time. We do not assume that the time it takes to collect these counts is measured, although in practice this could be useful information especially for certain estimation tasks. Thus, time is a random variable, and its distribution would have to be included in the analysis if it were being measured. The discussion in this section could be modified to take this into account.

A. Null Functions

We need to understand null functions for the system in order to make sense of asymptotic ideal observers in what follows. First we will review the basics of list-mode data for fixed counts. Then we will define the concept of a null function in this context and provide the conditions that null functions have to satisfy.

The conditional probability density function for a list entry  in a list-mode imaging system conditioned on the object function f is given by

pr(A^f)=Spr(A^r)s(r)f(r)drSs(r)f(r)dr=p(A^)fsf, (1)

where S is the support region for all object functions, s(r) is the sensitivity function (which gives the probability that a photon emitted from location r will be detected), and pr(Â|r) is the conditional probability that a detected photon emitted from location r will give the list entry Â. The space of list entries will be denoted by L. We will switch between the function notation and the vector notation at will. The vector notation may refer either to vectors in an infinite dimensional Hilbert space for theoretical purposes or to vectors in a finite dimensional space if the functions involved are discretized for computational purposes. In either case the space of objects will be denoted by U. Note that the vector p(Â) represents the function pr(Â|r)s(r). This fact gives us the relation

Lp(A^)dA^=s, (2)

which will be used below. Also note that the adjoint dagger symbol notation is used to denote the Hilbert space inner products represented by the integrals in Eq. (1).

Without loss of generality, we can assume that s(r) > 0 everywhere in the support region S. If the sensitivity function vanishes in some subregion of S, we might as well remove that subregion from consideration since photons being emitted from there will never contribute to the data. The quantity

s(r)f(r)Ss(r)f(r)dr (3)

is the conditional probability that a detected photon was emitted from location r given that the object function is f(r). Note that this fact implies that f(r) ≥ 0 everywhere in S and

Ss(r)f(r)dr>0. (4)

We will assume that all object functions satisfy these two positivity constraints.

The conditional probability for entire list is given by the product

pr(A^f)=j=1Jpr(A^jf), (5)

where J is the number of entries in the list. In this section J is assumed to be fixed and known. In Section 3, the exposure time will be fixed and known, and J will become a random number. If the time to reach J counts is measured, then this PDF would be multiplied by the PDF for this time, which would be a gamma distribution.

In the following fa will usually represent the actual static activity distribution, the object that produced the given data list. A null function at fa is a function fn that satisfies the condition

pr(A^fa+fn)=pr(A^fa) (6)

for all arbitrary list entries Â. Our first task is to see what conditions define the null functions at fa. To this end, let f = fa + fn and note that this function must satisfy the two positivity constraints. Define the normalization integrals for each object by

A=Ss(r)f(r)dr>0, (7)
Aa=Ss(r)fa(r)dr>0. (8)

We next define the scaled objects by fa=Aa-1fa and f′ = A−1f. The conditional probability for a list entry is invariant under scaling, so we have pr(A^f)=pr(A^fa). The scaled object functions therefore satisfy

Ss(r)f(r)dr=Ss(r)fa(r)dr=1, (9)

and, for all Â,

Spr(A^r)s(r)f(r)dr=Spr(A^r)s(r)fa(r)dr. (10)

We may write f=fa+fn and derive the equivalent conditions for fn:

Ss(r)fn(r)dr=0, (11)

and, for all Â,

Spr(A^r)s(r)fn(r)dr=0. (12)

If we integrate the second of these equations over  we arrive at the first equation. Thus, the second equation for fn(r) is all that we need. Now reverting to the original unscaled objects, we find that

f=(AAa)fa+Afn=fa+[(AAa)-1]fa+Afn. (13)

Therefore, we may say that the null functions at fa all have the form fn = Cfa + , where, for all Â,

Spr(A^r)s(r)f(r)dr=0 (14)

and C > −1. Of the two positivity constraints, the second one is automatically satisfied at this point. The first one gives us (r) ≥ −(1 + C)fa(r) for all r in S.

An equivalent description can be formulated in terms of the space of constant conditional probability at fa, i.e., the set of objects f satisfying the two positivity constraints and pr(Â|f) = pr(Â|fa). These functions have the form f = Bfa + with B > 0 and as above. Again the second positivity constraint is automatically satisfied, and the first positivity constraint reduces to (r) ≥ −Bfa(r) for all r in S. We will call this space U(fa), and it will make an appearance below when we discuss the asymptotic values of the likelihood ratio. It is straightforward to show that U(fa) is a convex cone, although this will not play a role in the discussions below.

B. Ideal Observer for Large Counts

The ideal observer for a detection task computes the likelihood ratio and compares this number to a threshold to decide the presence or absence of a signal. We will show that, in the asymptotic limit of a large number of counts, the likelihood ratio has a limiting value that depends only on the conditional probability pr(Â|f). We start by looking at asymptotic values for the conditional probability of the list.

The conditional PDF for the list Inline graphic = {Â1, …, ÂJ} is given by the product

pr(A^f)=j=1Jpr(A^jf). (15)

The unconditional PDF for the list is the integral

pr(A^)=Upr(A^f)pr(f)df. (16)

Both of these PDFs are defined on the space of lists Inline graphic = LJ. We assume that the support for the prior PDF pr(f) is contained within the set of objects satisfying the two positivity constraints. In the case where the object space U is infinite dimensional, the concept of a prior PDF is problematic since stochastic processes, in general, do not have PDFs. In this case we would replace this last equation with the expectation

pr(A^)=pr(A^f)f, (17)

which has a well-defined meaning for any stochastic process. We will use the integral notation throughout, but all expressions may be easily converted to the expectation notation as needed.

As a first step toward the asymptotic ideal observer for signal detection, we rewrite the conditional probability in the form

pr(A^f)=exp{j=1Jln[pr(A^jf)]}. (18)

For a large number of counts, we will get, with high probability, a good approximation by replacing the sum with an expectation value:

pr(A^f)exp{JLpr(A^fa)ln[pr(A^f)]dA^}. (19)

The object fa is the actual object that produced the data. This object is often specified in simulations used to evaluate imaging approaches but is not directly known in real clinical scans. This type of approximation will be used often below and has a rigorous meaning from information theory. That is, we can show that, for any ε > 0, the set Inline graphic(ε) of lists Inline graphic that satisfy

|1Jln[pr(A^f)]-Lpr(A^fa)ln[pr(A^f)]dA^|<ε, (20)

has probability

L(ε)pr(A^fa)dA^>1-ε (21)

if J is sufficiently large. This type of convergence is called convergence in probability. Note that the quantity on the right-hand side in the approximation for pr( Inline graphic|f) does not depend on the actual list itself but only on the object function that generated it. This means that, for large J, most lists, in the probabilistic sense, are approximately equally probable. In information theory this fact is called the asymptotic equipartition property [8]. The integral in the exponent in the approximation to pr( Inline graphic|f) is the cross entropy, and it can be expressed in terms of the Kullback–Leibler distance,

D[pr(A^fa),pr(A^f)]=Lpr(A^fa)ln[pr(A^fa)pr(A^f)]dA^, (22)

and the differential entropy,

H[pr(A^fa)]=-Lpr(A^fa)ln[pr(A^fa)]dA^. (23)

Notice that each of these two expressions involves an average over attribute vector Â. We will use a more concise notation DL(fa, f) for the Kullback–Leibler distance and HL(fa) for the differential entropy. The subscript L is used to emphasize that these quantities are calculated from the conditional PDFs in list space corresponding to the objects. In terms of these two quantities, the approximate value for pr( Inline graphic|f) is given by

pr(A^f)exp[-JDL(fa,f)]exp[-JHL(fa)]. (24)

The Kullback–Leibler distance is always nonnegative and equals zero only if the two conditional PDFs pr(Â|fa) and pr(Â|f) are the same. The differential entropy can be positive or negative.

The likelihood ratio is the test statistic for the ideal observer and is given by

Λ(A^)=Upr(A^f+Δf)pr(f)dfUpr(A^f)pr(f)df. (25)

We assume the signal Δf is a nonnegative function. This is needed to assure that pr( Inline graphic|f + Δf) is always well defined for objects f that satisfy the two positivity constraints. An alternative form for the likelihood is given by

Λ(A^)=Upr(A^f)pr(f-Δf)dfUpr(A^f)pr(f)df. (26)

Using the cross-entropy approximation, we may write

Λ(A^)Uexp[-JDL(fa,f)]pr(f-Δf)dfUexp[-JDL(fa,f)]pr(f)df. (27)

Note the quantity on the right-hand side in this expression is not itself a test statistic since it does not depend on the actual data. What this expression tells us is that, if the actual object is fa and we generate many data lists from this object, then with high probability the value for the corresponding likelihood ratio will be close to the quantity on the right-hand side.

We expect that, for a large number of counts, with high probability the objects that give a small value for DL(fa, f) will give the overwhelming contribution to the integrals in this last expression. Expanding the Kullback–Leibler distance around the actual object gives, to the lowest order, a quadratic function, and we then arrive at the approximation

Λ(A^)Uexp{-J2(f-fa)F(fa)(f-fa)}pr(f-Δf)dfUexp{-J2(f-fa)F(fa)(f-fa)}pr(f)df. (28)

The matrix F(fa) is the FIM for the conditional probability pr(Â|f) evaluated at the actual object. For an infinite dimensional object space, the FIM is an operator. We discuss this last approximation further in Subsection 2.C.

C. FIM and the Null Space

If we examine the approximation in Eq. (27), we can see that the quantity in the exponent is zero if f and fa differ by a null function at fa. We want to see whether the approximation in Eq. (28) has the same property. This will help us to see what happens to the likelihood ratio as J → ∞.

We can compute the FIM by using the definition of the conditional probability pr(Â|f) from above. Computing the score is the first step, and this vector is defined by sc(Â|f) = ∇f ln[pr(Â|f)]. A short calculation leads to

sc(A^f)=p(A^)p(A^)f-ssf. (29)

The two positivity constraints on f ensure that the components of this vector are always finite numbers. The FIM is the covariance of the score F(f) = 〈sc(Â|f)sc(Â|f)〉Â|f and can be written as

F(f)=p(A^)p(A^)[p(A^)f]2A^f-ss(sf)2. (30)

In practice this matrix can be estimated by drawing samples from the conditional PDF pr(Â|f).

Now we want to relate the FIM to the null functions described above. Let us define the difference function fd = ffa. Then the quadratic form in the exponent in the FIM approximation to the likelihood ratio in Eq. (28) is given by

fdF(fa)fd=[p(A^)fdp(A^)fa]2A^fa-(sfdsfa)2. (31)

This quantity vanishes if fd is a null function at fa. On the other hand, we may write the quadratic form as

fdF(fa)fd=[p(A^)fdp(A^)fa-sfdsfa]2A^f. (32)

If this quantity vanishes, then, for almost all Â, we have the equality

p(A^)fdp(A^)fa=sfdsfa. (33)

If the numerator on the right in this equation is zero, the numerator on the left is also, and fd is a null function. If the numerator on the right is nonzero, we have the equality

p(A^)fdsfd=p(A^)fasfa. (34)

As above this gives us fd = Bfa + with p(Â) = 0 for all Â. Therefore, fd is also a null function at fa in this case. Note that the B = 0 situation is covered in the first case, where the numerators are zero. On the other hand, the exponent in Eq. (28) will be zero when B = −1 even though fd is not a null function at fa strictly speaking for this value of B. Other than that caveat, which is not important for any of the integrals involved, we can say that the quadratic form in the exponent in Eq. (28) is zero if and only if fd is a null function at fa. This means that the quadratic expansion used in Subsection 2.B for the Kullback–Leibler distance is an expansion for f near the space of constant conditional probability U(fa) within the object space. In particular f does not have to be near fa for the quadratic expansion to be a valid approximation to the KL distance. All that is required is that ffa is near a null function at fa. Now we will return to the likelihood ratio and see what happens as J → ∞.

D. Asymptotic Form for the Ideal Observer

Now we return to the FIM approximation of the likelihood ratio in Eq. (28) and rewrite the denominator as

Uexp[-J2fdF(fa)fd]pr(fd+fa)dfd. (35)

Define U0 = U0(fa) as the null space of F(fa) and U1 = U1(fa) as the orthogonal complement to U0. Also let fd = fd0 + fd1 be the corresponding orthogonal decomposition for the difference vector. We have shown above that U0 is the linear span of the functions in U(fa), and that these functions are essentially all of the nonnegative functions in U0. Define the eigenvalues and eigenvectors of the FIM by F(fa)un = λnun and consider the eigenvector decomposition of the difference vector:

fd=n=1Nfdnun. (36)

The quadratic form in the exponent of the FIM approximation to Λ can be written as

fdF(fa)fd=n=1Rλnfdn2=fd1F(fa)fd1, (37)

where R is the rank of F(fa). Now the denominator in Eq. (28) can be written as

U1exp[-J2fd1F(fa)fd1]U0pr(fd0+fd1+fa)dfd0dfd1. (38)

By normalizing the Gaussian function that appears in both the numerator and denominator of the FIM approximation to Λ in Eq. (28), we can see that, as J → ∞, the value of the likelihood ratio asymptotically approaches a limit:

Λ(A^)U0pr(fd0+fa-Δf)dfd0U0pr(fd0+fa)dfd0. (39)

Since the vector fa is in the subspace U0, we may write this limit as

Λ(A^)U0pr(fd0-Δf)dfd0U0pr(fd0)dfd0. (40)

Finally, we use the fact that we have a positive signal and that the prior PDF is supported on the set of functions satisfying the two positivity constraints to write

Λ(A^)U(fa)pr(f-Δf)dfU(fa)pr(f)df, (41)

as J → ∞. The ratio in this last equation is the likelihood ratio for the observer that is directly observing the subspace of objects whose conditional probabilities are identical with the conditional probability of the actual object. The performance of this observer defines the maximum performance for the original ideal observer. Another way of thinking about this observer is that it is observing the conditional PDF pr(Â|fa) directly instead of J samples from this PDF.

The performance of this asymptotic ideal observer provides an upper limit for the performance of the ideal list-mode observer for the given signal-detection task. It is possible that the AUC of the asymptotic ideal observer is unity, in which case the detectability for the list-mode ideal observer will increase without bound as J → ∞. To see how this is possible, suppose first that the signal is absent. Then fa is in the support of the prior distribution. If f − Δf is not within this support for any f in U(fa), then the numerator in the asymptotic likelihood ratio in Eq. (41) will be zero. If this property of the signal holds for any signal-absent fa, then the asymptotic likelihood ratio is always zero when the signal is absent. Now suppose that the signal is present. Then fa − Δf is in the support of the prior distribution. If f is not in this support for any f in U(fa), then the denominator in the asymptotic likelihood ratio in Eq. (41) is zero. If this property holds for any signal-present fa, then the asymptotic likelihood ratio is always infinite when the signal is present. If the signal satisfies both of these properties, then the AUC for the asymptotic ideal observer will be unity.

As an example of how the scenario described in the previous paragraph can happen, suppose that the we have a planar optical imaging system and the actual signal-free objects are flat with unknown amplitude. The support of the prior is then the one-dimensional space of flat objects. Suppose that the signal has frequency components within the passband of the imaging system. For any signal-absent object fa, the space U(fa) consists of sums of flat objects and objects whose frequency components are all higher than the band limit of the imaging system. Then f − Δf is not flat for any f in U(fa). On the other hand, for any signal-present object fa, the space U(fa) consists of sums of flat objects, the signal, and objects whose frequency components are all higher than the band limit of the imaging system. Then f is not flat for any f in U(fa). This is an extreme example, but it illustrates that an AUC of unity can occur for the asymptotic likelihood ratio if the prior is supported on a relatively low-dimensional parameterized space, the signal is not a null object, and the signal is not in the support of the prior.

E. SFOMs

We could use the AUC or detectability of the ideal list-mode observer as an FOM for the imaging system on the detection task, but he likelihood ratio is difficult to compute accurately. We have developed Markov chain Monte Carlo (MCMC) methods to do this computation, but they are time consuming. An SFOM is a quantity that can be computed more easily and will correlate with the ideal-observer detectability when system or signal parameters are varied. We will examine some SFOMs that arise from a relation between ideal-observer detectability and Fisher information that we have derived elsewhere.

To derive an SFOM, we first introduce an amplitude parameter α for the signal Δf and write

pr(A^α)=Upr(A^f+αΔf)pr(f)df. (42)

For convenience of notation we will define pr( Inline graphic) = pr( Inline graphic|0). The SFOM will arise from the relation between the detectability for the ideal observer and the Fisher information for the amplitude parameter α. Therefore, the first step is to compute the score for α. The Fisher information is then the expectation of the square of the score evaluated at α = 0.

The score for the amplitude parameter at α = 0 is defined in the usual way as

sc(A^)=ddαln[pr(A^α)]|α=0. (43)

A short calculation leads to an integral expression

sc(A^)=UΔfsc(A^f)pr(fA^)df, (44)

where the local score is defined by sc( Inline graphic|f) =∇f ln[pr( Inline graphic|f)] and the posterior PDF pr(f| Inline graphic) satisfies the equation pr(f| Inline graphic)pr( Inline graphic) = pr( Inline graphic|f)pr(f). Thus, the (global) score sc( Inline graphic) is the inner product of the signal with the posterior mean of the local score. The word “local” here refers to the fact that the local score is conditioned on a given object f.

Now we can calculate the local score and get an expression for the score. The result is

sc(A^)=UΔf[j=1Jsc(A^jf)]pr(fA^)df, (45)

where sc(Âj|f) is given by Eq. (29). The FIM approximation to ideal-observer detectability as a function of amplitude follows from the Taylor series expansion of the square of the ideal-observer detectability as a function of α:

d2(α)=α2[sc(A^)]2A^+. (46)

The next term in the expansion is cubic in the amplitude parameter and involves the derivative of the FIM as a function of α. The fourth-order term is not defined unless the kurtosis of the score is zero. We will use the first nonzero term, i.e., the quadratic term shown above, as our approximation for the detectability:

d2(α)α2[sc(A^)]2A^fafa. (47)

In this expression we have written out the expectation as a two-stage process, where we average over lists given an object and then average over the objects. This approximation will be good for weak signals. Note that the ideal-observer detectability is related to the ideal-observer AUC by the relation 2[AUC(α)] = 1 + erf[1/2d(α)].

We could stop here, but our expression for the score involves a posterior mean, which will require MCMC techniques to evaluate. MCMC methods tend to be computationally intensive, so we will examine approximations to the FIM approximation and hope they lead us to something easier to compute. The ultimate test of any resulting SFOM is to simulate a list-mode system where the ideal-observer detectability can be computed (as noted earlier, we have to use MCMC methods for this) and see if the proposed SFOM correlates with this quantity over a range of signals and/or system parameters.

1. Fixed-Counts SFOM 1

For the first SFOM, we use the law of large numbers to approximate the sum of local scores with an expectation:

sc(A^)JUΔfsc(A^f)A^fapr(fA^)df. (48)

Ideal-observer detectability can then be approximated by

d2(α)α2J2[UΔfsc(A^f)A^fapr(fA^)df]2A^fafa. (49)

So far this does not help much as we will still need MCMC methods to compute the integral in square brackets. However, if we interchange the expectation and the integration within the square brackets, we arrive at

d2(α)α2J2[UΔfsc(A^f)pr(fA^)dfA^fa]2A^fafa. (50)

The integral is the posterior mean (PM) estimate of the quantity Δfsc(Â|f) from the data list. One option at this point to eliminate the need for MCMC methods is to replace this PM estimate with a maximum likelihood (ML) estimate. We get an ML estimate of the local score by producing an ML estimate of the object and then inserting it into the local score. This replacement would then result in the approximation

d2(α)α2J2[Δfsc[A^f^ML(A)]A^fa]2A^fafa. (51)

This approximation requires computation of an ML estimate of the object from the list Inline graphic and then ordinary Monte Carlo sampling to compute the expectation of Δfsc[Â|ML( Inline graphic)] over  given fa. Then we square and use Monte Carlo sampling to average over Inline graphic given fa, and finally average over fa. This is a lot of Monte Carlo averaging, but at least there are no MCMC computations.

2. Fixed-Counts SFOM 2

For this SFOM, we start with the same expectation approximation for the score as in SFOM 1, but now we write the expectation in terms of cross entropy:

Δfsc(A^f)A^fa=ΔffLln[pr(A^f)]pr(A^fa)dA^. (52)

For large values of J, the posterior distribution is localized near the space U(fa), so we will use the quadratic approximation for the KL distance:

Δfsc(A^f)A^fa-12Δff[(f-fa)F(fa)(f-fa)]. (53)

Now we take the gradient

Δfsc(A^f)A^fa-ΔfF(fa)(f-fa). (54)

This gives an approximation for the score in terms of the PM estimate of the actual object: sc( Inline graphic) ≅ JΔfF(fa)[PM( Inline graphic) − fa].

This score approximation gives an approximation for the detectability via the Fisher information. The approximate detectability is expressed as a nested pair of expectations:

d2(α)α2J2ΔfF(fa)[f^PM(A^)-fa][f^PM(A^)-fa]A^faF(fa)Δffa. (55)

MCMC methods are still needed here to compute the PM object estimate. To get a more tractable SFOM, we replace the PM estimate with an ML estimate and write

d2(α)α2J2ΔfF(fa)[f^ML(A^)-fa][f^ML(A^)-fa]A^faF(fa)Δffa. (56)

We will see below that, in the asymptotic limit, there is actually no difference in this approximation when we replace the PM estimate with an ML estimate. Note that we will still have three expectations to compute in this approximation since the matrix F(fa) requires an expectation using samples from pr(Â|fa) to evaluate.

3. Fixed-Counts SFOM 3

For this SFOM, we first find the asymptotic form of the vector that appears in SFOM 2:

F(fa)[f^PM(A^)-fa]F(fa)U(f-fa)exp[-J2(f-fa)F(fa)(f-fa)]pr(f)dfUexp[-J2(f-fa)F(fa)(f-fa)]pr(f)df. (57)

With the difference vector fd as above, we can rewrite these integrals and insert them into the Fisher information detectability approximation. The result is

d2(α)α2J2{UΔfF(fa)fdexp[-J2fdF(fa)fd]pr(fd+fa)dfdUexp[-J2fd4F(fa)fd]pr(fd+fa)dfd}2fa. (58)

Both integrals in the curly brackets are inner products between a shifted prior PDF and a function whose Fourier transform is known. This inner product could be performed in Fourier space if the characteristic function for the prior PDF is known, which is often the case. One complication of this approach is that the matrix F(fa) is singular due to the presence of null functions at fa, but this does not present insurmountable difficulties if the pseudoinverse of F(fa) can be computed. If the Fourier space approach is viable, then there will only be two Monte Carlo calculations involved in the computation of the SFOM: one for computing F(fa) and the other for the expectation over fa. This approximation to the ideal-observer detectability should be good for large J.

What happens to this SFOM in the limit as J → ∞? To answer this last question, we decompose object space using the singular value decomposition of F(fa) as we did for the likelihood ratio. The end result is

F(fa)[f^PM(A^)-fa]F(fa)U1fd1exp{-J2fd1F(fa)fd1}U0pr(fd0+fd1+fa)dfd0dfd1U1exp{-J2fd1F(fa)fd1}U0pr(fd0+fd1+fa)dfd0dfd1. (59)

By normalizing the Gaussian in the numerator and denominator and letting J → ∞, we find that F(fa)[PM( Inline graphic) − fa] → 0 independently of the actual list Inline graphic. This shows that the quantity in the angle brackets in our third approximation to detectability is approaching zero as J → ∞. We know that the detectability is bounded by the detectability of the asymptotic ideal observer discussed above. Therefore, the decreasing value of the expectation must cancel the J2 dependence when the asymptotic ideal detectability is finite.

We cannot say that the PM estimator is asymptotically unbiased due to the existence of null functions. It does however follow from this discussion that, as J → ∞, we have pr(Â|PM) → pr(Â|fa). This relation is as close as we can get to an asymptotic unbiased condition. Now we will find the asymptotic form for the PM estimator itself. For large J we have the approximation

f^PM(A^)-faU1exp{-J2fd1F(fa)fd1}U0(fd0+fd1)pr(fd0+fd1+fa)dfd0dfd1U1exp{-J2fd1F(fa)fd1}U0pr(fd0+fd1+fa)dfd0dfd1. (60)

Normalizing the Gaussian in the usual way, we find that, in the limit as J → ∞, we get

f^PM(A^)-faU0fd0pr(fd0+fa)dfd0U0pr(fd0+fa)dfd0. (61)

This shows that the PM estimator is biased asymptotically. The bias is given by the null function on the right in this relation. Note that the bias depends on the actual object fa. We may also write this relation in the form

f^PM(A^)U(fa)fpr(f)dfU(fa)pr(f)df. (62)

This shows that the PM estimator asymptotically depends only on the prior PDF pr(f) and the conditional PDF pr(Â|fa), which is what we expect.

From the asymptotic form of the conditional probability as J → ∞,

pr(A^f)exp{-J2(f-fa)F(fa)(f-fa)}, (63)

we can see that any ML estimate will have the asymptotic form MLfa + fn, where fn is a null function at fa. Since the PM estimate has the same form asymptotically, the replacement of F(fa)[PM( Inline graphic) − fa] with F(fa)[ML( Inline graphic) − fa] in Approximation 2 results in no penalty in the limit as J → ∞. If the PM of the local score was asymptotically the same as sc(Â|PM), we could say something similar about Approximation 1, but these two estimates of the local score are probably not the same.

4. Fixed-Counts SFOM 4

For this last SFOM for fixed counts, we go back to SFOM 2 in Eq. (56) and use asymptotic properties of ML estimators to simplify the expression for approximate detectability. An ML estimator is a solution to the equation

j=1Jfln[pr(A^jf^ML)]=0. (64)

Using Taylor’s theorem, we can expand to the first order about fa and write this as

j=1J fln[pr(A^jfa)]+j=1Jffln[pr(A^jfβ)](f^ML-fa)=0, (65)

where fβ = βML + (1 − β)fa and 0 ≤ β ≤ 1. These conditions on fβ place it on the line segment connecting fa and ML.

Now define a random vector y by

y=j=1Jfln[pr(A^jfa)]. (66)

Each component in the sum for the vector y is the local score evaluated at a list entry. These are independent identically distributed random vectors with mean zero and covariance F(fa). The covariance of y is therefore JF(fa). Next, using the law of large numbers, we get the first approximate equality in

-1Jj=1Jffln[pr(A^jfβ)]F(fβ)F(fa) (67)

for large J. The second approximate equality here follows from the fact noted above that, as J → ∞, MLfa + fn, where fn is a null function at fa. Since f1 is between fa and ML, it will have the same asymptotic limit.

Putting these facts together, we have, for large J, the approximation yJF(fa)(MLfa), and the covariance of the vector on the right in this approximate equality is therefore approximately JF(fa). Together with Eq. (56), this gives us a simpler SFOM:

d2(α)α2JΔfF(fa)Δffa. (68)

In explicit form this SFOM can be written as

d2(α)α2J{[p(A^)Δfp(A^)fa]2A^fafa-(sΔfsfa)2fa}. (69)

In terms of computation, this SFOM requires the least amount of Monte Carlo sampling. The dependence on J is somewhat troubling since we know that the detectability cannot increase indefinitely with J when the asymptotic ideal observer has finite detectability. At present we have no convincing solution to this difficulty. However, this may not detract form the use of this approximation as an SFOM since we only need something correlated with ideal-observer performance for this application.

Using Eq. (1), we may write SFOM 4 as a more analytic expression:

d2(α)α2J{1sfaL[p(A^)Δf]2p(A^)fadA^fa-(sΔfsfa)2fa}. (70)

If the integral over attribute space can be performed, then we only need to sample from the prior to calculate this quantity.

3. FIXED TIME

Now we fix the exposure time and let the number of counts become a Poisson random variable. The number of counts is now part of the data and gives us a little more information about the object than we have in the fixed-counts case. We will be concerned with the same issues in this section as in the previous one. We want to know the asymptotic form of the likelihood ratio, and we want to use the asymptotic analysis to develop SFOMs for ideal-observer performance.

A. Null Functions

As with the fixed-counts case, we need to know what the null functions are in order to proceed. The null space is less complicated for the fixed-time case and does not depend on the actual abject.

The mean number of counts involves the exposure time τ:

J¯(f)=τSs(r)f(r)dr=τsf. (71)

The data are the list Inline graphic as in Section 2. plus the total count J. We write G = ( Inline graphic, J) for the data. The conditional probability for the data given the object is given by

pr(Gf)=[J¯(f)]JJ!exp[-J¯(f)]j=1Jpr(A^jf)=Pr(Jf)pr(A^f,J). (72)

The conditions for a null function fn at fa are pr(Â|fa + fn) = pr(Â|fa) for all attribute vectors Â, and (fa + fn) = (fa). These conditions are satisfied if and only if fn satisfies Eq. (12) for all Â. The denominator in the conditional probability pr(Â|fa) is fixed by the mean count and the time, so there is no scale invariance as there was for the fixed-count case. The null space is therefore independent of fa for fixed time. This null space will be denoted by Ũ0 and its orthogonal complement by Ũ1.

B. Ideal Observer for Long Time Interval

The PDF for the data is given by the integral

pr(G)=Upr(Gf)pr(f)df. (73)

As with the fixed-count case, an approximation to the value of the likelihood ratio that is accurate with high probability is given by

Λ(G)UPr(Jf)exp{-JDKL[pr(A^fa),pr(A^f)]}pr(f-Δf)dfUPr(Jf)exp{-JDKL[pr(A^fa),pr(A^f)]}pr(f)df. (74)

Note that, for long exposure times, the mean (fa) of the Poisson distribution for the count J is very large, and this means that, with very high probability, J itself is very large since the standard deviation of a Poisson distribution is the square root of the mean. This fact justifies the use of this approximation for the value of the likelihood ratio. As with the fixed-counts case, the quantity on the left in this relation is not itself a test statistic but represents with very high probability an approximate value for the likelihood ratio.

Using the same FIM F(fa) as before, we may also write an approximation to the approximation:

Λ(G)UPr(Jf)exp{-J2(f-fa)F(fa)(f-fa)}pr(f-Δf)dfUPr(Jf)exp{-J2(f-fa)F(fa)(f-fa)}pr(f)df. (75)

We will make use of this expression to examine the asymptotic form of the likelihood ratio.

C. Limiting Forms for Ideal Observer

Introducing the difference vector fd as before, the denominator in the approximation to Λ in Eq. (75) can be written as

UPr(Jfd+fa)exp[-J2fdF(fa)fd]pr(fd+fa)dfd. (76)

By decomposing the object space U into the null space U0 of F(fa) and its orthogonal complement U1, the denominator can be written

U1exp[-J2fd1F(fa)fd1]U0Pr(Jfd0+fd1+fa)×pr(fd0+fd1+fa)dfd0dfd1. (77)

Note that U0 is not the null space discussed in Subsection 3.A; it is the same U0 as we used in the fixed-counts case.

By normalizing the Gaussian function in the numerator and the denominator, we can see that, as J → ∞, with high probability, the value of the likelihood ratio converges to a limit:

Λ(A^)U0Pr(Jfd0+fa-Δf)pr(fd0+fa-Δf)dfd0U0Pr(Jfd0+fa)pr(fd0+fa)dfd0. (78)

The invariance of the subspace U0 under translation by fa reduces this expression to

Λ(A^)U0Pr(Jfd0-Δf)pr(fd0-Δf)dfd0U0Pr(Jfd0)pr(fd0)dfd0. (79)

Now we can again use the fact that we have a positive signal to write:

Λ(A^)U(fa)Pr(Jf-Δf)pr(f-Δf)dfU(fa)Pr(Jf)pr(f)df. (80)

In the numerator and denominator of this last relation, we are again integrating over the space of object functions that satisfy pr(Â|f) = pr(Â|fa), but, in contrast to the fixed-counts case, the prior probability at each point in this space is now multiplied by the corresponding probability for the observed value of J. This is the likelihood ratio for the observer who has direct access to the normalized mean of the Poisson point process in attribute space that generates the list and to the number of entries in that list.

We can take this a step further and expand the probability for the number of counts to the second order about the actual object:

Pr(Jf)1J!exp{-J¯(fa)+Jln[J¯(fa)]+[J-J¯(fa)J¯(fa)]J¯(f-fa)-J2J¯2(fa)[J¯(f-fa)]2}. (81)

Define a modified Fisher matrix by the equation

Fs(fa)=F(fa)+1(sfa)2ss=p(A^)p(A^)[p(A^)fa]2A^fa. (82)

After canceling some terms, we may write an approximation for the likelihood ratio:

Λ(G)Uexp{-J2fdFs(fa)fd+[J-J¯(fa)J¯(fa)]J¯(fd)}pr(fd+fa-Δf)dfdUexp[-J2fdFs(fa)fd+[J-J¯(fa)J¯(fa)]J¯(fd)]pr(fd+fa)dfd. (83)

Decompose object space using the SVD of Fs(fa). A null vector fn for this matrix satisfies

[p(A^)fnp(A^)fa2]2A^fa=0, (84)

which is true if and only if fn is in Ũ0. Also note that (fn) = 0 in this case. Let fd = d0 + d1 be the corresponding decomposition for the difference vector. We may rewrite the denominator in this expression for the approximate likelihood ratio as

U1exp{-J2fd1Fs(fa)fd1+[J-J¯(fa)J¯(fa)]J¯(fd1)}U0pr(fd0+fd1+fa)dfd0dfd1. (85)

Note that, for a given d1, the linear term in the exponent is of the order of τ1/2 as τ → ∞ since J is a Poisson random variable with mean (fa). The quadratic term, on the other hand, will be of the order of τ. Thus, the quadratic term will dominate in the limit and produce a delta function at zero when the Gaussian is normalized. By treating the numerator the same way, we find that

Λ(A^)U0pr(fd0+fa-Δf)dfd0U0pr(fd0+fa)dfd0 (86)

as τ → ∞. We define the space Ũ(fa) to be those objects f that satisfy the two positivity constraints, pr(Â|f) = pr(Â|fa) for all  and (f) = (fa). Now we can again use the fact that we have a positive signal to write:

Λ(A^)U(fa)pr(f-Δf)dfU(fa)pr(f)df. (87)

This is the likelihood ratio for the observer who has access to the normalized mean to the Poisson process for the data list and to the mean number of counts. By multiplying this function by this number, this observer has access to the mean

u¯(A^)=Spr(A^r)s(r)fa(r)dr (88)

of the Poisson process that generates the data list. This function in attribute space contains all of the possible information that the imaging system can deliver about the object in question.

If we write the integral relationship in Eq. (88) as ū = Inline graphicfa, we can formulate the condition under which the asymptotic ideal observer will have a finite value for detectability. In order for this to occur, there must be a set CU of signal-absent objects such that the prior probability assigned to C is nonzero and for each object f(r) in C there is another signal-free object f′(r) such that Inline graphicf = Inline graphicf′ + Inline graphicΔf. If the function Inline graphicf is observed by the asymptotic ideal observer, both fa = f and fa = f′ + Δf are possibilities for the actual object. In this case, the numerator and denominator of the likelihood ratio in Eq. (87) are nonzero numbers, and the decision of asymptotic ideal observer will depend on the threshold. Since the set C has finite prior probability, there is a finite probability that the asymptotic ideal observer will make the wrong decision, at least for thresholds in a certain range of values. Thus, the detectability for this observer will be finite. If the set of confusion C is empty, then the asymptotic ideal observer will always make the correct decision, and the detectability will be infinite. In terms of the likelihood ratio in Eq. (87), this means that the numerator will be zero when the signal is absent and the denominator zero when the signal is present. The condition for finite detectability can also be stated in object space: there is a set of signal-free objects CU with finite prior probability such that, for any fC, there is a signal-free object f′ and a null function fn satisfying f + fn = f′ + Δf.

D. SFOMs

The SFOMs will again be based on the Fisher information approximation to the ideal-observer detectability. As in the fixed-counts case, we introduce amplitude parameter α,

pr(Gα)=Upr(Gf+αΔf)pr(f)df, (89)

and define pr(G) = pr(G|0). The score for the amplitude parameter at α = 0 is defined as

sc(G)=ddtln[pr(Gα)]|α=0, (90)

and a short calculation leads to the integral

sc(G)=UΔfsc(Gf)pr(fG)df. (91)

As before, this integral involves the local score, sc(G|f) = ∇f ln[pr(G|f)], and the posterior PDF pr(f|G). The global score is again the PM of the local score.

Now we can calculate the local score and get an expression for the score:

sc(G)=UΔf[j=1Jsc(A^jf)+JτJ¯(f)s-τs]pr(fG)df. (92)

By simplifying this expression, we arrive at

sc(G)=UΔf[j=1Jp(A^j)p(A^j)f]pr(fG)df-τΔfs. (93)

The FIM approximation to ideal-observer detectability as a function of amplitude is then given by d2(α) ≅ α2 〈[sc(G)]2G. We can also write this SFOM in terms of nested expectations as

d2(α)α2[sc(G)]2Gfafa. (94)

As we did for the fixed-counts case, we will seek to find asymptotic approximations to the expectation on the right in this relation.

1. Fixed-Time SFOM 1

For this approximation, we replace the sum in the score with an expectation:

sc(G)JUp(A^)Δfp(A^)fA^fapr(fG)df-τΔfs. (95)

Interchanging the expectation with the integral gives us

sc(G)JU[p(A^)Δfp(A^)f]pr(fG)dfA^fa-τΔfs. (96)

Once again the integral represents a PM estimate of the quantity in the square brackets. To avoid MCMC computations, we could replace the PM estimate with an ML estimate and use

sc(G)Jp(A^)Δfp(A^)fML(G)A^fa-τΔfs (97)

in the expression for the approximate ideal-observer detectability. Computation of this approximation would then require only Monte Carlo evaluations of the expectations and an algorithm for finding an ML estimate of the object. The covariance of the score then gives the Fisher information.

2. Fixed-Time SFOM 2

For this approximation, we use the unsimplified form for the score and replace the sum with an expectation:

sc(G)JUΔfsc(A^f)A^fapr(fG)df+[JUpr(fG)sfdf-τ]sΔf. (98)

Now we follow the same steps that we used in the fixed-counts case to approximate the expectation in the first term of this expression with a quadratic function. The end result is that the score can be written in terms of the PM estimate of the object and what can be viewed as a PM estimate of the exposure time:

sc(G)-JΔfF(fa)[f^PM(G)-fa]+[τ^PM(G)-τ]Δfs. (99)

The PM estimate of the exposure time is the integral

τ^PM(G)=UJsfpr(fG)df. (100)

Of course we do not need to estimate the exposure time since we measure it, but since the actual exposure time is given by

τ=J¯(fa)sfa, (101)

this PM expression is not an unreasonable estimate of τ if it was not measured. We could approximate this estimate of the exposure time by

τ^PM(G)JsfPM(G). (102)

Since we want to avoid MCMC calculations, we will replace PM estimates with ML estimates to produce an approximate score to use in our SFOM:

sc(G)-JΔfF(fa)[f^ML(G)-fa]+[τ^ML(G)-τ]Δfs, (103)

where

τ^ML(G)=JsfML(G). (104)

The use of the ML estimate in the first term of the approximate score can actually be rigorously justified in the limit as τ → ∞ because, as we will see below, the PM estimate of the object and the ML estimate of the object differ by a null vector of F(fa) in this limit. As in the first approximation, the resulting approximation to the ideal-observer detectability gives an SFOM that requires only Monte Carlo evaluations of expectations and an ML reconstruction algorithm.

3. Fixed-Time SFOM 3

For this approximation, we expand the second term in the score approximation to the first order about the actual object:

[τ^PM(G)-τ]JsfPM(G)-τJsfa-τ-Js(sfa)2[f^PM(G)-fa]. (105)

Now the score is approximated by

sc(G)-JΔfFs(fa)[f^PM(G)-fa]+(Jsfa-τ)Δfs. (106)

To proceed further, we use the approximation

Fs(fa)[f^PM(G)-fa]Fs(fa)U(f-fa)exp{-J2(f-fa)Fs(fa)(f-fa)+[J-J¯(fa)J¯(fa)]J¯(f-fa)}pr(f)dfUexp[-J2(f-fa)Fs(fa)(f-fa)+[J-J¯(fa)J¯(fa)]J¯(f-fa)]pr(f)df. (107)

With the difference vector fd as above, we can write this as

Fs(fa)[f^PM(G)-fa]Fs(fa)Ufdexp{-J2fdFs(fa)fd+[J-J¯(fa)J¯(fa)]J¯(fd)}pr(fd+fa)dfdUexp[-J2fd4Fs(fa)fd+[J-J¯(fa)J¯(fa)]J¯(fd)]pr(fd+fa)dfd. (108)

Again, both integrals in the curly brackets are inner products between a shifted prior PDF and a function whose Fourier transform is known. This inner product could be performed in Fourier space if the characteristic function for the prior PDF is known and the pseudoinverse of Fs(fa) can be computed. This approximation should be good for large τ. What happens in the limit as τ → ∞?

Using the decomposition of object space into Ũ0 and Ũ1, we can write

Fs(fa)[f^PM(G)-fa]Fs(fa)U1fd1exp{-J2fd1Fs(fa)fd1+[J-J¯(fa)J¯(fa)]J¯(fd1)}U0pr(fd0+fd1+fa)dfd0dfd1U1exp{-J2fd1Fs(fa)fd1+[J-J¯(fa)J¯(fa)]J¯(fd1)}U0pr(fd0+fd1+fa)dfd0dfd1. (109)

By normalizing the Gaussians and letting τ → ∞, we find that Fs(fa)[PM(G) − fa] → 0 independently of G.

We cannot say that the PM estimator is asymptotically unbiased due to the existence of null functions. It does however follow from this discussion that, as τ → ∞, we have, for all Â, the relation p(Â) fPMp(Â)fa. This last equation is as close as we can get to an asymptotic unbiased condition for the fixed-time case.

We find as in the fixed-counts case that the asymptotic form as τ → ∞ for the PM estimator itself is determined by

f^PM(G)-faU0fd0pr(fd0+fa)dfd0U0pr(fd0+fa)dfd0. (110)

This shows that the PM estimator is biased asymptotically. The bias is given by the null function on the right in this relation. Note that the bias depends on the actual object fa. As with the fixed-count case, we may also write this relation in the form

f^PM(A^)U(fa)fpr(f)dfU(fa)pr(f)df. (111)

This shows that the PM estimator asymptotically depends only on the prior PDF pr(f), the conditional PDF pr(Â|fa), and the exposure time τ, which is again what we expect.

In the asymptotic limit, we are using the fact that

pr(Gf)1J!exp{-J¯(fa)+Jln[J¯(fa)]-J2(f-fa)Fs(fa)(f-fa)+[J-J¯(fa)J¯(fa)]J¯(f-fa)}. (112)

Therefore, any ML estimate must satisfy fMLfa + fn as τ → ∞ where fn is a null function. This shows that the PM estimate of the object and the ML estimate differ by a null function in this limit.

4. Fixed-Time SFOM 4

For this last SFOM for fixed counts, we go back to SFOM 2 in Eq. (103) and use asymptotic properties of ML estimators to simplify the expression for approximate detectability. An ML estimator is a solution to the equation

(Jsf^ML-τ)s+j=1Jfln[pr(A^jf^ML)]=0. (113)

Using Taylor’s theorem, we can expand to the first order about fa and write this as

(Jsfa-τ)s+j=1Jfln[pr(A^jfa)]+{-J(sfβ)2ss+j=1Jffln[pr(A^jfβ)]}(f^ML-fa)0, (114)

where fβ = βML + (1 − β)fa and 0 ≤ β ≤ 1 as before.

Define a random vector y as before via the equation

y=j=1Jfln[pr(A^jfa)]. (115)

For a fixed J, we know that the mean of this vector is 0, and therefore the mean remains 0 when the averaging over J is included. For fixed J the vector y has covariance JF(fa). After averaging over J, we find that the overall covariance of y is F(fa). For large exposure time τ, the count J is also large with high probability. Using the law of large numbers, we get, with high probability,

1J{J(sfa)2ss-j=1Jffln[pr(A^jfβ)]}Fs(fβ)Fs(fa). (116)

The second approximate equality here follows from the fact noted above that, as J → ∞, MLfa + fn, where fn is a null function. Since fβ is between fa and ML, it will have the same asymptotic limit.

Putting these facts together, we have, for large τ, the approximation

yJFs(fa)(f^ML-fa)-(Jsfa-τ), (117)

and the covariance of this vector is approximately F(fa). By taking this relation together with Eq. (106) and the fact that the ML estimate of the object and the PM estimate differ by a null function in the asymptotic limit, we get

d2(α)α2τ(sfa)(ΔfF(fa)Δf)fa. (118)

Note the similarity with SFOM 4 from the fixed-counts case. If we write this approximation out explicitly, we have a convenient expression for this SFOM:

d2(α)α2τ(sfa){[p(A^Δf)p(A^)fa]2A^fa-(sΔfsfa)2}fa. (119)

Once again we have the problem that the detectability increases without bound as the exposure time increases, which we know cannot be the case if the asymptotic ideal observer has finite detectability. It must be the case that some or all of the approximations that lead to this final approximation to the detectability become worse as the exposure time increases, but we have not been able to determine where the error enters into the calculation.

If we use Eq. (1), we may rewrite this SFOM in the more analytic form,

d2(α)α2τL[p(A^)Δf]2p(A^)fadA^-(sΔf)2sfafa. (120)

As with the fixed-counts SFOM 4, if the integral over attribute space can be performed, then we only need to sample from the prior to evaluate this quantity.

4. CONCLUSION

In view of the similarity between the result in Eq. (70) and the result in Eq. (120), we propose that the most useful SFOM to investigate at this time is the quantity

SFOM4=L[p(A^)Δf]2p(A^)fadA^-(sΔf)2sfafa.

There are several advantages that this SFOM possesses over other SFOMs discussed in this paper and in the work on SFOMs for binned data in [2]. The first advantage is that the computation of SFOM 4 does not require any reconstructions. SFOMs 1 and 2 in this paper and all of the SFOMs in [2] require an object reconstruction from simulated data for each sample drawn from the object ensemble. The second advantage for SFOM 4 is that the integral is over the low-dimensional attribute space as opposed to SFOM 3 for fixed-count and fixed-time cases, which requires an integral over the large-dimensional object space. Another advantage for SFOM 4 is that it is nearly identical to an approximation to ideal-observer detectability derived in [3] by a completely different approach. In that paper this approximation was shown to correlate more closely with ideal-observer performance as calculated by MCMC methods [9] than the Hotelling signal-to-noise ratio. The Hotelling observer is the ideal observer when all of the statistics are Gaussian, and the Hotelling SNR is equal to the ideal-observer detectability in that case.

Our next step is therefore to compute SFOM 4 in simulations where the ideal-observer detectability can be directly calculated. By varying background, signal, and system parameters, we can then observe the degree of correlation between SFOM 4 and ideal-observer detectability. If these studies have positive results, then we can start applying SFOM 4 to the optimization of list-mode imaging systems. For systems in the design stage, SFOM 4 could be used to rapidly assess various design options in terms of signal detectability. Since SFOM 4 is based on the FIM, these systems would be simultaneously optimized for estimating signal amplitude. For adaptive imaging systems [10], the initial or scout scan can be used to reduce the object ensemble to those objects compatible with the scout data. Then the average over objects in SFOM 4 can be performed over this reduced ensemble to select the optimum system configuration for the diagnostic scan.

Acknowledgments

The author wishes to acknowledge Harrison Barrett and Matthew Kupinski for their help with this work, and National Institutes of Health grants R37 EB000803, P41 EB002035, and RC1 EB010974 for financial support.

Footnotes

OCIS codes: 110.3000, 110.3055, 110.4280.

References

  • 1.Barrett HH, Abbey CK, Clarkson E. Objective assessment of image quality. III. ROC metrics, ideal observers, and likelihood generating functions. J Opt Soc Am A. 1998;15:1520–1535. doi: 10.1364/josaa.15.001520. [DOI] [PubMed] [Google Scholar]
  • 2.Clarkson E, Shen F. Fisher information and surrogate figures of merit for the task-based assessment of image quality. J Opt Soc Am A. 2010;27:2313–2326. doi: 10.1364/JOSAA.27.002313. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Caucci L, Barrett HH. Objective assessment of image quality. V. Photon counting detectors and list-mode data. J Opt Soc Am A. 2012;29:1003–1016. doi: 10.1364/JOSAA.29.001003. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Barrett HH, White T, Parra LC. List-mode likelihood. J Opt Soc Am A. 1997;14:2914–2923. doi: 10.1364/josaa.14.002914. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.Shao J. Mathematical Statistics. Springer: 1999. [Google Scholar]
  • 6.Kay SM. Fundamentals of Statistical Signal Processing II: Detection Theory. Prentice Hall; 2008. [Google Scholar]
  • 7.Shen F, Clarkson E. Using Fisher information to approximate ideal-observer performance on detection tasks for lumpy-background images. J Opt Soc Am A. 2006;23:2406–2414. doi: 10.1364/josaa.23.002406. [DOI] [PubMed] [Google Scholar]
  • 8.Cover TM, Thomas JA. Elements of Information Theory. Wiley; 1991. [Google Scholar]
  • 9.Kupinski MA, Hoppin JW, Clarkson E, Barrett HH. Ideal-observer computation in medical imaging with use of Markov-chain Monte Carlo techniques. J Opt Soc Am A. 2003;20:430–438. doi: 10.1364/josaa.20.000430. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Clarkson E, Kupinski MA, Barrett HH, Furenlid L. A task-based approach to adaptive and multimodality imaging. Proc IEEE. 2008;96:500–511. doi: 10.1109/JPROC.2007.913553. [DOI] [PMC free article] [PubMed] [Google Scholar]

RESOURCES