Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2008 Oct 29.
Published in final edited form as: IEEE Trans Med Imaging. 2008 Jun;27(6):775–788. doi: 10.1109/TMI.2007.913241

Adaptive SPECT

Harrison H Barrett 1,*, Lars R Furenlid 2, Melanie Freed 3, Jacob Y Hesterman 4, Matthew A Kupinski 2, Eric Clarkson 2, Meredith K Whitaker 2
PMCID: PMC2575754  NIHMSID: NIHMS74317  PMID: 18541485

Abstract

Adaptive imaging systems alter their data-acquisition configuration or protocol in response to the image information received. An adaptive pinhole single-photon emission computed tomography (SPECT) system might acquire an initial scout image to obtain preliminary information about the radiotracer distribution and then adjust the configuration or sizes of the pinholes, the magnifications, or the projection angles in order to improve performance. This paper briefly describes two small-animal SPECT systems that allow this flexibility and then presents a framework for evaluating adaptive systems in general, and adaptive SPECT systems in particular. The evaluation is in terms of the performance of linear observers on detection or estimation tasks. Expressions are derived for the ideal linear (Hotelling) observer and the ideal linear (Wiener) estimator with adaptive imaging. Detailed expressions for the performance figures of merit are given, and possible adaptation rules are discussed.

Index Terms: Adaptive imaging, Hotelling observer, image quality, single-photon emission computed tomography (SPECT), Wiener estimator

I. Introduction

There is considerable interest in many fields of imaging in adaptive systems that autonomously alter their data-acquisition configuration or protocol in real time in response to the image information being received. The goal of the adaptation may be to correct for random, dynamic imperfections in the imaging system or to optimize the system for the particular (unknown) object being imaged.

Adaptive imaging is most highly developed in ground-based astronomical imaging [1]–[5], where the spatial resolution is severely limited by rapidly changing phase distortions produced by the turbulent atmosphere. All modern observatories attempt to overcome this problem with some form of adaptive optics.

Most commonly, an auxiliary device called a wavefront sensor [6] is used to analyze the wavefront produced by a bright point source (a so-called guide star) in the field-of-view, and a control element such as a deformable mirror is used to correct the phase distortions. It is also possible, however, to analyze short-exposure images of an unknown extended object and deduce the wavefront distortions [7], [8] or derive various sharpness metrics [9], [10] that can be maximized by changing the signals applied to the control element.

Spurred by the success of astronomical adaptive optics, researchers in another phase-sensitive imaging modality, medical ultrasound, soon began using similar adaptive techniques [11]–[17]. In addition, adaptive control of the pulse sequence in magnetic resonance imaging, first proposed in 1993 by Cao and Levin [18], is currently an active area of study [19]–[25].

Much less has been done with adaptive data acquisition with ionizing radiation. A European consortium [26], [27] is currently developing a digital radiography system that allows active control of beam filtration. Some commercial computed tomography (CT) systems adapt the current in the X-ray tube in response to measured attenuation variations through the body, but with this exception adaptive data-acquisition methods have not been used in single-photon emission computed tomography (SPECT), positron emission tomography (PET), or CT.

Adaptive data acquisition in tomography is difficult for several reasons. First, the computational requirements of a reconstruction algorithm may not be compatible with rapid control of acquisition parameters, and second, mechanical control of the imaging configuration is slower and more complex than the electronic control used in adaptive optics, ultrasound, or medical resonance imaging (MRI). More importantly, there has been no compelling argument that adaptation would be helpful and no rationale for performing the adaptation.

Dedicated parallel processors should solve the computational problems, and we have recently proposed and built [28]–[30] a prototype adaptive SPECT system, described briefly in Section II, to address the mechanical issues. The main thrust of the present paper is to present a framework for assessing the performance of adaptive SPECT systems and optimizing the adaptation strategy.

The optimization is based on the paradigm of objective or task-based assessment of image quality in which image quality is defined by the performance of a specific observer performing a specific task on a predetermined class or ensemble of patients [31], [32]. Important tasks include detection of a signal such as a tumor, estimation of parameters associated with the signal, or a combination of detection and estimation. The observer can be a human analyzing the images or a computer algorithm. Particular attention has been given in the literature to linear numerical observers which compute linear functions of the data and use them to perform the task. Optimal linear observers for both detection and estimation have been studied thoroughly for conventional, nonadaptive imaging systems, and we shall extend that theory to adaptive imaging in this paper. Then we shall use the results to discuss practical adaptation strategies and show how to assess the improvement in task performance.

For context, we begin in Section II by discussing the tradeoffs in SPECT imaging, and then we describe two specific hardware platforms for implementing task-based system optimization in small-animal SPECT: the prototype adaptive system mentioned above and a more conventional but very flexible system called the multimodule, multiresolution (M3R) system [33]–[35].

In Section III, we present the basic theory of image formation and task performance for adaptive systems, contrasting the results with more familiar ones for conventional, nonadaptive, systems. As we shall see there, implementation of the important linear observers and calculation of figures of merit require knowledge of the first-order and second-order statistics (mean and covariance) of the data. Building on recent work related to astronomical adaptive optics [36], Section IV develops a methodology for deriving these statistical properties of the data.

In Section V, we apply the theory from Section IV to calculating the performance of ideal linear observers with adaptive data. An important conclusion that will emerge is that the ideal linear observers require knowledge of object and data statistics when the objects are known to be consistent with a measured preliminary or scout image. We refer to this set of objects as the posterior ensemble.

In Section VI, we discuss the practical issues in estimating these posterior statistics and assessing the resulting task performance. Section VII summarizes the results presented and discusses some remaining challenges in adaptive imaging.

II. Hardware Platforms

A. Controllable Variables and Tradeoffs in Pinhole SPECT

Pinhole SPECT systems are reviewed in [37] and from the viewpoint of small-animal imaging in [38]. Conventional single-camera SPECT systems use just one imaging detector, but several groups [39]–[42] are building small-animal SPECT systems with multiple detectors. At the University of Arizona, we use four imaging detectors in M3R, eight in our SemiSPECT system [43], 16 in FastSPECT II [44], [45], and 24 in the original FastSPECT.

Conceptually, the simplest approach to pinhole SPECT is to have multiple image detectors, one pinhole per detector and no motion of the object or detector system. In this configuration, which we use routinely for FastSPECT and FastSPECT II, the controllable variables are the pinhole diameter, the distance from the center of the object to the pinhole plane, and the distance from the pinhole plane to the detector. The system magnification is defined as s2/s1, and the system sensitivity is controlled by the pinhole diameter and s1. Spatial resolution, which depends on pinhole diameter, detector resolution, and magnification, can be traded off for sensitivity and field-of-view.

If we also allow rotation of the object or detector/aperture assembly, the projection angles are additional controllable variables, and the object can be translated parallel to the axis of rotation if desired, for example to perform helical cone-beam scans.

Next, we can consider multiple pinholes for each detector. These pinholes can form an arbitrary pattern, and shutter assemblies can be used to control which pinholes are open at any specific time. If all pinholes are open simultaneously, we have a classic coded-aperture system with a large photon-collection efficiency, but there is ambiguity about which pinhole captured a particular photon. If they are opened individually, each for a fraction of the same total acquisition time, we have lower collection efficiency but no ambiguity. It is also possible to open subsets of the pinholes simultaneously if that should prove advantageous for task performance.

We assume that the number and type of detectors will be fixed in an operational adaptive SPECT system, but that all of the other options listed above will be available, as they are in and our adaptive prototype.

B. Multimodule, Multiresolution System, M3R

To investigate the tradeoffs listed above, we constructed the M3R system. It uses four modular scintillation cameras, each of which has a 12-cm square, 5-mm-thick sodium iodide crystal, 15-mm-thick light guide, and nine photomultiplier tubes. Each camera is individually contained in an aluminum housing with lead shielding.

To allow rapid modification of the hardware configuration, we designed and built a Cerrobend structure that allows for interchangeable pinhole plates. There are five pinhole plate slots for each camera, providing for either magnification or minification depending on plate selection. Varying magnification and/or pinhole configuration is as simple as swapping pinhole plates. There is a central area able to accommodate a mouse or a phantom, and a rotary stage situated beneath the structure allows for object rotation.

Current research with M3R includes comparison of overlapping with nonoverlapping pinhole images, study of the effect of pinhole diameter on tumor-detection performance, and choice of the optimum aperture for tumor-volume estimation [34], [35].

C. Adaptive SPECT Prototype

A prototype adaptive gamma-ray imaging system for small animals has been designed and constructed [30]. It places camera location, aperture location, aperture size, and integration times under flexible software control. The projections that are acquired and included for tomographic reconstruction have different magnifications, resolutions, sensitivities, fields-of-view, and dwell times.

The current prototype system has a single modular camera, a single on-axis aperture with an array of four selectable diameters, and a vertically-oriented object. The camera position, aperture position, aperture diameter, and object rotation angle are all controlled by motorized stages and thus can adopt an essentially continuous range of object-aperture and aperture-detector configurations allowing magnifications from less than one to about 21. The aperture diameter is selectable between four different pinhole diameters that are currently 0.25, 0.5, 1.0, and 1.5 mm.

In operation, the system acquires an initial scout image in order to obtain preliminary information on the distribution of radiotracer in the particular subject being imaged, and the system parameters are then altered in real time to optimize the system, within practical constraints, for best performance with that subject and for a specific clinical or biomedical task. Fixed-geometry data can also be taken with the system to allow a more accurate comparison of fixed versus adaptive geometry optimizations.

Preliminary simulations indicate that the system can achieve sensitivities and resolutions greatly improved over comparable fixed-geometry systems [30].

III. Image Formation and Task Performance

A. Image Formation

The raw projection data in a conventional (nonadaptive) SPECT system consist of a set of M measurements, {gm, m = 1, …, M}, where m is a composite index specifying a particular detector pixel, a particular projection angle, and possibly the temporal frame in dynamic studies. These measurements constitute the elements of an M × 1 data vector g.

Image formation in conventional SPECT can be described abstractly as [32]

g=Hf+n (1)

Where f is the radiotracer distribution being imaged, n is a zero-mean noise vector, and ℋ is the system operator mapping the object to the mean image. The object is a function of continuous spatial variables and possibly time, and the image is a discrete vector; we refer to ℋ, therefore, as a continuous-to-discrete (CD) operator. It is a linear operator if we ignore the effects of detector saturation at high count rate.

It is sometimes useful to separate the effects of the imaging aperture and detector from the patient-dependent effects of attenuation and scatter. We can describe the latter by a linear operator Inline graphic, so that the overall system operator is ℋ = ℋ0 Inline graphic and

g=H0Af+n. (2)

The operator Inline graphic maps the primary-energy radiation emitted by the object to a phase-space distribution function which specifies the photon density as a function of spatial position, photon propagation direction, and energy. The spectrum of energies and propagation directions results from Compton scattering in the object. The effective object Inline graphicf is thus a function of energy as well as spatial and angular variables. The operator ℋ0 as an energy sensitivity based on the energy resolution of the detector, and it has some finite exposure time, so any particular element of g is obtained as an integral of Inline graphicf over spatial position, direction, time, and energy. For details, see [32, Sec. 10.4 and Sec. 16.2].

Image reconstruction maps the M × 1 data vector g to a discrete image with N voxels, which we can regard as an N × 1 vector . Denoting this mapping by the operator O, we have

f^=Og=O[H0Af+n]. (3)

For linear algorithms such as filtered backprojection, O is an N × M matrix, and the overall mapping from object to mean reconstruction, O0 Inline graphic, is a linear CD operator.

Adaptive systems are inherently nonlinear because the system depends on the object being imaged. In principle, the system could be altered continuously during acquisition as in adaptive optics, but we assume here that there is only one adaptation step; a scout image gs is acquired, and some information about the object is derived from it and used to control the system for acquisition of the final data g.

If the scout image is obtained with a linear system ℋs, we can write

gs=Hsf+ns. (4)

The scout image is used to modify the characteristics of the final system, so we have

g=H(gs)f+n. (5)

As in (2), we can decompose the scout-system operator as ℋs = ℋ0s Inline graphic, where the patient-specific part, Inline graphic, is common to both ℋ and ℋs. The resulting raw data vector from the final system is thus given by

g=H0(gs)Af+n=H0(H0sAf+ns)Af=n. (6)

The operator-valued function ℋ0(gs) is called the adaptation rule.

All reconstruction algorithms require information about the system operator (usually in the form of a matrix approximation to ℋ0), so the final reconstructed image has the form

f^=O(gs)g=O(gs)[H0(H0sAf+ns)Af+n]. (7)

For linear reconstruction algorithms, O(gs) is an N × M matrix with elements dependent on the scout image gs in some way, at least because it uses ℋ0(gs).

B. Linear Observers for Detection Tasks

A linear observer for a detection task is one that computes a scalar product of the data vector with another vector w called the template or discriminant function.

If the task is to be performed on the raw data g from a conventional system, the test statistic has the form

t(g)=c+m=1Mwmgm=c+wtg (8)

Where c is a constant independent of the data, w is an M × 1 vector, and superscript t denotes transpose. A similar form is used when the observer has access only to a reconstructed image, but there the template is an N × 1 vector to be applied to .

For adaptive systems, the template must depend on the scout image, at least through knowledge of ℋ0(gs), so (8) is modified to

t(g,gs)=c(gs)+wt(gs)g. (9)

The constant in (8) must now depend on the scout image because different adapted systems have different sensitivities (see Section V-B).

Common choices for the template in the nonadaptive case include the matched filter, the prewhitening matched filter, or Hotelling observer discussed below, and various models based on preprocessing the data through spatial-frequency-selective filters or channels. Noteworthy in this last category is the channelized Hotelling observer (CHO), which is often an excellent predictor of human performance. When the signal location is random, the linear observer can be scanned to different positions, giving rise to scanning Hotelling observers, scanning CHOs, etc.

Each of these choices has a counterpart in the adaptive case. For example, the matched filter for a nonadaptive system is given by w = ℋΔ, where Δ is the difference between the mean objects for the two hypotheses. For an adaptive system, the matched filter generalizes to ℋ(gs.

A generally applicable figure of merit (FOM) for detection tasks is the area under the receiver operating characteristic (ROC) curve. A common surrogate FOM, however, is a scalar separability measure or signal-to-noise ratio (SNR) defined by

SNRw2[t1t0]212Var1(t)+12Var0(t) (10)

where 〈tj is the expectation of the test statistic when hypothesis Hj is true (where, for example, H0 is tumor-absent and H1 is tumor-present), and Varj(t) is the corresponding variance. If the test statistic is normally distributed under both hypotheses, as it often is for linear discriminants, the area under the ROC curve is a simple monotonic function of the SNR [32].

The SNR as defined in (10) is applicable to both adaptive and nonadaptive systems, so long as the means and variances include all relevant random effects (including the statistics of gs in the adaptive case).

The linear observer that maximizes the SNR is called the Hotelling observer [46]–[48]. If this observer operates on the raw data from a nonadaptive system, its template is given by

wHot=Kg1HΔf¯ (11)

Where Kg is the covariance matrix of the data (usually approximately the same under both hypotheses for weak signals) and Δ is the difference in the mean object under the two hypotheses. The corresponding FOM for the Hotelling observer is obtained by evaluating (10) for template (11); the result is

SNRHot2=[HΔf¯]tKg1[HΔf¯]. (12)

The form of the Hotelling observer in the case of adaptive imaging is discussed in Section V-B.

C. Linear Observers for Estimation Tasks

If P parameters are to be estimated, we can assemble them into a P × 1 vector θ. For a nonadaptive system, an estimate of this parameter vector g, derived from an image vector g, is denoted θ̂(g), and a common FOM is the ensemble mean-square error defined by

EMSEθ^(g)θ2 (13)

where the angle brackets denote an average over all sources of randomness in g and also over some prior probability density on θ itself.

For an adaptive system, the estimator depends also on the scout image, so we denote it as θ̂(g, gs) and write the EMSE as

EMSEθ^(g,gs)θ2 (14)

where now the angle brackets must include the average over noise in the scout image.

In the nonadaptive case, an estimator is said to be linear (or more properly, affine), if it has the form

θ^(g)=a+Wtg (15)

where is a P × 1 vector and Wt is a P × M matrix. The vector a can be chosen so that the estimator is globally unbiased, i.e., 〈θ̂(g)〉 = θ̄, where θ̄ is the average of the parameter vector over the ensemble of objects.

The linear estimator that minimizes the EMSE in the nonadaptive case is the generalized Wiener estimator (not to be confused with a Wiener filter), given by [32], [36]

θ^W(g)=θ¯+KθgKg1[gg] (16)

where 〈g〉 is the data vector averaged over all sources of randomness, and Kθg is the P × M cross covariance between the parameter being estimated (θ) and the data.

The resulting EMSE is given by [32]

EMSEW=trKθtrKθgKg1Kθgt (17)

where tr denotes the trace (sum of the diagonal elements) of the matrix, and Kθ is the prior covariance of θ, based on the prior PDF before any measurements are made.

As with the Hotelling observer for a detection task, the Wiener estimator requires that we know the overall data covariance matrix, and it also requires the cross covariance Kθg.

The adaptive counterpart of (15) is [cf. (9)]

θ^(g,gs)=a(gs)+Wt(gs)g (18)

which is a linear (affine) function of g but a nonlinear function of gs. Expressions for a(gs) and W(gs) are given in Section V-C.

IV. Image Statistics

The image g from an adaptive or nonadaptive system must be regarded as a random vector. In a nonadaptive system described by (1), g is random because of the noise vector n, because a random ensemble of objects f will be imaged, and also because the system itself can be considered random. Sources of randomness in ℋ include practical issues such as mechanical errors, electronic drift, and patient motion as well as scatter and attenuation; if we consider random patients to be imaged, and we treat the attenuation and scatter as part of the system, then each randomly chosen patient will correspond to a different random system acting on a different radiotracer distribution f.

For an adaptive system, there is the additional randomness from the noise ns in the scout image. That is, if we repeatedly imaged the same patient through the same scout system, we would get a different final system ℋ(gs) each time. Even if the adaptation rule ℋ0(gs) were completely deterministic (no noise or mechanical errors in setting the parameters of the acquisition system), the overall system would still be described by a random operator ℋ0(gs) Inline graphic since the patient-specific operator Inline graphic is unknown and must be treated as random.

A. Nested Averages

A general method for analyzing all of these sources of randomness is the use of nested conditional averages [31], [32], [36]. To illustrate the procedure and notation, consider a random vector x that depends on two random parameter vectors a and b. By elementary probability theory, the probability density function of x is given by

pr(x)=dadbpr(xa,b)pr(ab)pr(b) (19)

where the integrals run over all values of all components of the vectors involved.

Random vectors with PDFs of this form are said to be multiply stochastic; they are random (because of measurement noise, for example), even if the parameters a and b are fixed and the randomness of each of the parameters also influences the overall PDF.

The expectation of an arbitrary function φ(x) is defined by

φ(x)xdxpr(x)φ(x)=dxdadbpr(xa,b)pr(ab)pr(b)φ(x) (20)

which we can also write as

φ(x)x=φ(x)xa,babb. (21)

The advantage of this nesting is that the inner expectation might be very simple since both random parameter vectors are fixed.

We can define partial averages of x itself by

x¯(a,b)xxa,b (22)
x¯¯(b)x¯(a,b)ab=xxa,bab (23)
x¯¯¯x¯¯(b)b=xxa,babb. (24)

With (21), the covariance matrix of can be written as

Kx[xx¯¯¯][xx¯¯¯]tx=[xx¯¯¯][xx¯¯¯]txa,babb (25)

where an expression of the form xyt denotes an outer product, i.e., if x and y are M × 1 vectors, then xyt is an M × M matrix with elements [xyt]mn = xmyn.

Adding and subtracting the two partial means in each factor yields

Kx=[xx¯(a,b)+x¯(a,b)x¯¯(b)+x¯¯(b)x¯¯¯]×[xx¯(a,b)+x¯(a,b)x¯¯(b)+x¯¯(b)x¯¯¯]txa,babb. (26)

When we multiply this expression out, all cross-terms vanish identically. For example

[x¯(a,b)x¯¯(b)][x¯¯(b)x¯¯¯]txa,babb=[x¯(a,b)x¯¯(b)]ab[x¯¯(b)x¯¯¯]tb=0 (27)

where the last step follows from definition (23). By construction, x¯(a,b)x¯¯(b) and x¯¯(b)x¯¯¯ are uncorrelated even if a and b are correlated.

Thus, we have the rigorous decomposition

Kx=K¯¯xa,b+K¯x¯b+Kx¯¯ (28)

where

K¯¯xa,b=Kxabb=[xx¯(a,b)][xx¯(a,b)]txa,babb (29)
K¯x¯b=Kx¯bb=[x¯(a,b)x¯¯(b)][x¯(a,b)x¯¯(b)]tabb (30)
Kx¯¯=[x¯¯(b)x¯¯¯][x¯¯(b)x¯¯¯]tb. (31)

Note that all three terms in (28) are averaged over all three sources of randomness, though some of the averages are before forming the relevant covariance and some after. In (29), for example, Kx|a,b (with no overbars) is averaged over x conditional on a and b, so it is a function of those parameters. The contribution of this conditional average to Kx, i.e., the first term in (19), requires two additional averages, denoted by the two overbars on K¯¯xa,b, and the result does not depend on a and b. All three terms in (28) have two overbars somewhere, either on the subscript or on K, and all three terms involve a total of three averages, one for each of the overbars and one by the definition of a covariance.

B. Multiply Stochastic Analysis of Nonadaptive Images

To illustrate this formalism and derive some results needed below, we compute the mean image vectors and covariance matrices for nonadaptive systems. We begin with the general expression for a nonadaptive image in (2), and we assume for simplicity that the aperture and detector are well characterized so that ℋ0 is known and nonrandom, but the overall system operator ℋ is still random because of the object-specific part Inline graphic. In addition, the noise n is random, with statistics that depend on the object f, and f itself is random when we consider an ensemble of objects.

For a detection task, we are interested in the expectation of g conditional on hypothesis Hj, which we can write as

g¯¯¯jggHj=H0Af+ngA,fAf,HjfHj (32)

with the three overbars denoting the three levels of averaging. Note that we do not need to write the inner expectation as conditioned on hypothesis Hj; when we have fixed Inline graphic and f, it does not matter what ensemble they were drawn from.

The innermost expectation is easy since it is just an average over the zero-mean noise vector n, which is the only thing that is random when we condition on Inline graphic and f with ℋ0 assumed known. Thus, the result of the inner expectation is

g¯(A,f)H0Af+ngA,f=H0Af. (33)

The next level of averaging yields

g¯¯j(f)g¯(A,f)Af,Hj=H0AfAf,Hj=H0AfAf,Hj (34)

where the last step follows since ℋ0 is a nonrandom linear operator. The subscript notation Inline graphic|f, Hj in (34) indicates the average over Inline graphic conditioned on both f and Hj, so the result depends on both and the f index j.

Finally, the overall (triple-bar) average is

g¯¯¯j=g¯¯j(f)fHj=H0AfAf,HjfHj. (35)

Here, an additional average over f conditioned on Hj has been performed, so the result depends on j but not f. Thus, g¯¯j(f) loses its argument f and acquires another overbar in going from (34) to (35).

We might be able to assume that the attenuation and scattering are statistically independent of the activity distribution, though this is not guaranteed since they arise from the same patient. Sometimes, for example, a tumor seen in PET or SPECT may be denser than the surrounding normal tissue. We might also be able to assume that the statistics of Inline graphic are the same for signal-present and signal-absent cases. With those assumptions

g¯¯¯j=H0AffHj (36)

which shows that the average image for hypothesis Hj in a nonadaptive system is an overall average system operator ℋ̄ ≡ ℋ0Inline graphic〉, acting on the average object j ≡ 〈ff|Hj.

For classification tasks we need the covariance matrix of the data under hypothesis Hj. A decomposition like (28) shows that

KgHj=K¯¯gHj(noise)+K¯g¯Hj(sys)+K¯¯g¯Hj(obj). (37)

The first term accounts for the Poisson noise in g. For a fixed object and system, this noise is Poisson and its covariance matrix is diagonal, and it remains diagonal after averaging Inline graphic and f. When these quantities are statistically independent as in (36), we find [32]

K¯¯gHj(noise)=[gg¯(A,f)][gg¯(A,f)]tgA,fAf,Hjf,Hj=diag[H0AAHjffHj] (38)

where diag[v] denotes a diagonal matrix with vector v as its diagonal elements.

The second term accounts for the randomness in the imaging system, here the result of attenuation and scatter; it is given by

K¯g¯Hj(sys)=[g¯(A,f)g¯¯(f)][g¯(A,f)g¯¯(f)]tAf,HjfHj=H0[AA¯j]ff[AA¯j]Af,HjfHjH0 (39)

where Inline graphic¯j ≡ 〈 Inline graphicInline graphic|f,Hj and H0 is the adjoint or backprojection operator (corresponding to the transpose for a real matrix). Note that K¯g¯Hj(sys) vanishes if Inline graphic is not random, either because attenuation and scatter are negligible (as they might be in mouse studies) or because they have been accurately measured (for example, in dual-modality SPECT/CT). Any inaccuracies in the measurement of Inline graphic must, however, be accounted for with the system covariance term.

The final term in the covariance decomposition accounts for object variability. If is Inline graphic independent of f, it is given by

Kg¯¯Hj(obj)=[g¯¯(f)g¯¯¯][g¯¯(f)g¯¯¯]tf,Hj=H0A¯jKfHjA¯jH0 (40)

where Inline graphicf|Hj is the integral operator whose kernel is the spatial autocovariance function of the object [32].

V. Statistical Analysis of Linear Observers

In this section, we apply the results of Section IV to the analysis of linear discriminants and estimators. For simplicity, we assume throughout that the activity distribution f and the attenuation and scattering properties contained in Inline graphic are statistically independent, and that f can be decomposed into statistically independent signal and background components.

For detection problems, we assume further that the signal is weak in the sense that it does not influence variances or covariances appreciably and that an adaptive system responds just to the background. For estimation of parameters associated with the signal, however, we consider the contribution of the signal to the covariances, and we allow the system to adapt to the signal.

A. Linear Discriminants for Nonadaptive Systems

For a nonadaptive system and a general template w, the numerator in the SNR expression (10) is obtained from (36); with (8) and the assumptions listed above, we have

Δtt1t0=wtH0Af¯sig (41)

where the constant c in (8) has cancelled.

The denominator of SNR2 is the variance of t, now assumed to be the same under the two hypotheses; it contains three terms, one for each term in the covariance expression (37):

Var(t)=wtKgw=wt[K¯¯g(noise)+K¯g¯(sys)+Kg¯¯(obj)]w (42)

where again the constant c is irrelevant as it does not influence the variance.

Thus

SNRw2=[wtH0Af¯sig]2wtKgw. (43)

The optimal or Hotelling template wHot can be derived by several well-known methods; here we sketch a variational derivation that will be extended to adaptive imaging below.

If we differentiate (43) with respect to w, use the identities ∂(wta)/∂w = a and ∂(wtAw)/∂w = 2Aw, and set the result to zero, we obtain

Kgw=[wtKgwwtH0Af¯sig]H0Af¯sig. (44)

The factor in large brackets is a scalar, so

w=[wtKgwwtH0Af¯sig]Kg1H0Af¯sig. (45)

Moreover, the scalar factor affects the magnitude but not the direction of the vector w, and SNRw2 is invariant to changes in magnitude. Therefore, we can drop the scalar factor and write the optimum template as

wHot=Kg1H0Af¯sig. (46)

With this template, the numerator in (43) is the square of the denominator, and we find

SNRHot2=f¯sigA¯H0Kg1H0Af¯sig. (47)

B. Linear Discriminants for Adaptive Systems

With the general adaptive template w(gs) as in (9) and the assumptions listed at the beginning of this section, the difference of means, which appears in the numerator of SNR2, becomes [cf. (41)]

Δt=wt(gs)H0(gs)gsfbgfbgAf¯sig (48)

where the term c(gs) has cancelled on the assumption that the scout system responds just to the background.

For deriving the optimum template, it is convenient to recast (48) by use of Bayes’ rule: pr(gs|fbg)pr(fbg) = pr(fbg|gs)pr(gs). Thus

Δt=wt(gs)H0(gs)fbggsgsAf¯sig=wt(gs)H0(gs)gsAf¯sig (49)

where the last step follows because the quantity being averaged does not depend explicitly on fbg. The remaining average in (49) could be computed, in principle, by choosing many backgrounds at random, constructing the corresponding noisy gs by (4), forming wt(gs) and ℋ0(gs), and then averaging. The randomness of the background and the noise in the scout image thus appear, but the average is with respect to the resulting distribution of gs.

If we put the average over gs last, the variance of the test statistic can be written as

Var(t)=[t(g,gs)t¯¯]2ggsgs. (50)

We can define a partial average of the test statistic (under hypothesis H0) by

t¯(gs)=c(gs)+wt(gs)gggs=c(gs)+wt(gs)g¯(gs) (51)

where we have used (9). The quantity (gs) is the mean of g with respect to the conditional densityn pr(g|gs), which implicitly includes the conditional randomness from n, Inline graphic, and f; it gets a single overbar here because we do not yet need to decompose the conditional average into three separate averages. Explicitly, with our current assumptions

g¯=(gs)=H0(gs)Af¯bg(gs) (52)

where bg(gs) is the posterior mean background object, obtained from the density pr(fbg|gs).

Adding and subtracting the partial average (gs), as in Section IV-A, we obtain

Var(t)=[t(g,gs)t¯(gs)]2ggsgs+[t¯(gs)t¯¯]2gs (53)

where again the cross terms vanish by construction.

The second term in (53) is zero if we take

c(gs)=wt(gs)g¯(gs) (54)

in which case t¯(gs)=t¯¯=0. Neither the difference in means nor the first term in (53) depends on c(gs), so (54) is the optimum choice; choosing (gs) = 0 avoids spreading out the density pr(t|H0) when considering systems with different (gs), for example systems that collect differing numbers of photons.

With (54), (53) becomes

Var(t)=wt(gs)[gg¯(gs)][gg¯(gs)]tggsw(gs)gs=wt(gs)Kggsw(gs)gs (55)

where we have used [atb]2 = atbbta.

We have actually already computed Kg|gs in detail because fixing gs removes the randomness in g arising from the scout system, though it also replaces averages with conditional averages. By appropriately modifying (37)–(40), we obtain

Kggs=K¯¯ggs(noise)+K¯g¯gs(sys)+Kg¯¯gs(obj) (56)

where with the assumptions made at the beginning of this section

K¯¯ggs(noise)=diag[H0(gs)Af¯(gs)] (57)
K¯g¯gs(sys)=H0(gs)[AA¯]fffgs[AA¯]tAH0(gs) (58)

and

Kg¯¯gs(obj)=H0(gs)A¯KfgsA¯H0(gs). (59)

If the adaptation rule ℋ0(gs) and the template function w(gs) are specified, then in principle (49) and (56)–(59) can be used to compute the SNR. In practice, actually performing the calculation in this manner would require knowledge of the posterior mean object (gs) and the posterior autocovariance function Inline graphicf|gs.

We shall return to the question of obtaining or estimating the posterior mean object and covariance function in the next section. For now, we assume that all requisite means and covariances are known, and we use this information to find the Hotelling observer for adaptive imaging. The procedure is to retrace the steps used in Section V-A to derive the nonadaptive Hotelling observer, but with the appropriate adaptive expressions (49) and (55) for the difference in means and variance, respectively, of the test statistic.

In the nonadaptive case, the difference in means had the form wta and the variance had the form wtAw so we used the identities ∂(wta)/∂w = a and ∂(wtAw)/∂w = 2Aw. In the adaptive case, we must perform a functional or Fréchet derivative (see [32, Sec. 15.3.5]). Differentiating a scalar-valued function with respect to the M × 1 vector w gave another M × 1 vector, and differentiating with respect to the vector-valued function w(gs)gives another vector-valued function. Specifically [32]

w(gs)wt(gs)a(gs)gs=w(gs)dgspr(gs)wt(gs)a(gs)=a(gs)pr(gs) (60)

and

w(gs)wt(gs)A(gs)w(gs)gs=2A(gs)w(gs)pr(gs). (61)

When we now retrace the steps that led to (46), all factors of pr(gs) cancel, and the optimal template is given by

wHot(gs)=Kggs1H0(gs)Af¯sig. (62)

The randomness of the measurement noise, object, and system are included in Kg|gs, but the noise in the scout image does not enter explicitly because ℋ0(gs) is presumed to be known precisely by the ideal observer.

With this template, the numerator in SNR2 is again the square of the denominator, and we find [cf. (47)]

SNRHot2=f¯sigA¯H0(gs)Kggs1H0(gs)gsAf¯sig. (63)

If the scout image provides information about either the signal or the attenuation and scatter, this expression remains valid if we simply replace and Inline graphic¯ and sig; with posterior means and move them inside the expectation brackets.

C. Linear Estimators

In this subsection, we shall derive the optimum linear estimator for the nonadaptive case and then extend the procedure to adaptive imaging. In both cases, a two-step procedure is used. First, the affine term in the general linear estimator [a in (15) or a(gs) in (18)] is chosen to make the estimator globally unbiased, and then the estimator matrix W or W(gs)) is chosen to minimize the EMSE.

In the nonadaptive case, the estimator is globally unbiased if

θ^(g)gθθ=θ¯ (64)

which requires that

θ^(g)=θ¯+Wt(gg¯¯) (65)

Where g¯¯=ggθθ.

The EMSE can be stated as

EMSE=θ^(g)θ2gθθ=tr[θ^(g)θ][θ^(g)θ]tgθθ (66)

where we have used ||x||2 = xtx = tr xxt.

The expectation in (66) is not a covariance because θ is not the mean of θ̂(g). We can, however, add and subtract the mean θ̄ in each factor and do some algebra to obtain

EMSE=trWt(gg¯¯)(gg¯¯)tgθθW2tr(θθ¯)(gg¯¯)tgθθW+tr(θθ¯)(θθ¯)tθ. (67)

Unlike all of the covariance expansions above, the cross term does not cancel in this EMSE expansion.

Each of the terms in (67) is a covariance or cross covariance, and we can write

EMSE=trWtKgW2trKθgW+trKθ. (68)

Differentiating the EMSE with respect to the matrix W yields

ddWEMSE=2KgW2Kθgt (69)

where we have used the identities ∂[trWtKW]/∂W = 2KW (valid if K is real and symmetric) and ∂tr[BW]/∂W = Bt. The derivative in (69) must be zero for the optimum W, and it follows that the Wiener estimator is indeed given by (16) and the optimized EMSE is given by (17).

In the adaptive case, a general estimator of the form (18) is globally unbiased if

θ^(g,gs)g,gs=a(gs)gs+Wt(gs)gg,gs=θ¯. (70)

This condition will hold if

θ¯(g,gs)=θ¯(gs)+Wt(gs)[gg¯¯(gs)] (71)

Where

g¯¯(gs)=ggθ,gsθgs (72)

and similarly, θ̄(gs) = 〈θθ|gs. Notice that the outer average on the right-hand side of (72) must be calculated with the posterior PDF of θ conditioned on the scout data gs.

The EMSE of a globally unbiased estimator for adaptive data is given by [cf. (67)]

EMSE=trWt(gs)KggsW(gs)gs2trKθggsW(gs)gs+trKθ. (73)

The Fréechet derivative now yields a matrix-valued function [cf. (60) and (61)]

W(gs)EMSE=[2KggsW(gs)2Kθggst]pr(gs). (74)

This derivative will be zero if W(gs)=Kggs1Kθggst, which means that the Wiener estimator is

θ^(g,gs)=θ¯(gs)+KθggsKggs1[gg¯¯(gs)]. (75)

This estimator is identical to the nonadaptive Wiener estimator of (16) except that all averages are computed with PDFs conditional on the scout image.

The resulting EMSE is given by [cf. (17)]

EMSE=trKθtrKθggsKggs1Kθggstgs. (76)

To get more explicit expressions for the means and covariances in (75) and (76), let us assume that the object can be separated into background and signal parts and that the parameter θ is associated only with the signal (perhaps a tumor). Thus

f=fbg+fsig(θ). (77)

Continuing to assume that Inline graphic is independent of the object and not influenced by the scout image, we get

g¯¯(gs)=H0(gs)A¯[f¯bg(gs)+f¯sig(gs)] (78)

where sig;(gs) = 〈fsig(θ)〉θ|gs

The posterior cross covariance, which expresses the sensitivity of the data to changes in the parameters, is given by

Kθggs=[θθ¯(gs)]{H0(gs)A¯[fsig(θ)f¯sig(gs)]}tθgs=KθfgsA¯H0(gs) (79)

Where Inline graphicθf|gs is an operator defined by

Kθfgs[θθ¯(gs)][fsig(θ)f¯sig(gs)]θgs. (80)

This operator maps a function of continuous spatial variables in object space to a P × 1 vector in parameter space.

Finally, the posterior covariance Kg|gs, which accounts for measurement noise, system randomness and object variability, can still be expressed by (56) if we choose to include randomness in both the background and the parameterized signal in the object term. If, however, we use separate posterior averaging steps for background and signal, (56) takes the form

Kggs=K¯¯ggs(noise)+K¯g¯gs(sys)+Kg¯¯gs(bg)+Kg¯¯gs(sig). (81)

VI. Implementation and Assessment

In this section, we examine some of the issues that arise in implementing the mathematics above in practice and assessing the improvements attained.

A. Goals and Figures of Merit

The goal of adaptive imaging is to choose an adaptation rule ℋ0(gs) in order to maximize the performance of a detection or estimation task. For a detection task, we take the performance figure of merit as the Hotelling SNR2, given by (63) along with the posterior covariance decomposition of (56)–(59). To simplify the expressions, we ignore the system covariance term (58) by assuming that the attenuation and scatter are known to the observer, and we include the attenuation and scatter in the adapted system operator, ℋ(gs) = ℋ0(gs) Inline graphic. Then we can rewrite the FOM as

FOMdet=SNRHot2=FOMdet(gs)gs (82)

where

FOMdet(gs)=[H(gs)f¯sig]t{diag[H(gs)f¯bg(gs)]+H(gs)KfgsH(gs)}1[H(gs)f¯sig]. (83)

The goal of adaptation for an estimation task is to minimize the EMSE given in (76). The first term in that equation, trKθ, expresses the prior knowledge, before the scout image, so we want to make trKθggsKggs1Kθggst as large as possible, and this term can be taken as the FOM for estimation. Thus, by analogy to (82) and (83), we write

FOMest=FOMest(gs)gs (84)

where, with (79) and (81)

FOMest(gs)=trKθggsKggs1Kθggst=tr[KθfgsH(gs)]×{diag[H(gs)(f¯bg(gs)+f¯sig(gs))]+H(gs)[Kfgs(bg)+Kfgs(sig)]H(gs)}1×[H(gs)Kθfgs]. (85)

The trace operator (tr) applies to the P × P matrix that follows, with being the number of parameters to be estimated. (Recall that Inline graphicθf|gs(gs) is a P × M matrix, and of course the inverse covariance matrix is M × M.)

A scout imaging system and an adaptation rule will thus be effective in increasing task performance, compared to a nonadaptive system, if they either 1) increase the signal ℋ (gs)sig in (83), 2) increase the sensitivity of the data to the parameters as expressed by Inline graphicθf|gs(gs) in (85), or 3), reduce one of the covariance terms in either equation. With these observers, there are no other ways to improve task performance.

B. Prior Ensemble Information

Before making any measurements on the particular subject of interest, we will almost always have available some information about the ensemble from which that subject is drawn. One important piece of information is the prior mean , which enters into the noise covariance term [see (38)]. This 3-D function represents the average distribution of a given tracer for all patients or animals that might receive it. If we have previously imaged many subjects with this tracer and a well sampled but nonadaptive SPECT system, we can average the images to get an estimate of on a voxel grid. Because of the large intersubject variability, we anticipate that this estimate will be diffuse and relatively nondescriptive of the particular subject to which we wish to adapt.

We might also have some prior information about the fine structure or texture of the tracer distribution. Considerable work has been done on describing the fine structure seen in medical images, using models such as the lumpy or clustered lumpy backgrounds, Markov random fields, and wavelet-based descriptions (see [32, Ch. 8] and references cited therein). If one of these models is known to be applicable to the tracer under consideration, it will give a good approximation to the object covariance term in (40). If the model is applicable in general but contains some unknown parameters, a method described by Kupinski et al. [49] can be used to estimate the parameter values from a set of sample images.

Of course, object covariance also has contributions from large-scale variations in addition to the fine structure. Most texture models in the literature are applicable to spatially stationary random processes, but a quasistationary model will usually be an improvement. The approach here is to express the object autocovariance function, which is the kernel of the integral operator in (40), in the approximate form (see [32, Sec. 8.2])

k(r1,r2)=a(Δr)b(r0) (86)

where r0 = (1/2) (r1 + r2) and Δr = r1r2. The function a(Δr) describes the rapidly varying fine structure and can be represented by mathematical texture models as mentioned above, while b(r0), representing large-scale variations, can be estimated from sample images.

In tumor-detection problems, we will probably have a good idea what a tumor looks like, so we will have some representation for the function sig. Similarly, for estimation of tumor parameters, we will necessarily have to assume some prior mean and covariance of the parameters in order to employ Wiener estimation in the first place.

C. Scout Strategies

A scout image should use up only a small fraction of the available imaging time, so it will necessarily be limited in the number of projection angles and/or the exposure time per projection compared to the final, adapted system. To obtain reasonable image statistics with this limitation, it will probably be necessary to use relatively large pinholes. With these restrictions, what can we learn about the object?

One thing we can do with the scout system is to refine our estimate of the prior mean , which appears in the noise covariance term. If we take short-exposure projection images with large pinholes and an adequate number of projection angles, a heavily smoothed reconstruction should provide a much better approximation to f than the prior mean . Of course, we do not know the actual object, but we can regard the smoothed scout reconstruction as an estimate of the posterior mean (gs) (where the word “posterior” in this context means after the scout image but still prior to the final measurement).

We can also try to refine our estimate of the object covariance terms. For example, with the quasistationary model (82), we can estimate the long-range factor b(r0) from a smoothed scout reconstruction, thereby reducing the object covariance term even if we rely on the prior model for a(Δr).

Another thing we can get from a scout image is an estimate of the sparsity of the tracer distribution. A sparse object in nuclear medicine is one where the tracer is concentrated within a volume that is small compared to the field of view of the system.

If we know the object is sparse, it is probably advantageous to use a large number of pinholes, thereby reducing the relative size of the noise term in the covariance and increasing the norm of the signal, but the performance advantage is much less for large, diffuse objects [50]–[53]. In addition, sparse objects can be more readily reconstructed from a few projections than can nonsparse ones [54], [55], so fewer projection angles can be used in that case.

A quantitative measure of degree of object sparseness is its entropy, defined by

S=d3rp(r)ln[p(r)] (87)

where p(r) ≡ f(r)/∫ d3rf(r′) is the tracer distribution normalized as a probability density function. Estimates of can be obtained from a low-resolution reconstruction or directly from the projection images in the scout data set.

From a reconstructed scout image, we can also estimate the overall support of the tracer distribution, defined as the set of points over which f(r) is nonzero, or where it exceeds some small threshold. If this support set is small, we can use large magnifications in the final image without truncating the projections, and if we know the 3-D coordinates of the centroid of the support, we can choose the pinhole location in each projection view to center the support on the detector.

For a pure detection problem, we cannot use the scout image to localize the signal as this would imply that the task could be performed from the scout data alone, obviating the adaptation. If, however, we wish to estimate the volume or activity of a previously detected tumor, say to assess response to therapy, we can use the scout image to get an estimate of the tumor location. Then we can adjust the magnification and pinhole locations so that the tumor projections optimally fill the detector field, thereby increasing cross-covariance Inline graphicθg and reducing the signal covariance term in (81). For this task, we might allow truncated projections so long as the final adapted system operator ℋ (gs) is correctly modeled and used in the estimation.

A different scout strategy can be used for rapid dynamic studies, where the task might be estimation of the activity in a region of interest as a function of time. We can administer a low-activity bolus of tracer and image it with large pinholes and a small number of projection angles (maybe just two) in order to determine the general trajectory of the bolus and how rapidly the activity varies in the region of interest. With this prior information, we can configure the system for optimal estimation performance and then administer a larger bolus for the final data acquisition.

D. Modes of Adaptation

There are several distinct ways in which the information from the scout images can be used to optimize the final imaging system. All of them are based on approximating the system operator ℋ with a matrix H and making further approximations in order to compute figures of merit rapidly. These computational issues are discussed below, but here we discuss the kinds of system variations to be considered in choosing the final system.

1) Choice Among Known Systems

First, we might have some relatively small number of candidate systems and merely want to choose the one that will give best task performance for a particular subject. For example, if the only variable is the magnification, we can precompute the H matrices for a series of magnifications and then compute the corresponding figures of merit.

2) Linear Combinations of Known Systems

If we have multiple precalibrated systems with known H, we can consider using combinations of them. There are two distinct forms of linear combination, which we can refer to as operator addition and concatenation. As an example, suppose we have pinhole plates with shutters, as mentioned in Section II for the adaptive prototype. Each pinhole in a plate can have its own, precomputed H matrix. If some subset of the pinholes is open for the entire exposure, we simply add the matrices. If, however, the shutters operate L times during the exposure and J projection angles are used, and the detector has K pixels and the object representation uses N voxels, the final concatenated matrix is JKL × N.

Another example of concatenation is when we choose some combination of projection angles, each of which has its own precomputed H. If the object representation consists of N voxels, and J angles are selected for a detector with K pixels, then each individual matrix Hj is K × N, but the overall matrix with concatenated rows is JK × N.

3) One-Parameter Continuous Optimization

As noted in Section II-A, key parameters in pinhole SPECT include the pinhole diameter and the magnification. If multiple pinholes are used but they are all constrained to have the same diameter D , then D is an important single parameter to vary for best task performance. Similarly, if all projection angles are constrained to the same magnification, then that is an important parameter to vary.

4) Multiparameter Optimization

In the example just given, the pinhole diameter and the magnification can be varied together, giving us a two-parameter optimization space. Similarly, if we use J projection angles and vary the exposure time and magnification at each, we have a total of 2J parameters to vary, greatly increasing the computational complexity of the optimization.

5) Empirical Rules

Rather than trying to find a true optimum system configuration, we can also develop empirical rules that allow us to choose broad system characteristics. One example already mentioned is to choose the degree of multiplexing based on some measure of sparsity of the object. Another example would be a rule for choosing the allowable degree of image truncation for an estimation task where we try to fill the detector field of view by placing a pinhole close to the signal of interest [30].

E. Computational Issues

Methods for computing task-based figures of merit in nonadaptive imaging systems are discussed in [32, Ch. 14]. The greatest difficulty is that the covariance matrices are enormous (M × M, where M is the total number of measurements, ~105 – 106 in SPECT), but there are two important techniques that allow a huge dimensionality reduction, often with little loss of accuracy in the final estimates of figures of merit.

First, we can use spatial-frequency-selective channels, similar to those known to exist in the human visual system, and use the channel outputs as new data values. The dimensionality reduction can be extreme; with so-called efficient channels, as few as 5–10 channels sometimes give excellent approximations for figures of merit [56], [57], and the size of the covariance matrix is then just 5 × 5 or 10 × 10.

The second trick, especially useful for the object term in the covariance expansion, is to use a simulation program to create noise-free images of randomly generated objects through a specified H matrix. The object covariance is then estimated by a low-rank sample covariance matrix, but the overall covariance matrix remains invertible because of the diagonal noise term. The Woodbury matrix inversion lemma [58] can then be used to reduce the size of the matrix to be inverted to S × S, where S is the number of sample images (~100–1000), instead of M × M.

The computational tasks needed to optimize system performance for an adaptive system are not fundamentally different from those required for any task-based system optimization, but they have to be done very rapidly in order to be able to make system changes before the patient leaves the room, and they have to be done many times in any iterative optimization scheme. The tools that give us any hope of meeting these demands are precomputation, interpolation, approximation, and parallelization.

1) Precomputation

As noted in Section VI-D, H matrices for candidate systems can be precomputed or measured, but other key components of the FOMs can be precomputed as well. For example, a voxel representation of the object covariance function of (86) can be computed ahead of time, and the slowly varying factor can then be modified after the scout data are obtained.

Similarly, the cross-covariance operator Inline graphicθf|gs defined in (80) can be computed in a voxel representation and stored as a set of images, one for each of the parameters to be estimated. These images are readily modified after preliminary parameter estimates are obtained from the scout data; for example, in tumor-volume estimation, a refined estimate of the location of a tumor requires only a recentering of the images that make up the rows of Inline graphicθf|gs.

2) Interpolation

Suppose the H matrix is known from precomputation or measurements for a sparse set of projection angles, but the optimization algorithm needs matrices for intermediate angles. It is easy to see that conventional linear or spline interpolation does not work by considering a single pinhole; a point in object space produces a blob at location r1 on the detector for projection angle φ1 and another blob of different size or shape at location r2on the detector for projection angle φ2. Linear interpolation would produce a linear combination of the two blobs, while the correct H matrix for an intermediate angle would exhibit a single blob at an intermediate location.

Chen [45], [49] has developed two algorithms to overcome this problem and provide highly accurate interpolated H matrices. One of them describes the blobs parametrically and interpolates the parameters, and the other uses 2-D Fourier transforms of the blob images and interpolates the Fourier magnitude and phase.

Similarly, if H is known for a restricted set of pinhole diameters, interpolation schemes can be devised to obtain it for intermediate diameters.

These interpolation schemes are particularly valuable when we wish to optimize the system over a continuous range of parameters. The only alternative to interpolation of the H matrix in this case is to represent it in a parametric form and interpolate the parameters. For example, we could use a geometrical model parameterized by the distances from the center of rotation to the pinhole plate and from the pinhole plate to the detector. When combined with a measured model of the scintillation detector, this geometrical model would then provide an overall H that is readily evaluated for any desired value of the two distances. This approach requires a high degree of mechanical precision (which we indeed tried to build in to the adaptive prototype) and distortion-free performance of the scintillation camera (which we achieve by using nearly unbiased maximum-likelihood position estimation).

3) Approximation

After the scout data are obtained, the goal is to modify the final imaging system so as to optimize the figure of merit (83) or (85). Doing so requires not only operational approximations to the posterior means and covariances, but also very rapid ways of performing the necessary inverses. As with nonadaptive systems, channels and use of the matrix inversion lemma can be enormously helpful, but further computational benefit can be obtained by replacing the actual FOM, (83) or (85), with some surrogate function which is easier to calculate. If it can be established (by offline simulation studies, for example) that system modifications that increase the surrogate function also increase the actual task performance, then the real-time computational requirements are eased.

A prime example of this strategy is when the scout image is obtained with large pinholes and short exposures at a large number of projection angles. A rapid, highly regularized reconstruction can be performed to get a blurred version of object, and the surrogate optimization function can be obtained by taking this reconstruction as the mean object and neglecting the object covariance term completely. In other words, a BKE (background known exactly) is used as a surrogate for a BKS (background known statistically) task. The requisite matrix inverse in this case is very easy because the noise covariance matrix is diagonal and readily computed for any proposed system config-uration.

Other potential surrogate functions include expressions based on a Neumann series expansion of the overall covariance (see [32, Ch. 14] and ones based on Fisher information [60]; neither of these methods requires inversion of nondiagonal matrices.

4) Parallelization

The search for an optimum—or at least improved—system configuration requires evaluating the figure of merit for many candidate systems. If multiple processors with adequate memory are available online, the computations can be performed in parallel.

F. Performance Validation

The figures of merit in (82) and (84) are defined as posterior averages, where here the term posterior refers to ensembles of objects consistent with the scout image. Averages with respect to this posterior ensemble are decidedly Bayesian in the sense that we cannot know the PDF pr(f|gs) analytically; we must regard it as a measure of our degree of belief in the properties of the object after we have obtained the scout image.

In a pure Bayesian paradigm, only this subjective density would be used in assessing different adaptation strategies. By contrast, a pure frequentist view would require that all probabilities be regarded as frequencies of occurrence, which is surely not possible for either the prior pr(f) or the posterior pr(f|gs), both of which are infinite-dimensional densities.

To resolve this dilemma, we adopt a hybrid view espoused in [32]; we allow subjective or approximate densities in performing the task, justifying them where we can by appeal to experiment, but in the end we estimate performance metrics by repeated experiments.

The experiments are best carried out in simulation. If we have an accurate model for the prior object class, including variations in object size and shape, object texture or signal parameters, we can create multiple sample objects from this prior on a voxel grid. We can then evaluate a particular combination of scout system ℋs and adaptation rule ℋ0(gs) by generating scout and final images for each sample object and performing the task. For a detection task, that means generating many sample objects with and without signals, computing the test statistic t(g,gs) for each, and then either estimating SNR2 from the definition (10) or varying the decision threshold and generating a sample ROC curve. For an estimation task, many random backgrounds would be generated, each with a signal having a parameter vector θ drawn from the prior pr(θ), and forming an estimate θ̂(g,gs) for each; the EMSE can then be estimated directly from its definition, (14).

Obviously the approach just outlined requires an enormous amount of computation, but it is offline and does not have to done as part of the adaptation process.

VII. Summary and Conclusion

Task-based methods provide rigorous definitions of image quality, which can be used to optimize a system for a predefined ensemble of patients. In this paper, we have shown that preliminary scout measurements on a given patient can be used to narrow down the ensemble and hence reduce task-related uncertainty.

We considered in detail two specific linear observers, the ideal linear detector known as the Hotelling observer and the ideal linear estimator known as the generalized Wiener estimator. For nonadaptive systems, we showed that the performance of these observers is limited by three sources of randomness in the data, which in SPECT are mainly Poisson measurement noise, randomness in the radionuclide distribution being images, and unknown attenuation and scatter in the patient’s body. We showed that the data covariance matrix could be decomposed rigorously into a sum of three component matrices, one associated with each source of randomness.

With this decomposition, we derived expressions for the figures of merit for the Hotelling and Wiener observers for nonadaptive systems, and then we showed how to extend the derivation to account for the adaptation. The expressions found for the observers and their figures of merit for the adaptive case were surprisingly similar to those for nonadaptive SPECT, but the mean vectors and covariance matrices had to be interpreted as posterior averages, conditional on the scout image, in the adaptive case.

Some of the practical issues in implementing adaptive SPECT and assessing its value were discussed qualitatively; more detailed studies based on this theory and the two imaging systems described in Section II will be published separately.

Two important issues that were not treated here are the use of nonlinear observers for detection or estimation and the problem of assessing and optimizing systems for joint detection-estimation tasks.

Nonlinear observers, such as the Bayesian ideal observer for detection or the maximum-likelihood estimator for estimation tasks are harder to analyze than the linear observers treated here. There is no closed-form rule for performing the task and no counterpart to the linear template or estimation matrix, so the Fréchet derivative on which we relied is not applicable. Nevertheless, there are well-established methods for implementing nonlinear observers and assessing the outcome, and these methods are applicable even when different systems are used with different objects as in the adaptive paradigm proposed here. A reasonable way of evaluating adaptation rules and scout strategies for nonlinear observers, therefore, would be to select several candidate adaptation methods that exhibit good performance with linear observers and then evaluate them with nonlinear observers; the approach to performance validation outlined in Section VI-F is still applicable with nonlinear observers.

Recent work by Clarkson [61] provides a basis for extending the methodology of this paper to joint detection-estimation tasks. He extended the familiar concept of localization receiver operating characteristic (LROC) curves to estimation ROC (EROC) curves and determined the form of the optimum observer for maximizing the area under the EROC curve. For Gaussian noise assumptions, this optimum observer is the scanning Hotelling observer, now scanned over all parameters of interest and not just over spatial coordinates as in the LROC problem. The adaptive Hotelling observer derived in the present paper can thus be extended to any detection-estimation problem, and the area under the EROC curve is a relevant figure of merit for assessing scout strategies and adaptation rules.

Finally, we note that the methodology in this paper is applicable, in principle, to any imaging system, and it is also applicable to multimodality imaging.

Acknowledgments

This work was supported by the National Institutes of Health under Grant P41 EB 002035 and Grant R37 EB 000803.

The authors would like to thank K. Myers for numerous interactions on the subjects of image quality and adaptive imaging and C. Dainty and N. Devaney for discussions on adaptive optics.

References

  • 1.Beckers JM. Adaptive optics for astronomy: Principles, performance, and applications. Annu Rev Astron Astrophys. 1993;31:13–62. [Google Scholar]
  • 2.Enard D, Marechal A, Espiard J. Progress in ground-based optical telescopes. Rep Prog Phys. 1996;59:601–656. [Google Scholar]
  • 3.Tyson RK. Principles of Adaptive Optics. Boston, MA: Academic; 1998. [Google Scholar]
  • 4.Roddier F, editor. Adaptive Optics in Astronomy. Cambridge, U.K: Cambridge Univ. Press; 1999. [Google Scholar]
  • 5.Guyon O. Limits of adaptive optics for high-contrast imaging. As-trophys J. 2005;629:592–614. [Google Scholar]
  • 6.Rousset G. Wavefront sensing. In: Roddier F, editor. Adaptive Optics in Astronomy. Cambridge, U.K: Cambridge Univ. Press; 1999. [Google Scholar]
  • 7.Gonsalves RA. Phase retrieval and diversity in adaptive optics. Opt Eng. 1982;21(5):829–832. [Google Scholar]
  • 8.Paxman RG, Schulz TJ, Fienup JR. Joint estimation of object and aberrations using phase diversity. J Opt Soc Amer. 1992;7:1072–1085. [Google Scholar]
  • 9.Muller RA, Buffington A. Real-time correction of atmospherically degraded telescope images through image sharpening. J Opt Soc Amer. 1974;64(9):1200–1210. [Google Scholar]
  • 10.Kim DH, Kolesnikov K, Kostrzewski AA, Savant GD, Vasiliev AA, Vorontsov MA. Adaptive imaging system using image quality metric based on statistical analysis of speckle fields. In: Casasent DP, Tescher AG, editors. Proc SPIE. Vol. 4044. Jul, 2000. pp. 177–186. [Google Scholar]
  • 11.Ng GC, Freiburger PD, Walker WF, Trahey GE. A speckle target adaptive imaging technique in the presence of distributed aberrations. IEEE Trans Ultrason Ferroelect Freq Control. 1997 Jan;44(1):140–151. doi: 10.1109/58.585209. [DOI] [PubMed] [Google Scholar]
  • 12.Gauss RC, Trahey GE, Soo MS. Adaptive imaging in the breast. Proc. Int. Symp. IEEE Ultrason. Symp.; Caesars Tahoe, NV. 1999. pp. 1563–1569. [Google Scholar]
  • 13.Fernandez AT, Dahl JJ, Gammelmark K, Dumont DM, Trahey GE. High resolution ultrasound beamforming using synthetic and adaptive imaging techniques. 2002 IEEE Int. Symp. Biomed. Imag.; Washington, DC. 2002. pp. 433–436. [Google Scholar]
  • 14.Li PC, Li ML. Adaptive imaging using the generalized coherence factor. IEEE Trans Ultrason Ferroelect Freq Control. 2003 Feb;50(2):128–141. doi: 10.1109/tuffc.2003.1182117. [DOI] [PubMed] [Google Scholar]
  • 15.Huang SW, Li PC. Computed tomography sound velocity reconstruction using incomplete data. IEEE Trans Ultrason Ferroelect Freq Contol. 2004 Mar;51(3):329–342. [PubMed] [Google Scholar]
  • 16.Claudon M, Tranquart F, Evans DH, Lefèvre F, Correas JM. Advances in ultrasound. Eur Radiol. 2002;12:7–18. doi: 10.1007/s00330-001-1185-1. [DOI] [PubMed] [Google Scholar]
  • 17.Li ML, Huang SW, Ustiner K, Li PC. Adaptive imaging using an optimal receive aperture size. Ultrason Imag. 2005 Apr;27(2):111–127. doi: 10.1177/016173460502700205. [DOI] [PubMed] [Google Scholar]
  • 18.Cao Y, Levin DN. Feature-guided acquisition and reconstruction of MR images. In: Barrett HH, Gmitro AF, editors. Lecture Notes in Computer Science; Proceedings of the International Conference on Information Processing in Medical Imaging; New York: Springer; 1993. Springer pp. 687pp. 278–292. [Google Scholar]
  • 19.Zientara GP, Panych LP, Jolesz FA. Dynamically adaptive MRI with encoding by singular value decomposition. Magn Reson Med. 1994;32(2):268–274. doi: 10.1002/mrm.1910320219. [DOI] [PubMed] [Google Scholar]
  • 20.Zientara GP. Fast Imaging Techniques for Interventional MRI. In: Young I, Jolesz FA, editors. Interventional MR. ch. 2. London, U.K: Martin Dunitz; 1995. pp. 25–52. [Google Scholar]
  • 21.Cao Y, Levin DN. Using prior knowledge of human anatomy to constrain MR image acquisition and reconstruction: Half k-space and full k-space techniques. Magn Reson Imag. 1997;15(6):669–677. doi: 10.1016/s0730-725x(97)00027-1. [DOI] [PubMed] [Google Scholar]
  • 22.Zientara GP, Panych LP, Jolesz FA. Applicability and efficiency of near-optimal spatial encoding for dynamically adaptive MRI. Magn Reson Med. 1998 Feb;39(2):204–213. doi: 10.1002/mrm.1910390207. [DOI] [PubMed] [Google Scholar]
  • 23.Wendt M. Dynamic tracking in interventional MRI using wavelet-encoded gradient-echo sequences. IEEE Trans Med Imag. 1998 Oct;17(10):803–809. doi: 10.1109/42.736048. [DOI] [PubMed] [Google Scholar]
  • 24.Yoo S, Guttmann CRG, Zhao L, Panych LP. Real-time adaptive functional MRI. NeuroImage. 1999;10:596–606. doi: 10.1006/nimg.1999.0494. [DOI] [PubMed] [Google Scholar]
  • 25.Ablitt NA, Gao JX, Keegan J, Firmin DN, Stegger L, Yang GZ. Predictive motion modelling for adaptive imaging. Proc Int Soc Mag Reson Med. 2003;11:1570. [Google Scholar]
  • 26.Griffiths JA, Metaxas MG, Royle GJ, Venanzi C, Esbrand C, van der Stelt PF, Verheij H, Li G, Turchetta R, Fant A, Gasiorek P, Theodoridis S, Georgiou H, Cavouras D, Hall G, Noy M, Jones J, Leaver J, Machin D, Greenwood S, Khaleeq M, Schulerud H, Ostby J, Triantis F, Asimidis A, Bolanakis D, Manthos N, Longo R, Bergamaschi A, Speller RD. A multi-element detector system for intelligent imaging: I-ImaS. Conf. Rec. IEEE Med. Imag. Conf; Nov. 2006; pp. 2554–2558. [Google Scholar]
  • 27.Fant A, Gasiorek P, Turchetta R, Avset B, Bergamaschi A, Cavouras D, Evangelou I, French MJ, Galbiati A, Georgiou H, Hall G, Iles G, Jones J, Longo R, Manthos N, Metaxas MG, Noy M, Ostby JM, Psomadellis F, Royle GJ, Schulerud H, Speller RD, van der Stelt PF, Theodoridis S, Triantis F, Venanzi C. I-IMAS: A 1.5D sensor for high-resolution scanning. Nucl Instrum Meth Res. 2007;1573:27–29. [Google Scholar]
  • 28.Freed M, Kupinski MA, Furenlid LR, Barrett HH. Design of an adaptive SPECT imager. presented at the Acad. Molecular Imag. Annu. Conf; Orlando, FL. Mar. 2006. [Google Scholar]
  • 29.Freed M, Kupinski MA, Furenlid LR, Whitaker MK, Barrett HH. A prototype instrument for adaptive SPECT imaging. Proc SPIE. 2007;6510:6510–6530. doi: 10.1117/12.708818. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30.Freed M, Kupinski MA, Furenlid LR, Wilson DW, Barrett HH. A prototype instrument for single-pinhole small-animal adaptive SPECT imaging. Med Phys. 2008;35(5):1912–1925. doi: 10.1118/1.2896072. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 31.Barrett HH. Objective assessment of image quality: Effects of quantum noise and object variability. J Opt Soc Amer. 1990 Jul;7(7):1266–1278. doi: 10.1364/josaa.7.001266. [DOI] [PubMed] [Google Scholar]
  • 32.Barrett HH, Myers KJ. Foundations of Image Science. New York: Wiley; 2004. [Google Scholar]
  • 33.Hesterman JY, Kupinski MA, Furenlid LR, Wilson DW, Barrett HH. The multi-module, multi-resolution system—a novel small-animal SPECT system. Med Phys. 2007;34:983–987. doi: 10.1118/1.2432071. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34.Hesterman JY, Kupinski MA, Clarkson E, Barrett HH. Hardware assessment using the multi-module, multi-resolution system (M3R). A signal-detection study. Med Phys. 2007;34(7):3034–3044. doi: 10.1118/1.2745920. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 35.Hesterman J. Ph.D. dissertation. Univ. Arizona; Tucson: 2007. The multimodule multiresolution SPECT system: A tool for variable-pinhole small-animal imaging. [Google Scholar]
  • 36.Barrett HH, Myers KJ, Devaney N, Dainty C. Objective assessment of image quality. IV. Application to adaptive optics. J Opt Soc Amer. 2006;23:3080–3105. doi: 10.1364/josaa.23.003080. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 37.Meikle SR, Kench P, Kassiou M, Banati RB. Small animal SPECT and its place in the matrix of molecular imaging technologies. Phys Med Biol. 2005;50:R45–R61. doi: 10.1088/0031-9155/50/22/R01. [DOI] [PubMed] [Google Scholar]
  • 38.Kupinski MA, Barrett HH, editors. Small-Animal SPECT Imaging. New York: Springer; 2005. [Google Scholar]
  • 39.Schramm N, Ebel G, Engeland U, Schurrat T, Behe M, Behr T. High-resolution SPECT using multipinhole collimation. IEEE Trans Nucl Sci. 2003 Jun;50(3):315–320. [Google Scholar]
  • 40.Beekman FJ, van der Have F, Vastenhouw B, van der Linden AJA, van Rijk PP, Burbach JPH, Smidt MP. U-SPECT-I: A novel system for submillimeter-resolution tomography with radiolabeled molecules in mice. J Nucl Med. 2005;46:1194–1200. [PubMed] [Google Scholar]
  • 41.Funk T, Depres P, Barber WC, Shah KS, Hasegawa BH. A multipinhole small animal SPECT system with submillimeter spatial resolution. Med Phys. 2006;33(5):1259–1268. doi: 10.1118/1.2190332. [DOI] [PubMed] [Google Scholar]
  • 42.Zimmerman R, Moore S, Mahmood A. Performance of a triple-detector, multiple-pinhole SPECT system with iodine and indium isotopes. IEEE Nucl. Sci. Symp. Conf. Rec. 2004 IEEE; Rome, Italy. 2004. pp. 2427–2429. [Google Scholar]
  • 43.Kim H, Furenlid LR, Crawford MJ, Wilson DW, Barber HB, Peterson TE, Hunter WCJ, Liu Z, Woolfenden JM, Barrett HH. SemiSPECT: A small-animal SPECT imager based on eight CZT detector array. Med Phys. 2006 Feb;33:465–474. doi: 10.1118/1.2164070. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 44.Furenlid LR, Wilson DW, Yi-chun C, Hyunki K, Pietraski PJ, Crawford MJ, Barrett HH. FastSPECT II: A second-generation high-resolution dynamic SPECT imager. IEEE Trans Nucl Sci. 2004 Jun;51(3):631–635. doi: 10.1109/TNS.2004.830975. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 45.Chen Y. Ph.D. dissertation. Univ. Arizona; Tucson: 2006. System calibration and image reconstruction for a new small-animal SPECT system. [Google Scholar]
  • 46.Smith WE, Barrett HH. Hotelling trace criterion as a figure of merit for the optimization of imaging systems. J Opt Soc Amer. 1986;3:717–725. [Google Scholar]
  • 47.Fiete RD, Barrett HH, Smith WE, Myers KJ. Hotelling trace criterion and its correlation with human-observer performance. J Opt Soc Amer. 1987;4:945–953. doi: 10.1364/josaa.4.000945. [DOI] [PubMed] [Google Scholar]
  • 48.Hotelling H. The generalization of student’s ratio. Ann Math Stat. 1931;2:360–378. [Google Scholar]
  • 49.Kupinski MA, Clarkson E, Hoppin J, Chen L, Barrett HH. Experimental determination of object statistics from noisy images. J Opt Soc Amer. 2003;20:421–429. doi: 10.1364/josaa.20.000421. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 50.Barrett HH, DeMeester GD. Quantum noise in fresnel zone plate imaging. Appl Opt. 1974;13(5):1100–1109. doi: 10.1364/AO.13.001100. [DOI] [PubMed] [Google Scholar]
  • 51.Barrett HH, Horrigan FA. Fresnel zone plate imaging of gamma rays: Theory. Appl Opt. 1973;12(11):2686–2702. doi: 10.1364/AO.12.002686. [DOI] [PubMed] [Google Scholar]
  • 52.Barrett HH, Swindell W. Radiological Imaging. New York: Academic; 1981. [Google Scholar]
  • 53.Myers KJ, Rolland JP, Barrett HH, Wagner RF. Aperture optimization for emission imaging: Effect of a spatially varying background. J Opt Soc Am A. 1990;7(7):1279–1293. doi: 10.1364/josaa.7.001279. [DOI] [PubMed] [Google Scholar]
  • 54.Candes EJ, Tao T. Near-optimal signal recovery from random projections: Universal encoding strategies. IEEE Trans Inform Theor. 2006 Dec;52(12):5406–5425. [Google Scholar]
  • 55.Donoho DL, Johnstone IM, Hoch JC, Stern AS. Maximum entropy and the nearly black object. J Roy Stat Soc. 1992;B54:41–81. [Google Scholar]
  • 56.Barrett HH, Abbey C, Gallas BD, Eckstein M. Stabilized estimates of hotelling-observer detection performance in patient-structured noise. Proc SPIE, Med Imag 1998: Image Perception. 1998;3340:27–43. [Google Scholar]
  • 57.Gallas BD, Barrett HH. Validating the use of channels to estimate the ideal linear observer. J Opt Soc Amer. 2003;20:1725–1739. doi: 10.1364/josaa.20.001725. [DOI] [PubMed] [Google Scholar]
  • 58.Tylavsky DJ, Sohie GR. Generalization of the matrix inversion lemma. Proc IEEE. 1986 Jul;74(7):1050–1052. [Google Scholar]
  • 59.Chen YC, Wilson DW, Furenlid LR, Barrett HH. Calibration of scintillation cameras and pinhole SPECT imaging systems. In: Kupinski M, Barrett H, editors. Small-Animal SPECT Imag. ch. 12. New York: Springer; 2005. pp. 195–201. [Google Scholar]
  • 60.Shen F, Clarkson E. Using Fisher information to approximate ideal-observer performance on detection tasks for lumpy-background images. J Opt Soc Amer. 2006;23:2406–2414. doi: 10.1364/josaa.23.002406. [DOI] [PubMed] [Google Scholar]
  • 61.Clarkson E. Estimation ROC curves and their corresponding ideal observers. Proc SPIE. 2007;6515:651504. [Google Scholar]

RESOURCES