Research Papers: Imaging

Reconstruction-classification method for quantitative photoacoustic tomography

[+] Author Affiliations
Emma Malone, Ben T. Cox

University College London, Department of Medical Physics and Biomedical Engineering, Gower Street, WC1E 6BT London, United Kingdom

Samuel Powell, Simon Arridge

University College London, Department of Computer Science, Gower Street, WC1E 6BT London, United Kingdom

J. Biomed. Opt. 20(12), 126004 (Dec 14, 2015). doi:10.1117/1.JBO.20.12.126004
History: Received August 4, 2015; Accepted October 30, 2015
Text Size: A A A

Open Access Open Access

Abstract.  We propose a combined reconstruction-classification method for simultaneously recovering absorption and scattering in turbid media from images of absorbed optical energy. This method exploits knowledge that optical parameters are determined by a limited number of classes to iteratively improve their estimate. Numerical experiments show that the proposed approach allows for accurate recovery of absorption and scattering in two and three dimensions, and delivers superior image quality with respect to traditional reconstruction-only approaches.

Figures in this Article

Photoacoustic tomography (PAT) is an emerging technique for in vivo imaging of soft biological tissue.1 This hybrid modality uses ultrasound to detect optical contrast, combining the high resolution of acoustic methods with the spectroscopic capability of optical imaging. To generate a PA image, a short laser pulse is shined into the object, the ultrasonic waves emitted following the heating of the tissue are measured, and an image of the absorbed optical energy field is recovered. Whereas purely optical methods suffer from poor spatial resolution, acoustic waves propagate with minimal scattering, and PAT can achieve 100μm resolution at depths of several centimeters. However, PA images provide only qualitative information about the tissue, and are not directly related to tissue morphology and functionality. The principal difficulty is that the PA image is the product of both the optical absorption coefficient (which is directly related to underlying tissue composition) and the light distribution (which is not). This severely restricts the range of applications for which PAT is suitable.

Quantitative photoacoustic tomography (QPAT) aims to provide clinically valuable images of the optical absorption and scattering coefficients, or chromophore (light-absorbing molecules) concentrations from conventional PA images via an image reconstruction method.2,3 A model of light propagation is required to relate the absorbed optical energy to the light fluence and tissue parameters. The primary challenge of QPAT is solving the nonlinear imaging problem. In particular, recovering the scattering coefficient is especially difficult due to the weak dependence of the absorbed energy density on scattering.

In this paper, we develop a method for solving the image reconstruction problem for QPAT by alternating reconstruction and segmentation steps in an automated iterative process. We introduce a probabilistic model that describes optical properties in terms of a limited number of optically distinct classes, which may correspond to tissues or chromophores. These are identified and characterized by a classification, or segmentation, algorithm. This approach allows for the use of information retrieved by the classification in the reconstruction stage and vice versa. The aim of the reconstruction is to choose solutions for which the image parameters take values close to a finite set of discrete points. The aim of the classification algorithm is to progressively improve the parametric optical model and correct for errors in the initial assumptions. Multinomial models have been employed previously in the related fields diffuse optical tomography4 and electrical impedance tomography.5 For QPAT, the main advantage is that this approach enables accurate recovery of both the absorption and scattering coefficients simultaneously.

Quantitative Photoacoustic Imaging

A conventional PAT image is proportional to the absorbed optical energy Display Formula

H(r)=Γ^(r)μa(r)φ[μa(r),μs(r)]rΩ,(1)
where r is a position vector within the domain Ω, μa and μs are the optical absorption and reduced scattering coefficients, φ is the optical fluence, and Γ^ is the Grüneisen parameter. The Grüneisen parameter represents the efficiency with which the tissue converts heat into acoustic pressure, and is often taken to be the constant Γ^(r)=1rΩ. The fluence is dependent on the optical parameters and illumination pattern in the whole domain. The problem of recovering the optical parameters (μa,μs) from a conventional PAT image is known as the “quantitative” problem. The optical absorption μa is of particular interest because it is fundamentally related to underlying tissue physiology and functionality, and encodes clinically useful information such as tissue oxygenation levels and chromophore concentrations. Conversely, the absorbed energy density H depends nontrivially on optical absorption and thus is not directly related to tissue morphology because it is distorted structurally and spectrally by the nonuniform light fluence.

Diffusion Model of Light Transport

In order to recover the optical parameters (μa,μs), a model of light propagation within the tissue is required. For highly scattering media and those far from boundaries and sources, a low-order spherical harmonic approximation to the “radiative transfer equation” is suitable. The “diffusion approximation” is given by6Display Formula

[μa·κ(r)]φ(r)=q(r),(2)
where q(r) is an isotropic source term and κ=1/3μs is the diffusion coefficient.

We set Robin boundary conditions Display Formula

φ(r)+12Aκ(r)n^·φ(r)=0rδΩ,(3)
where A accounts for the refractive index mismatch at the boundary.

Minimization-Based Quantitative Photoacoustic Tomography Imaging

In this paper, we adopt a gradient-based minimization approach to image reconstruction. Typically, both μa and μs are unknown and need to be recovered simultaneously from the absorbed energy density. An objective function is defined, which measures the distance between the conventional PAT image Hm and the data predicted by the model for the current estimates H(μa,μs)Display Formula

E=12Ω[HmH(μa,μs)]2dΩ.(4)
In order to treat the problem for a generic geometry, the finite element method is employed, whereby a weak formulation of the diffusion approximation [Eq. (2)] is considered. A discretization of the domain is defined, and the fluence and optical parameters are expressed in terms of the same piecewise-linear basis functions ui(r): χiχiui(r) for χ{μa,μs,φ}, where χi are nodal coefficients and i=1,,N.

We assume that the data dm is the absorbed energy density Hm, projected onto a particular basis {Ψj}, Display Formula

dm={djm,j=1,,N},djm=ΩHm(r)Ψj(r)dΩ=Ψj,Hm.(5)
Choices for {Ψj} include:

  1. Point sampling Ψj(r)=δ(rrj),
  2. Piecewise-linear sampling Ψj=uj,
  3. Sinc sampling Ψj=sinc(|rrj|).

Substituting into the objective function [Eq. (4)] leads to the discrete form of the objective function Display Formula

E=12j[djmΨj,H(μa,μs)]2=12j[djmΨj,μaφ]2.(6)

If a single illumination source is used and both absorption and scattering are undetermined, the problem is ill posed.2 In this study, the nonuniqueness of the solution was removed by using multiple illumination patterns,79 thus the objective function must be summed over the number of sources. In the following, we have omitted this sum for ease of notation. Prior information regarding the solution can be included by adding a regularization term Display Formula

E=12j(djmΨj,μaφ)2+R(μa,μs).(7)
In the Bayesian framework, an image is obtained by maximizing the posterior probability of the parameters, given the data Display Formula
p(μa,μs|dm)p(dm|μa,μs)p(μa,μs).(8)
Under this interpretation, the regularization term R is given by the negative log of the prior probability distribution Display Formula
R(μa,μs)=logp(μa,μs).(9)

Gradient Calculations

Cox et al.10 have shown that, for the continuous case, the gradient of Eq. (4) with respect to μa at position r0 is given by Display Formula

Eμa|r0=φ(HmH)|r0+φ·φ*|r0,(10)
where the “adjoint” light field φ* is the solution to the equation Display Formula
[μa·κ(r)]φ*(r)=μa(HmH).(11)

In the following, we derive the expression for the gradient in the discrete case. The sampled forward model can be expressed as a vector H={Hj,j=1,,N}Display Formula

Hj=ΩH(r)Ψj(r)dΩ=Ψj,H.=ikμaiφkΩΨj(r)ui(r)uk(r)dΩ=φTCjμa,(12)
where Cj is a sparse matrix indexed by i, k where the support of the basis functions Ψj(r), ui(r), uk(r) overlap. Taking the derivative of Eq. (6) with respect to μai, we have Display Formula
Eμai=j(Hjμai)(djmHj).(13)
Using the expression for the absorbed energy density [Eq. (12)], Display Formula
Hjμai=eiTCjφ+μaTCjφμai,(14)
where ei is a vector of zeros with a single 1 in position i. Substituting into Eq. (13) gives Display Formula
Eμai=j(eiTCjφ+μaTCjφμai)(djmHj).(15)
The first term in Eq. (15) is Display Formula
jeiTCjφ(djmHj)=j,i,keiCikjφk(djmHj)=j,kφk(djmHj)ΩΨj(r)ui(r)uk(r)dΩ=φTEi(dmH),(16)
where Ei is given by a reordering of CikjDisplay Formula
Ekji=ΩΨj(r)ui(r)uk(r)dΩ.(17)
Note that while Cj is symmetric, in general, Ei is not.

It remains to determine φ/μai. The discrete form of the DA model [Eq. (2)] assumes the form11Display Formula

(M+K+F)φ=Q,(18)
where Display Formula
Mjk=iμaiΩuiujukdΩ,(19)
Display Formula
Kjk=iκiΩuiuj·ukdΩ,(20)
Display Formula
Fjk=i12AΩujukdS,(21)
Display Formula
Qj=iqiΩuiujdΩ.(22)
Taking the derivative of Eq. (18) with respect to the i’th coefficient of μa, Display Formula
(M+K+F)φμai=Vμaiφ,(23)
where Display Formula
Vμa,jki=ΩuiujukdΩ(24)
is given by the derivative of the system matrix. We define the adjoint field φ* as the solution to the equation Display Formula
(M+K+F)φ*=Q*,(25)
where Display Formula
Q*=jμaTCj(djmHj)(26)
is the adjoint source. Taking φ* Equation (23) (φ/μai). Eq. (25), we obtain Display Formula
jμaTCjφμai(djmHj)=φTVμaiφ*.(27)
Substituting into Eq. (15) gives the expression for the derivative with respect to μaiDisplay Formula
Eμai=φT[Vμaiφ*Ei(dmH)].(28)
The derivative with respect to μsi can be derived analogously Display Formula
Eμsi=κiμsiφTVμsiφ*,(29)
where Display Formula
Vμs,jki=Ωuiuj·ukdΩ(30)
and (κi/μsi)=1/3μsi2. Note that calculation of the gradient requires only two runs of the forward model. The forward problem was solved using the Toast++ software package.11

Choosing point-sampling Ψj(r)=δ(rrj) gives simply Cj=Ei=I. In this study, we chose piecewise-linear sampling Ψj=uj, so we had Cj=Ei=Vμai and Display Formula

Eμai=φTVμai(φ*dm+H).(31)

A reconstruction-classification scheme is devised, which enables the recovery μa and μs by approaching the image reconstruction and segmentation problems simultaneously. At each reconstruction step, we minimize a regularized objective function, where the regularization term is given by a mixture model. At each classification step, the result of the previous reconstruction step is employed to update the class parameters for the multinomial model. We alternate between reconstruction and classification steps for a fixed number of iterations (Fig. 1).

Mixture Model for μa and μs′

In this section, we introduce a probability model for μa and μs, which encodes prior knowledge about the optical parameters and allows us to bias the solution of the imaging problem accordingly. We assume that an array of labels ζi can be determined for each node, such that Display Formula

ζij={1iftheithnode is assigned to thejthclass,0otherwise.(32)
The labels constitute “hidden variables” on which the image parameters are dependent. For each class j=1,,J, a mean vector mj=(μ¯aj,μ¯sj)R2 is defined, and the covariance of each class is described by matrix ΣjR2×2.

We assume that if ζij=1, the probability distribution for xi=(μai,μsi) is given by a multivariate Gaussian distribution Display Formula

p(xi|θj)=N(mj,Σj),(33)
where θj indicates the set of class parameters (mj,Σj).

The prior probability distribution of the class properties θj is given by the conjugate prior to the Gaussian distribution. Prior information about the distribution of the class means or covariances can be encoded by choosing the parameters of the conjugate prior accordingly. Using a noninformative prior for the class means we have p(mj)1. The conjugate prior distribution for the covariance of a normal distribution is given by the normal inverse Wishart distribution (NIW): Display Formula

NIW(νj,Γj)=|Σj|(ν+d+1)/2exp[12Tr(ΓjΣj1)],(34)
where d is the dimension of the domain, νj indicates the number of degrees of freedom, and Γj is a scaling matrix. If the prior is noninformative, then νj=0 and Γj=0, and the probability distribution of the class parameters becomes Display Formula
p(θj)|Σj|(d+1)/2,(35)
which is known as Jeffreys prior.

The probability that the set of labels ζi={ζi1,,ζij,,ζiJ} is assigned to the i’th node is given by a multinomial distribution Display Formula

p(ζi|λ)=jλjζij,(36)
where λj is the overall probability that a node is assigned to the j’th class. Therefore, the joint probability for (xi,ζi) is given by the product Display Formula
p(xi,ζi|θ,λ)=p(xi|ζi,θ)p(ζi|λ)=j[λjp(xi|θj)]ζij.(37)
By marginalizing over all possible values of the indicator variables ζij, a “mixture of Gaussians” model for the optical parameters is obtained Display Formula
p(xi|θ,λ)=ζip(xi,ζi|θ,λ)dζi=jλjp(xi|θj).(38)
Finally, for independent nodes, the prior of the image is given by Display Formula
p(x|θ,λ)=ijλjp(xi|θj).(39)

Reconstruction step

The objective function takes the form of Eq. (7), where at iteration t of the reconstruction-classification algorithm, the regularization is given by Eqs. (9) and (39) Display Formula

Rt(μa,μs)=logp(x|θt,λt)=logN(x¯,Σx¯)=τ2Lx¯(xx¯)2,(40)
where τ is a regularization parameter and Display Formula
x¯i=jζij·mj|MAP(ζ)=mjR2(41)
is obtained by fixing the labels to the “maximum a posteriori” estimate, given the results of the previous iteration Display Formula
MAP(ζ)=argmaxζp(ζ|xt1,θt1,λt1),(42)
which is calculated in the classification step (see Sec. 3.1.2). The weighting matrix Lx¯ is the Cholesky decomposition of Σx¯1, where Σx¯R2N×2N is a sparse matrix of which the i’th 2×2 block along the diagonal is Σj if the i’th element belongs to the j’th class.

In order to sphere the solution space, that is, to render the space dimensionless, we performed a change of variables μaμa/μa0 and μsμa/μs0, where (μa0,μs0) is the initial guess for the optical parameters (in this study, we initialized to the homogeneous background). Given the size of the problem, we chose a gradient-based optimization method in order to reduce memory use and computational expense.12 The minimization was performed using the limited-memory Broyden–Fletcher–Goldfarb–Shanno (L-BFGS) method,13 with a storage memory of six iterations.

Classification

The purpose of the classification step is to update the multinomial model using the result of the previous reconstruction step. First, the expected values of the labels ζt+1 are computed for the current class parameters (θt,λt) and image xt=(μat,μst) (E-step). Then the model parameters are updated by maximizing the posterior probability (M-step) Display Formula

p(θ,λ|xt)p(xt|θ,λ)p(θ,λ).(43)

  • The “responsibility” rijt is a measure of the probability that the i’th node is assigned to the j’th class. Using Bayes’ theorem and the Gaussian mixture model [Eq. (38)], we have Display Formula
    p(ζij=1|xit,θt,λt)=p(xi|ζij=1,θt)p(ζij=1)p(xi|θ,λ)=λjtp(xit|θjt)jλjtp(xit|θjt)=rnjt.(44)
    The expectation for the indicator values is Display Formula
    E(ζij|xit,θt,λt)=ζijp(ζij=1|xit,θt,λt)dζij=0×p(ζij=0|xit,θt,λt)+1×p(ζij=1|xit,θt,λt)=rijt.(45)
    Therefore, the MAP estimate for the labels is Display Formula
    ζijt+1={1ifrijtismaximumj,0otherwise,(46)
    which can be used in Eq. (42).
  • The parameters (θ,λ) are chosen in order to maximize the log posterior Display Formula
    (θt+1,λt+1)=argmax(θ,λ)logp(xt|θ,λ)+logp(θ,λ).(47)
    Averaging over all possible values of ζ gives Display Formula
    logp(xt|θ,λ)+logp(θ,λ)=ζlogp(xt,ζ|θ,λ)dζ+logp(θ,λ).(48)
    Using “Jensen’s inequality”14 and ignoring terms that do not depend on (θ,λ), we obtain a lower bound for the log prior Display Formula
    B(θ,λ)=ijrijtlog[λjp(σn|θj)]+logp(λ)+logp(θ)=ijrijt[log(λj)+log(|Σj|)12(xi(n)mj)Σj1(xi(n)mj)]+j[(αj1)log(λj)νj+d+12log|Σj|].(49)
    Maximizing B(θ,λ) for jλj=1 and using noninformative priors, we obtain the update rules for the model parameters Display Formula
    λjt+1=irijtN,(50)
    Display Formula
    mjt+1=irijtxiirijt,(51)
    Display Formula
    Σjt+1=irijt(ximj)(ximj)T+Γjirijt+νj+d+1.(52)

Class Means Initialization

The number of classes J and the class means mj were initialized by automatically segmenting the result of the first reconstruction step and averaging over the segmented areas. To segment the image [e.g., see Fig. 2(a)], we looked at a binned histogram of the image of μa and chose the value μah for which the number of occurrences was highest [Fig. 2(c), column 1]. We found the first node index h for which the value μah occurs, and identified the corresponding scattering value μsh. Having chosen a covariance matrix Σh, we computed a map of the multivariate normal probability of the (μa,μs) images, with mean (μah,μsh) [Fig. 2(c), column 2]. A suitable choice for Σh is the initial covariance of the classes. Then we selected a tolerance level tolh at which to truncate the probability map, and selected all nodes with probability higher than the tolerance as belonging to the same class as node h [Fig. 2(c), column 3]. We repeated this process on the remaining nodes until all nodes were classified. Thus, the number of classes was set to the number of iterations, and the average of the optical parameters over each class was used to initialize the class means [Fig. 2(b)].

Graphic Jump Location
Fig. 1
F1 :

Reconstruction-classification algorithm outline.

Graphic Jump Location
Fig. 2
F2 :

Class initialization example: (a) original image of μa to which we apply the segmentation; (b) result of taking average image values over the segmented areas; (c) first column, histogram of occurrences of values of μa in the portion of the image requiring segmentation—value with highest number of occurrences is μah (indicated by a red cross); second column, probability density function with mean (μah,μsh) and covariance Σh; third column, labels identifying nodes with probability density higher than tolerance value tolh; each row corresponds to an iteration and a distinct class, so in this case, J=3.

Visualization of the Results

Results obtained using the reconstruction-classification method are displayed alongside scatter plots of the nodal values recovered in the two-dimensional (2-D) feature space (μa,μs) [e.g., see Fig. 2(c), final column in 4]. The positions of the class means mj=(μ¯aj,μ¯sj) are identified by a cross, and the class covariances Σj are represented by ellipses. These are color-coded by class, and are indicative of the clustering of image nodal values around the class means.

Two-Dimensional Validation and Reconstruction

We chose a numerical phantom defined on a 2-D circular mesh with 1331 nodes and radius 25 mm. Four illumination sources were placed on the boundary at angles 0, π/2, π, and 3π/2rad. In all cases, the illumination profile was a normalized Gaussian with radius (distance from the center at which the profile drops to 1/e) 6 mm. The background optical parameters were set to μa=0.01mm1 and μs=1mm1. Two circular perturbations of radius 6 mm were added in positions (6 mm, 10 mm) and (6mm,10mm) [Fig. 3(a)]. The values of the perturbations were μa=0.02mm1, μs=1.5mm1 and μa=0.03mm1, μs=1.25mm1, respectively. The absorbed energy field was simulated for each illumination, and 1% white Gaussian noise was added [Fig. 3(b)]. The class covariances were initialized to Display Formula

Σj=(10600101)j=1,,3,(53)
where the first variable was the absorption and the second was the reduced scattering. The parameters of the Jeffreys prior were set to Γj=Σjj, ν(1)=1 for the background class and ν(2,3)=10 for the perturbation classes. The number of classes and optical parameters was initialized using the class means initialization method (Sec. 3.2) with tolh=105 and Σh=Σj [Eq. (53)], and the labels were initialized to 1 for the background class and zero for all other classes. The tolerance of the L-BFGS algorithm was set to tol=1011, and the total number of reconstruction-classification iterations was set to MaxIt=10 (Fig. 4). The regularization parameter τ=1010 was chosen by inspection. For comparison, images were reconstructed without introducing a prior (Fig. 5); the images were reconstructed by minimizing Eq. (6) using the L-BFGS method with tol=1012.

Graphic Jump Location
Fig. 3
F3 :

Two-dimensional (2-D) model: (a) circular mesh and (b) absorbed energy for each illumination pattern.

Graphic Jump Location
Fig. 4
F4 :

2-D reconstruction-classification results at iteration 1 (first row), 5 (second row), and 10 (third row). Reconstructed values of μa and μs (first and second columns), labels recovered for perturbation classes (third and fourth columns), and scatter plot (fifth column).

Graphic Jump Location
Fig. 5
F5 :

2-D model and reconstruction: first column, model of μs and μs; second column, reconstructed values of μa and μs without multinomial prior; third column, reconstructed values of μa and μs with multinomial prior.

Three-Dimensional Validation and Reconstruction

We chose a three-dimensional (3-D) phantom analogous to the 2-D case, defined on a cylinder with 27,084 nodes, radius 25 mm, and height 25 mm. Two spherical inclusions of radius 6 mm were placed in (6, 10, and 0 mm) and (6, 10, and 0 mm) [Fig. 6(a)]. Illumination sources were Gaussian in the xy-plane constant in the z-axis, with radius 6 mm and length 25 mm [Figs. 6(b) and 6(c)]. PAT images were simulated for four illuminations at the cardinal points, and 1% noise was added to the absorbed energy [Fig. 6(d)]. The optical, covariance, and reconstruction parameters were set to the same values used in the 2-D case. The class initialization parameters were set to tolh=107 and Σh=Σj. Images were reconstructed by performing 10 iterations of the reconstruction-classification method (Fig. 7).

Graphic Jump Location
Fig. 6
F6 :

Three-dimensional (3-D) model: (a) numerical phantom and perturbation locations, (b) all illumination sources, (c) cross section of optical parameters used to simulate the data for z=0, (d) cross section of absorbed energy for each illumination pattern.

Graphic Jump Location
Fig. 7
F7 :

3-D reconstruction-classification results at iteration 1 (first row), 5 (second row), and 10 (third row). Reconstructed values of μa and μs (first and second columns), labels recovered for perturbation classes (third column), and scatter plot (fourth column).

Summary of Findings

We applied the proposed reconstruction-classification algorithm to a 2-D numerical phantom with three tissues, a background, and two perturbations (Fig. 3). The optical absorption was recovered reliably within a small number of iterations, and the scattering was recovered with sufficient accuracy after approximately 10 iterations (Fig. 4). We compared the optical model with images obtained by the reconstruction-classification method and by a traditional reconstruction-only (no regularization) method (Fig. 5). We found that the reconstruction-classification method delivered superior image quality, particularly with regards to the scattering parameter. We applied the reconstruction-classification algorithm to a much larger 3-D problem (Fig. 6) and observed similar results (Fig. 7) as in the 2-D case.

Choice of Parameters

The parametric optical model and classification algorithm introduce a number of parameters that require tuning by the user. In addition to the regularization parameter, the parameters of the Jeffreys prior Γ and ν and the initial guess of the class variances Σj must be set before performing the classification. However, their significance is fairly intuitive, and with experience of a certain type of problem, the choice of parameters becomes natural. Visualizing the class covariance matrix Σj as an ellipse, changing the value of Γ varies its eccentricity, and changing ν varies the length of its axes. Further, given that in the first iteration the optical absorption is recovered with superior accuracy to the scattering, it is preferable to initialize the variance of the former to a smaller value than the latter, indicating greater confidence in the imaging solution.

Initialization of the Class Means

The purpose of the means initialization scheme is to increase automation of the method so that minimum user intervention and no prior knowledge of the number of tissues or their optical properties is required. The algorithm simply performs a segmentation of the image, then takes averages over the segmented areas to initialize the class properties (Fig. 1). Alternative segmentation techniques could have been employed; however, the advantage of the proposed approach is that it directly exploits the mixture of Gaussians model to identify the tissues. Our choice to investigate a node h with μa belonging to the bin with a maximum number of occurrences leads to the background tissue being identified first, followed by the perturbation tissues. The choice of the node index h could have been randomized so that tissues would be identified in random order. This approach is equally valid; however, we found that in cases where tissue values were close together (such as after a single reconstruction-classification iteration), it was preferable to identify the largest classes first because the mean was estimated with greater accuracy for the classes with a larger number of samples. Further, for a given image and tolerance level, our choice renders the result of the segmentation process unique and reproducible.

Recovery of the Scattering

From the comparison with the reconstruction-only case with no regularization (Fig. 5), it is evident that the introduction of the parametric prior enables better recovery of the scattering. The inconsistency between the quality of the recovered absorption and scattering parameters in the nonregularized case is due to the weaker dependence of the latter on the absorbed energy density with respect to the former. This results in the scattering gradient being approximately an order of magnitude smaller than the absorption gradient. Although the problem can be mitigated by sphering the solution space, variations in the data due to the scattering often fall below the noise floor. In the reconstruction-classification case, typically the absorption is recovered with good accuracy within a small number of iterations. Thus, the absorption takes values very close to the class means (resulting in small clusters), and the variance along the μa direction converges to a small value. Given that the regularization term is weighted by the inverse of the covariance matrix, the dependence of the absorption gradient on the data becomes weaker at each iteration, until its magnitude is comparable or smaller to that of the scattering. In the iterations that follow, the descent of the data term of the objective function is primarily due to updates to the scattering, which converges to the correct values.

Computational Demands

Computational performance was found to be strongly dependent on the problem size. In the 2-D case with 1331 nodes (Fig. 4), the total reconstruction time (10 outer reconstruction-classification iterations) using MATLAB on a 16-processor PC with 128 GB RAM was only 77 s. In the 3-D case with 27,084 nodes (Fig. 7), the total reconstruction time increased linearly with the number of nodes and was approximately 3.7 h on the same workstation. The increase in computation time was mostly due to much longer processing times for the L-BFGS algorithm in the reconstruction step.

Experimental Application

In experimental situations, prior information on tissue properties may be held, such as knowledge of the characteristic optical absorption and scattering spectra of chromophores of interest. These may be obtained from the literature15 or gained through tissue sample measurements. This information could be used in one of two ways. First, a library of typical chromophores could be used to initialize the class parameters instead of the proposed class means initialization method. The classification process could then perform the function of correcting for uncertainty, errors, or local variations in the real optical properties with respect to the prior information. Alternatively, it could be used to label the chromophores found by the segmentation process and identify these as certain tissues such as, e.g., “oxygenated blood” or “fat,” on the basis of the closeness of the recovered means to the characteristic properties.

Additional Priors

In this study, we assumed independence between nodal values; however, the mixture of Gaussian models could be used in conjunction with a spatial prior. Knowledge of smoothness or sparsity properties of the solution could be employed to introduce a homogeneous spatial regularizer such as first-order Tikhonov16 or total variation.7,17 Knowledge of structural information, such as that provided by an alternative imaging method or anatomical library, could be exploited by introducing a spatially varying probability map for the optical properties.

In this paper, we proposed a method for performing image reconstruction in QPAT. We introduced a parametric class model for the optical parameters and implemented a minimization-based reconstruction algorithm. We suggested an automated method by which to initialize the parameters of the class model and proposed a classification algorithm by which to progressively update and improve those parameters after each reconstruction step. We demonstrated though 2-D and 3-D numerical examples that the reconstruction-classification method allows for the simultaneous recovery of optical absorption and scattering. In particular, we found that this approach delivered superior accuracy in the recovery of the scattering with respect to traditional gradient-based reconstruction.

This work was funded by the EPSRC Doctoral Prize Fellowship EP/M506448/1.

Beard  P., “Biomedical photoacoustic imaging,” Interface Focus. 1, , 602 –631 (2011).CrossRef
Cox  B.  et al., “Quantitative spectroscopic photoacoustic imaging: a review,” J. Biomed. Opt.. 17, , 061202  (2012). 1083-3668 CrossRef
Gao  H., , Osher  S., and Zhao  H., “Quantitative photoacoustic tomography,” Lect. Notes Math.. 2035, , 131 –158 (2012).CrossRef
Hiltunen  P., , Prince  S. J. D., and Arridge  S., “A combined reconstruction-classification method for diffuse optical tomography,” Phys. Med. Biol.. 54, , 6457 –6476 (2009). 0031-9155 CrossRef
Malone  E.  et al., “A reconstruction-classification method for multifrequency electrical impedance tomography,” IEEE Trans. Med. Imaging. 34, (7 ), 1486 –1497 (2015). 0278-0062 CrossRef
Arridge  S., “Optical tomography in medical imaging,” Inverse Prob.. 15, , R41  (1999). 0266-5611 CrossRef
Bal  G., and Ren  K., “Multiple-source quantitative photoacoustic tomography in a diffusive regime,” Inverse Prob.. 27, (7 ), (2011). 0266-5611 CrossRef
Shao  P., , Cox  B., and Zemp  R. J., “Estimating optical absorption, scattering, and Grueneisen distributions with multiple-illumination photoacoustic tomography,” Appl. Opt.. 50, (19 ), 3145 –3154 (2011). 0003-6935 CrossRef
Gao  H., , Feng  J., and Song  L., “Limited-view multi-source quantitative photoacoustic tomography,” Inverse Prob.. 31, (6 ), 065004  (2015). 0266-5611 CrossRef
Cox  B. T., , Arridge  S. R., and Beard  P. C., “Gradient-based quantitative photoacoustic image reconstruction for molecular imaging,” Proc. SPIE. 6437, , 64371T  (2007). 0277-786X CrossRef
Schweiger  M., and Arridge  S., “The Toast++ software suite for forward and inverse modeling in optical tomography,” J. Biomed. Opt.. 19, , 040801  (2014). 1083-3668 CrossRef
Saratoon  T.  et al., “3D quantitative photoacoustic tomography using the δ-Eddington approximation,” Proc. SPIE. 8581,  85810V  (2013). 0277-786X CrossRef
Nocedal  J., and Wright  S., “Numerical optimization,” in Springer Series in Operations Research and Financial Engineering. ,  Springer-Verlag ,  New York  (1999).
Prince  S., Computer Vision: Models, Learning, and Inference. ,  Cambridge University Press ,  Cambridge, United Kingdom  (2012).
Jacques  S. L., “Optical properties of biological tissues: a review,” Phys. Med. Biol.. 58, , 5007 –5008 (2013). 0031-9155 CrossRef
Saratoon  T.  et al., “A gradient-based method for quantitative photoacoustic tomography using the radiative transfer equation,” Inverse Prob.. 29, , 075006  (2013). 0266-5611 CrossRef
Tarvainen  T.  et al., “Reconstructing absorption and scattering distributions in quantitative photoacoustic tomography,” Inverse Prob.. 28, , 084009  (2012). 0266-5611 CrossRef
© 2015 The Authors

Citation

Emma Malone ; Samuel Powell ; Ben T. Cox and Simon Arridge
"Reconstruction-classification method for quantitative photoacoustic tomography", J. Biomed. Opt. 20(12), 126004 (Dec 14, 2015). ; http://dx.doi.org/10.1117/1.JBO.20.12.126004


Figures

Graphic Jump Location
Fig. 3
F3 :

Two-dimensional (2-D) model: (a) circular mesh and (b) absorbed energy for each illumination pattern.

Graphic Jump Location
Fig. 4
F4 :

2-D reconstruction-classification results at iteration 1 (first row), 5 (second row), and 10 (third row). Reconstructed values of μa and μs (first and second columns), labels recovered for perturbation classes (third and fourth columns), and scatter plot (fifth column).

Graphic Jump Location
Fig. 5
F5 :

2-D model and reconstruction: first column, model of μs and μs; second column, reconstructed values of μa and μs without multinomial prior; third column, reconstructed values of μa and μs with multinomial prior.

Graphic Jump Location
Fig. 2
F2 :

Class initialization example: (a) original image of μa to which we apply the segmentation; (b) result of taking average image values over the segmented areas; (c) first column, histogram of occurrences of values of μa in the portion of the image requiring segmentation—value with highest number of occurrences is μah (indicated by a red cross); second column, probability density function with mean (μah,μsh) and covariance Σh; third column, labels identifying nodes with probability density higher than tolerance value tolh; each row corresponds to an iteration and a distinct class, so in this case, J=3.

Graphic Jump Location
Fig. 1
F1 :

Reconstruction-classification algorithm outline.

Graphic Jump Location
Fig. 6
F6 :

Three-dimensional (3-D) model: (a) numerical phantom and perturbation locations, (b) all illumination sources, (c) cross section of optical parameters used to simulate the data for z=0, (d) cross section of absorbed energy for each illumination pattern.

Graphic Jump Location
Fig. 7
F7 :

3-D reconstruction-classification results at iteration 1 (first row), 5 (second row), and 10 (third row). Reconstructed values of μa and μs (first and second columns), labels recovered for perturbation classes (third column), and scatter plot (fourth column).

Tables

References

Beard  P., “Biomedical photoacoustic imaging,” Interface Focus. 1, , 602 –631 (2011).CrossRef
Cox  B.  et al., “Quantitative spectroscopic photoacoustic imaging: a review,” J. Biomed. Opt.. 17, , 061202  (2012). 1083-3668 CrossRef
Gao  H., , Osher  S., and Zhao  H., “Quantitative photoacoustic tomography,” Lect. Notes Math.. 2035, , 131 –158 (2012).CrossRef
Hiltunen  P., , Prince  S. J. D., and Arridge  S., “A combined reconstruction-classification method for diffuse optical tomography,” Phys. Med. Biol.. 54, , 6457 –6476 (2009). 0031-9155 CrossRef
Malone  E.  et al., “A reconstruction-classification method for multifrequency electrical impedance tomography,” IEEE Trans. Med. Imaging. 34, (7 ), 1486 –1497 (2015). 0278-0062 CrossRef
Arridge  S., “Optical tomography in medical imaging,” Inverse Prob.. 15, , R41  (1999). 0266-5611 CrossRef
Bal  G., and Ren  K., “Multiple-source quantitative photoacoustic tomography in a diffusive regime,” Inverse Prob.. 27, (7 ), (2011). 0266-5611 CrossRef
Shao  P., , Cox  B., and Zemp  R. J., “Estimating optical absorption, scattering, and Grueneisen distributions with multiple-illumination photoacoustic tomography,” Appl. Opt.. 50, (19 ), 3145 –3154 (2011). 0003-6935 CrossRef
Gao  H., , Feng  J., and Song  L., “Limited-view multi-source quantitative photoacoustic tomography,” Inverse Prob.. 31, (6 ), 065004  (2015). 0266-5611 CrossRef
Cox  B. T., , Arridge  S. R., and Beard  P. C., “Gradient-based quantitative photoacoustic image reconstruction for molecular imaging,” Proc. SPIE. 6437, , 64371T  (2007). 0277-786X CrossRef
Schweiger  M., and Arridge  S., “The Toast++ software suite for forward and inverse modeling in optical tomography,” J. Biomed. Opt.. 19, , 040801  (2014). 1083-3668 CrossRef
Saratoon  T.  et al., “3D quantitative photoacoustic tomography using the δ-Eddington approximation,” Proc. SPIE. 8581,  85810V  (2013). 0277-786X CrossRef
Nocedal  J., and Wright  S., “Numerical optimization,” in Springer Series in Operations Research and Financial Engineering. ,  Springer-Verlag ,  New York  (1999).
Prince  S., Computer Vision: Models, Learning, and Inference. ,  Cambridge University Press ,  Cambridge, United Kingdom  (2012).
Jacques  S. L., “Optical properties of biological tissues: a review,” Phys. Med. Biol.. 58, , 5007 –5008 (2013). 0031-9155 CrossRef
Saratoon  T.  et al., “A gradient-based method for quantitative photoacoustic tomography using the radiative transfer equation,” Inverse Prob.. 29, , 075006  (2013). 0266-5611 CrossRef
Tarvainen  T.  et al., “Reconstructing absorption and scattering distributions in quantitative photoacoustic tomography,” Inverse Prob.. 28, , 084009  (2012). 0266-5611 CrossRef

Some tools below are only available to our subscribers or users with an online account.

Related Content

Customize your page view by dragging & repositioning the boxes below.

Related Book Chapters

Topic Collections

PubMed Articles
Advertisement
  • Don't have an account?
  • Subscribe to the SPIE Digital Library
  • Create a FREE account to sign up for Digital Library content alerts and gain access to institutional subscriptions remotely.
Access This Article
Sign in or Create a personal account to Buy this article ($20 for members, $25 for non-members).
Access This Proceeding
Sign in or Create a personal account to Buy this article ($15 for members, $18 for non-members).
Access This Chapter

Access to SPIE eBooks is limited to subscribing institutions and is not available as part of a personal subscription. Print or electronic versions of individual SPIE books may be purchased via SPIE.org.