Open Access
29 August 2017 Optical coherence tomography image denoising using Gaussianization transform
Zahra Amini, Hossein Rabbani
Author Affiliations +
Abstract
We demonstrate the power of the Gaussianization transform (GT) for modeling image content by applying GT for optical coherence tomography (OCT) denoising. The proposed method is a developed version of the spatially constrained Gaussian mixture model (SC-GMM) method, which assumes that each cluster of similar patches in an image has a Gaussian distribution. SC-GMM tries to find some clusters of similar patches in the image using a spatially constrained patch clustering and then denoise each cluster by the Wiener filter. Although in this method GMM distribution is assumed for the noisy image, holding this assumption on a dataset is not investigated. We illustrate that making a Gaussian assumption on a noisy dataset has a significant effect on denoising results. For this purpose, a suitable distribution for OCT images is first obtained and then GT is employed to map this original distribution of OCT images to a GMM distribution. Then, this Gaussianized image is used as the input of the SC-GMM algorithm. This method, which is a combination of GT and SC-GMM, remarkably improves the results of OCT denoising compared with earlier version of SC-GMM and even produces better visual and numerical results than the state-of-the art works in this field. Indeed, the main advantage of the proposed OCT despeckling method is texture preservation, which is important for main image processing tasks like OCT inter- and intraretinal layer analysis. Thus, to prove the efficacy of the proposed method for this analysis, an improvement in the segmentation of intraretinal layers using the proposed method as a preprocessing step is investigated. Furthermore, the proposed method can achieve the best expert ranking between other contending methods, and the results show the helpfulness and usefulness of the proposed method in clinical applications.

1.

Introduction

Optical coherence tomography (OCT) is a recently created imaging technique to illustrate different information about the internal structures of an object. As in ultrasound imaging, OCT is based on interferometric techniques, except that OCT uses light beams instead of sound profiles.1 Use of OCT images is more popular in the field of ophthalmology, where it has become a key diagnostic technology in the areas of retinal diseases.2 OCT images usually suffer from speckle noise, which restricts the ability of image interpretation. Hence, noise reduction techniques are an essential process for these images. In a wide division, OCT denoising methods can be classified into two basic groups: methods that reduce the noise by some modifications on OCT device imaging such as Refs. 3 and 4 (based on hardware) and the methods that work on the recorded data (based on software). Because of our limitation in access to hardware of devices, we focus more on the second group of denoising methods. Based on our categorization in Ref. 5, these methods can be applied on the spatial or transform domain. For reducing noise in OCT images in the spatial domain, usually some simple methods like low-pass filters, linear smoothing, mean, median, and Wiener filters are used. However, some developed methods like multiple uncorrelated B-scan averaging,6,7 nonlinear anisotropic filter,8,9 complex diffusion,10,11 directional filtering,12,13 adaptive vector-valued kernel function,14 support vector machine approach,15 and Bayesian estimations16 are also applied in the spatial domain.

In the transform based group, some data adaptive methods like principle component analysis;17 dictionary learning;18,19 some nondata adaptive methods that are based on wavelets,20,21 dual-tree complex wavelet transform22,23 or curvelets;24,25 and a circular symmetric Laplacian mixture model in wavelet diffusion26 have been applied for denoising.

Recently, some groups of investigators27 improved the result of denoising by combining dictionary learning and wavelet thresholding and by defining a proper data-dependent initial dictionary.

In another point of view, the denoising problem may be solved in a deterministic or statistical framework.23 In the first framework, each pixel is considered an unknown deterministic variable, and non-Bayesian techniques are used for denoising. In the statistical framework, a random field is employed to model the data, and Bayesian methods are applied to estimate noise-free data. Thus, finding a suitable prior probability distribution function (pdf) for noise-free data plays an important role in the denoising problem.23 Although there is a large body of work on OCT image analysis, few studies have been done on OCT images from a statistical modeling point of view. One of these studies is the work of Grzywacz et al.28 in which a probability model for OCT data was suggested and used for early detection of diabetic retinopathy. The denoising method that is proposed in this paper is the second type (statistical framework), so it first needs to obtain an appropriate statistical model for OCT images.

Another issue is that assumption of the Gaussianity plays a key role in many signal processing methods. For example, a Wiener filter is a linear minimum-mean-square error filter in which the observed and noise-free data are jointly Gaussian. It is proven that, for a Gaussian process, nonlinear filters do not lead to better results compared with a Wiener filter. However, sometimes researchers use this Wiener filtering method without considering this Gaussianity assumption, so they cannot fully benefit from the ability of the algorithm. In this paper, it is illustrated that holding this assumption has a notable effect on the denoising results. Therefore, in this work, a Gaussianization transform (GT) that maps the original distribution of the OCT data to Gaussian components is introduced, and, with this transform, all of the good characteristics of the Gaussian process can be used.

In the rest of the paper, a powerful patch-based denoising method, namely the spatially constrained Gaussian mixture model (SC-GMM), is first explained in Sec. 2.1. Since SC-GMM is based on the Gaussian distribution assumption of noise-free data, which is not satisfied for OCT images, in Sec. 2.2 GT is applied on OCT data. By applying GT, the distribution of transformed OCT data is expressed by a combination of Gaussian pdfs. In this base, the proposed denoising method, which is a combination of Gaussianization and SC-GMM, is demonstrated in Sec. 2.3. The visual and numerical results of the proposed OCT denoising method are presented in Sec. 3, and Sec. 4 presents the discussion of our results in comparison with the state of the art methods. Finally, a summary of this paper is presented in Sec. 5.

2.

Method

Today, patch-based image denoising methods are known as an effective way for denoising natural images.2931 The proposed denoising method in this paper is also a patch-based technique; it is a modified version of the SC-GMM algorithm introduced by Niknejad et al.32 In Sec. 2.1, the SC-GMM method that is based on a Gaussian distribution assumption of similar image patches in a neighborhood is described. However, it is clear that this assumption is not satisfied for all kinds of datasets like OCT images. Thus, after investigation of the distribution of the original OCT image, it is necessary to convert it to a Gaussian distribution for which GT is used. These steps are explained in Sec. 2.2 and finally the proposed OCT denoising method, which is a combination of GT and SC-GMM, is introduced in Sec. 2.3.

2.1.

SC-GMM Denoising Method

As mentioned above, SC-GMM denoising is a patch-based method in which the whole of the image is partitioned into overlapping patches and it is assumed that each set of neighbor patches can be modeled using a multivariate Gaussian pdf with a specific mean vector and covariance matrix. For this purpose, some exemplar patches are uniformly chosen from the whole image; then, similar patches in the vicinity of each exemplar patch are grouped together to construct a region using the K-nearest neighbors (KNN) algorithm. To determine these clusters of patches and estimate the parameters of Gaussian pdfs (covariance matrices and mean vectors), a two-step approach is applied.33 This estimation approach includes a clustering step and a denoising step.

Assume y=x+n is a noisy image, where x is noise-free data and n is an independent zero-mean Gaussian noise with known variance σn2. Also, let xr denote the r’th exemplar patch with mean vector μr and covariance matrix Σr for r=1,,R. After initialization, x^=y, the following steps are iteratively implemented:

Clustering step. In this step, for each exemplar patch x^r, KNN patches x^i, i=1,,K, that have minimum dissimilarity (based on l2 norm metric, i.e., d=x^rx^i22) with the exemplar patch are chosen. These x^r and x^is are estimated in the previous iteration or the initialization in the first step.

Denoising step. The goal of this step is to restore the image using the clusters obtained in the clustering step. Thus, first the parameters of Gaussian distribution for each set of patches are estimated using the maximum likelihood method

Eq. (1)

μ^r=1KiSrx^i,

Eq. (2)

Σ^r=1KiSr(x^iμ^r)(x^iμ^r)T,
where Sr shows the set of KNN patches in the r’th region. It should be noted that in our implementation the above sample covariance matrix is not invertible, so as with similar works,34 an eigenvalue regularization is applied using

Eq. (3)

Σ^r=Σ^r+δI,
where δ is a small constant (here δ=0.1) and I is the identity matrix.

After estimating the parameters of the Gaussian pdf, the denoised patch x^i in the r’th region is found by maximizing a posteriori function logp(x|yi,μ^r,Σ^r), i.e.,

Eq. (4)

x^i=argmaxxlogp(x|yi,μ^r,Σ^r)=argminxyix22+σn2(xμ^r)TΣ^r1(xμ^r),
where yi is the corresponding noisy observed patch and Eq. (4) is derived from the assumption of the Gaussian pdf for noise-free patch and noise, i.e., xiN(μ^r,Σ^r) and nN(0,σn2I), respectively.

By setting the derivative of Eq. (4) to zero, this optimization problem can be solved, which leads to the linear Wiener filter as follows:

Eq. (5)

x^i=(I+σn2Σ^r1)1(yi+σn2Σ^r1μ^r).

The last step for reconstructing the whole image is averaging the weighted restored patches. In this work, based on Ref. 32, these weights are considered proportional to similarity to Gaussian distributions. To be more explicit, patches that are more likely to be generated from the estimated Gaussian distribution of their allocated clusters benefit from higher weights. Thus, the weight of the patch x^i in the r’th region is obtained by

Eq. (6)

w(i,r)=exp[γ2(x^iμ^r)TΣ^r1(x^iμ^r)],
where γ is an appropriate scaling constant.

Using these weights and averaging on all patches, the denoised image is found. Figure 1 shows the block diagram of the SC-GMM method for image denoising.

Fig. 1

Block diagram of spatially constrained GMM (SC-GMM) method for image denoising.

JBO_22_8_086011_f001.png

2.2.

Gaussianization as a Preprocessing Step

In fact, the assumption of Gaussianity plays an important role in many signal processing methods. For example, in this work, the Gaussian distribution assumption for a dataset causes the derivation of the Wiener filter as an optimal linear filter for denoising. Thus, the best denoising result is reached when the initial assumption of Gaussianity is satisfied. However in Ref. 32 and other related works,34 holding this Gaussianity assumption for the pdf of data is not investigated, and here we demonstrate the importance of this assumption by converting the pdf of OCT intraretinal layers to Gaussian distribution and comparing denoising results with and without Gaussianization. For this purpose, the distribution of the OCT images should first be determined, and then an appropriate GT is designed to convert the OCT original distribution to a GMM and satisfy the initial assumption for our denoising algorithm. The block diagram of the Gaussianization process is displayed in Fig. 2.

Fig. 2

Block diagram of GT.

JBO_22_8_086011_f002.png

2.2.1.

OCT distribution estimation

Since the speckle noise in the OCT images is a multiplicative noise and working with the additive noise is simpler and more common, first, a logarithm operator is used to convert this noise to an additive Gaussian noise. Thus, all the processes are implemented in the logarithmic domain. In Ref. 35, we discussed that, because of the layered structure of the retina, a mixture model can be well fitted to a whole OCT B-scan.

Moreover, a monotonically decaying behavior of the OCT intensities in each layer means that leptokurtic distributions such as the Laplace pdf are good candidates for describing the statistical properties of layers.35 On the other hand, this image is corrupted with the additive white Gaussian noise in the logarithmic space, so an appropriate statistical model for an OCT image is a mixture distribution. Each component of this mixture model is composed of a Laplace random variable xj, plus a Gaussian noise n. We called this model a normal-Laplace mixture (NLM) model and described it in detail in Ref. 35.

Indeed, for each component of the mixture model, we would have yj=xj+n, where the pdfs of xj and n are

Eq. (7)

fXj(x)=12σjexp(2σj|xμj|),

Eq. (8)

fN(n)=12πσnexp(n22σn2).

μj shows the mean of the Laplace variable and σj and σn display the scaling parameters of the Laplace and Gaussian distributions, respectively.

Since yj=xj+n and the noise signal n is assumed to be independent from the noise-free signal xj, the pdf of the normal-Laplace variable yj is given by the convolution

Eq. (9)

fj(y)=(fXj*fN)(y)=+fXj(x)fN(yx)dx.

After some mathematical computations, we have35

Eq. (10)

fj(y)=122σj{exp[σn2σj2+2(yμj)σj]erfc[(yμj)2σnσnσj]exp[σn2σj22(yμj)σj]erfc[(yμj)2σn+σnσj]},
where erfc(x)=ex2erfcx(x)=2π0ex2t22txdt.

Thus, yj has a normal-Laplace distribution, and we show it by yjNL(μj,σj,σn). Additionally, the pdf of the NLM model is presented by gY(y) as follows:

Eq. (11)

gY(y)=j=1Jajfj(y),
where fj(y) is the pdf of each component, aj0, j=1Jaj=1, and J is the number of mixture components. In Ref. 35, it has been shown that the choice of J=5 causes the best results.

After finding a proper model, its parameters are estimated based on a used dataset. The EM algorithm36 is used to estimate parameters iteratively as follows:

  • E-step: In each iteration, the responsibility factors rj (as an auxiliary variable that for each observed data y represents how likely that observed data was produced by each component {fj(y)}j=1:J) are updated by

    Eq. (12)

    rjajfj(y)j=1Jajfj(y),for  j=1,,J.

  • M-step: Then, the parameters aj are updated by

    Eq. (13)

    ajl=1Nrj(l)/N,
    where N shows the number of pixels in image y.

In the M-step, the updated equations for {μj,σj}j=1J are approximately estimated as37

Eq. (14)

μjl=1Ny.  rj(l)/l=1Nrj(l),

Eq. (15)

σj2max{l=1N|yμj|2rj(l)/l=1Nrj(l),δ},
where δ is a small positive constant number used to avoid numerical errors.

2.2.2.

Gaussianization transform

As mentioned before, SC-GMM has been designed based on Gaussianity assumption. However, the distribution of each OCT intraretinal layer can be modeled as a normal-Laplace pdf.35 Thus, if a GT is designed to map the original distribution to a Gaussian distribution, it is expected to obtain better denoising results. The core of this transform is solving the equation that compares the cumulative distribution function (CDF) of original dataset and Gaussian distribution. Assume FW(w)=0.5[1+erf(wμw2σw)] expresses the CDF of a Gaussian variable (w) with μw and σw as the mean and standard deviation of w, respectively, and erf(x)=2π0xet2dt. Also, by computing the integral of fj(y) in Eq. (10), the CDF of a normal-Laplace variable is found as

Eq. (16)

Fj(y)=12{1+erf[(yμj)2σn]exp[σn2σj2+2(yμj)σi]+12e[(yμj)2σn]2{erfcx[(yμj)2σn+σnσj]+erfcx[(yμj)2σnσnσj]}}.

By equating these two CDFs, one can find an equation that expresses each value of variable y in the initial distribution, in terms of the equivalent value of w in Gaussian space

Eq. (17)

FY(y)=FW(w)=>w=μw+2σw.erfinv[2FY(y)1],
where erfinv is the inverse of the error function.

2.2.3.

Construct a GMM distribution

After Gaussianization, the distribution of each component of the NLM model, wj, j=1,,J, becomes Gaussian; by combining these Gaussian components, an image with the GMM pdf is constructed and can be used as the input of the SC-GMM denoising method. In Ref. 38, it is shown that the averaged maximum a posterior (AMAP) method is a suitable way for defining the weights of summation for mixture models. In this method, the weight of each component is defined as ajfj(y)/j=1Jajfj(y), so y^, as the Gaussianized image with GMM distribution, is obtained as follows:

Eq. (18)

y^=j=1Jaj.wj.fj(y)/j=1Jajfj(y).

Also, in Ref. 39, it was discussed that instead of using the univariate weighting function ajfj(y)/j=1Jajfj(y), employing multivariate weighting function leads to better results. Thus, for an OCT image of size S×T, a local P×Q window (where P and Q are odd) is placed around each pixel ys,t for s=1,,S, t=1,,T; and the multivariate function f¯j(y) for the central pixel of window (yp,q) is calculated by multiplying all univariate fj(y)s in that window i.e.

Eq. (19)

f¯j(yp,q)=k=P12P12l=Q12Q12fj(yp+k,q+l).

Figure 3 shows how the multivariate function f¯j(y) is produced.

Fig. 3

Using local window around pixels to make a multivariate weighting function f¯j(y).

JBO_22_8_086011_f003.png

Hence, Eq. (18) is replaced with

Eq. (20)

y^=j=1Jaj.wj.f¯j(y)j=1Jajf¯j(y).

2.3.

Proposed Denoising Method

According to Secs. 2.2.1 and 2.2.2, our proposed OCT denoising method is obtained by combining GT and the SC-GMM method as shown in Fig. 4. In this method, the OCT image in the log domain is first Gaussianized using the block diagram of Fig. 2, and then the Gaussianized image is denoised by the SC-GMM method using the block diagram of Fig. 1.

Fig. 4

Block diagram of the proposed combined denoising method, in this figure, third block “GT” includes all the blocks in Fig. 2 (except the Log block) and fourth block “SC-GMM denoising method” is equivalent to the block diagram of Fig. 1.

JBO_22_8_086011_f004.png

The proposed denoising algorithm can be summarized in the following steps:

  • 1. Transform image to the logarithmic domain

  • 2. Fit an NLM model to the data using the EM algorithm:

    • 2.1 Initialization: choose initial values for aj, μj, σj

    • 2.2 E-step: compute responsibility factors using Eq. (12)

    • 2.3 M-step: update parameters aj, μj, σj using Eqs. (13)–(15)

    • 2.4 Iteration: substitute the updated parameters in the previous step for calculating fj(y)

  • 3. Calculate the CDF of each normal-Laplace component by substituting final values of μj, σj, and σn in Eq. (16)

  • 4. Gaussianize each normal-Laplace component using Eq. (17)

  • 5. Use AMAP method in Eq. (20) to obtain the image y^ with GMM distribution

  • 6. Apply SC-GMM algorithm on image y^:

    • 6.1 Partition image into overlapping patches and choose exemplar patches

    • 6.2 For each exemplar patch:

      • 6.2.1 Select KNN patches (a region) based on l2 norm metric in a finite sized window around exemplar patch

      • 6.2.2. Estimate Gaussian parameters in each region by Eqs. (1)–(3)

      • 6.2.3. Denoise the patches in each region by the Wiener filter Eq. (5)

    • 6.3. Obtain reconstructed image by weighted average of denoised patches using Eq. (6)

  • 7. Apply exponential operator to find the denoised image.

3.

Results

In this section, some simulation results performed on our dataset are presented. This dataset consists of thirteen three-dimensional (3-D) macular spectral domain OCT images obtained from eyes without pathologies using a Topcon 3DOCT-1000 imaging system in the Ophthalmology Department of Feiz Hospital, Isfahan, Iran. The x, y, z size of the obtained volumes was 512×650×128  voxels, 7×3.125×3.125  mm3, voxel size 13.67×4.81×24.41  μm3.

From this dataset, 130 B-scans were selected randomly (10 B-scans from each subject), and these randomly selected B-scans were denoised with the methods described previously in the paper: once images were denoised only using the SC-GMM method and they were then despeckled by a combination of GT and the SC-GMM according to our method.

Considering the constant values and parameters, for each region, K=40 nearest neighbor patches were accumulated. The exemplar patches were selected every 10 pixels along both the column and row directions of the image. The size of the constraining window around each exemplar patch was set to 30×30, and the size of patches was selected as 10×10. Also, the size of local window around each pixel for obtaining multivariate f¯j(y) was 3×3. All of these values were selected based on previous related work and our implementation experiences.

Both the qualitative and quantitative performance of the proposed combined method and SC-GMM were provided. The results of denoising with a simple Wiener filter was also prepared to compare with the proposed approach and prove the need and benefit of GT in the proposed method. Contrast to noise ratio (CNR), edge preservation (EP), and texture preservation (TP) as quantitative measures were computed for each denoising method. The mentioned values were computed based on methods discussed in Ref. 40, and regions for analysis of each criterion are shown in Fig. 5.

Fig. 5

Selected ROIs for calculation of CNR, EP, and TP. The larger elliptical region outside the retinal layers is used as background ROI and other circles represent foreground ROIs for CNR calculation. Rectangular ROIs are used for TP and EP.

JBO_22_8_086011_f005.png

CNR measures the contrast between a feature of interest and background noise. EP computes the correlation between edges in the denoised image and the noisy image over the locally selected region of interest (ROI). TP shows the degree of preserving texture in an ROI. Both the TP and EP measures range between 0 and 1 and, in the best condition, go to 1.

In addition, to illustrate the power of the proposed method, its ability in OCT denoising was compared with the BM3D as a benchmark denoising algorithm and recently reported state-of-the-art algorithm in this field.27 Kafieh et al.27 proposed a strong method in speckle reduction of OCT datasets in two-dimensional and 3-D by the combination of dictionary learning and wavelet thresholding. They improved the performance of simple dual-tree complex wavelets by taking advantage of adaptability in dictionary learning methods and introducing complex wavelet-based K-SVD. They also compared their algorithm by other previous outstanding researches in this filed like Fang et al.,18 Yang et al.,41, and Luan and Wu17 and demonstrated the superiority of their method. Thus, this complex wavelet-based K-SVD method (3DCWT-KSVD)27 is chosen as a proper reference to evaluate our proposed method. Figures 6 and 7 show some samples of datasets after application of each one of the mentioned denoising methods (simple Wiener, SC-GMM, 3DCWT-KSVD, BM3D, and the proposed method). Also, the value of the quantitative measurements for each denoising method on all randomly selected slices is summarized in Table 1. Furthermore, to have a comparison between time complexity of the evaluated denoising methods, the mean of the needed computation time for each slice on a PC with Microsoft Windows X32 edition, Intel core2 Duo CPU at 3.00 GHz, 4 GB RAM is reported in Table 1.

Fig. 6

Denoising results in a sample B-scan. The second and third columns show an enlargement of the specified regions of first column images. (a), (g), (m) Original noisy image; (b), (h), (n) simple Wiener; (c), (i), (o) SC-GMM; (d), (j), (p) BM3D; (e), (k), (q) 3DCWT-KSVD; (f), (l), (r) proposed method.

JBO_22_8_086011_f006.png

Fig. 7

Denoising results in a sample B-scan. The second and third columns show an enlargement of the specified regions of first column images. (a), (g), (m) Original noisy image; (b), (h), (n) simple Wiener; (c), (i), (o) SC-GMM; (d), (j), (p) BM3D; (e), (k), (q) 3DCWT-KSVD; and (f), (l), (r) proposed method.

JBO_22_8_086011_f007.png

Table 1

Evaluation measure results; averaged over 130 OCT B-scans from the Topcon device.

MethodMeasure
CNREPTPComputation time (s)
Original4.92011.00001.0000
SC-GMM7.29870.98950.599584.13
BM3D13.06510.98260.72784.20
3DCWT-KSVD39.79330.93160.115087.09a
Proposed method22.92700.96530.934483.22
Simple Wiener5.05661.00000.99690.54

aFor 3DCWT-KSVD, computation time is a combination of training and testing time; i.e., first a dictionary has to be learned (on average 77.72 s per B-scan) and then the test image is denoised using the learned dictionary (on average 9.37 s per B-scan).

Before analyzing the results of Table 1, it should be noted that visual results of a simple Wiener filter (like those shown in Figs. 6 and 7) indicate the weakness of it in OCT image denoising, and the high value of TP and EP for the Wiener filter in Table 1 is because of the definition of these two measures in which the similarity of the denoised image to the original one plays an important role. However, the low value of CNR for the Wiener filter emphasizes the weakness of this denoising method. Regarding the other mentioned denoising algorithms, according to Table 1, the performance of the proposed method is considerably better than the other methods in TP, which is obviously visible in Figs. 6 and 7. Performance of all of the studied methods is close to each other in EP, and the SC-GMM has the highest EP, which is obtained because of the least difference appearing in the resulting image visually. However, it cannot be considered a positive point lonely, since the values of CNR and TP for this method are very low and indeed this method does not show the proper features as a powerful method for OCT denoising. Regarding the CNR measure, the proposed method and 3DCWT-KSVD have remarkably better results than others. Although 3DCWT-KSVD has the highest CNR among all studied methods, it cannot be the winner in this denoising competition because of its weak results in TP and EP; visual results in Figs. 6 and 7 also confirm this matter.

4.

Discussion

The main objective of this paper is to prove the effect of Gaussianization in the result of the OCT denoising by a Wiener-based denoising method. As our results in Table 1 and Figs. 6 and 7 show, it is obviously clear that using the proposed combined method has been much more successful than the SC-GMM method, which assumes Gaussianity for the noise-free image without any justification for proposing this assumption on OCT datasets.

Our second goal in this work is to generally propose a significant method for OCT denoising. To investigate the power of the proposed method, a comparison with the 3DCWT-KSVD algorithm as a recent state of the art in this field and BM3D as a powerful denoising method was done. Visual results in Figs. 6 and 7 demonstrate better noise reduction using the proposed method while desired textural features are preserved simultaneously. Also, in Table 1, this matter is emphasized by checking the high value of TP in the proposed method compared with the 3DCWT-KSVD and BM3D algorithms. Regarding the CNR measures, 3DCWT-KSVD could provide higher CNR because of the very good speckle removing in the background area, but it does not mean that this method is better for next OCT processing because the main goal behind the OCT denoising is providing an appropriate dataset for the next OCT processing, such as layer segmentation42 or lesions identification, while the denoised image obtained by the 3DCWT-KSVD method does not show this ability very well.

To demonstrate this superiority in our proposed method, A-scans of each denoised image were analyzed because many OCT segmentation methods perform based on gradient information from B-scans or analysis of A-scans in OCT datasets.43 The high level of speckle noise in the original noisy OCT image does not allow features of the image to be detectable in segmentation processes; however, applying a good denoising algorithm that suppresses much of the speckle noise and preserves features of the image can provide more reliable segmentation. Figure 8 shows the ability of the proposed method to reduce the speckle noise and protect identifiable feature peaks and valleys. In fact, because of producing the clean and almost noiseless B-scan images in the proposed method [Fig. 8(f)], its A-scans obviously include a peak or valley for each layer of the retina, and the signal amplitude at the boundary layers can be greatly altered. In the other words, an important benefit of the proposed method is using contrast enhancement and denoising together. As depicted in our previous work,35 applying GT on OCT images leads to contrast enhancement of the image and combination of GT and SC-GMM is a proper algorithm for simultaneously denoising and contrast enhancement.

Fig. 8

The cross-section signal along the white solid line. (a) Original noisy image, (b) simple Wiener, (c) SC-GMM, (d) BM3D, (e) 3DCWT-KSVD, (f) proposed method. Red stars on the A-scans show the place of each of 12 boundaries of retinal layers. The name of these layers is ILM, internal limiting membrane; NFL, nerve fiber layer; GCL, ganglion cell layer; IPL, inner plexiform layer; INL, inner nuclear layer; OPL, outer plexiform layer; ONL, outer nuclear layer; IS, inner segment of rods and cones; CL, connecting cilia layer; OS, outer segment of rods and cones, RPE, retinal pigment epithelium, respectively, from top to down. RPE includes two surfaces and three last boundaries in our figures. Green indicators in (f) display some examples of the superiority of the proposed method in reducing the speckle noise and protecting identifiable feature peaks and valleys, compared with 3DCWT-KSVD [orange indicators in (e)].

JBO_22_8_086011_f008.png

Another way to evaluate the performance of the proposed method is by analyzing the intraretinal layer segmentation results. For this purpose, a diffusion map-based method introduced in Ref. 43 for OCT boundaries segmentation was used. This segmentation method was tested once on original OCT B-scans and another time on images denoised by each of the aforementioned denoising algorithms.

For each case, the error values were calculated as the difference of computed layer position by a diffusion map segmentation method and manually labeled values by an expert observer.

Tables 2 and 3 show the mean signed and unsigned border positioning errors for each boundary, which is obtained for 130 randomly selected B-scans from the Topcon dataset. The numbering of retinal surfaces in these tables is based on Fig. 8.

Table 2

Summary of mean signed border positioning errors on 12 boundaries from the Topcon dataset.

Boundary numberOriginal imageSimple WienerSC-GMM3DCWT-KSVDBM3DProposed method
1 (ILM)0.12120.13020.10970.30680.12850.2237
2 (NFL/GCL)1.84802.05692.16202.35371.75821.8392
3 (GCL/IPL)1.17631.24561.51971.83030.95731.3748
4 (IPL/INL)1.18101.19511.80783.33821.32082.0366
5 (INL/OPL)1.52981.41101.11820.62191.32440.7859
6 (OPL/ONL)1.14941.32851.72342.50951.22042.0787
7 (ONL/IS)1.76951.97331.70581.50621.94721.9390
8 (IS/CL)0.72520.73840.68320.58290.70300.5460
9 (CL/OS)0.85410.86830.80660.58460.85150.6888
10 (RPE complex)0.82940.84570.70740.30900.75770.6303
11 (RPE complex)0.79960.81100.70230.07230.77700.5778
12 (RPE complex)0.84710.88220.65890.61240.86300.1574

Table 3

Summary of mean unsigned border positioning errors on 12 boundaries from the Topcon dataset.

Boundary numberOriginal imageSimple WienerSC-GMM3DCWT-KSVDBM3DProposed method
1 (ILM)1.24231.23801.23671.19561.18851.0726
2 (NFL/GCL)2.24482.38542.36162.49212.47772.3974
3 (GCL/IPL)3.19383.13103.73235.10353.38274.4923
4 (IPL/INL)3.22523.13493.88145.76982.99004.9480
5 (INL/OPL)3.40903.50003.70804.98333.51594.4966
6 (OPL/ONL)3.37833.38383.58204.82863.10654.4527
7 (ONL/IS)2.38962.54742.28522.05322.53092.4022
8 (IS/CL)1.13141.11701.12731.03601.12650.9387
9 (CL/OS)1.14061.13961.13620.96801.15020.9646
10 (RPE complex)1.11321.11581.04410.91121.06940.9036
11 (RPE complex)1.14681.14691.17851.31181.20801.0101
12 (RPE complex)1.04151.06870.96151.01881.06500.7117

Based on Tables 2 and 3, the proposed method seems superior to the other mentioned algorithms in the segmentation of some boundaries (8th to 12th boundaries). However, to illustrate the statistically significant improvement of the proposed method over the other algorithms, Table 4 shows the obtained p-values. Based on Table 4, it can be seen that the 8th to 12th boundaries, which are more difficult to identify because of lack of contrast, are significantly better localized in the proposed method compared with the SC-GMM algorithm. Regarding the BM3D and 3DCWT-KSVD algorithms, this superiority is significant just for the 11th and 12th boundaries (p-value <0.05).

Table 4

Improvement of the proposed method compared with SC-GMM, BM3D, and 3DCWT-KSVD algorithm, on the 8th to 12th boundaries of retinal OCT.

Boundary numberp-value (our alg. versus SC-GMM)p-value (our alg. versus BM3D)p-value (our alg. versus 3DCWT-KSVD)
8 (IS/CL)0.00440.05910.1306
9 (CL/OS)0.00750.06640.4699
10 (RPE complex)0.02210.09000.4441
11 (RPE complex)0.01170.04570.0021
12 (RPE complex)<0.00010.00300.0005

Furthermore, given that the ultimate goal of all the works in the medical image processing field is to help physicians determine better and more accurate diagnosis and treatment of disease, ranking mentioned denoising methods by an expert was considered another acceptable criterion for evaluating our method. Accordingly, an ophthalmologist was asked to score the denoised images by each method (raw data, denoised images by SC-GMM, 3DCWT-KSVD, and the proposed method) between one and four; the score of one to four is assigned “best,” “very good,” “good,” and “worst,” respectively. This ranking was done on all 130 randomly selected B-scans (some sample images are illustrated in Figs. 6 and 7) and the average score on all B-scans for our method was 1.2846 (for 110 slices of all 130 B-scans, the proposed method achieved score one), while this value for raw images was 1.8538, for SC-GMM was 2.8231, and for 3DCWT-KSVD was 3.9615. Based on this result, the superiority of the proposed method in clinical applications is also obvious.

Another noticeable issue is that in this research we worked on single-frame images provided by Topcon imaging system; hence the proposed method may become a useful step in OCT image processing approaches, where single frame detection is a must (where the data acquisition time or system restrictions do not allow for multiframe acquisition). However, to make a good judgment on the power of the method in general, it should be compared with results from a multiframe approach (as angular compounding for example), and hopefully this research can be used as a starting point for further and deeper discussions in this field in the OCT community.

5.

Conclusion

In this paper, a developed denoising method for speckle reduction in OCT images was introduced. Our main issue was the effect of holding the assumptions in different statistical methods. Particularly, in this work, the employed denoising algorithm (SC-GMM) was based on the Wiener filter, and in this method both noise-free data and noise are assumed to be Gaussian; however, holding this assumption has not been investigated in previous research. Since OCT datasets do not satisfy this assumption and an NLM model is fitted on these datasets, a transform to Gaussianize each component of this mixture model was needed. After constructing an image underlying GMM, SC-GMM was used for denoising. The visual and numerical results show the superiority of this combined method and encourage us to compare the proposed method with BM3D and 3DCWT-KSVD algorithms as the state-of-the-art studies in this field. Indeed, the advantage of our method, in addition to appropriate noise reduction, is in preserving and highlighting the features of the image like textures and edges that provide a suitable dataset for the next OCT processing such as layer segmentation or object detection. This matter was demonstrated by investigating the behavior of A-scans in all mentioned denoising methods. In addition, the method was shown to improve segmentation accuracy of some intraretinal layers. Furthermore, as another way to prove the superiority of the proposed method in clinical applications, the result of ranking different methods by an ophthalmologist shows predominance of the proposed method.

Disclosures

No conflicts of interest, financial or otherwise, are declared by the authors.

Acknowledgments

The authors would like to thank Dr. MR Akhlaghi from Feiz Eye Hospital for his contribution in clinical evaluation of denoising algorithms and Prof. Sina Farsiu from Duke Eye Center for his valuable comments for improving the technical content of this paper.

References

1. 

D. Huang et al., “Optical coherence tomography,” Science, 254 (5035), 1178 –1181 (1991). http://dx.doi.org/10.1126/science.1957169 SCIEAS 0036-8075 Google Scholar

2. 

M. R. Hee et al., “Optical coherence tomography of the human retina,” Arch. Ophthalmol., 113 (3), 325 –332 (1995). http://dx.doi.org/10.1001/archopht.1995.01100030081025 AROPAW 0003-9950 Google Scholar

3. 

M. Pircher et al., “Speckle reduction in optical coherence tomography by frequency compounding,” J. Biomed. Opt., 8 (3), 565 –569 (2003). http://dx.doi.org/10.1117/1.1578087 JBOPFO 1083-3668 Google Scholar

4. 

B. F. Kennedy et al., “Speckle reduction in optical coherence tomography by strain compounding,” Opt. Lett., 35 (14), 2445 –2447 (2010). http://dx.doi.org/10.1364/OL.35.002445 OPLEDP 0146-9592 Google Scholar

5. 

Z. Amini and H. Rabbani, “Classification of medical image modeling methods: a review,” Curr. Med. Imaging Rev., 12 (2), 130 –148 (2016). Google Scholar

6. 

T. M. Jørgensen et al., “Enhancing the signal-to-noise ratio in ophthalmic optical coherence tomography by image registration—method and clinical examples,” J. Biomed. Opt., 12 (4), 041208 (2007). http://dx.doi.org/10.1117/1.2772879 JBOPFO 1083-3668 Google Scholar

7. 

D. Alonso-Caneiro, S. A. Read and M. J. Collins, “Speckle reduction in optical coherence tomography imaging by affine-motion image registration,” J. Biomed. Opt., 16 (11), 116027 (2011). http://dx.doi.org/10.1117/1.3652713 JBOPFO 1083-3668 Google Scholar

8. 

D. C. Fernández et al., “Comparing total macular volume changes measured by optical coherence tomography with retinal lesion volume estimated by active contours,” Invest. Ophthalmol. Visual Sci., 45 (13), 3072 (2004). IOVSDA 0146-0404 Google Scholar

9. 

P. Puvanathasan and K. Bizheva, “Interval type-II fuzzy anisotropic diffusion algorithm for speckle noise reduction in optical coherence tomography images,” Opt. Express, 17 (2), 733 –746 (2009). http://dx.doi.org/10.1364/OE.17.000733 Google Scholar

10. 

R. Bernardes et al., “Improved adaptive complex diffusion despeckling filter,” Opt. Express, 18 (23), 24048 –24059 (2010). http://dx.doi.org/10.1364/OE.18.024048 Google Scholar

11. 

H. M. Salinas and D. C. Fernández, “Comparison of PDE-based nonlinear diffusion approaches for image enhancement and denoising in optical coherence tomography,” IEEE Trans. Med. Imaging, 26 (6), 761 –771 (2007). http://dx.doi.org/10.1109/TMI.2006.887375 ITMID4 0278-0062 Google Scholar

12. 

A. M. Bagci et al., “Thickness profiles of retinal layers by optical coherence tomography image segmentation,” Am. J. Ophthalmol., 146 (5), 679 –687.e1 (2008). http://dx.doi.org/10.1016/j.ajo.2008.06.010 Google Scholar

13. 

J. Rogowska and M. E. Brezinski, “Evaluation of the adaptive speckle suppression filter for coronary optical coherence tomography imaging,” IEEE Trans. Med. Imaging, 19 (12), 1261 –1266 (2000). http://dx.doi.org/10.1109/42.897820 Google Scholar

14. 

A. Mishra et al., “Intra-retinal layer segmentation in optical coherence tomography images,” Opt. Express, 17 (26), 23719 –23728 (2009). http://dx.doi.org/10.1364/OE.17.023719 OPEXFF 1094-4087 Google Scholar

15. 

A. R. Fuller et al., “Segmentation of three-dimensional retinal image data,” IEEE Trans. Visual Comput. Graphics, 13 (6), 1719 –1726 (2007). http://dx.doi.org/10.1109/TVCG.2007.70590 Google Scholar

16. 

A. Wong et al., “General Bayesian estimation for speckle noise reduction in optical coherence tomography retinal imagery,” Opt. Express, 18 (8), 8338 –8352 (2010). http://dx.doi.org/10.1364/OE.18.008338 OPEXFF 1094-4087 Google Scholar

17. 

F. Luan and Y. Wu, “Application of RPCA in optical coherence tomography for speckle noise reduction,” Laser Phys. Lett., 10 (3), 035603 (2013). http://dx.doi.org/10.1088/1612-2011/10/3/035603 1612-2011 Google Scholar

18. 

L. Fang et al., “Sparsity based denoising of spectral domain optical coherence tomography images,” Biomed. Opt. Express, 3 (5), 927 –942 (2012). http://dx.doi.org/10.1364/BOE.3.000927 BOEICL 2156-7085 Google Scholar

19. 

L. Fang et al., “Fast acquisition and reconstruction of optical coherence tomography images via sparse representation,” IEEE Trans. Med. Imaging, 32 (11), 2034 –2049 (2013). http://dx.doi.org/10.1109/TMI.2013.2271904 ITMID4 0278-0062 Google Scholar

20. 

D. C. Adler, T. H. Ko and J. G. Fujimoto, “Speckle reduction in optical coherence tomography images by use of a spatially adaptive wavelet filter,” Opt. Lett., 29 (24), 2878 –2880 (2004). http://dx.doi.org/10.1364/OL.29.002878 OPLEDP 0146-9592 Google Scholar

21. 

V. Gupta et al., “Computerized automation of wavelet based denoising method to reduce speckle noise in OCT images,” in Int. Conf. on Information Technology and Applications in Biomedicine (ITAB 2008), (2008). http://dx.doi.org/10.1109/ITAB.2008.4570582 Google Scholar

22. 

M. A. Mayer et al., “Wavelet denoising of multiframe optical coherence tomography data,” Biomed. Opt. Express, 3 (3), 572 –589 (2012). http://dx.doi.org/10.1364/BOE.3.000572 BOEICL 2156-7085 Google Scholar

23. 

H. Rabbani, M. Sonka and M. D. Abramoff, “Optical coherence tomography noise reduction using anisotropic local bivariate Gaussian mixture prior in 3D complex wavelet domain,” J. Biomed. Imaging, 2013 417491 (2013). http://dx.doi.org/10.1155/2013/417491 Google Scholar

24. 

Z. Jian et al., “Speckle attenuation in optical coherence tomography by curvelet shrinkage,” Opt. Lett., 34 (10), 1516 –1518 (2009). http://dx.doi.org/10.1364/OL.34.001516 OPLEDP 0146-9592 Google Scholar

25. 

Z. Jian et al., “Three-dimensional speckle suppression in optical coherence tomography based on the curvelet transform,” Opt. Express, 18 (2), 1024 –1032 (2010). http://dx.doi.org/10.1364/OE.18.001024 OPEXFF 1094-4087 Google Scholar

26. 

R. Kafieh et al., “Curvature correction of retinal OCTs using graph-based geometry detection,” Phys. Med. Biol., 58 (9), 2925 –2938 (2013). http://dx.doi.org/10.1088/0031-9155/58/9/2925 PHMBA7 0031-9155 Google Scholar

27. 

R. Kafieh, H. Rabbani and I. Selesnick, “Three dimensional data-driven multi scale atomic representation of optical coherence tomography,” IEEE Trans. Med. Imaging, 34 (5), 1042 –1062 (2015). http://dx.doi.org/10.1109/TMI.2014.2374354 ITMID4 0278-0062 Google Scholar

28. 

N. M. Grzywacz et al., “Statistics of optical coherence tomography data from human retina,” IEEE Trans. Med. Imaging, 29 (6), 1224 –1237 (2010). http://dx.doi.org/10.1109/TMI.2009.2038375 ITMID4 0278-0062 Google Scholar

29. 

A. Buades, B. Coll and J.-M. Morel, “A non-local algorithm for image denoising,” in IEEE Computer Society Conf. on Computer Vision and Pattern Recognition (CVPR 2005), (2005). http://dx.doi.org/10.1109/CVPR.2005.38 Google Scholar

30. 

M. Elad and M. Aharon, “Image denoising via sparse and redundant representations over learned dictionaries,” IEEE Trans. Image Process., 15 (12), 3736 –3745 (2006). http://dx.doi.org/10.1109/TIP.2006.881969 IIPRE4 1057-7149 Google Scholar

31. 

K. Dabov et al., “Image denoising by sparse 3-D transform-domain collaborative filtering,” IEEE Trans. Image Process., 16 (8), 2080 –2095 (2007). http://dx.doi.org/10.1109/TIP.2007.901238 IIPRE4 1057-7149 Google Scholar

32. 

M. Niknejad, H. Rabbani and M. Babaie-Zadeh, “Image restoration using Gaussian mixture models with spatially constrained patch clustering,” IEEE Trans. Image Process., 24 (11), 3624 –3636 (2015). http://dx.doi.org/10.1109/TIP.2015.2447836 IIPRE4 1057-7149 Google Scholar

33. 

X. Li, “Exemplar-based EM-like image denoising via manifold reconstruction,” in 17th IEEE Int. Conf. on Image Processing (ICIP 2010), (2010). http://dx.doi.org/10.1109/ICIP.2010.5652529 Google Scholar

34. 

G. Yu, G. Sapiro and S. Mallat, “Solving inverse problems with piecewise linear estimators: from Gaussian mixture models to structured sparsity,” IEEE Trans. Image Process., 21 (5), 2481 –2499 (2012). http://dx.doi.org/10.1109/TIP.2011.2176743 IIPRE4 1057-7149 Google Scholar

35. 

Z. Amini and H. Rabbani, “Statistical modeling of retinal optical coherent tomography,” IEEE Trans. Med. Imaging, 35 (6), 1544 (2016). http://dx.doi.org/10.1109/TMI.2016.2519439 Google Scholar

36. 

A. P. Dempster, N. M. Laird and D. B. Rubin, “Maximum likelihood from incomplete data via the EM algorithm,” J. Royal Stat. Soc. Ser. B (Methodol.), 39 (1), 1 –38 (1977). Google Scholar

37. 

H. Rabbani et al., “Speckle noise reduction of medical ultrasound images in complex wavelet domain using mixture priors,” IEEE Trans. Biomed. Eng., 55 (9), 2152 –2160 (2008). http://dx.doi.org/10.1109/TBME.2008.923140 IEBEAX 0018-9294 Google Scholar

38. 

H. Rabbani, M. Vafadust and S. Gazor, “Image denoising based on a mixture of Laplace distributions with local parameters in complex wavelet domain,” in IEEE Int. Conf. on Image Processing, (2006). http://dx.doi.org/10.1109/ICIP.2006.313018 Google Scholar

39. 

H. Rabbani, R. Nezafat and S. Gazor, “Wavelet-domain medical image denoising using bivariate Laplacian mixture model,” IEEE Trans. Biomed. Eng., 56 (12), 2826 –2837 (2009). http://dx.doi.org/10.1109/TBME.2009.2028876 IEBEAX 0018-9294 Google Scholar

40. 

A. Pizurica et al., “Multiresolution denoising for optical coherence tomography: a review and evaluation,” Curr. Med. Imaging Rev., 4 (4), 270 –284 (2008). http://dx.doi.org/10.2174/157340508786404044 Google Scholar

41. 

Q. Yang et al., “Automated layer segmentation of macular OCT images using dual-scale gradient information,” Opt. Express, 18 (20), 21293 –21307 (2010). http://dx.doi.org/10.1364/OE.18.021293 OPEXFF 1094-4087 Google Scholar

42. 

L. Fang et al., “Segmentation based sparse reconstruction of optical coherence tomography images,” IEEE Trans. Med. Imaging, 36 (2), 407 –421 (2016). http://dx.doi.org/10.1109/TMI.2016.2611503 Google Scholar

43. 

R. Kafieh et al., “Intra-retinal layer segmentation of 3D optical coherence tomography using coarse grained diffusion map,” Med. Image Anal., 17 (8), 907 –928 (2013). http://dx.doi.org/10.1016/j.media.2013.05.006 Google Scholar

Biography

Zahra Amini received her BSc degree from Isfahan University of Technology (IUT), Iran, in 2007, her MSc degree from the University of Yazd, Iran, in 2011, both in electrical engineering (communications), and her PhD in biomedical engineering from Isfahan University of Medical Sciences, Iran, in 2016. She is currently an assistant professor in the Department of Biomedical Engineering, School of Advanced Technologies in Medicine, Isfahan University of Medical Sciences. Her research interests include medical image modeling, signal and image processing, sparse transforms, and brain computer interface.

Hossein Rabbani received his BSc degree in electrical engineering from Isfahan University of Technology in 2000 with the highest honors, and his MSc degree and PhD in bioelectrical engineering in 2002 and 2008, respectively, from Tehran Polytechnic. In 2007, he was at Queen’s University, as a visiting researcher; in 2011 at University of Iowa as a postdoctoral research scholar; and from 2013 to 2014 at Duke University Eye Center as a postdoctoral fellow. He is now senior member of IEEE and an associate professor in Biomedical Engineering Department and MISP Research Center, Isfahan University of Medical Sciences. His main research interests are medical image analysis and modeling, sparse transforms, and image restoration. He has published more than 110 papers as an author or coauthor in these areas.

© 2017 Society of Photo-Optical Instrumentation Engineers (SPIE) 1083-3668/2017/$25.00 © 2017 SPIE
Zahra Amini and Hossein Rabbani "Optical coherence tomography image denoising using Gaussianization transform," Journal of Biomedical Optics 22(8), 086011 (29 August 2017). https://doi.org/10.1117/1.JBO.22.8.086011
Received: 5 March 2017; Accepted: 3 August 2017; Published: 29 August 2017
Lens.org Logo
CITATIONS
Cited by 25 scholarly publications.
Advertisement
Advertisement
RIGHTS & PERMISSIONS
Get copyright permission  Get copyright permission on Copyright Marketplace
KEYWORDS
Optical coherence tomography

Denoising

Image denoising

Image segmentation

Filtering (signal processing)

Data modeling

Image processing

Back to Top