Noise Reduction Based on Multilevel DTCWT Using Bootstrap Method

International Journal of Advanced Research in Computer Engineering & Technology (IJARCET)
Volume 3 Issue 11, November 2014
Noise Reduction Based on Multilevel DTCWT
Using Bootstrap Method
T.Eswar Reddy and K.M.Hemambaran

Abstract- The main aim of this project is to reduce a
noise introduced by new method which is known as
bootstrap method. Image enhancement technique is
clearly based on the random spray sampling technique.
The output image of this exhibits noise with some
unknown statistical distribution. Non enhanced image is
consider to be free of noise or effected by certain levels of
noise. By taking an advantage of highest sensitivity of
HVS to change in brightness. The whole analysis is done
on luma channel of both noise free and added noisy
images. For giving an importance to directional content in
human vision perception, the analysis is performed
through the DTCWT. With the replacement of DWT we
are using DTCWT for allowing data directionality. Every
level of transform the standard deviation of noise free
image is computed across the six orientations of the dtcwt.
The obtained results show a map of directional structures
present in the noise free image. According to data
directionality in the dtcwt the shrunk coefficients and also
the coefficients from the non enhanced image are then
mixed. Finally by doing the inverse wavelet transform we
are obtaining the resultant image as a noise reduced
version of the enhanced image.
Keywords: Bootstrap Method, Dual tree complex
wavelet transform (DTWCT), Image enhancement, noise
reduction, random sprays.
I. INTRODUCTION
Image enhancement is a very important application in
digital world, as per consumer status it has never stopped
evolving. A novel multi resolution denoising method is used
to find a specific image quality problem this can be avoided
by using image enhancement algorithm based on random
spray sampling.
Random sprays are implemented in spatial domain to
collect a two dimensional points and used to sample an image
support in place of other techniques, and have been
previously used in works such as Provenzi [2], [3] and Kolas
[4]. It can be implemented when image is Human Visual
System (HVS).
The existing Denoising algorithm is gives good noise
reduction via coefficient shrinkage. And also the noise free
image is introduced in the form of partial reference images.
Having a reference allows the shrinkage process to be
Manuscript received Nov, 2014.
T.Eswar Reddy, ECE, SITAMS/JNTUA, Chittoor, India.
K.M.Hemambaran, ECE, SITAMS/JNTUA, Chittoor, India
ISSN: 2278 – 1323
data-driven and strong advantage of this method is to work
based on Dual-tree Complex Wavelet Transform.
.
Finally, Proposing a Bootstrap method to prevent data
loss in noise images compare to other methods i.e. existing
denoised method and RSR(Random Spray Retinex) and
RACE (combination of RSR and ACE(Automatic Color
Equalization)) and main advantage in image analysis is that
no assumptions have to be previously stated about the
distribution of elements within the population. It is an
empirical nonparametric technique commonly used to obtain
several measures of a given statistic when the underlying
statistical model is not known. Fig 1 shows traditional method
for recovering noise free image from noisy image.
Fig 1: Traditional Denoising Method
The paper is organized as follows. Section II contains
brief overview of Dual-tree Complex Wavelet Transform.
After that, Section III and IV contain the proposed Denoising
method and The Bootstrap Method, experimental results is
implemented in Section V and finally, Section VI gives the
conclusion.
II. DUAL -TREE COMPLEX WAVELET TRANSFORM
Discrete wavelet transform is introduced for application
of digital image processing like image denoising to pattern
recognition, image encoding and more. Being a complete and
invertible transform of 2D data the Discrete Wavelet
Transform gives rise to „checker board‟ pattern phenomenon.
To avoid these two problems we use steerable filters
introduced by Adelson and Freeman [19]. Which can be used
to decompose an image into a Steerable Pyramid, by means of
the Steerable Pyramid Transform (SPT) [8]? While, the SPT
is an over-complete representation of data , data orientation
will be shift invariant And it also having drawbacks that are
perfect reconstruction is not possible and computational
efficiency can be a concern.
After SPT, thus a further development has been
accomplished the Complex Wavelet Transform in order to
compute the energy response by using the Hilbert pair of
filters. To recovering the whole Fourier spectrum similar to
SPT this transform needs to over complete by a factor of 4
means there are 3 complex coefficients for each real one. The
CWT is also efficient and it can be computed through
All Rights Reserved © 2014 IJARCET
3876
International Journal of Advanced Research in Computer Engineering & Technology (IJARCET)
Volume 3 Issue 11, November 2014
separable filters, even though it still lacks the perfect
reconstruction property.
Later the Dual tree complex wavelet transform is
proposed by Kingsbury and referred as work by Selesnick
[21] for a comprehensive coverage on the DTCWT. The 2D
DTCWT is implemented using the scaling and wavelet
coefficients as shown in the below equations,
=
,
=
=
,
=
,
=
,
,
,
,
=
=
=
(1)
,
,
the chroma channel, and the whole operation is done only on
the luma channel by using this wavelet space. Here we are
using only the luminance channel because it does not provide
any visible color.
At final, we have to assume that the input is consider to
be free of noise or affected by certain levels of noise. If we
assume like this the original image contains some amount of
information for reduction of noise successfully. For using the
reference input image we have to reduce the noise in a noise
affected image perfectly by using DTCWT. The algorithm for
the proposed method is
ALOGRITHM: Algorithm for Proposed Noise Reduction
Method
(2)
The low pass and high pass wavelet filters h and g is as
shown below
(
repeat
Here j is the decomposition level.
is iteration dependent
When combining DTCWT coefficients the bases give rise
to two sets of real and 2D oriented wavelets
for j=1
do
for k=1
do
end for
for k=1
do
+
Rank of
if
then
Shrunk coefficients from
else
Coefficients from
end if
end for
end for
Fig. 2. Quasi-Hilbert pairs wavelets using in dual-tree complex wavelet
transform. Even part shows on top and odd part on bottom and also each
Pair is shown in a column
The most important characteristic of wavelets is they are
approximately Hilbert pairs and the coefficients deriving
from one tree is imaginary and the other tree is real and also
we obtain the desired 2D DTCWT.
)
until ssim ( ,
=concat (
ycbcr2rgb (
Inverse DTCWT
,
,
)
III PROPOSED NOISE REDUCTION METHOD
Main idea of this method can be simplified as follows:
user information conveys to the Human Visual System.
According to data directionality, the proposed method
using DTCWT produces wavelet coefficients by using those
coefficients we have to shrink the coefficients from non
enhanced image. The propose algorithm is as shown below.
HVS has been more sensitive to changes occurred in the
luma channel than the chroma channel [35]. And hence the
proposed method first converts the original image into
YCbCr. image and then the luma channel is separated from
ISSN: 2278 – 1323
All Rights Reserved © 2014 IJARCET
(a)
3877
International Journal of Advanced Research in Computer Engineering & Technology (IJARCET)
Volume 3 Issue 11, November 2014
a normalized map of directionally sensitive weights for a
given level j can be obtained as
Where the choice of γ depends on j as explained later on.
A shrunk version of the enhanced image‟s coefficients,
according to data directionality, is then computed as
+
+
Since the main interest is retaining directional information,
we obtain a rank for each of the non-enhanced coefficients as
,
(b)
where ord is the function that returns the rank according to
natural ordering. The output coefficients are then computed as
follows
(c)
Fig.3. Proposed method flowchart. (a) Luma channels for both non enhanced
and the enhanced images that are transformed using DTCWT and the
obtained coefficients are enlarged. By using inverse DTCWT the output
coefficients are transformed into the output image‟s luma channel (b) and (c)
Computation indicated by the box in Fig. 3(a) is performed per level of the
decomposition. (b) Directional energy map is first computed as the standard
deviation of sum-of-squares of the coefficients. A weight map is then
obtained by using the Bootstrap Method. The even coefficients of the
enhanced image are also ranked according to their magnitude. (c) Weight
map is used to scale the coefficients of the enhanced image. The resulting
scaled coefficients and the coefficients from the non enhanced image are
mixed according to the ranking. The process in (c) is illustrated for even
coefficients only, but it is repeated identically for odd coefficients.
A. Wavelet Coefficients Shrinkage
Assuming level j of the wavelet pyramid, one can compute
the energy for each direction of the non-enhanced image k
ε{1, 2, . . . , 6} as the sum of squares of the real coefficients
and the complex ones
Coefficients associated with non-directional data will have
similar energy in all directions. On the other hand, directional
data will give rise to high energy in one or two directions,
according to its orientation. The standard deviation of energy
across the six directions k = 1, 2. . . 6 is hence computed as a
measure of directionality.
=
Since the input coefficients are not normalized, it naturally
follows that the standard deviation is also non-normalized.
The Bootstrap Method is thus applied to normalize data range
and also for estimating standard errors. Such function is
sigmoid-like and it has been used to model the cones
responses of many species. The equation is as follows:
BM(x,
=
Where x is the quantity to be compressed, γ a real-valued
exponent and μ the data expected value or its estimate. Hence,
ISSN: 2278 – 1323
where ord is the function that returns the index of a
coefficient in
=1,2,...,6 when the set is sorted in a
descending order. The meaning of the whole sequence can be
roughly expressed as follows: where the enhanced image
shows directional content, shrink the two most significant
coefficients and replace the four less significant ones with
those from the non enhanced image.
The reason why only the two most significant coefficients
are taken from the shrunk ones of the enhanced image is to be
found in the nature of “directional content”. For an content of
an image to be directional, the responses across the six
orientations of the DTCWT need to be highly skewed. In
particular, any data orientation can be represented by a strong
response on two adjacent orientations, while the remaining
coefficients will be near zero. This will make it so that the two
significant coefficients are carried over almost un-shrunk. To
help for understanding the energies of the decompositions is
as shown in equation 9.
B. Parameter Tuning
When dealing with functions with free parameters, a
fundamental problem is finding optimum parameters values.
While this can often be attempted with optimization
techniques, such methods proved unfeasible in the case. To at
least provide a reasonable default value for γj and J (the
parameter of the Bootstrap method and the depth of the
complex wavelet decomposition, respectively) tests were
performed on three images from the USC-SIPI Image
Database [37].We are taking Lenna image, Splash image and
Girl image for showing the experimental results. In different
rounds, Gaussian, Poissonian and Speckle noise was added to
the luma channel of said images and the proposed noise
reduction method was run with a J = 2 and values for γj
varying from − 5 to 10 in unit steps for the two levels of the
decomposition. This allowed us to determine that values of
γ1, γ2 and γ3 equal to 1 represent a reasonable choice,
although non-optimal for all inputs. It should be noticed that
All Rights Reserved © 2014 IJARCET
3878
International Journal of Advanced Research in Computer Engineering & Technology (IJARCET)
Volume 3 Issue 11, November 2014
by doing so the Michaelis-Menten function is reduced to the
Naka–Rushton formulation [38]. Since J proved to be
drastically more dependent on the input image than γj, it was
impossible to determine a single optimum value. Therefore, J
was left as the only user set parameter of the method, with
reasonable bounds of Jmin = 1, Jmax = 2
IV PROPOSED METHOD
The Bootstrap Method:
The advantage of using bootstrap method with DTCWT
for analyzing the image is, there is no assumption on previous
statements about distribution of pixels within the original
image. This bootstrap is commonly used for which is
generally used for obtaining several measurements like mean
and median values. This is used to collect number of data
pixels in an image and then resample those collecting data
pixels. The proposed data set is formed by the pixels of the
original image is resample. Hence the several bootstrap data
sets are generated for a previously number of times and also
data collection is explained. The coefficients that are having
in the bootstrap data sets are generally used to estimate mean
and median values and also standard errors and unknown
values. The noise is to be determined for a typical number of
collected data sets and the bootstrapping value ranges from up
to 300.The bootstrap standard error (SEBn) represents the
standard deviation of replication of the bootstrap method
SEBn=
(18)
Where b represents the measured parameters that is mean
and median and so on. B represents the bootstrap number that
means the number of random data sets and n represents the
sub sampling used in this method and
(19)
The bootstrap coefficient of variation (CVBn) is estimated as
follows
CVBn =
(20)
Generally the bootstrap method is obtained for represents
of this sampling areas and here we assume that the area of the
whole image occupied is consider to be an object for
analyzing purpose and it contains the total pixels of original
image and hence the sampling area is considering as equal
value of sizes. Every bootstrap data set is formed for
measuring the sampling area of original image and the
calculation of unknown statistical parameters that is mean and
median etc. This will results that a large number of
estimations of the unknown statistics parameters i.e. one for
every bootstrap data range. It is used for determining the
coefficient of variation (CVBn) and standard error (SEBn) for
the particular sampling area is as shown in the below Figure 1.
The minimum value for coefficient of variation or the
standard error is defined as wit increasing the size of the
sampling area and it is possible to determine the
representative sampling area size. The advantage of the
bootstrap method is its simplicity. It is straight forward way
for estimating the standard errors of complex coefficients.
And also it is an appropriate way for controlling and checking
stability of the results. The general applications of bootstrap
method are data missing problems and censored data.
ISSN: 2278 – 1323
Estimation of Noise Statistics by Bootstrapping:
Bootstrapping provides a practical procedure for estimating
the parameters such as the mean and standard deviation etc.
Depending on resampling process the parameters can be
measured from an approximated distribution. This
distribution is obtained from the available bootstrap data set.
From the general Gaussian distribution we have to assume
that a set of observations and it is implemented by using a
number of resample from the bootstrap data that are all equal
size. Every resample is obtained by sampling one by one witjh
replacement from the original image pixel values. Another
advantage of bootstrapping is it provides easy procedure for
estimating the noise pixels in an enhanced image from
complex transform coefficients of the observed bootstrap
resampling pixel values. This bootstrap method is useful
while satisfying the below two that is (i) the theoretical
distribution of one statistic is unknown (ii) for statistical noise
the size of the sample is not enough.
Here we are using bootstrap method for estimating the
noise intensity level
according to the noise corrupted
pixels. We are taking the noise pixel values while sample the
image at the first time the noise intensity level
is only one
parameter of that image. Therefore the value of
is to be
estimate from the resampling original image. Our main idea of
implementing the bootstrap method is more number of
samples from the same pixels that can be efficiently replaced
by resampling one that is by doing first sampling. By using the
mathematical notations and equations let be the mean based
on one random bootstrapping sample of size
m
. The estimated noise
intensity level
is obtained by selecting the median value
of
based on Z different bootstrapping samples
where each sample is of size m. The following formula can be
used to calculate different
and based on calculated
the
can be obtained.
Where
is the absolute value of
element in
bootstrapping sample pixel. The estimation of
is clearly
obtained by using bootstrapping process. This methodology is
a flexible and robust method for deriving sampling
distributions and also standard errors.
All Rights Reserved © 2014 IJARCET
3879
International Journal of Advanced Research in Computer Engineering & Technology (IJARCET)
Volume 3 Issue 11, November 2014
V EXPERIMENTAL RESULTS
In this proposed Bootstrap Method with DTCWT the peak
signal to noise ratio is increased compared to other denoising
methods. Here we are adding different noise like Gaussian
noise, poission noise and speckle noise for the enhanced
luminance image. We are effectively reducing the all type of
noises and increase PSNR value. The peak signal to noise
ratio (PSNR) is defined as,
PSNR=10
The below figures shows the comparison results of DTCWT
and DTCWT using Bootstrap Method with different noise
affected images.
(c)
(d)
Figure 5: (a) Input image
(b) poission noise image
(c) Denoised luminance image (d) Denoised RGB image
PROPOSED METHOD:
Lenna image:
EXISTING METHOD:
Lenna image:
(a)
(b)
(a) (b)
(c)
(d)
Figure 6: (a) Input image
(b) Gaussian noise image
(c) Denoised luminance image (d) Denoised RGB image
(c)
(d)
Figure 4: (a) Input image,
(b) Gaussian noise image
(c) Denoised luminance image (d) Denoised RGB image
Girl image:
Girl image:
(a)
(a)
ISSN: 2278 – 1323
(b)
(b)
All Rights Reserved © 2014 IJARCET
3880
International Journal of Advanced Research in Computer Engineering & Technology (IJARCET)
Volume 3 Issue 11, November 2014
(c)
(d)
Figure 7: (a) Input image
(b) Poission noise image
(c) Denoised luminance image (d) Denoised RGB image
The below table shows the comparisons of PSNR values of
existing and proposed method of different images at different
noises.
TABLE 1: EXISTING METHOD
Noise type
Image name
Gaussian
Noisy
Denoised
PSNR
PSNR
Lenna
27.68
33.90
Girl
28.24
35.39
Splash
27.27
37.15
VI CONCLUSION&FUTURE WORK
In this paper, we are using wavelet based transforms for image
denoising. Noise is one of the major problems in image
processing that occurs while capturing the image and the
image is transmitting through a channel. Here we are focused
on different types of noises like Gaussian, poission and
speckle noises. In this propose method we are using Bootstrap
Method with DTCWT. The DTCWT using Bootstrap Method
has several advantages over traditional denoising methods
that are it is having shift invariance property and good
directional selectivity.
The above experimental results show that different images
are affected by different type of noises. By using our proposed
algorithm we have to denoised the noise affected images
effectively and also calculate the image quality parameter
which is known as PSNR value. Comparison results for
existing and proposed methods are as shown in the above
tables 1 and 2. The proposed method produces good quality
output and removing noise without changing the directional
structures in the image.
Furthermore, improve the speed of the algorithm by
avoiding iterations and also DTCWT is a tool for other
activities such as image quality measures. Thus in future
applied for audio and video signals denoising.
REFERENCES
Poission
Speckle
Lenna
28.50
33.92
Girl
29.08
34.38
Splash
30.26
37.13
Lenna
27.66
33.89
Girl
28.27
35.47
Splash
27.28
36.98
Noisy
Denoised
PSNR
PSNR
Lenna
27.68
37.94
Girl
28.24
42.23
Splash
27.26
47.10
Lenna
28.50
37.82
Girl
29.08
40.50
Splash
30.26
45.08
Lenna
27.66
40.87
Girl
28.27
44.03
Splash
27.28
41.10
TABLE 2: PROPOSED SYSTEM
Noise type
Image name
Gaussian
Poission
Speckle
ISSN: 2278 – 1323
[1] Kingsbury, selesnick, „‟Noise reduction based on partial-reference, Dual
Tree Complex Wavelet Transform Shrinkage‟‟, IEEE Transaction on
IMAGE PROCESSING, vol. 22, no. 5, May 2013
[2] M. Fierro, W.-J. Kyung and Y.-H. Ha, “Dual-tree complex wavelet
transform based denoising for random spray image enhancement
methods,” in Proc. 6th Eur. Conf. Color Graph, Imag. Vis., 2012, pp.
194–199.
[3] E. Provenzi, M. Fierro, A. Rizzi, L. De Carli, D. Gadia, and D.
Marini,“Random spray retinex: A new retinex implementation to
investigate the local properties of the model,” vol. 16, no. 1, pp.
162–171, Jan. 2007.
[4] E. Provenzi, C. Gatta, M. Fierro, and A. Rizzi, “A spatially variant white
patch and gray-world method for color image enhancement driven by
local contrast,” vol. 30, no. 10, pp. 1757–1770, 2008.
[5] Ø. Kolas‟, I. Farup, and A. Rizzi, “Spatio-temporal retinex-inspired
envelope with stochastic sampling: A framework for spatial color
algorithms,” J. Imag. Sci. Technol., vol. 55, no. 4, pp. 1–10, 2011.
[6] A. Chambolle, R. De Vore, N.-Y. Lee, and B. Lucier, “Nonlinear
wavelet image Processing: Variational problems, compression, and
noise removal through wavelet shrinkage,” IEEE Trans. Image
Process., vol. 7, no. 3, pp. 319–335, Mar. 1998.
[7] E. P. Simoncelli and W. T. Freeman, “The steerable pyramid: A flexible
architecture for multi-scale derivative computation,” in Proc. 2nd Annu.
Int. Conf. Image Process., Oct. 1995, pp. 444–447.
[8] H. Rabbani, “Image denoising in steerable pyramid domain based on a
local laplace prior,” Pattern Recognit., vol. 42, no. 9, pp. 2181–2193,
Sep. 2009.
[9] K. Li, J. Gao, and W. Wang, “Adaptive shrinkage for image denoising
based on contourlet transform,” in Proc. 2nd Int. Symp. Intell. Inf.
Technol. Appl., vol. 2. Dec. 2008, pp. 995–999.
[10] Q. Guo, S. Yu, X. Chen, C. Liu, and W. Wei, “Shearlet-based image
denoising using bivariate shrinkage with intra-band and opposite
orientation dependencies,” in Proc. Int. Joint Conf. Comput. Sci.
Optim., vol. 1. Apr. 2009, pp. 863–866.
[11] X. Chen, C. Deng, and S. Wang, “Shearlet-based adaptive shrinkage
threshold for image denoising,” in Proc. Int. Conf. E-Bus.
E-Government, Nanchang, China, May 2010, pp. 1616–1619.
[12] J. Zhao, L. Lu, and H. Sun, “Multi-threshold image denoising based on
shearlet transform,” Appl. Mech. Mater., vols. 29–32, pp. 2251–2255,
Aug. 2010.
All Rights Reserved © 2014 IJARCET
3881
International Journal of Advanced Research in Computer Engineering & Technology (IJARCET)
Volume 3 Issue 11, November 2014
[13] N. G. Kingsbury, “The dual-tree complex wavelet transform: A new
technique for shift invariance and directional filters,” in Proc. 8th IEEE
Digit. Signal Process. Workshop, Aug. 1998, no. 86, pp. 1–4.
[14] D. L. Donoho and I. M. Johnstone, “Threshold selection for wavelet
shrinkage of noisy data,” in Proc. 16th Annu. Int. Conf. IEEE Eng Adv.,
New Opportunities Biomed. Eng. Eng. Med. Biol. Soc., vol. 1. Nov.
1994, pp. A24–A25.
[15] W. Freeman and E. Adelson, “The design and use of steerable filters,”
IEEE Trans. Pattern Anal. Mach. Intell., vol. 13, no. 9, pp. 891–906,
Sep. 1991.
[16] N. G. Kingsbury, “Image Processing with complex wavelets,” Philos.
Trans. Math. Phys. Eng. Sci., vol. 357, no. 1760, pp. 2543–2560, 1999.
[17] I. W. Selesnick, R. G. Baraniuk, and N. G. Kingsbury, “The dual-tree
complex wavelet transform: A coherent framework for multiscale signal
and image Processing,” vol. 22, no. 6, pp. 123–151, Nov. 2005.
[18] S. Cai and K. Li. Matlab Implementation of Wavelet Transforms.
Polytechnic
Univ.,
Brooklyn,
NY
[Online].
Available:
http://eeweb.poly.edu/iselesni/WaveletSoftware/index.html
[19] E. H. Land and J. J. McCann, “Lightness and retinex theory,” J. Opt.
Soc. Amer. A, vol. 61, no. 1, pp. 1–11, 1971.
[20] D. Marini and A. Rizzi, “A computational approach to color adaptation
effects,” Image Vis. Comput., vol. 18, no. 13, pp. 1005–1014, 2000.
[21] R. C. Gonzales and R. E. Woods, Digital Image Processing, 2nd ed.
Englewood Cliffs, NJ: Prentice-Hall, 2002.
[22] J. van de Weijer, T. Gevers, and A. Gijsenij, “Edge-based color
constancy,” IEEE Trans. Image Process., vol. 16, no. 9, pp. 2207–2214,
Sep. 2007.
[23] USC-SIPI Image Database. (2009) [Online]. Available:
http://sipi.usc.edu/database/
ABOUT THE AUTHORS
T.Eswar Reddyreceived B.Tech degree from
JNT University, Anantapur and is pursuing his
M.tech
degree
from
JNT
University,
Anantapur.Hisareas of interest are digital image
processing and networking.
Mr. K.M Hemambaran working as Assistant
Professor in the department of ECE in SITAMS,
chittoor. He received AMIE degree from
institution of engineering (India) and he received
M.tech degree from JNT University, Hyderabad.
His areas of interest are Digital systems and
Digital image processing.
ISSN: 2278 – 1323
All Rights Reserved © 2014 IJARCET
3882