Upload
anusha-yangali
View
228
Download
0
Embed Size (px)
Citation preview
8/2/2019 Anusha Final Tech Doc
1/23
1
Chapter 1
INTRODUCTION TO IMAGE TRANSFORMATION AND COMPRESSION
USING WAVELETS1. Introduction
Often signals we wish to process are in the time-domain, but in order to process them more
easily other information, such as frequency, is required. Mathematical transforms translate the
information of signals into different representations. For example, the Fourier transform converts a
signal between the time and frequency domains, such that the frequencies of a signal can be seen.
However the Fourier transform cannot provide information on which frequencies occur at specific
times in the signal as time and frequency are viewed independently. To solve this problem the
Short Term Fourier Transform (STFT) introduced the idea of windows through which different
parts of a signal are viewed. For a given window in time the frequencies can be viewed. However
Heisenburgs Uncertainty Principle states that as the resolution of the signal improves in the time
domain, by zooming on different sections, the frequency resolution gets worse. Ideally, a method
of multiresolution is needed, which allows certain parts of the signal to be resolved well in time,
and other parts to be resolved well in frequency.
The power and magic of wavelet analysis is exactly this multiresolution. Images contain
large amounts of information that requires much storage space, large transmission bandwidths and
long transmission times. Therefore it is advantageous to compress the image by storing only the
essential information needed to reconstruct the image. An image can be thought of as a matrix of
pixel (or intensity) values. In order to compress the image, redundancies must be exploited, for
example, areas where there is little or no change between pixel values. Therefore images having
large areas of uniform colour will have large redundancies, and conversely images that have
frequent and large changes in colour will be less redundant and harder to compress. The
approximation subsignal shows the general trend of pixel values, and three detail subsignals show
the vertical, horizontal and diagonal details or changes in the image. Ideally, during compression
the number of zeros and the energy retention will be as high as possible. However, as more zeros
are obtained more energy is lost, so a balance between the two needs to be found.
Chapter 2
LITERATURE SURVEY
8/2/2019 Anusha Final Tech Doc
2/23
2
2.1 The need for Wavelets
Often signals we wish to process are in the time-domain, but in order to process them more
easily other information, such as frequency, is required. A good analogy for this idea is given by
Hubbard. In order to do this calculation we would find it easier to first translate the numerals in toour number system, and then translate the answer back into a roman numeral. The result is the
same, but taking the detour into an alternative number system made the process easier and quicker.
Similarly we can take a detour into frequency space to analysis or process a signal.
2.1.1 Multiresolution and Wavelets
The power of Wavelets comes from the use of multiresolution. High frequency parts of the
signal use a small window to give good time resolution, low frequency parts use a big window to
get good frequency information. As frequency resolution gets bigger the time resolution must get
smaller. .
In Fourier analysis a signal is broken up into sine and cosine waves of different
frequencies, and it effectively re-writes a signal in terms of different sine and cosine waves.
Wavelet analysis does a similar thing, it takes a mother wavelet, then the signal is translated into
shifted and scale versions of this mother wavelet.
8/2/2019 Anusha Final Tech Doc
3/23
3
Figure 2.1 The different transforms provided different resolutions of time and frequency.
2.1.2 The Continuous Wavelet Transform (CWT)
The continuous wavelet transform is the sum over all time of scaled and shifted versions of
the mother wavelet . Calculating the CWT results in many coefficients C, which are functions of
scale and translation.
The translation, , is proportional to time information and the scale, s, is proportional to the
inverse of the frequency information. To find the constituent wavelets of the signal, the coefficients
should be multiplied by the relevant version of the mother wavelet. The scale of a wavelet simply
means how stretched it is along the x-axis, larger scales are more stretched:
8/2/2019 Anusha Final Tech Doc
4/23
4
Figure 2.2 The db8 wavelet shown at two different scales
The translation is how far it has been shifted along the x-axis. Figure 2.3 shows a wavelet,
figure 2.4 shows the same mother wavelet translated by k:
Figure 2.3 The translated wavelet.
8/2/2019 Anusha Final Tech Doc
5/23
5
Figure 2.4 The wavelet translated by k
The Wavelet Toolbox Users Guide [7] suggests five easy steps to compute the CWT
coefficients for a signal.
1. Choose a wavelet, and compare this to the section at the start of the signal.
2. Calculate C, which should measure how similar the wavelet and the section of the signal are.
3. Shift the wavelet to the right by translation , and repeat steps 1 and 2, calculating values of C
for all translations.
4. Scale the wavelet, and repeat steps 1-3.
5. Repeat 4 for all scales.
Chapter 3
WAVELET TRANSFORMS
Wavelet transform is capable of providing the time and frequency information
simultaneously, hence it gives a time-frequency representation of the signal. When we are
interested in knowing what spectral component exists at any given instant of time, we want to
know the particular spectral component at that instant. In these cases it may be very beneficial to
know the time intervals these particular spectral components occur. For example, in EEGs, the
latency of an event-related potential is of particular interest (Event-related potential is the response
of the brain to a specific stimulus like flash-light, the latency of this response is the amount of time
elapsed between the onset of the stimulus and the response)examplecomes from a reference.
8/2/2019 Anusha Final Tech Doc
6/23
6
Wavelets (small waves) are functions defined over a finite interval and having an average
value of zero. The basic idea of the wavelet transform is to represent any arbitrary function (t) as
a superposition of a set of such wavelets or basis functions. These basis functions are obtained
from a single wave, by dilations or contractions (scaling) and translations (shifts). For pattern
recognition continuous wavelet transform is better. The Discrete Wavelet Transform of a finite
length signal x (n) having N components, for example, is expressed by an N x N matrix similar to
the Discrete cosine transform (DCT).
3.1 WAVELET-BASED COMPRESSION
Digital image is represented as a two-dimensional array of coefficients, each coefficient
representing the brightness level in that point. We can differentiate between coefficients as more
important ones, and lesser important ones. Most natural images have smooth color variations, with
the fine details being represented as sharp edges in between the smooth variations. Technically, the
smooth variations in color can be termed as low frequency variations and the sharp variations as
high frequency variations.
The basic difference wavelet-based and Fourier-based techniques is that short-time Fourier-
based techniques use a fixed analysis window, while wavelet-based techniques can be considered
using a short window at high spatial frequency data and a long window at low spatial frequency
data. This makes DWT more accurate in analyzing image signals at different spatial frequency, and
thus can represent more precisely both smooth and dynamic regions in image. The compressor
includes forward wavelet transform, Quantizer, and Lossless entropy encoder. The corresponding
decompressed is formed by Lossless entropy decoder, de-Quantizer, and an inverse wavelet
transform. Wavelet-based image compression has good compression results in both rate and
distortion sense.
3.2 PRINCIPLES OF IMAGE COMPRESSION
Image Compression is different from data compression (binary data). When we apply
techniques used for binary data compression to the images, the results are not optimal. In Lossless
compression the data (binary data such as executables, documents etc) are compresses such that
when decompressed, it give an exact replica of the original data. They need to be exactly
reproduced when decompressed. On the other hand, images need not be reproduced 'exactly'.
8/2/2019 Anusha Final Tech Doc
7/23
7
In images the neighboring pixels are correlated and therefore contain redundant
information. Before we compress an image, we first find out the pixels, which are correlated. The
fundamental components of compression are redundancy and irrelevancy reduction. Redundancy
means duplication and Irrelevancy means the parts of signal that will not be noticed by the signal
receiver, which is the Human Visual System (HVS).
There are three types of redundancy can be identified:
1) Spatial Redundancy i.e. correlation between neighboring pixel values.
2) Spectral Redundancy i.e. correlation between different color planes or spectral bands.
3) Temporal Redundancy i.e. correlation between adjacent frames in a sequence of images (in
video applications).
Image compression focuses on reducing the number of bits needed to represent an image
by removing the spatial and spectral redundancies.
3.3 PROCEDURE FOR COMPRESSING AN IMAGE
Just as in any other image compression schemes the wavelet method for image
compression also follows the same procedure:
i) Decomposition of signal using filters banks.
ii) Down sampling of Output of filter bankiii) Quantization of the above
iv) Finally encoding
For decoding
i) Up samples
ii) Re composition of signal with filter bank.
3.4 PRINCIPLES OF USING TRANSFORM AS SOURCE ENCODER
Before going into details of wavelet theory and its application to image compression, I willgo ahead and describe few terms most commonly used in this report.
Mathematical transformations are applied to signals to obtain further information from that
signal that is not readily available in the raw signal.
Various wavelet transforms
Discrete Wavelet Transform
8/2/2019 Anusha Final Tech Doc
8/23
8
Haar Wavelet Transform
3.5 DISCRETE WAVELET TRANSFORM
At the heart of the analysis (or compression) stage of the system is the forward discrete
wavelet transform (DWT). A block diagram of a wavelet based image compression system is
shown in Figure 3.1. Here, the input image is mapped from a spatial domain, to a scale-shift
domain. This transform separates the image information into octave frequency subbands. The
expectation is that certain frequency bands will have zero or negligible energy content; thus,
information in these bands can be thrown away or reduced so that the image is compressed without
much loss of information.
The DWT coefficients are then quantized to achieve compression. Information lost during
the quantization process cannot be recovered and this impacts the quality of the reconstructed
image. Due to the nature of the transform, DWT coefficients exhibit spatial correlation, that are
exploited by quantization algorithms like the embedded zero-tree wavelet (EZW) and set
partitioning in hierarchical trees (SPIHT) for efficient quantization. The quantized coefficients may
then be entropy coded; this is a reversible process that eliminates any redundancy at the output of
the quantizer.
Figure 3.1: A wavelet based image compression system.
In the synthesis (or decompression) stage, the inverse discrete wavelet transform recovers
the original image from the DWT coefficients. In the absence of any quantization the reconstructed
image will be identical to the input image. However, if any information was discarded during the
8/2/2019 Anusha Final Tech Doc
9/23
9
quantization process, the reconstructed image will only be an approximation of the original image.
Hence this is called lossy compression. The more an image is compressed, the more information is
discarded by the quantizer; the result is a reconstructed image that exhibits increasingly more
artifacts. Certain integer wavelet transforms exist that result in DWT coefficients that can be
quantized without any loss of information. These result in lossless compression, where the
reconstructed image is an exact replica of the input image. However, compression ratios achieved
by these transforms are small compared to lossy transforms (e.g. 4:1 compared to 40:1).
3.6 2-D DISCRETE WAVELET TRANSFORM
A 2-D separable discrete wavelet transform is equivalent to two consecutive 1-D
transforms. For an image, a 2-D DWT is implemented as a 1-D row transform followed by a 1-D
column transform. Transform coefficients are obtained by projecting the 2-D input image x(u, v)
onto a set of 2-D basis functions that are expressed as the product of two 1-D basis.
The 2-D DWT (analysis) can be expressed as the set of equations shown below. The scaling
function and wavelets and (and the corresponding
transform coefficients X(N, j,m), X(1)(N, j,m), X(2)(N, j,m) and X(3)(N, j,m)) correspond to different
subbands in the decomposition. X(N, j,m) are the coarse coefficients that constitute the LL
subband. The X(1)(N, j,m) coefficients contain the vertical details and correspond to the LH
subband. The X(2)(N, j,m) coefficients contain the horizontal details and correspond to the HL
subband. The X(3)(N, j,m) coefficients represent the diagonal details in the image and constitute the
HH subband.
The histograms for coefficients in all subbands shown in, illustrate the advantage of
transform coding. The original 512 512 lighthouse image had few zero pixel values. Thus a
single-level decomposition at scale (N +1) has four subbands of coefficients as shown in Figure
3.2.
8/2/2019 Anusha Final Tech Doc
10/23
10
Figure 3.2: Single-level 2-D wavelet decomposition (analysis).
The synthesis bank performs the 2-D IDWT to reconstruct . 2-D IDWT is
given in equation below.
A single stage of a 2-D filter bank is shown in Figure 3.3. First, the rows of the input image
are filtered by the highpass and lowpass filters. The outputs from these filters are downsampled by
8/2/2019 Anusha Final Tech Doc
11/23
11
two, and then the columns of the outputs are filtered and downsampled to decompose the image
into four subbands. The synthesis stage performs upsampling and filtering to reconstruct the
original image.
Multiple levels of decomposition are achieved by iterating the analysis stage on only the
LL band.
Figure 3.3: One level filter bank for computation of 2-D DWT and IDWT.
For l levels of decomposition, the image is decomposed into 3l + 1 subbands. Figure 3.4
shows the filter bank structure for three levels of decomposition of an input image, and Figure 3.5
shows the subbands in the transformed image.
Figure 3.4: Analysis section of a three level 2-D filter bank.
8/2/2019 Anusha Final Tech Doc
12/23
12
Figure 3.5: Three level wavelet decomposition of an image.
Chapter 4
HAAR WAVELETS
4.1 THE HAAR WAVELET
The Haar wavelet algorithms published here are applied to time series where the number of
samples is a power of two (e.g., 2, 4, 8, 16, 32, 64...) The Haar wavelet uses a rectangular window
to sample the time series. The first pass over the time series uses a window width of two. The
window width is doubled at each step until the window encompasses the entire time series. Each
pass over the time series generates a new time series and a set of coefficients. The new time series
is the average of the previous time series over the sampling window. The coefficients represent the
average change in the sample window. For example, if we have a time series consisting of the
values v0, v1, ... vn, a new time series, with half as many points is calculated by averaging the pointsin the window. If it is the first pass over the time series, the window width will be two, so two
points will be averaged:
for (i = 0; i < n; i = i + 2)
si = (vi + vi+1)/2;
The 3-D surface below graphs nine wavelet spectrums generated from the 512 point AMAT
close price time series. The x-axis shows the sample number, the y-axis shows the average value at
that point and the z-axis shows log2 of the window width.
8/2/2019 Anusha Final Tech Doc
13/23
13
Figure 4.1: The Wavelet Spectrum
The wavelet coefficients are calcalculated along with the new average time series values.
The coefficients represent the average change over the window. If the windows width is two this
would be:
for (i = 0; i < n; i = i + 2)
ci = (vi - vi+1)/2;
The graph below shows the coefficient spectrums. As before the z-axis represents the log2
of the window width. The y-axis represents the time series change over the window width.
Somewhat counter intutitively, the negative values mean that the time series is moving upward
vi is less than vi+1
so vi - vi+1 will be less than zero
Positive values mean the the time series is going down, since vi is greater than vi+1. Note
that the high frequency coefficient spectrum (log2(windowWidth) = 1) reflects the noisiest part of
the time series. Here the change between values fluctuates around zero.
8/2/2019 Anusha Final Tech Doc
14/23
14
Figure 4.2: Haar coefficient spectrum
Plot of the Haar coefficient spectrum. The surface plots the highest frequency spectrum in
the front and the lowest frequency spectrum in the back. Note that the highest frequency spectrum
contains most of the noise.
4.2 FILTERING SPECTRUM
The wavelet transform allows some or all of a given spectrum to be removed by setting the
coefficients to zero. The signal can then be rebuilt using the inverse wavelet transform. Plots of the
AMAT close price time series with various spectrums filtered out are shown here.
4.3 NOISE FILTERS
Each spectrum that makes up a time series can be examined independently. A noise filter
can be applied to each spectrum removing the coefficients that are classified as noise by setting the
coefficients to zero.
4.4 WAVELETS VS. SIMPLE FILTERS
How do Haar wavelet filters compare to simple filters, like windowed mean and median
filters. Whether a wavelet filter is better than a windowed mean filter depends on the application.
8/2/2019 Anusha Final Tech Doc
15/23
15
The wavelet filter allows specific parts of the spectrum to be filtered. For example, the entire high
frequency spectrum can be removed Or selected parts of the spectrum can be removed, as is done
with the gaussian noise filter. The scaling function (which uses the hn coefficients) produces N/2
elements that are a smoother version of the original data set. In the case of the Haar transform,
these elements are a pairwise average. The last step of the Haar transform calculates one scaling
value and one wavelet value. In the case of the Haar transform, the scaling value will be the
average of the input data.
The power of Haar wavelet filters is that they can be efficiently calculated and they
provide a lot of flexibility. They can potentially leave more detail in the time series, compared to
the mean or median filter. To the extent that this detail is useful for an application, the wavelet
filter is a better choice.
4.5 LIMITATIONS OF THE HAAR WAVELET TRANSFORM
The Haar wavelet transform has a number of advantages:
It is conceptually simple.
It is fast.
It is memory efficient, since it can be calculated in place without a temporary array.
It is exactly reversible without the edge effects that are a problem with other wavelettrasforms.
The Haar transform also has limitations, which can be a problem for some applications. In
generating each set of averages for the next level and each set of coefficients, the Haar transform
performs an average and difference on a pair of values. Then the algorithm shifts over by two
values and calculates another average and difference on the next pair. The high frequency
coefficient spectrum should reflect all high frequency changes. The Haar window is only two
elements wide. If a big change takes place from an even value to an odd value, the change will notbe reflected in the high frequency coefficients.
For example, in the 64 element time series graphed below, there is a large drop between
elements 16 and 17, and elements 44 and 45.
8/2/2019 Anusha Final Tech Doc
16/23
16
Figure 4.3: frequency coefficient spectrum
Chapter 5
IMAGE COMPRESSION USING WAVELETS
Images require much storage space, large transmission bandwidth and long transmission
time. The only way currently to improve on these resource requirements is to compress images,
such that they can be transmitted quicker and then decompressed by the receiver. In image
processing there are 256 intensity levels (scales) of grey. 0 is black and 255 is white. Each level isrepresented by an 8-bit binary number so black is 00000000 and white is 11111111. An image can
therefore be thought of as grid of pixels, where each pixel can be represented by the 8-bit binary
value for grey-scale.
Figure 5.1 Intensity level representation of image
8/2/2019 Anusha Final Tech Doc
17/23
17
The resolution of an image is the pixels per square inch. (So 500dpi means that a pixel is
1/500th of an inch). To digitise a one-inch square image at 500 dpi requires 8 x 500 x500 = 2
million storage bits. Using this representation it is clear that image data compression is a great
advantage if many images are to be stored, transmitted or processed. According to [6] "Image
compression algorithms aim to remove redundancy in data in a way which makes image
reconstruction possible." This basically means that image compression algorithms try to exploit
redundancies in the data; they calculate which data needs to be kept in order to reconstruct the
original image and therefore which data can be thrown away. By removing the redundant data,
the image can be represented in a smaller number of bits, and hence can be compressed. During
compression however, energy is lost because thresholding changes the coefficient values and hence
the compressed version contains less energy.But what is redundant information? Redundancy reduction is aimed at removing
duplication in the image. According to Saha there are two different types of redundancy relevant to
images:
(i) Spatial Redundancy . correlation between neighbouring pixels.
(ii) Spectral Redundancy - correlation between different colour planes and spectral bands.
Where there is high correlation, there is also high redundancy, so it may not be necessary to record
the data for every pixel.
There are two parts to the compression:
1. Find image data properties; grey-level histogram, image entropy, correlation functions etc..
2. Find an appropriate compression technique for an image of those properties.
5.1 Image Data Properties relavent to wavelets
In order to make meaningful comparisons of different image compression techniques it is
necessary to know the properties of the image. One property is the image entropy; a highly
correlated picture will have a low entropy. For example a very low frequency, highly correlated
image will be compressed well by many different techniques. A compression algorithm that is good
for some images will not necessarily be good for all images, it would be better if we could say
what the best compression technique would be given the type of image we have. One way of
calculating entropy is suggested by [6] Information redundancy, r, is r = b - He [6], where b is the
smallest number of bits for which the image quantisation levels can be represented.
8/2/2019 Anusha Final Tech Doc
18/23
18
Information redundancy can only be evaluated if a good estimate of image entropy is
available, but this is not usually the case because some statistical information is not known. More
specifically this means that upsampling can be used to remove every second sample. The scale has
now been doubled. The resolution has also been changed, the filtering made the frequency
resolution better, but reduced the time resolution. An estimate of He can be obtained from a grey-
level histogram. If h(k) is the frequency of grey-level k in an image f, and image size is MxN then
an estimate of P(k) can be made:
The Compression ratio K = b / He.
5.2 WAVELETS AND COMPRESSION
Wavelets are useful for compressing signals but they also have far more extensive uses.They can be used to process and improve signals, in fields such as medical imaging where image
degradation is not tolerated they are of particular use. They can be used to remove noise in an
image, for example if it is of very fine scales, wavelets can be used to cut out this fine scale,
effectively removing the noise.
5.2.1 The Fingerprint example
The FBI have been using wavelet techniques in order to store and process fingerprint
images more efficiently. The problem that the FBI were faced with was that they had over 200
Million sets of fingerprints, with up to 30,0000 new ones arriving each day, so searching through
them was taking too long. The FBI thought that computerizing the fingerprint images would be a
better solution, however it was estimated that checking each fingerprint would use 600Kbytes of
memory and even worse 2000 terabytes of storage space would be required to hold all the image
data. The FBI then turned to wavelets for help, adapting a technique to compress each image into
just 7% of the original space. Even more amazingly, according to Kiernan[8], when the images are
decompressed they show "little distortion". However this was unsatisfactory, trying to compressimages this way into less than 10% caused "tiling artefacts" to occur, leaving marked boundaries in
the image. The basic steps used in the fingerprint compression were:
(1) Digitise the source image into a signals
(2) Decompose the signal s into wavelet coefficients
8/2/2019 Anusha Final Tech Doc
19/23
19
(3) Modify the coefficients from w, using thresholding to a sequence w.
(4) Use quantisation to convert w to q.
(5) Apply entropy encoding to compress q to e.
5.3 WAVELETS AND DECOMPRESSION
Whenever we compress an image, at the receiver end to get the original form of the image
we have to decompress it. This can be done by image decompression models. The main blocks in
this decompression models are channel decoder and source decoder. A source decoder consists of
two stages, namely symbol decoder and inverse mapper. These blocks do the exact opposite
operation of channel as well as source encoder blocks.
The source decoder contains only two components: a symbol decoder and an inverse
mapper. These blocks perform, in reverse order, the inverse of the source encoders symbol
encoder and mapper blocks. Because quantization results in irreversible information loss, an
inverse quantizer block is not included in the general decoder model.
Chapter 6
SUMMARY
8/2/2019 Anusha Final Tech Doc
20/23
20
The importance of the threshold value on the energy level was something of which we did not have
an appreciation before collecting the results. To be more specific we understood that hresholding
had an effect but didn.t realise the extent to which thresholding could change the energy retained
and compression rates. Therefore when the investigation was carried out more attention was paid
to choosing the wavelets, images and decomposition levels than the thresholding strategy.
ADVANTAGES:
We can avoid creating of blocking artefacts from an transformed image.
Multi-Resolution is possible with Wavelet transformation.
Image compression.
DISADVANTAGES:
When compared with DCT(Discrete Cosine Transformation) algorithm, Performance is
low.
Blockiness can be occurred after this transformation.
Chapter 7
CONCLUSION
The general aim of the project was to investigate how wavelets could be used for image
compression. We believe that we have achieved this. We have gained an understanding of whatwavelets are, why they are required and how they can be used to compress images. We understand
the problems involved with choosing threshold values. While the idea of thresholding is simple and
effective, finding a good threshold is not an easy task. We have also gained a general
understanding of how decomposition levels, wavelets and images change the division of energy
8/2/2019 Anusha Final Tech Doc
21/23
21
between the approximation and detail subsignals. So we did achieve an investigation into
compression with wavelets.
Using global thresholding is not incorrect, it is a perfectly valid solution to threshold, the
problem is that using global thresholding masked the true effect of the decomposition levels in
particular on the results. This meant that the true potential of a wavelet to compress an image
whilst retaining energy was not shown. The investigation then moved to local thresholding which
was better than global thresholding because each detail subsignal had its own threshold based on
the coefficients that it contained. This meant it was easier to retain energy during compression.
However even better local thresholding techniques could be used. These techniques would be
based on the energy contained within each subsignal rather than the range of coefficient values and
use cumulative energy profiles to find the required threshold values. If the actual energy retained
value is not important, rather a near optimal trade off is required then a method called BayesShrink
could be used. This method performs a denoising of the image which thresholds the insignificant
details and hence produces Zeros while retaining the significant energy.
We were perhaps too keen to start collecting results in order to analyse them when we
should have spent more time considering the best way to go about the investigation. Having
analysed the results it is clear that the number of thresholds analysed (only 10 for each
combination of wavelet, image and level) was not adequate to conclude which is the best wavelet
and decomposition level to use for an image. There is likely to be an optimal value that the
investigation did not find. So it was difficult to make quantative predictions for the behaviour of
wavelets with images, only the general trends could be investigated. We feel however that the
reason behind my problems with thresholding was that thresholding is a complex problem in
general. Perhaps it would have been better to do more research into thresholding strategies and
images prior to collecting results. This would not have removed the problem of thresholding but
allowed us to make more informed choices and obtain more conclusive results. At the start of the
project our aim was to discover the effect of decomposition levels, wavelets and images on the
compression. We believe that we have discovered the effect each have.
How well a wavelet can compact the energy of a subsignal into the approximation
subsignal depends on the spread of energy in the image. An attempt was made to study image
properties but it is still unclear as to how to link image properties to a best wavelet basis function.
Therefore an investigation into the best basis to use for a given image could be another extension.
8/2/2019 Anusha Final Tech Doc
22/23
22
Only one family of wavelets was used in the investigation, the Daubechies wavelets. However
there are many other wavelets that could be used such as Meyer, Morlet and Coiflet. Wavelets can
also be used for more than just images, they can be used for other signals such as audio signals.
They can also be used for processing signals not just compressing them. Although compression
and denoising is done in similar ways, so compressing signals also performs a denoising of the
signal. Overall we feel that we have achieved quite a lot given the time constraints considering that
before we could start investigating wavelet compression we had to first learn about wavelets and
how to use VC++. We feel that we have learned a great deal about wavelets, compression and how
to analyse image.
REFERENCES
1. D. Donoho and I. Johnstone, Adapting to unknown smoothness via wavelet shrinkage,
JASA, v.19, pp. 1200-1224, 1995
2. A. Chambolle, R. DeVore, N. Lee and B. Lucier, Nonlinear Wavelet Image Processing, IEE
Transactions on Image Processing, v.7, pp319-335, 1998
3. K. Castleman, Digital Image Processing', Prentice Hall 1996
4. V. Cherkassky and F. Mulier, Learning from Data, Wiley, N.Y. 1998
5. V. Cherkassky and X. Shao, Model Selection for Wavelet-based Signal Estimation, Proc.
International Joint Conference on Neural Networks, Anchorage, Alaska, 1998
6. Amir Said and William A. Pearlman, A New Fast and Efficient Image Code Based on Set
Partitioning in Hierarchical Trees, IEEE Transactions on Circuits and Systems for Video
Technology, vol. 6, pp. 243-250, June 1996
8/2/2019 Anusha Final Tech Doc
23/23
23
7. Amir Said and William A. Pearlman, An Image Multi-resolution Representation for
Lossless and Lossy Image Compression, first submitted July 1994, and published in IEEE
Transactions on Image Processing, vol. 5, pp. 1303-1310, Sept. 1996