Image Compression Using Discretewaveletstransforms.: G.Raja Sekhar
Image Compression Using Discretewaveletstransforms.: G.Raja Sekhar
Image Compression Using Discretewaveletstransforms.: G.Raja Sekhar
USING
DISCRETEWAVELETSTRANSFORMS.
G.RAJA SEKHAR
INTRODUCTION
The advent of multimedia computing has lead to an increased demand for
digital images. The storage and manipulation of these images in their raw
form is very expensive; for example, a standard 35mm photograph digitized
at 12µm per pixel requires about 18Mbytes of storage and one second NTSC
quality color video requires almost 23Mbytes of storage. To make
widespread use of digital imagery practical, some form of data compression
must be used.
Digital images can be compressed by eliminating redundant information.
There are three types of redundancy that can be exploited by image
compression systems:
Image enhancement:
Image enhancement refers to accentuation, or sharpening of image
features such as edges, boundaries, or contrast to make a graphic display
more useful for analysis. The enhancement process does not increase the
inherent information content in the data. But it does increase the dynamic
range of the chosen features so that they can be detected easily. Image
enhancement includes gray level and contrast manipulation, noise reduction,
edge crisping and sharpening, filtering, interpolation and magnification,
pseudo coloring, etc.
Image enhancement techniques can be broadly classified into four
groups, Point operations
1. Contrast stretching.
2. Clipping and thresholding.
3. Digital negative.
4. Intensity level slicing.
5. Range compression
6. Image subtraction and change detection.
Histogram modeling
Histogram equalization, modification & specification
Spatial operators
Image restoration:
Image analysis:
The amount of data associated with visual information is so large that its
storage requires enormous storage capacity. Although the storage capacities
of several media are substantial, their access speeds are usually inversely
proportional to their capacity. Typical television images are generating data
rates exceeding 10 million bytes per second. There are other images that
generate even higher data rates. Storage and/or band width, which could be
expensive. Image data compression techniques are concerned with reduction
in the number of bits required to store or transmit images without any
predictable loss of information. Image transmission applications are in
broadcast television; remote sensing via satellite, aircraft, radar, or sonar;
teleconferencing; computer communications; and facsimile transmission.
Image storage is required most commonly for educational and business
documents, medical images, etc. Because of their wide applications, data
compression is of great importance in digital image processing.
IMAGE TRANSFORMS
The image transforms usually refers to a class of unitary matrices used for
representing images. Just as one-dimensional signal can be represented by an
orthogonal series of basis functions, an image can also be expanded in terms
of a discrete set of basis functions, an image can also be expanded in terms
of a discrete set of basis arrays called basis images.
WN=exp {-j2∏/N}
The DFT is one of the most important transforms in digital signal and
image processing. It has several properties that make it deal for image
processing applications.
Properties of DFT
F^-1=F*
2. The DFT is sampled spectrum of the finite sequence
u(n) extended by zeros outside the interval [0, N-1]
TWO-DIMENSIONAL DFT
F1=F*
־
2. The two dimensional DFT is sampled spectrum of the finite sequence
u(n) extended by zeros outside the interval [0,N-1]
(or)
C=C*≥C1=C^T
־
2. The cosine transform is not real part of the unitary DFT. However, the
The use of digital images has increased at a rapid pace over the past
decade. Photographs, printed text, and other hard-copy media are now
routinely converted into digital form, and the direct acquisition of digital
images is becoming more common as sensors and associated electronics
improve. Many recent imaging modalities in medicine such as MRI, CT
generate images directly in digital form. Representing images in digital
form allows visual information to be easily manipulated in useful and
novel ways. Compression is required in order to
1. Reduce the memory required for storage,
2. Improve the data access rate from storage devices.
3. Reduce the band width and/or the time required for transfer across
communication channels.
Symbol Encoding
In these techniques each pixel is processed independently, ignoring the
inter pixel dependencies.
PCM
In PCM the incoming video signal is sampled, quantized, and coded by
a suitable code word (before feeding it to a digital modulator for
transmission). The quantizer output is generally coded by a fixed length
binary code word having B bits. Commonly, 8 bits are sufficient for
monochrome broadcast or video conferencing quality images; whereas
medical images or color video signals may require 10 to 12 bits per pixel.
Entropy Coding
If the quantized pixels are not uniformly distributed, then their entropy
will be less than B, and there exists a code that uses les than B bits per
pixel. In entropy will be less than B, and there exists a code that uses less
than B bits per pixel. In entropy coding the goal is to encode a block of
M pixels containing MB bits with probabilities pi, i=0, 1,…….., L-1,
L=2^MB, by -log2pi bits, so that the average bit rate is
∑pi (-log2pi) =H
This gives a variable-length code for each block, where highly probable
blocks (or symbols) are represented by small-length codes, and vice
versa.
Merge the two nodes with smallest probability to form a new node
whose probability is the sum of the two merged nodes.
Arbitrarily assign 1 and 0 to each pair of branches merging into a
node.
3. Read sequentially from the root node to the leaf node
where the symbol is located.
Run-Length Coding
Bit-Plane Encoding
A 256 gray level image can be considered as a set of eight 1-bit planes,
each of which can be run-length encoded. For 8-bit monochrome images,
compression ratios of 1.5 to 2 can be achieved. This method becomes
very sensitive to channel errors unless the significant bit planes are
carefully protected.
PREDICTIVE TECHINQUES
Basic Principle
Arithmetic coding
QUANTIZATION
Scalar Quantization
di=xi-1+ri /2
ri =∫xp(x)dx/∫p(x)dx
The Lloyd-Max quantizer has non uniform decision regions. Since its
objective is to minimize the average distortion, it tends to allocate more
levels to those regions where the signal pdf is large.
2. Entropy-Constrained Quantizers
Vector Quantization
PACKING
And N and M are the width and height, respectively, of the images in
pixels, f is the original image and f is the reconstructed image. Note that
the original and reconstructed images must be same size.
WAVELET TRNSFORM TECHINQUE
Wavelet Transform Technique
The purpose of this section is to provide an intuitive understanding
of what wavelets are and why they are useful for signal compression.
One of the most commonly used approaches for analyzing a signal f(x) is
to represent it as a weighted sum of simple building blocks, called basis
functions:
f(x) = ∑ ciΨi(x)
i
Where Ψi(x) are basis functions and the ci are coefficients, or weights.
Since the basis functions Ψi are fixed, it is the coefficients which contain
the information about the signal. The simplest such representation uses
translates of the impulse functions as its only bases, yielding a
representation that reveals information only about the signal’s frequency
domain behavior.
Δω x Δx ≥ 0.5
This inequality means that we must trade off time resolution for
frequency resolution, or vice versa. Thus, it is possible to get very good
resolution in time if you are willing to settle for low resolution in
frequency, and you can get very good resolution in frequency if you are
willing to settle for low resolution in time.
The situation is really not all that bad from a compression standpoint.
By their very nature, low frequency events are spread out (or non-local)
in time and high frequency events are concentrated (or localized) in time.
Thus, one way that we can live within the confines of the Heisenberg
inequality and yet still get useful time-frequency information about a
signal is if we design our basis functions to act like cascaded octave band
pass filters, which repeatedly split the signal’s bandwidth in half.
To gain insight into designing a set of basis function that will satisfy
both our desire for information and the Heisenberg inequality, let us
compare the impulse function and the sinusoids. The impulse function
cannot provide information about the frequency behavior of the signal
because its support- the interval over which it is non-zero is
infinitesimally small. At the opposite extreme are the sinusoids, which
cannot provide information about comprise between these two extremes:
a set of basis functions {ψi}, each with finite support of a different width.
The different support widths allow us to trade off time and frequency
resolution in different ways; for example, a wide basis function can
examine a large region of the signal and resolve low frequency details
accurately, while a short basis function can examine a small region of the
signal to resolve time details accurately.
Note that this really means that we are translating ψ in steps of size
2^vk. Putting this altogether gives us a wavelet decomposition of the
signal,
Where
Gn = (-1)^1-n H1-n
Gn=G-n
Hn=H-n
H0 = 0.33267055290
H1 = 0.806891509311
H2 =0.459877502118
H3=-0.135011020010
H4 =-0.085441273882
H5=0.035226291882
From which the coefficients for G,H and G can be derived using
equations 1,2,and 3. The impulse response of H and G are shown in fig 3.
h5 h4 h3 h2 h1 h0
S0 S1 S2 S3 S4 S5 S6 ………….
g5 g4 g3 g1 g0
Note that G filter extends before the signal in time; if S is finite, the H
filter will extend beyond the end of the signal. A similar situation is
encountered with the inverse wavelet transform filters H and G. In an
implementation; one must make some choice about what values to pad
the extension with. A choice which works well in practice, is to wrap the
signal about its endpoints, i.e.,
fll
Inverse Transform tree
After the forward wavelet transform is completed, we are left