0% found this document useful (0 votes)
76 views6 pages

Lossless Image Compression Algorithm For Transmitting Over Low Bandwidth Line

This paper presents a new near-lossless image compression algorithm. It can provide low average compression rate (2. Bits / pixel) with high image quality (larger than 53. DB) for endoscopic images. The proposed technique produces a bit stream that results in a progressive and ultimately lossless reconstruction of an image similar to what one can obtain with a reversible wavelet codec.

Uploaded by

editor_ijarcsse
Copyright
© Attribution Non-Commercial (BY-NC)
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
76 views6 pages

Lossless Image Compression Algorithm For Transmitting Over Low Bandwidth Line

This paper presents a new near-lossless image compression algorithm. It can provide low average compression rate (2. Bits / pixel) with high image quality (larger than 53. DB) for endoscopic images. The proposed technique produces a bit stream that results in a progressive and ultimately lossless reconstruction of an image similar to what one can obtain with a reversible wavelet codec.

Uploaded by

editor_ijarcsse
Copyright
© Attribution Non-Commercial (BY-NC)
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 6

Volume 2, Issue 2, February 2012 ISSN: 2277 128X

International Journal of Advanced Research in


Computer Science and Software Engineering
Research Paper
Available online at: www.ijarcsse.com
Lossless Image Compression Algorithm For Transmitting
Over Low Bandwidth Line
Dr.E.KANNAN G. Murugan
Registrar & Dean(Academics) Research Scholar,CMJ University
Vel Tech Dr.RR & Dr.SR Technical University, Meghalaya
Chennai,TN-INDIA INDIA

Abstract-In this paper to decrease the communication bandwidth and save the transmitting power in the wireless
endoscopy capsule, this paper presents a new near-lossless image compression algorithm based on the Bayer
format image suitable for hardware design. This algorithm can provide low average compression rate (2.12
bits/pixel) with high image quality (larger than 53.11 dB) for endoscopic images. Especially, it has low
complexity hardware overhead (only two line buffers) and supports real-time compressing. In addition, the
algorithm can provide lossless compression for the region of interest (ROI) and high-quality compression for
other regions. The ROI can be selected arbitrarily by varying ROI parameters. The proposed technique produces
a bit stream that results in a progressive and ultimately lossless reconstruction of an image similar to what one
can obtain with a reversible wavelet codec. In addition, the proposed scheme provides near-lossless
reconstruction with respect to a given bound after decoding of each layer of the successively refundable bit
stream. We formulate the image data compression problem as one of successively refining the probability density
function (pdf) estimate of each pixel. Experimental results for both lossless and near-lossless cases indicate that
the proposed compression scheme, that innovatively combines lossless, near-lossless and progressive coding
attributes, gives competitive performance in comparison to state-of-the-art compression schemes.
Key word: Image compression, Image quality, Image compressor configuration Huffman coding


I. INTRODUCTION
Lossless or reversible compression refers to compression
techniques in which the reconstructed data exactly
matches the original. Near-lossless compression denotes
compression methods, which give quantitative bounds
on the nature of the loss that is introduced. Such
compression techniques provide the guarantee that no
pixel difference between the original and the compressed
image is above a given value. Both lossless and near-
lossless compression find potential applications in
remote sensing, medical and space imaging, and
multispectral image archiving. In these applications the
volume of the data would call for lossy compression for
practical storage or transmission. However, the necessity
to preserve the validity and precision of data for
subsequent recognized diagnosis operations, forensic
analysis, as well as scientific or clinical measurements,
often imposes strict constraints on the reconstruction
error. In such situations near-lossless compression
becomes a viable solution, as, on the one hand, it
provides significantly higher compression gains lossless
algorithms, and on the other hand it provides guaranteed
bounds on the nature of loss introduced by compression.
Another way to deal with the lossy-lossless dilemma
faced in applications such as medical imaging and
remote sensing is to use a successively refundable
compression technique that provides a bit stream that
leads to a progressive reconstruction of the image. Using
wavelets, for example, one can obtain an embedded bit
stream from which various levels of rate and distortion
can be obtained. In fact with reversible integer wavelets,
one gets a progressive reconstruction capability all the
way to lossless recovery of the original. Such techniques
have been explored for potential use in tele-radiology
where a physician typically requests portions of an
image at increased quality (including lossless
reconstruction) while accepting initial renderings and
unimportant portions at lower quality, and thus reducing
the overall bandwidth requirements. In fact, the new still
image compression standard, JPEG 2000, provides such
features.
1.1 METHODS FOR LOSSY COMPRESSION
Reducing the color space the most common colors in the
image. The selected colors are specified in the color
Volume 2, issue 2, February 2012 www.ijarcsse.com
2012, IJARCSSE All Rights Reserved
palette in the header of the compressed image. Each
pixel just references the index of a color in the color
palette. This method can be combined with dithering to
avoid pasteurizations. Chrome sub sampling. This takes
advantage of the fact that the human eye perceives
spatial changes of brightness more sharply than those of
color, by averaging or dropping some of the
chrominance information in the image. Transform
coding. This is the most commonly used method.
A Fourier-related transform such as DCT or the wavelet
transform are applied, followed
by quantization and entropy coding.
Fig 1. LOSSY IMAGE COMPRESSIONRESULT


Original Image 257Kb
Compressed Image 256Kb
1.2. Image Compression
Digital Image compression addresses the problem of
reducing the amount of data required to represent a
digital image. The underlying basis of the reduction
process is removal of redundant data. From the
mathematical viewpoint, this amounts to transforming a
2D pixel array into a statically uncorrelated data set. The
data redundancy is not an abstract concept but a
mathematically quantifiable entity. If n1 and n2 denote
the number of information-carrying units in two data sets
that represent the same information, the relative data
redundancy
D
R of the first data set (the one
characterized by n1) can be defined as,
R
D
C
R
1
1 =
Where
R
C called as compression ratio. It is defined as
R
C =
2
1
n
n

In image compression, three basic data
redundancies can be identified and exploited:
redundancy. Image compression is achieved when one or
more of these redundancies are reduced or eliminated.
The image compression is mainly used for image
transmission and storage. Image transmission
applications are in broadcast television; remote sensing
via satellite, air-craft, radar, or sonar; teleconferencing;
computer communications; and facsimile transmission.
Image storage is required most commonly for
educational and business documents, medical images
that arise in computer tomography (CT), magnetic
resonance imaging (MRI) and digital radiology, motion
pictures, satellite images, weather maps, geological
surveys, and so on.
II. PROPOSED SYSTEM
2.1. SIGNIFICANCE OF THIS WORK
In this project, Image compression based on adaptive
wavelet decomposition is presented. Adaptive wavelet
decomposition is very useful in various applications,
such as image analysis, compression, and feature
extraction and denoising. For such task, it is important
that multiresolution representations take into account the
characteristics of the underlying signal and do leave
intact important signal characteristics, such as sharp
transitions, edges, singularities, and other region of
interests. The adaptive lifting technique includes an
adaptive update lifting and fixed prediction lifting step.
The adaptively hereof consists that, the system can
choose different update filters in two ways; i) the choice
is triggered by combining the different norms, ii) Based
on the arbitrary Threshold.
This image compression based on adaptive wavelet
decomposition is implemented using MATLAB
programs, and the results compared with Non-adaptive
(Harr) decomposition.
III. . HUFFMAN CODING
Huffman coding is based on the frequency of occurrence
of a data item (pixel in images). The principle is to use a
lower number of bits to encode the data that occurs more
frequently. Codes are stored in a Code Book which may
be constructed for each image or a set of images. In all
cases the code book plus encoded data must be
transmitted to enable decoding.
Decoding for the above two algorithms is trivial as long
as the coding table (the statistics) is sent before the data.
(There is a bit overhead for sending this, negligible if the
data file is big.) Unique Prefix Property. No code is a
prefix to any other code (all symbols are at the leaf
nodes) great for decoder, unambiguous. If prior statistics
are available and accurate, then Huffman coding is very
good.
Volume 2, issue 2, February 2012 www.ijarcsse.com
2012, IJARCSSE All Rights Reserved

IV. WAVELET APPROACH
Storage constrains and bandwidth limitations in
communication systems have necessitated the search for
efficient image compression techniques. For real time
video and multimedia applications where a reasonable
approximation to the original signal can be tolerated,
lossy compression is used. In the recent past, wavelet
based image compression schemes have gained wide
popularity. The characteristics of the wavelet transform
provide compression results that outperform other
transform techniques such as discrete cosine transform
(DCT). Consequently, the JPEG2000 compression
standard and FBI fingerprint compression system have
adopted a wavelet approach to image compression.
The wavelet coding techniques is based on the idea that
the co-efficient of a transform that decor relates the
pixels of an image can be coded more efficiently than the
original pixels themselves. If the transforms basis
functions in this case wavelet- packs most of the
important visual information into small number of co-
efficient, the remaining co-efficient can be coarsely
quantized or truncated to zero with little image
distortion.
The still image compression, modern DWT based coders
have outperformed DCT based coders providing higher
compression ratio and more peak signal to noise ratio
(PSNR) due to the wavelet transforms multi-resolution
and energy compaction properties and the ability to
handle signals.
The basis set of wavelets is generated from the mother or
basic wavelet is defined as:
+
a,b
(t) =
|
.
|

\
|
a
b t
a

1
; a, b e 9 and a>0 --- (1)
5.1. 1-D Continuous wavelet transforms
The 1-D continuous wavelet transform is given by:
W
f
(a, b) =
}


dt t t x
b a
) ( ) (
,
------------ (2)
The inverse 1-D wavelet transform is given by:
x (t) =
} }

0
2
,
) ( ) , (
1
a
da
db t b a W
C
b a f
----- (3)
Where C = e
e
e
}


d
2
)
<
( ) e + is the Fourier transform of the mother wavelet
+(t). C is required to be finite, which leads to one of the
required properties of a mother wavelet. Since C must be
finite, then 0 ) 0 ( = + to avoid a singularity in the
integral, and thus the ) (t + must have zero mean. This
condition can be stated as
}


dt t) (
= 0 and known
as the admissibility condition.
5.2. 1-D Discrete wavelet transforms
The discrete wavelets transform (DWT), which
transforms a discrete time signal to a discrete wavelet
representation. The first step is to discredit the wavelet
parameters, which reduce the previously continuous
basis set of wavelets to a discrete and orthogonal /
orthonormal set of basis wavelets.

m,n
(t) = 2
m/2
(2
m
t n) ; m, n e Z such that - < m, n
< -------- (4)
The 1-D DWT is given as the inner product of the signal
x(t) being transformed with each of the discrete basis
functions.
W
m,n
= < x(t),
m,n
(t) > ; m, n e Z
------------ (5)
The 1-D inverse DWT is given as:
x (t) =

m n
n m n m
t W ) (
, ,
; m, n e Z
------------- (6)
5.3. 2-D wavelet transform
The 1-D DWT can be extended to 2-D transform using
separable wavelet filters. With separable filters, applying
a 1-D transform to all the rows of the input and then
repeating on all of the columns can compute the 2-D
transform. When one-level 2-D DWT is applied to an
image, four transform coefficient sets are created. As
depicted in Figure 1(c), the four sets are LL, HL, LH,
and HH, where the first letter corresponds to applying
either a low pass or high pass filter to the rows, and the
second letter refers to the filter applied to the columns.



Volume 2, issue 2, February 2012 www.ijarcsse.com
2012, IJARCSSE All Rights Reserved
Figure 2. Block Diagram of DWT (a)Original Image (b)
Output image after the 1-D applied on Row input (c)
Output image after the second 1-D applied on row input

Figure 3. DWT for Lena image (a)Original Image (b)
Output image after the 1-D applied on column input (c)
Output image after the second 1-D applied on row input

The Two-Dimensional DWT (2D-DWT) converts
images from spatial domain to frequency domain. At
each level of the wavelet decomposition, each column of
an image is first transformed using a 1D vertical analysis
filter-bank. The same filter-bank is then applied
horizontally to each row of the filtered and sub sampled
data. One-level of wavelet decomposition produces four
filtered and sub sampled images, referred to as sub
bands. The upper and lower areas of Fig. 3(b),
respectively, represent the low pass and high pass
coefficients after vertical 1D-DWT and sub sampling.
The result of the horizontal 1D-DWT and sub sampling
to form a 2D-DWT output image is shown in Fig.2(c).
V. ADAPTIVENESS BASED ON
COMBINING NORMS
The PSNR value for different bit rates and
different decomposition levels of the sharp edge
preserved images and image without sharp edge as
shown in table III.
TABLE I. PEAK SIGNAL TO NOISE RATIO
IMAGE: 2. Circles. raw
(256 X 256)
De
co
m
Le
vel
/
Bit
rat
e
1 2 3 4 5 6 7 8
0.1
5.9
648
9.4
727
13.
379
9
16.
014
9
15.
228
6
13.
002
7
12.
801
1
1
2.
2
1
0
3
0.2
6.8
893
14.
096
5
17.
030
2
19.
202
0
19.
324
3
17.
574
0
17.
682
3
1
6.
0
2
7
7
0.3
8.2
989
14.
134
8
17.
805
8
21.
380
7
20.
047
3
18.
380
4
18.
445
3
1
6.
5
8
7
2
0.4
9.6
339
14.
217
6
21.
970
0
28.
518
2
26.
758
0
25.
020
2
24.
011
6
2
2.
5
0
1
5
0.5
11.
821
5
14.
254
5
22.
281
6
31.
913
4
31.
776
4
30.
666
4
29.
902
1
2
7.
5
2
8
9
0.6
14.
658
6
17.
837
8
30.
271
9
33.
818
0
34.
575
4
33.
757
8
33.
074
0
3
0.
4
9
1
6
0.7
14.
664
9
18.
093
6
31.
263
6
40.
867
9
38.
458
6
37.
091
3
36.
144
1
3
3.
6
6
4
0
0.8
14.
737
8
18.
289
6
35.
151
4
42.
587
0
42.
354
0
42.
442
7
41.
684
6
4
1.
4
8
7
4
0.9
14.
760
1
18.
390
2
35.
363
2
43.
422
2
48.
481
3
47.
838
3
47.
176
6
4
6.
7
1
6
1
1.0
14.
775
9
22.
104
6
43.
510
8
49.
717
4
50.
118
1
49.
895
1
50.
257
4
4
8.
8
5
7
1
Volume 2, issue 2, February 2012 www.ijarcsse.com
2012, IJARCSSE All Rights Reserved
IMAGE: 3. Crosses. raw (256 X 256)
De
co
m
Le
vel
/
Bit
rat
e
1 2 3 4 5 6 7 8
0.1
15.
447
1
15.
908
4
15.
183
8
15.
432
9
15.
500
5
15.
489
0
15.
482
9
15.
482
9
0.2
16.
100
7
15.
994
7
16.
036
7
16.
852
7
16.
326
1
16.
326
4
16.
344
9
16.
329
9
0.3
16.
818
5
16.
368
7
17.
873
1
17.
479
2
16.
792
4
16.
853
6
16.
848
7
16.
846
7
0.4
17.
539
0
16.
662
8
18.
506
1
19.
966
0
19.
433
5
18.
733
6
18.
126
0
17.
608
9
0.5
18.
040
4
17.
151
5
20.
088
0
20.
628
1
20.
104
8
19.
349
5
18.
639
6
18.
182
8
0.6
18.
102
9
18.
123
8
21.
063
4
20.
834
2
20.
310
2
19.
532
6
18.
788
7
18.
292
1
0.7
18.
316
7
18.
809
6
21.
220
5
29.
466
4
27.
939
2
27.
763
6
25.
062
0
26.
431
6
0.8
18.
828
9
19.
358
8
25.
818
4
30.
039
0
28.
179
3
27.
991
8
25.
196
2
26.
565
2
0.9
19.
356
6
19.
727
8
30.
743
1
33.
135
9
32.
596
1
32.
208
5
31.
075
8
29.
537
2
1.0
19.
521
0
20.
419
4
30.
766
0
33.
684
2
32.
981
8
32.
461
9
31.
344
9
29.
674
3

7.1. IMAGE COMPRESSION BASED ON
ADAPTIVE LIFTING
PSNR Vs DECOMPOSTION LEVEL (CIRCLES)
0
5
10
15
20
25
30
35
40
45
0 2 4 6 8 10
DECOMPOSTION LEVEL
P
S
N
R
CR=10
CR=20
CR=40
CR=80

PSNR Vs DECOMPOSTION LEVEL (CROSSES)
0
5
10
15
20
25
30
35
0 2 4 6 8 10
DECOMPOSTION LEVEL
P
S
N
R
CR=10
CR=20
CR=40
CR=80

PSNR Vs DECOMPOSTION LEVEL (HORIZ)
0
5
10
15
20
25
30
35
40
0 2 4 6 8 10
DECOMPOSTION LEVEL
P
S
N
R
CR=10
CR=20
CR=40
CR=80

PSNR Vs DECOMPOSTION LEVEL (CAMERAMAN)
0
5
10
15
20
25
30
0 2 4 6 8 10
DECOMPOSTION LEVEL
P
S
N
R
CR=10
CR=20
CR=40
CR=80

.
VI. EXPERIMENTAL
SIMULATION RESULTS
7.1Opening Function:

7.2. Validation




Volume 2, issue 2, February 2012 www.ijarcsse.com
2012, IJARCSSE All Rights Reserved
VII. . CONCLUSION
This work has shown that the compression of image can
be improved by considering spectral and temporal
correlations as well as spatial redundancy. The efficiency
of temporal prediction was found to be highly dependent
on individual image sequences. Given the results from
earlier work that found temporal prediction to be more
useful for image, we can conclude that the relatively
poor performance of temporal prediction, for some
sequences, is due to spectral prediction being more
efficient than temporal. Another Conclusions and Future
Work finding from this work is that the extra
compression available from image can be achieved
without necessitating a large increase in decoder
complexity. Indeed the presented scheme has a decoder
that is less complex than many lossless image
compression decoders, due mainly to the use of forward
rather than backward adaptation. The results of adaptive
and Non-adaptive based image compression are
compared. From the results the adaptive wavelet
decomposition works better than non-adaptive (Haar)
wavelet decomposition. Future work aims at extending
this frame work for color images, video compressions,
and Denoising applications.

VIII. . REFERENCES
[1] N. Memon and K. Sayood. Lossless image
compression: A comparative study. Proc.SPIE Still-
Image Compression, 2418:820, March 1995.
[2] N. Memon and K. Sayood. Lossless compression of
rgb color images. Optical Engineering, 34(6):1711
1717, June 1995.
[3] S. Assche, W. Philips, and I. Lemahieu. Lossless
compression of pre-press images using a novel color
decorrelation technique. Proc. SPIE Very High
Resolution and Quality Imaging III, 3308:8592,
January 1998.
[4] N. Memon, X. Wu, V. Sippy, and G. Miller. An
interband coding extension of the new lossless jpeg
standard. Proc. SPIE Visual Communications and Image
Processing, 3024:4758, January 1997.
[5] N. Memon and K. Sayood. Lossless compression of
video sequences. IEEE Trans. On Communications,
44(10):13401345, October 1996.
[6] S. Martucci. Reversible compression of hdtv images
using median adaptive prediction and arithmetic coding.
Proc. IEEE International Symposium on Circuits and
Systems, pages 13101313, 1990.
[7] M. Weinberger, G. Seroussi, and G. Sapiro. LOCO-
I: A low complexity, context-based, lossless image
compression algorithm. Proc. Data Compression
Conference, pages 140149, March 1996.
[8] ITU-T Rec. T.84 - Information Technology. Digital
compression and coding of continuous-tone still images:
Extensions, July 1996.
[9] M. Weinberger, J. Rissanen, and R. Arps.
Applications of universal context modeling to lossless
compression of gray-scale images. IEEE Trans. on
Image Processing, 5(4):575 586, April 1996.
[10] G. Wallace. The jpeg still picture compression
standard. Comms. of the ACM, 34(4):30 44, April
1991.
[11] ISO/IEC 14495-1, ITU Recommendation T.87,
Information technology - Lossless and near-lossless
compression of continuous-tone still images, 1999.
[12] M. J. Weinberger, G. Seroussi, and G. Sapiro,
LOCO-I: A low complexity lossless image compression
algorithm. ISO/IEC JTC1/SC29/WG1 document N203,
July 1995.
[13] J. Rissanen and G. G. Langdon, Jr., Universal
modeling and coding, IEEE Trans. Inform. Theory, vol.
IT-27, pp. 1223, Jan. 1981.
[14] J. Rissanen, Generalized Kraft inequality and
arithmetic coding, IBM Jl. Res. Develop., vol. 20 (3),
pp. 198203, May 1976.
[15] M. J. Weinberger, J. Rissanen, and R. B. Arps,
Applications of universal context modeling to lossless
compression of gray-scale images, IEEE Trans. Image
Processing, vol. 5, pp.575586, Apr. 1996.



G.Murugan was born on 15
th

May 1974 India. He is working
as Lecturer in Al Musanna
College of Technology,Muscat
and pursuing Ph.D. in the
Information Technology, CMJ
University, Meghalaya, India..
He has also published papers
on image processing, Computer
networks in National and
International Conferences. His current research interests
are image processing, computer networks

Dr.Kannan finished his Phd in
National Institute of
Technology,(N.I.T.T) ,Trichy
2006, M.Tech in International
University of Comtemprory
Sciences, California,2008 M.E in Computer Science in
Sathiyabama University, Chennai 2011 and M.Sc in
Bharathidasan University,Trichy 1989. He has published five
papers in international journals in High Voltage.

You might also like