Você está na página 1de 7

1

LOW BIT-RATE IMAGE COMPRESSION VIA ADAPTIVE DOWN


SAMPLING AND CONSTRAINED LEAST SQUARES UPCONVERSION
R. Ashok Kumar Reddy, Project Associate, ashokreddy02@gmail.com

Abstract 1.1.1 Digitizer


In this paper, we are going to use a practical A digitizer converts an image into a
approach of uniform down sampling in image space and numerical representation suitable for input into a
yet making the sampling adaptive by spatially varying, digital computer. Some common digitizers are
directional low-pass pre-filtering. The resulting down-
sampled pre-filtered image remains a conventi onal square
conventional 1. Microdensitometer
sample grid, and, thus, it can be compressed and 2. Flying spot scanner
transmitted without any change to current image coding 3. Image dissector
standards and systems. The decoder first decompresses 4. Videocon camera
the low-resolution image and then up-converts it to the
5. Photosensitive solid- state arrays.
original resolution in a cons trained least squares
constrained
restoration process, using a 2-D piecewise autoregressive 1.1.2 Image Processor
model and the knowledge of directional low-pass pre-
An image processor does the functions of
filtering. The proposed compression approach of
collaborative adaptive down-sampling and up-conversion image acquisition, storage, preprocessing,
(CADU) outperforms JPEG 2000 in PSNR measure at low segmentation, representation, recognition and
to medium bit rates and achieves superior visual quality, interpretation and finally displays or records the
as well. The superior low bit-rate performance of the resulting image. The following block diagram
CADU approach seems to suggest that over-sampling not gives the fundamental sequence involved in an
only wastes hardware resources and energy, and it could image processing system.
be counterproductive to image quality given a tight bit
budget.

1. Introduction:
The term digital image refers to processing of a
two dimensional picture by a digital computer. In a
broader context, it implies digital processing of any two
dimensional data. A digital image is an array of real or
complex numbers represented by a finite number of bits.
An image given in the form of a transparency, slide,
photograph or an X-rays is first digitized and stored as a As detailed in the diagram, the first step in
matrix of binary digits in computer memory. This the process is image acquisition by an imaging
digitized image can then be processed and or displayed sensor in conjunction with a digitizer to digitize the
on a high-resolution television monitor. For display, the image. The next step is the preprocessing step
image is stored in a rapid-access buffer memory, which where the image is improved being fed as an input
refreshes the monitor at a rate of 25 frames per second to the other processes. Preprocessing typically
to produce a visually continuous display. deals with enhancing, removing noise, isolating
regions, etc. Segmentation partitions an image into
1.1 The Image Processing System: its constituent parts or objects. The output of
A typical digital image processing system is segmentation is usually raw pixel data, which
given in fig.1.1 consists of either the boundary of the region or the
pixels in the region themselves. Representation is
the process of transforming the raw pixel data into
a form useful for subsequent processing by the
computer. Description deals with extracting
features that are basic in differentiating one class of
objects from another. Recognition assigns a label
to an object based on the information provided by
its descriptors. Interpretation involves assigning
meaning to an ensemble of recognized objects. The
knowledge about a problem domain is incorporated
into the knowledge base. The knowledge base
guides the operation of each processing module
and also controls the interaction between the
modules. Not all modules need be necessarily
present for a specific function. The composition of
the image processing system depends on its
2

application. The frame rate of the image processor is 2.2.1 Image Enhancement
normally around 25 frames per second. Image enhancement operations improve the
qualities of an image like improving the image’s
1.1.3 Digital Computer:
contrast and brightness characteristics, reducing its
Mathematical processing of the digitized image noise content, or sharpen the details. This just
such as convolution, averaging, addition, subtraction, enhances the image and reveals the same
etc. are done by the computer. information in more understandable image. It does
not add any information to it.
1.1.4 Mass Storage
2.2.2 Image Restoration
The secondary storage devices normally used are Image restoration like enhancement improves
floppy disks, CD ROMs etc. the qualities of image but all the operations are
mainly based on known, measured, or degradations
1.1.5 Hard Copy Device of the original image. Image restorations are used
The hard copy device is used to produce a to restore images with problems such as geometric
permanent copy of the image and for the storage of the distortion, improper focus, repetitive noise, and
software involved. camera motion. It is used to correct images for
known degradations.
1.1.6 Operator Console 2.2.3 Image Analysis
The operator console consists of equipment Image analysis operations produce numerical
and arrangements for verification of intermediate or graphical information based on characteristics of
results and for alterations in the software as and when the original image. They break into objects and
require. The operator is also capable of checking for then classify them. They depend on the image
any resulting errors and for the entry of requisite data. statistics. Common operations are extraction and
description of scene and image features, automated
2. Image processing fundamentals: measurements, and object classification. Image
2.1 Introduction analyze are mainly used in machine vision
Digital image processing refers processing of applications.
the image in digital form. Modern cameras may directly 2.2.4 Image Compression
take the image in digital form but generally images are Image compression and decompression reduce
originated in optical form. They are captured by video the data content necessary to describe the image.
cameras and digitalized. The digitalization process Most of the images contain lot of redundant
includes sampling, quantization. Then these images are information, compression removes all the
processed by the five fundamental processes, at least redundancies. Because of the compression the size
any one of them, not necessarily all of them. is reduced, so efficiently stored or transported. The
2.2 Image Processing Techniques: compressed image is decompressed when
This section gives various image processing displayed. Lossless compression preserves the
techniques. exact data in the original image, but Lossy
compression does not represent the original image
but provide excellent compression.
2.2.5 Image Synthesis
Image synthesis operations create images from
other images or non-image data. Image synthesis
operations generally create images that are either
physically impossible or impractical to acquire.
2.3 Applications
Digital image processing has a broad spectrum
of applications, such as remote sensing via
satellites and other spacecrafts, image transmission
and storage for business applications, medical
processing, radar, sonar and acoustic image
Fig 2.2.1 Image processing Techniques processing, robotics and automated inspection of
industrial parts.
3

2.3.1 Medical Applications • Digital audio is a series of sample values;


In medical applications, one is concerned with image is a rectangular array of pixel values;
processing of chest X-rays, cineangiograms, projection video is a sequence of images played out at a
images of transaxial tomography and other medical certain rate
images that occur in radiology, nuclear magnetic • Neighboring sample values are correlated
resonance (NMR) and ultrasonic scanning. These • Transmission and storage of uncompressed
images may be used for patient screening and video would be extremely costly and
monitoring or for detection of tumors or other disease impractical.
in patients. • Frame with 352x288 contains 202,752 bytes of
information.
2.3.2 Satellite Imaging • Recoding of uncompressed version of this
Images acquired by satellites are useful in tracking video at 15 frames per second would require 3
of earth resources; geographical mapping; prediction of MB. One minute�180 MB storage. One 24-
agricultural crops, urban growth and weather; flood and hour day�262 GB
fire control; and many other environmental applications. • Using compression, 15 frames/second for 24
Space image applications include recognition and hour�1.4 GB, 187 days of video could be
analysis of objects contained in image obtained from stored using the same disk space that
deep space-probe missions. uncompressed video would use in one day
2.3.3 Communication Redundancy
Image transmission and storage applications occur • Adjacent audio samples are similar (predictive
in broadcast television, teleconferencing, and encoding); samples corresponding to silence
transmission of facsimile images for office automation, (silence removal)
communication of computer networks, closed-circuit • In digital image, neighboring samples on a
television based security monitoring systems and in scanning line are normally similar (spatial
military communications.. redundancy)
2.3.4 Radar Imaging Systems • In digital video, in addition to spatial
Radar and sonar images are used for detection and redundancy, neighboring images in a video
recognition of various types of targets or in guidance sequence may be similar (temporal
and manoeuvring of aircraft or missile systems. redundancy)

2.3.5 Document Processing Principles of Compression


It is used in scanning, and transmission for Spatial Correlation:
converting paper documents to a digital image form, • Redundancy among neighboring pixels
compressing the image, and storing it on magnetic tape. Spectral Correlation:
It is also used in document reading for automatically • Redundancy among different color planes
detecting and recognizing printed characteristics. Temporal Correlation:
• Redundancy between adjacent frames in a
2.3.6 Defense/Intelligence sequence of image
It is used in reconnaissance photo-interpretation for
automatic interpretation of earth satellite imagery to Human Perception Factors
look for sensitive targets or military threats and target • Compressed version of digital audio, image,
acquisition and guidance for recognizing and tracking video need not represent the original information
targets in real-time smart-bomb and missile-guidance exactly
systems. • Perception sensitivities are different for
different signal patterns.
Need for compression: • Human eye is less sensitive to the higher
• To reduce the volume of data to be transmitted spatial frequency components than the lower
(text, fax, images) frequencies (transform coding)
• To reduce the bandwidth required for transmission
and to reduce storage requirements (speech, audio, Classification
video) • Lossless compression
• How is compression possible? – lossless compression for legal and medical
• Redundancy in digital audio, image, and video data documents, computer programs
• Properties of human perception – Exploit only data redundancy.
4

• Lossy compression Disadvantages:


– digital audio, image, video where some errors or • The number of Fourier samples we need to
loss can be tolerated acquire must match the desired resolution of the
– Exploit both data redundancy and human image, i.e. the number of pixels in the image.
perception properties.
Proposed System
Predictive coding: We propose a new, standard-compliant
• Information already received (in transmission) is approach of coding uniformly down-sampled
used to predict future values images, which outperforms JPEG 2000 in both
• Difference between predicted and actual is stored. PSNR and visual quality at low to modest bit rates.
• Easily implemented in spatial (image) domain This success is due to the novel up-conversion
• Example: Differential Pulse Code Modulation process of least square non-causal predictive
(DPCM). decoding, constrained by adaptive directional low-
pass pre-filtering. Our findings suggest that a lower
Transform Coding:
sampling rate can actually produce higher quality
• Transform signal from spatial domain to other
images at certain bit rates. By feeding the standard
space using a well-known transform methods down-sampled images, the new approach
• Encode signal in new domain (by string reduces the workload and energy consumption of
coefficients) the encoders, which is important for wireless visual
• Higher compression, in general than predictive, but communication.
requires more computation (apply quantization).
Advantages
• The proposed compression approach of
Sub band Coding: collaborative adaptive down-sampling and up-
• Split the frequency band of a signal in various conversion (CADU) outperforms JPEG 2000 in
subbands. PSNR measure at low to medium bit rates and
• The filters used in subband coding are known as achieves superior visual quality, as well.
quadrature mirror filter(QMF)
• Use octave tree decomposition of an image data
Domain: Digital Image Processing
into various frequency subbands. Digital Image processing is the use of
• The output of each decimated subbands quantized computer algorithms to perform image processing
and encoded separately on Digital Images. As a subfield of digital signal
• Constant bit rate versus variable bit rate coding processing, digital image processing has many
advantages over analog image processing; it allows
Block diagram: a much wider range of algorithms to be applied to
the input data, and can avoid problems such as the
build-up of noise and signal distortion during
processing.
Software Requirement
MATLAB 7.0 and above MATLAB is a high-
performance language for technical computing. It
integrates computation, visualization, and
programming in an easy-to-use environment where
Fig. 1. Block diagram of the proposed CADU problems and solutions are expressed in familiar
image compression system. mathematical notation. Typical uses include:
Existing system: • Math and computation Algorithm development
Compressive Sampling: This paper surveys an • Modeling, simulation, and prototyping
emerging theory which goes by the name of • Data analysis, exploration, and visualization
“compressive sampling” or “compressed sensing,” and • Scientific and engineering graphics
which says that this conventional wisdom is inaccurate. • Application development, including Graphical
Perhaps surprisingly, it is possible to reconstruct images User Interface building.
or signals of scientific interest accurately and MATLAB is an interactive system whose
sometimes even exactly from a number of samples basic data element is an array that does not require
which is far smaller than the desired resolution of the dimensioning. This allows you to solve many
image/signal, e.g. the number of pixels in the image. technical computing problems, especially those
with matrix and vector formulations, in a fraction
5

of the time it would take to write a program in a scalar technique of predictive image coding For the
non-interactive language such as C or FORTRAN. CADU decoder; the PAR model plays a role of
adaptive noncausal predictor. The CADU approach
Introduction:
is very novel and unique that the predictor is only
The prevailing engineering practice of image/video
used at the decoder side, and the noncausal
compression usually starts with a dense 2-D sample
predictive decoding is performed in collaboration
grid of pixels. Compression is done by transforming the
with the prefiltering of the encoder.
spatial image signal into a space (e.g., spaces of Fourier
or wavelet bases) in which the image has a sparse Module 1:
representation and by entropy coding of transform The CADU image compression technique,
coefficients. Recently, researchers in the emerging field although operating on down-sampled images,
of compressive sensing introduced a new method called obtains some of the best PSNR results and visual
“over sampling followed massive dumping” approach. quality at low to medium bit rates. CADU
They showed, quite surprisingly, it is possible, at least outperforms the JPEG 2000 standard, even though
theoretically, to obtain compact signal representation by the latter is fed images of higher resolution and is
a greatly reduced number of random samples widely regarded as an excellent low bit-rate image
This project investigates the problem of compact codec. Since the down-sampled image has the
image representation in an approach of sparse sampling conventional form of square pixel grid and can be
in the spatial domain. The fact that most natural images fed directly to any existing image codec, standard
have an exponentially decaying power spectrum or proprietary, the CADU up conversion process is
suggests the possibility of interpolation-based compact entirely up to the decoder the proposed CADU
representation of images. A typical scene contains image coding approach can work in tandem with
predominantly smooth regions that can be satisfactorily any third party image/video compression
interpolated from a sparsely sampled low-resolution techniques. This flexibility makes standard
image. The difficulty is with the reconstruction of high compliance a non issue for the new CADU method.
frequency contents. Of particular importance is faithful We envision that CADU becomes a useful
reconstruction of edges without large phase errors, enhancer of any existing image compression
which is detrimental to perceptual quality of a decoded standard for improved low bit-rate performance.
image. For all these drawbacks, new image We make a more compact representation of an
compression methodology of collaborative adaptive image by decimating every other row and every
down-sampling and up conversion (CADU). other column of the image. This simple approach
Scope Of The Project: has an operational advantage that the down-
The main objective is to propose a new, sampled image remains a uniform rectilinear grid
standard-compliant approach of coding uniformly of pixels and can readily be compressed by any of
down-sampled images, which outperforms JPEG 2000 existing international image coding standards. To
in both PSNR and visual quality at low to modest bit prevent the down-sampling process from causing
rates by using the novel up conversion process of least aliasing artifacts, it seems necessary to low-pass
square noncausal predictive decoding, constrained by prefilter an input image to half of its maximum
adaptive directional low-pass prefiltering. It is to frequency. However, on a second reflection, one
estimate that a lower sampling rate can actually produce can do somewhat better. In areas of edges, the 2-D
higher quality images at certain bit rates. spectrum of the local image signal is not isotropic.
Thus, we seek to perform adaptive sampling,
Module: within the uniform down-sampling
Module 1: Decomposition of low –resolution image. framework, by judiciously smoothing the image
Module 2: Upconversion of the image to its resolution with directional low-pass prefiltering prior to
by PAR . down-sampling.
Module 3: Reverse the directional low-pass prefiltering To this end, we design a family of 2-D
operation of the encoder. directional low-pass prefilters under the criterion of
preserving the maximum 2-D bandwidth without
Module Description:
the risk of aliasing. Let WL(θ) and WH(θ) be the
The CADU decoder first decompresses the low-
side lengths of the rectangular low-passed region of
resolution image and then upconverts it to the original
the 2-D filter in the low- and high-frequency
resolution in constrained least squares restoration
directions of an edge of angle , respectively. The
process, using a 2-D piecewise autoregressive model
maximum area of this low-passed region without
(PAR) and by reversing the directional low-pass
aliasing is
prefiltering operation of the encoder. Two-dimensional
A= WL (θ) .WH (θ) = π2
autoregressive modeling was a known effective
6

as that between its reference and its neighbors in


It is easy to show that there are only eight values of the low resolution version. Therefore, the PAR
(corresponding to three combinations of and values) to model parameters can be estimated from the
achieve while avoiding aliasing. These eight cases are decoded image by solving the following two least-
tabulated in Table I. Fig. 2 illustrates the above square estimation problems.
directional low-pass filter design for , (the lowpassed Under the assumption of piecewise stationarity
frequency range for other angles can be obtained by of the image signal, the decoder carries out the
rotation). The spectra in the diagrams are those of the proposed up sampling process on a block of pixels
straight line of angle .In addition, the directional low- at a time. To prevent possible block visual artifacts,
pass filter design serves two other purposes: 1) most the neighboring blocks should have some overlaps.
efficient packing of signal energy in presence of edges; The choice of block size and the degree of spatial
2) preservation of subjective image quality for the edge overlap are design details related to
is an important semantic construct. Moreover, as we implementation complexity and viewer preference
will see in the next section, the use of low-pass on image attributes. The CADU system design is
prefilters establishes sample relations that play a central asymmetric: the encoder is a simple and
role in the decoding process of constrained least squares inexpensive process, while the decoder involves
upconversion. Many implementations of directional solving a rather large-scale optimization problem
low-pass prefilters are possible. In the directional described . The computation bottleneck is in
prefiltering step, the CADU encoder first computes the inverting an matrix in , where is the number of
gradient at the sampled position. pixels to be jointly recovered. Instead of inverting
If the amplitude of the gradient is below a the matrix directly, we solve numerically via
threshold, the isotropic low-pass filter is applied. differentiation using the conjugate gradient method.
Otherwise, the gradient direction is quantized and the The solution is guarantied to be globally optimal
corresponding filter is selected and applied. Despite its for the objective function is convex.
simplicity, the CADU compression approach via
uniform down-sampling is not inherently inferior to Results:
other image compression techniques in rate-distortion
performance, as long as the target bit rate is below a
threshold.
Module 2:
In the directional prefiltering step, the CADU
encoder first computes the gradient at the sampled
position. If the amplitude of the gradient is below a
threshold, the isotropic low-pass filter is applied.
Otherwise, the gradient direction is quantized and the
corresponding filter is selected and applied. Despite its
simplicity, the CADU compression approach via
uniform down-sampling is not inherently inferior to
other image compression techniques in rate-distortion
performance, as long as the target bit rate is below a
threshold. The argument is based on the classical water-
filling principle in rate-distortion theory. To encode a
set of independent Gaussian random variables, the rate-
distortion bounds, the total bit rate and the total mean-
squares distortion are calculated.
Most natural images have a rapidly (e.g.,
exponentially) decaying power spectrum . Suppose that
the input image is in the Fourier domain then its power
spectrum is monotonically decreasing.
Module 3:
Then the next step is to estimate the model CONCLUSION:
parameters. Recall that our design aims to reconstruct We proposed a new, standard-compliant
significant edges. Because edges have high correlation approach of coding uniformly down-sampled
in the scale space, the directional correlation of that images, which outperforms JPEG 2000 in both
pixel and its neighbors in the original image is the same PSNR and visual quality at low to modest bit
7

rates. This success is due to the novel up conversion


process of least square noncausal predictive decoding,
constrained by adaptive directional low-pass
prefiltering. Our findings suggest that a lower sampling
rate can actually produce higher quality images at
certain bit rates. By feeding the standard methods down
sampled images, the new approach reduces the
workload and energy consumption of the encoders,
which is important for wireless visual communication.

REFERENCES:
[1] E. Cands, “Compressive sampling,” in Proc. Int.
Congr. Mathematics, Madrid, Spain, 2006, pp.
1433–1452. WU et al.: LOW BIT-RATE IMAGE
COMPRESSION VIA ADAPTIVE DOWN-
SAMPLING 561
[2] X.Wu,K.U. Barthel, and W. Zhang, “Piecewise 2-D
autoregression for predictive image coding,” in Proc.
IEEE Int. Conf. Image Processing, Chicago, IL, Oct.
1998, vol. 3, pp. 901–904.
[3] X. Li and M. T. Orchard, “Edge-direted prediction
for lossless compression of natural images,” IEEE
Trans. Image Process., vol. 10, no. 6, pp. 813–817, Jun.
2001.
[4] D. Santa-Cruz, R. Grosbois, and T. Ebrahimi, “Jpeg
2000 performance evaluation and assessment,” Signal
Process.: Image Commun., vol. 1, no. 17, pp. 113–130,
2002.
[5] A. M. Bruckstein, M. Elad, and R. Kimmel, “Down-
scaling for better transform compression,” IEEE Trans.
Image Process., vol. 12, no. 9, pp. 1132–1144, Sep.
2003.
[6] Y. Tsaig, M. Elad, and P. Milanfar, “Variable
projection for near-optimal filtering in low bit-rate
block coders,” IEEE Trans. Circuits Syst. Video
Technol., vol. 15, no. 1, pp. 154–160, Jan. 2005.
[7] W. Lin and D. Li, “Adaptive downsampling to
improve image compression at low bit rates,” IEEE
Trans. Image Process., vol. 15, no. 9, pp. 2513–2521,
Sep. 2006.
[8] L. Gan, C. Tu, J. Liang, T. D. Tran, and K.-K. Ma,
“Undersampled boundary pre-/post-filters for low bit-
rate dct-based coders,” IEEE Trans. Image Process. ,
vol. 16, no. 2, pp. 428–441, Feb. 2007.
[9] X. Zhang, X. Wu, and F. Wu, “Image coding on
quincunx lattice with adaptive lifting and
interpolation,” in Proc. IEEE Data Compression Conf.,
Mar. 2007, pp. 193–202.
[10] B. Zeng and A. N.Venetsanopoulos, “A jpeg-based
interpolative image coding scheme,” in Proc. IEEE
ICASSP, 1993, vol. 5, pp. 393–396.
[11] D. Tabuman and M. Marcellin , JPEG2000: Image
Compression Fundamentals, Standards and Parctice.
Norwell, MA: Kluwer, 2002.

Você também pode gostar