Você está na página 1de 11

IAETSD JOURNAL FOR ADVANCED RESEARCH IN APPLIED SCIENCES ISSN NO: 2394-8442

A STUDY OF COMPRESSION AND CLASSIFICATION TECHNIQUES


OF IMAGE RETRIEVAL SYSTEM USING NEURAL NETWORKS
1
Shaik.Jakeer Hussain, 2Dr. R. Kiran Kumar
1 Research Scholar, Department of Computer Science and Engineering,, Krishna University, Machilipatnam, A.P, India.
2 Asst. Professor, Department of Computer Science and Engineering,, Krishna University, Machilipatnam, A.P, India.
jakeerhussinsk@gmail.com

ABSTRACT: In image processing has main two objectives, more suitable image for people to observe and identify and
computers can recognise the images. X-Rays, MRI, ECG and CT Scan are produce digital images for diagnosis and treatment
to the patients. CBIR technique used to extract images which are visually like a specified target image. Today, digital images
storage in databases has been growth for feature references in medical sectors. Images can have occupied more space and
high bandwidth. Image compression techniques applied on the medical databases, to occupy less space to store and low
bandwidth utilization for transfer the images one place to another. Image compression and decompression are two
techniques, to reduce the image size and retrieve the original image.

Compression are two types -Lossless compression and Lossy Compression. In Lossless compression, all the information is
completely restored from compressed file and lossy compression lossy compression reduces a file by permanently and all
the information is not restored completely from compressed file. This paper is to study the image compression and
classification techniques and comparison between them. Compression techniques like Haar wavelet transform, Sobel edge
detection and Gabor filter. Classification techniques are accuracy evaluated using Neural Networks like Navie Bayesian
classifier, Support Vector Machines (SVM) and instance based learning (IBL).

Index Terms—Compressiontechniques, Classification Techniques, Haar wavelet transform, Sobel edge detection and
Gabor Filter, Neural Networks like Navie Bayesian classifier, Support Vector Machines (SVM) and instance based learning
(IBL).

INTRODUCTION
Digital images like X-Rays, MRI, ECG and CT Scan are used for diagnosis and treatment due to the medical image
technology. Content based image retrieval (CBIR) is also known as Query by Image Content (QBIC) or content-based visual
information retrieval (CBVIR) is the application of image processing to searching for digital images in large databases. To
store and transmit the data for future references, we need to reduce the sizes of images by Compression technique. It
compresses the images without degrading the image [1]. The main objective of CBIR is to provide right information to the
right time to improve quality and efficiency of care process. It plays key role in decision making system in medical sector.

Hence compressed images take less space and transmission of the images are required low bandwidth. Image compression
is two types like - lossy compression and lossless compressions. In lossless compression, restored the original images after
decompression. Medical Images and technical drawings used lossless compressions. In lossy data compression original data
is not exactly restored. Natural images used lossy compression [2].

Images are retrieved by features like - colour, texture and shape. Feature selection searches and extracts a subset of relevant
features. Effectiveness of classification can be optimised. By applying feature selection, a subset of relevant features is
retained, to simplify classification process to make it more effective.

VOLUME 4, ISSUE 7, DEC/2017 190 http://iaetsdjaras.org/


IAETSD JOURNAL FOR ADVANCED RESEARCH IN APPLIED SCIENCES ISSN NO: 2394-8442

METHODOLOGY
THE HAAR WAVELET TRANSFORM
The Haar Wavelet Transform is memory efficient and exactly reversible without the edge effect. It is very simple and fast
[3]. The Haar Transform (HT) is one of the simplest and basic transformations from the space domain to a local frequency
domain. A HT decomposes each signal into two components, one is called average (approximation) or trend and the other
is known as difference (detail) or fluctuation.

A precise formula for the values of first average subsignal = , , , …, at one level for a signal of length N

i.e. = , , ,… is
+
= , = 1, 2, 3 …
√2 2

and
= , , , …, at the same level for a signal of length N i.e. = , , ,… is

= , = 1, 2, 3 …
√2 2

To apply the Haar wavelet Transform to the below of matrix L. After applying to matrix L, matrix L has divided into four
parts.

Approximation Area is denoted by A, includes information about the global properties of analysed image,removal of spectral
coefficients from this area to the biggest direction in original image.

Horizontal Area is denoted by H, includes information about vertical lines hidden in image, removal of spectral image,
removal of spectral coefficients from this area excludes horizontal details from original image.

Vertical Area is denoted by V contains information about the horizontal lines hidden in image removal of spectral
coefficients from this area eliminates vertical details from original image.

Diagonal Area is denoted by D contains informationabout the diagonal details hidden in image. Removal of spectral
coefficients from this area leads to minimumdistortions in original image.

1 5 36
L= 2 6 48
3 7 83
4 86 2

By applying 1D HT on first row, then approximation coefficients of matrix L are

(1 + 5) and (3 + 6)
√ √

and detail coefficients are

(1 − 5) and (3 − 6)
√ √

VOLUME 4, ISSUE 7, DEC/2017 191 http://iaetsdjaras.org/


IAETSD JOURNAL FOR ADVANCED RESEARCH IN APPLIED SCIENCES ISSN NO: 2394-8442

The same transform is applied to the other rows of matrix L.

1 5 63 6 9 ⋮ −4 3
2 3 48 1
5 12 ⋮ −1 −4
⎯⎯⎯⎯⎯⎯⎯⎯⎯
3 6 83 √2 9 11 ⋮ −3 5
4 86 2 12 8 ⋮ −4 4

By applying the 1D HT on columns of the matrix, the resultant matrix at first level.

15 −1 ⋮ −4 3
6 9 ⋮ −4 3 ⎡17 −5 ⋮ −7 3 ⎤
5 12 ⋮ −1 −4 1 ⎢ ⎥
⎯⎯⎯⎯⎯⎯⎯⎯⎯ ⎢… … … … …⎥
9 11 ⋮ −3 5 √2 ⎢20 2 ⋮ −2 −8⎥
12 8 ⋮ −4 4 ⎣20 8 ⋮ 4 0⎦

From the above resultant matrix, the values of A, H, V and D are

15 −1 −4 3
= =
17 −5 −1 −4

20 2 −2 −8
= =
20 8 4 0

In, Fast Haar transform (FHT) involves addition, subtraction and division by 2, then it becomes faster and reduces the
calculation work in comparison to HT. First apply 1D FHT on each row of an input image matrix. These transformed rows
are themselves an image and we apply the 1D FHT to each column. The resulting values are all detail coefficients except for
a single overall average coefficient.

In Modified Fast Haar wavelet transform (MFHWT), the values of first average subsignal = , , , …, ⁄ at
one level for a signal of length N i.e. = , , ,… is

+ + +
= , = 1, 2, 3 …
4 4

and first detail subsignal = , , , …, ⁄ at the same level for a signal of length N i.e. = , , ,… is
( + )−( + )
= 4 , = 1, 2, 3 …
0 , = 2, … 4

Four nodes are considered in Fast Haar Wavelet Transform. The author has considered the values of N/2 detail coefficient
zero in each step than to find the N/2 detail coefficient by FHT as in [4].

SOBEL EDGE DETECTION

Edge is a collection of pixels, associated with the boundaries of objects in image.Edge Detection is a measurement and
detection and location of the image changes in image gray. It is basic feature of the image. In edge detection, first edge
enhancement operator to highlight the local edge of the image, then define the pixel edge detection and set the threshold to
extract the edge point set.

VOLUME 4, ISSUE 7, DEC/2017 192 http://iaetsdjaras.org/


IAETSD JOURNAL FOR ADVANCED RESEARCH IN APPLIED SCIENCES ISSN NO: 2394-8442

The Sobel operator is widely used in image processing, particularly within edge detection algorithms.Sobel edge detection
has two advantages.
1)Average factor, has something effect to the random noise of the image.
2) Edges seems thick and bright, because it is differential of two rows and columns, so the elements of the edge on both
sides has been enhanced.

Sobel operator is a kind of orthogonal gradient operator.It generates a series of gradient magnitudes and simple convection
kernel. For the continuous function ( , ) in the position ( , ). The gradient can be expressed as a vector.

∇ ( , )= =
The magnitude and direction angle of the vectors are
(∇ ) =

∅( , ) =

The Sobel edge detector uses a pair of 3 x 3 convolution masks, one estimating gradient of the x-direction and the other
gradient is y–direction. It is easy to implement than the other operators [5].

−1 0 +1
= −2 0 +2
−1 0 +1

+1 +2 +1
= 0 0 0
−1 −2 +1

Pseudo-codes for Sobel edge detection method .


Input: A Sample Image.
Output: Detected Edges.
Step 1: Accept the input image.
Step 2: Apply mask , to the input image
Step 3: Apply Sobel edge detection algorithm and the gradient.
Step 4: Masks manipulation of , , separately on the input image.
Step 5: Results combined to find the absolute magnitude of the gradient.

| |= +
Step 6: the absolute magnitude is the output edges.

GABOR FILTER

Gabor filters are used in image processing for texture analysis. Several researchers have considered the use of such texture
feature for pattern retrieval [7] [8]. The Gabor filter uses extracting textured image features motivated by various factors.
Gabor filter and Gabor transform provides an accurate time-frequency location by “Uncertainty Principle” [9].

Gabor filter uses in several image analysis applications like texture classification and segmentation [10] [11], image
recognition [12] [13] [14], image registration and motion tracking [15].

VOLUME 4, ISSUE 7, DEC/2017 193 http://iaetsdjaras.org/


IAETSD JOURNAL FOR ADVANCED RESEARCH IN APPLIED SCIENCES ISSN NO: 2394-8442

A two dimensional Gabor function ( , ) and its Fourier Transform ( , ) can be written as
1 1
( , )= exp − + +2
2 2

1 ( − )
( , )= − +
2

Here = and =

Let ( , ) be the mother Gabor wavelet, then this self-similar filter dictionary can be obtained by appropriate dilations and
rotations ( , )through the generating function is
( , )= ( , ), > 1, , =

= ( + ), = (− + )

Here = ⁄ , and K is the total number of orientation. The scale factor in above equation is meant to ensure that
the energy is independent of m.

Pseudo-code for Gabor Filter:


1. For each frequency f in the frequency list flist:
2. For each orientation ∅ in the orientation list flist
3. Construct a Gabor filter ( , ∅),
4. Convolve ( , ∅)with original image I, get response image R,
5. Compute the mean response in R, denotes as r,
6. Count # of pixels that have a larger value than r,
7. Divide R into n x m frames,
8. For i = 1 to n:
9. For j = 1 to m:
10. Count the # of strong responses and compute the ratio r:
11. = /
12. Append r to the feature vector x:
13. Finally, = ( , , … | |∗| |∗ ∗ )

CLASSIFICATION
Classification is a process in which individual items like objects, patterns, image regions, pixels are grouped based on the
similarity between the item and the description of the group.For example, the classification may be presence of heart disease
(positive) or absence of heart disease (negative).

Figure 1: Classificaion image

VOLUME 4, ISSUE 7, DEC/2017 194 http://iaetsdjaras.org/


IAETSD JOURNAL FOR ADVANCED RESEARCH IN APPLIED SCIENCES ISSN NO: 2394-8442

CLASSIFICATION TECHNIQUES
NAVIE BAYES CLASSIFIER
The Navie Bayes classifier is based on Bayes theorem of probability [16]. The Navie Bayes is a probabilistic method used
for prediction. This classifier provides good results and easy to implementation of the results. The Navie Bayes classifier is
refered as the maximum a posterioiri (MAP) decision rule.

As per Bayes theorem, the conditional probability that an event x belongs to class k can be calculated by using the following
equation.

( | |)
( | )= ( )
( )

The above equation shows the calculation of ( | ) is a pattern classification problem, to find the probability that given
data x belongs to class k.

To estimate ( | ) and assume that any particular value x conditional on is statistically independent of each dimension
and can be written as

( | )= ( | )

[17] The comparison between classification accuracy on different classifier with DCT feature vectors.

Table 1: Accuracy of Classification Techniques


Accuracy (%)
Classifier
Average Std. dev
Centroid Neural Networks 71.8 7.68
Fuzzy C-Means 66.5 9.62
Multilayer Perception
72.6 8.04
Neural Networks
Navie Bayes 77.2 1.16

Graph
90
80
70
60
50
40
30
20
10
0
Centroid Neural Fuzzy C-Means Multilayer Navie Bayes
Networks Perception Neural
Networks

Accuracy (%) Average Accuracy (%) Std. dev

VOLUME 4, ISSUE 7, DEC/2017 195 http://iaetsdjaras.org/


IAETSD JOURNAL FOR ADVANCED RESEARCH IN APPLIED SCIENCES ISSN NO: 2394-8442

Table 2: comparison of training time on different classifier.


Training
Classifier
time
Centroid Neural Networks 1.22
Fuzzy C-Means 1.86
Multilayer Perception Neural
4.22
Networks
Navie Bayes 0.42

Graph2:
Training time

Centroid Neural Networks


Fuzzy C-Means
Multilayer Perception Neural Networks
Navie Bayes

SUPPORT VECTOR MACHINES


Support Vector Mechines are useful technique for classification. Support Vector Machinehave strong theoretical foundations
and excellent empirical successes. They have been applied to tasks such as handwritten digit recognition, object recognition,
and text classification.

SVM are based on structural Risk Minimisation principle from computional learning theory [18]. The idea of structural risk
minimization is to find a hypothesis h for which we can guarantee the lowest true error. The true error h is the probability
that h will make an error on an unseen and randomly seleted test example. An upper bound can used to connect the ture
error of a hypothesis h with the error of h on the training set and the complexity H, the hypothesis space containing h. SVM
find hypothesis h which minimizes this bound o,n the true error by effectively and efficient controlling the VC-Dimension
of H [19],[20].

The aim of SVM is production of a model (based on training data) that predicts target values of test data when given only
test data attributes.

The SVM
1
+
, , 2
Subject to
( ( )+ ) ≥1− ,
≥0

VOLUME 4, ISSUE 7, DEC/2017 196 http://iaetsdjaras.org/


IAETSD JOURNAL FOR ADVANCED RESEARCH IN APPLIED SCIENCES ISSN NO: 2394-8442

Here training vectors are mapped into higher dimensional space (may be infilnity) by a fuction ∅. Support Vector
Mechines finds a separating hyperlane with maximal margins in the higher dimensional space. >0 is he penalty parameter
of the error term. Furthermore , ≡ ∅( ) ∅ is called the kernel function [22].

SVM have four kernals.


1. Linear
, =
2. Polynomial
, = + , >0
3. Radial Basis Function (BSF)
, = − || − || , > 0
4. Sigmoid
, = ℎ +

Here , and are kernel parameters.

CLASSIFICATION AND REGRESSION TREE MODEL


CART model is developed by Breiman, Friedman, Olshen and Stone in 1984. Classification and regression trees are
machine-learning methods for constructing prediction models from data.

CART is non-parametric procedure for predicting continuous variable with categorical and/or continuous predictor variable
where data is partitioned into nodes on the basis of conditional binary responses to involving the predictor variable .

The models are obtained by recursively partitioning the data space and fitting a simple prediction model within each partition.
As a result, the partitioning can be represented graphically as a decision tree.

CART uses binary tree, to recursively partition the predictor space into subsets in successively more homogenous. For
eample, CART procedure derives conditional distribution of given , where x is a vector predictions like =
( , , , … ). A decision Tree with terminal nodes is used for communicating the classification decision. A
parameter ∅ = ( , , …. ) associated with the parameter = (1, 2, 3, … ) with the terminal node.

This procedure searches through all values of predictor values to find the variable that provides best partitioned into the
child nodes.The search is performed to determine the best split. It is given by

∅ =2 | − |

Here , are probability that a tuple in the training set will be on the left or right side of the tree. This is defined as :

| , | is the probabilities that a tuples is in class and in the left and right subtree. It is defined as

VOLUME 4, ISSUE 7, DEC/2017 197 http://iaetsdjaras.org/


IAETSD JOURNAL FOR ADVANCED RESEARCH IN APPLIED SCIENCES ISSN NO: 2394-8442

INSTANCE-BASED LEARNING
Instance-based learning refers to a family of techniques for classification and regression, which produce a class
label/predication based on the similarity of the query to its nearest neighbor(s) in the training set. Instance-based learning
algorithms do not create an abstraction from specific instances rather to store all the data, and at query time derive an answer
from an examination of the query’s nearest neighbor(s). [24].

Advantages of IBL is Training is very fast, Learn complex target functions and dontt lose information. Disadvantages are
Slow at query time and Easily fooled by irrelevant attributes.

IBL algorithms consist of the following three components: 1. Similarity Function: Given two normalized instances, this
yields their numeric-valued similarity.
2. Classification Function: An instance i to be classified and its similarity with each saved instance, this yields a classification
for i.
3. Memory Updating Algorithm: Given the instance being classified and the results of the other two components, this
updates the set of saved instances and their classification records.
Usually the Euclidean distance between two intances and is

, ≡ ( )−

Attributes are normalised by = here is the actal value of the attribute is output space and ( , ) ∈ ×
is called as a label instance, a case or an example. Y is the finite set comprised of classes { , ,… } whereas =
in regression.

Table 3: Accuracy of CART and IBL


Technique Accuracy (%)
CART 88
IBL 93

ACCURACY (%)

CART IBL

Precision, recall and f-measure for various classification techniques.

Table3: Precission, Recall, F-measure of CART and IBL

Technique Precission Recall f-Measure


CART 0.913 0.84 0.875
IBL 0.0957 0.9 0.928

VOLUME 4, ISSUE 7, DEC/2017 198 http://iaetsdjaras.org/


IAETSD JOURNAL FOR ADVANCED RESEARCH IN APPLIED SCIENCES ISSN NO: 2394-8442

Classification Techniques
1

0.5

0
Precission Recall f-Measure

CART IBL

CONCLUSION
Lossless images compression is obtained used the Haar Wavelet. Sobel Edge Detection and Gabor Filter transforms
extraction the edge and texture features from the compressed medical images. The classification is used the CART obtained
accuracy is 88% and IBL is 93%.

REFERENCES
1] Dezhgosha, K., Sylla, A. K., &Ngouyassa, E. (1994, May). Lossless and lossy image compression algorithms for on-board processing in
spacecrafts. InAerospace and Electronics Conference, 1994. NAECON 1994., Proceedings of the IEEE 1994 National (pp. 416423).
IEEE.

[2] Srikanth, R., &Ramakrishnan, A. G. (2005). Contextual encoding in uniform and adaptive mesh-based lossless compression of MR
images. Medical Imaging, IEEE Transactions on, 24(9), 1199-1206.

[3] Image Compression Using Modified Fast Haar Wavelet Transform, Anuj Bhardwaj and Rashid Ali, World Applied Sciences Journal 7
(5): 647-653, 2009 ISSN 1818-4952

[4] Chang, P. and P. Piau, 2007. Modified Fast and Exact Algorithm for Fast Haar Transform. Proceedings of World Academy of Science,
Engineering and Technology, 26: 509-512.

[5] An Improved Sobel Edge Detection, Wenshuo Gao Digital Media Department. Communication University of China ICICT 2010
Beijing,China gws8620@sina.com, Xiaoguang Zhang Digital Media Department. Communication University of China ICICT 2010,
Beijing,China zxgprc 1988@yahoo.cn, Lei Yang Digital Media Department. Communication University of China ICICT 2010 Beijing,China
young-Iad@263.net, Huizhong Liu Digital Media Department. Communication University of China ICICT 2010 Beijing,China
liujieI621@163.com.

[6] A Descriptive Algorithm for Sobel Image Edge Detection, O. R. Vincent, Clausthal University of Technology, Germany and University of
Agriculture, Abeokuta, Nigeria, vincent.rebecca@gmail.com, O. Folorunso Department of Computer Science, University of Agriculture,
Abeokuta, Nigeria, folorunsolusegun@yahoo.com, Informing Science & IT Education Conference (InSITE) 2009.

[7] W. Niblack et al, “The QBIC Project”, Proc. SPIE, Vol. 1, 908, pp.173-181, Feb 1993.

[8] Alexander P. Pentland, Rosalind W. Picard, Stan Sclaroff, "Photobook: tools for content-based manipulation of image databases", Proc.
SPIE 2368, 23rd AIPR Workshop: Image and Information Systems: Applications and Opportunities.

VOLUME 4, ISSUE 7, DEC/2017 199 http://iaetsdjaras.org/


IAETSD JOURNAL FOR ADVANCED RESEARCH IN APPLIED SCIENCES ISSN NO: 2394-8442

[9] C J Setchell, N W Campbell ,”Using Colour Gabor Texture Features For Scene Understanding.” In Proc. 7th Internat Conf. on image
processing applications. Vol. 67(5), pp. 372-376.

[10] A.C. Bovik, M. Clark and W.S. Geisier, “Multichannel Texture Analysis Using Localized Spatial Filters”, IEEE. Trans. Pattern
Analysis and Machine Intelligence, Vol.12, No. 1, pp.55-73, Jan 1990.

[11] B.S. Manjunath, and R.Chellappa, “A Unified Approach To Boundary Detection”, IEEE Trans. Neural Networks, Vol.4, No. 1,
pp-96-108, Jan 1993.

[12] J.G. Daugman, “High Confidence Visual Recognition of Persons by a Test of Statistical Indepence”, IEEE Trans. Pattern Analysis and
Machine Intelligence”, Vol. 15, No.11, pp.1, 148-161, Nov. 1993.

[13] M. Lades et. Al, “Distortion Invariant Object Recognition in the Dynamic Link Architecture”, IEEE Trans. Computer, Vol.42, No.3,
pp.300-311, Mar-1993.

[14] B.S. Manjunath, and R.Chellappa, “A Feature Based Approach to Face Recognition”, Proc. IEEE Conf. CVPR ’92,
pp.373-378, Champaign, III, June 1992.

[15] B.S. Manjunath, C. Sekhar and R.Chellappa, A New Approach to Image Feature Detection with Application”, Pattern Recognition,
Apr. 1996.

[16] M. Jang, D.Park, “Stochastic Classifier Integration Model,” International Journal of Applied Engineering Research, vol. 11, no.2, pp.
809-814, 2016.

[17] Dong-Chul Park, “Image Classification Using Naïve Bayes Classifier”, International Journal of Computer Science and Electronics
Engineering (IJCSEE) Volume 4, Issue 3 (2016) ISSN 2320–4028 (Online), pp. 135-139.

[18] Vladimir N. Vapnic, The Nature Of Statistic To Text Categorisation, New York.

[19] V S Giridhar Akula, Surya Bhupal Rao, “Data Embedding in a moving Compressed Video” , Scholars Journal of Engineering and
Technology, 1(1):44-48,2013

[20]Text categorization with support vector machines: Learning with many relevant features, T Joachims - Machine learning: ECML-98, 1998

[21] A Practical Guide to Support Vector Classification Chih-Wei Hsu, Chih-Chung Chang, and Chih-Jen Lin Department of Computer
Science National Taiwan University, Taipei 106, Taiwan.

[22]A comparative predictive analysis of neural networks (NNs), nonlinear regression and classification and regression tree (CART) models,
MA Razi, K Athappilly - Expert Systems with Applications, 2005 - Elsevier

[23] Instance-Based Learning, E Keogh - Encyclopedia of Machine Learning, 2011 - Springer

[24] Noise-Tolerant Instance-Based Learning Algorithms Davi d W. Ah a and Dennis Kibler* Department of Information and Computer
Science University of California, Irvine, CA 92717 aha@ics.uci.edu kiblerOics.uci.edu.

VOLUME 4, ISSUE 7, DEC/2017 200 http://iaetsdjaras.org/

Você também pode gostar