Escolar Documentos
Profissional Documentos
Cultura Documentos
Neha Sharma
Valentina Emilia Balas Editors
Advances in
Computing
Applications
Advances in Computing Applications
Amlan Chakrabarti Neha Sharma
Valentina Emilia Balas
Editors
Advances in Computing
Applications
123
Editors
Amlan Chakrabarti Valentina Emilia Balas
A.K. Choudhury School of Information Faculty of Engineering, Department of
Technology Automatics and Applied Software
University of Calcutta Aurel Vlaicu University of Arad
Kolkata, West Bengal Arad, Arad
India Romania
Neha Sharma
Zeal Institute of Business Administration,
Computer Application and Research
Zeal Education Society
Pune, Maharashtra
India
Globalization has influenced almost every facet of human life due to the emergence
of new and affordable digital and computing technologies as well as current trends
in management. At the same time, informatics with its strong focus on providing
fast and ready access of information is the backbone of many of the present-day
intelligent applications serving our mankind. All these create a perfect landscape for
widespread research interest in information and communication technologies along
with management policies and thus impacting peoples lives from entertainment to
health care and from databases to e-governance.
This edited volume of Advances in Computing and Management presents the
latest high-quality technical contributions and research results in the areas of
computing, informatics, management, and information management. It deals with
state-of-the-art topics and provides challenges and solutions for and explores future
research directions. Original, unpublished research work highlighting specic
research domains from all viewpoints is contributed from scientists throughout the
globe. The main goal of this volume is not only to summarize new research ndings
but also to place these in the context of past work. This volume is designed for
professional audience, composed of researchers, practitioners, scientists, and
engineers in academia and industry.
The following is the brief summary extracted from the respective chapters and
their abstracts:
Chapter 1, by Vamsi Krishna Myalapalli, proposes an approach to overcome
these befalling sundry wait events in the internal database engine through Query
and PL/SQL rewrite methodologies, and additional overhauling approaches to
increase data hit ratio. The experimental progression and approaches evinced that
CPU impact, wait events, and other performance bottlenecks are minimized. This
paper could serve as tuning tool to boost query as well as database performance by
wait event tuning and also can oblige as a utility for Database Administrators, SQL
programmers, and database operators.
Chapter 2, by Snehalata Shirude and Satish Kolhe, proposes Agent based
Library Recommender System with the objective to provide effective and intelligent
v
vi Preface
use of library resources such as nding right books, relevant research journal
papers, and articles.
Chapter 3, by Kalyan Baital and Amlan Chakrabarti, presents a scheduling
algorithm where random tasks generated at different time interval with different
periodicity and execution time can be accommodated into a system, which is
already running a set of tasks, meeting the deadline criteria of the tasks.
Chapter 4, by Manoj K. Sabnis and Manoj Kumar Shukla, explains Model-based
Approach for Shadow Detection of Static Images using two methods, i.e.
color-based and texture-based.
Chapter 5, by Nitin Vijaykumar Swami et al., has covered the concepts of Li-Fi,
how the Li-Fi technology can be enhanced in the mobile communication, how it
works, the Li-Fi cellular network, some ubiquitous computing applications, com-
mon misconceptions about Li-Fi, Li-Fi in solar cell, and Internet of things (IoT).
Chapter 6, by Shraddha Oza and K.R. Joshi, analyzes the performance of
denoising lters like NLM (non-local mean (NLM) spatial domain lter), bilateral,
and linear Gaussian lters using PSNR, MSE, and SSIM for MR (Magnetic
Resonance) Images.
Chapter 7, by Gnter Fahrnberger, presents a detailed view on SecureString 3.0.
The homomorphic cryptosystem SecureString 3.0 remedies to recapture the cloud
users faith in secure cloud computing on encrypted character strings by combining
the paradigms blind computing and secret sharing. Implementation details of
this cryptosystem given in pseudocode allow researchers to realize their own pro-
totypes and practitioners to integrate SecureString 3.0 in their own security
solutions.
Chapter 8, by Sujay D. Mainkar and S.P. Mahajan paper, focuses on the
development of feature extraction and accurate classication of variety of acoustic
sounds in unstructured environments, where adverse effects such as noise and
distortion are likely to dominate. This chapter attempts to classify 10 different
unstructured real-world acoustic environments using empirical mode decomposi-
tion (EMD) which considers inherent non-stationarity of acoustic signals by
decomposing the signal into intrinsic mode functions (IMFs).
Chapter 9, by Mrunal Pathak and N. Srinivasu, presents the overview of different
multimodal biometric (multibiometric) systems and their fusion techniques with
respective to their performance.
Chapter 10, by Mohan S. Khedkar and Vijay Shelake, proposes a technique
based on the concept of dynamic secret key (DSK), which is used to generate
symmetric cryptography keys for designing an authentication and encryption
scheme in smart grid wireless communication. In this scheme, recently added
device (e.g., smart meter) is authenticated by a randomly chosen authenticated
device. This scheme enables mutual authentication between a control center situ-
ated in local management ofce and the randomly chosen device as an authenticator
to generate proper dynamic secret-based dynamic encryption key (DEK) for con-
sequent secure data communications.
Chapter 11, by Vyanktesh Dorlikar and Anjali Chandavale, proposes a smart
security framework with the integration of enhanced existing technologies.
Preface vii
This authentication framework provides security for the high-risk location, such as
government ofce buildings, airport, military bases, and space stations.
Chapter 12, by Ashwini Shewale et al., attempts to scale down the medical
image processing time, without getting arousing effect on the quality of image using
efcient computational methods for medical image processing.
Chapter 13, by Supriya Kunjir and Rajesh Autee, attempts to develop a
cost-effective security system based on radar sensor network to prevent terrorism to
a great extent. The system specically aims at the task of detecting obstacles by
means of ultrasonic radar sensor network and provides photograph of the detected
obstacles using camera and also provides total count of detected obstacles by means
of counter. The ultrasonic sensor network coupled with counter and display unit is
then totally coupled to the FM transceiver to get the voice announcement.
Chapter 14, by Moumita Acharya et al., proposes a low-resource and
energy-aware hardware design for DWT through dynamic bit width adaptation, thus
performing the computation in an inexact way. They have performed eld pro-
grammable gate array (FPGA)-based prototype hardware implementation of the
proposed design.
Chapter 15, by Valentina Emilia Balas et al., aims to minimize the possibility of
avalanche via systematically analyzing the cause behind avalanche. A novel and
efcient attack model is proposed to evaluate the degree of vulnerability in a
dependency-based system caused by its members. This model uses an algorithmic
approach to identify, quantify, and prioritizing, i.e., ranking the extent of vulner-
ability due to the active members in a dependency-based system.
Chapter 16, by Neha Sharma and Hari Om, presents a case study to predict the
survival rate of oral malignancy patients, with the help of two predictive models,
linear regression (LR), which is a contemporary statistical model, and multilayer
perceptron (MLP), which is an articial neural network model.
The main goal of this volume is to summarize new results but also place these in
the context of past work.
We are grateful to Springer, especially to Ms. Swati Meherishi (Senior Editor,
Applied Sciences & Engineering) and her team for the excellent collaboration,
patience, and help during the evolvement of this volume.
We hope that the volume will provide useful information to professors,
researchers, and graduated students in the area of intelligent transportation.
We, the editors of the book, Dr. Amlan Chakraborty, Dr. Neha Sharma, and
Dr. Valentina Emilia Balas, take this opportunity to express our heartfelt gratitude
toward all those who have contributed toward this book and supported us in one
way or the other. This book incorporates the work of many people, all over the
globe. We are indebted to all those people who helped us in the making of this
high-quality book which deals with state-of-the-art topics in the areas of computing,
informatics, management, and information management.
At the outset, we would like to extend our deepest gratitude and appreciation to
our afliations: Dr. Amlan Chakrabarti to University of Calcutta, India, Dr. Neha
Sharma to Zeal Institute of Business Administration, Computer Application, and
Research of S.P. Pune University, India, and Dr. Valentina Emilia Balas to
Department of Automatics and Applied Software, Faculty of Engineering of
University Aurel Vlaicu of Arad, Romania, for providing all the necessary
support throughout the process of book publishing. We are grateful to all the
ofcers and staff members of our afliated institutions who have always been very
supportive and have always been companions as well as contributed graciously in
the making of this book.
Our sincere appreciation goes to our entire family for their undying prayers,
love, encouragement, and moral support and for being with us throughout this
period, constantly encouraging us to work hard. Thank You for being our
backbone during this journey of compilation and editing of this book.
Amlan Chakrabarti
Neha Sharma
Valentina Emilia Balas
ix
Contents
xi
xii Contents
xiii
xiv About the Editors
of more than 160 research papers in refereed journals and international conferences.
Her research interests are in intelligent systems, fuzzy control, soft computing,
smart sensors, information fusion, modeling, and simulation. She is the editor in
chief to International Journal of Advanced Intelligence Paradigms (IJAIP) and
Editorial Board member of several national and international journals and is
evaluator expert for national and international projects. She is a member of
EUSFLAT, ACM, and a Senior Member IEEE, member in TCFuzzy Sys-
tems (IEEE CIS), member in TCEmergent Technologies (IEEE CIS), member in
TCSoft Computing (IEEE SMCS), and also a member in IFACTC 3.2
Computational Intelligence in Control.
Chapter 1
Wait Event Tuning in Database Engine
1.1 Introduction
Tuning a database is an activity with the goal of making the database run better than
it did earlier. Even a small database will, with time and use, grow in size and
complexity. A process or report that once performed with an adequate runtime will
run more slowly the longer the system is in usage. This is due to more demands
being made on the system as well as the increasing amount of data through which
the systems need to search to perform operations. All too often, this happens
gradually until the users have a critical deadline; in worst case it may take form of a
crisis. This realization triggers a reactive tuning session as the DBA attempts to
respond to sudden demand to enhance database performance.
V.K. Myalapalli ()
JNTUK University College of Engineering, Vizianagaram, Andhra Pradesh, India
e-mail: vamsikrishna.vasu@gmail.com
Tuning has always been an important part of DBA job next to backup and
recovery. It is sometimes deemed as Black Magic. The analytical process is
centered on using ratios to determine health of a database or component of the
database.
Further paper is organized in the following manner. Section 1.2 explains the
related work and background. Section 1.3 deals with proposed benchmark, i.e., the
methodologies. Section 1.4 demonstrates experimental set-up, i.e., implementing
methodologies in a pragmatic approach. Section 1.5 explains comparative analysis
and nally Sect. 1.6 concludes the paper.
Query optimization can take form of either Rule Based or Cost Based referred as
Rule-Based Optimization (RBO) or Cost-Based Optimization (CBO), respectively.
High-Performance SQL [1] explained RBO and CBO level tuning models. It made
part of the queries faster via CBO approach and rest others via RBO approach.
An appraisal to optimize SQL queries [2] explained basic model that rewrites
sundry queries to minimize CPU cost and raise the index utilization. It also reduced
rate of hard parsing and increased query reuse (soft parse).
Augmenting database performance via SQL tuning [3] explained query tuning
through index tuning, data access types, and hints. On the other hand, it delineated
database tuning through session handling.
High-performance PL/SQL [4] explained the tuning model to reduce the rate of
Context Switching (an overhead) among SQL and PL/SQL engines. Also it
explained gaining performance over implicit and explicit cursors.
This paper explains reactive and a few proactive-tuning approaches toward
minimizing wait events and other bottlenecks in database engine. In reactive tuning,
it has to be seen whether it is the buffer cache not containing enough information for
quick retrieval, execution plans being aged out of the shared pool (shared memory),
locking or latching issues, excessive parsing or any of the myriad of other com-
ponents in the DBMS engine.
This paper explains sundry query rewriting methodologies, choosing the precise
query based on the scenario and other overhauling approaches toward reducing
impact on database engine.
1 Wait Event Tuning in Database Engine 3
In the case of wait event tuning, there are plenty of v$ views (dynamic per-
formance views) in the database that contains comprehensive information about
wait events as they occur.
Query Tuning #(18):
Query tuning constitutes the major part (>70%) of the comprehensive database
tuning.
(1) Deter Full Table Scans (FTS): FTS retrieves each and every bit of data from
the table, i.e., triggers a lot of disk I/O. FTS may arise due to the following
reasons
(a) WHERE clause absence.
(b) Index and statistics of table or view are stale.
(c) Absence of row ltering, data type mismatch, not equal, like and null
operators in WHERE clause.
(d) Using functions in SELECT clause.
(2) Pivoting vs. Grouping: Grouping functions can be re-written as pivot state-
ments to reduce overhead in grouping. It also transforms rows in columns.
Succinctly, prefer pivoting to simple grouping when output format is not a
concern.
(3) Efciently handling function(s) in WHERE clause: Occasionally functions in
WHERE clause will lead to FTS. These are caused when a WHERE clause
predicate is invalidated with a pre-dened function.
Ex: Following query constraint(s) involving date range may lead to FTS.
Case1: where trunc (join_date) > trunc (sysdate-7);
Case2: where to_char (join_date,yyyy-mm-dd) = 1991-12-27;
Even though there exists an index for column join_date, the trunc() and
to_char() pre-dened function(s) would invalidate index, leading to sub
optimal execution and needless I/O.
(4) Logical vs. Physical Delete of a Column: In a huge database environment,
physical drop of a column would instigate higher resource consumption apart
from consuming more time.
Logical deletion of a column allows DBA or Developer to physically delete
the column at later time (during non-peak time).
Exploiting Index #(58):
(5) Efcient Truncate: Using truncate function on column prevents using
index [8]. Henceforth, query should be re-written to take advantage of index
for faster access.
(6) Deter using columns on both sides of the operator: If an indexed column is
present on both sides of the operator, then index is not used.
4 V.K. Myalapalli
(7) Enforcing CBO to Pick Best Access Path: This is performed to make operation
more or less index friendly, i.e., compelling optimizer less or more prone to
select an index access path over an FTS. This behavior can be leveraged
through the parameter OPTIMIZER_INDEX_COST_ADJ.
On an (OLTP) Online Transaction Processing system, altering this parameter
to lesser value (preferably 1030) would lead to massive performance gain.
(8) Index Merge: Index merging permits merging separate indexes and use the
result, instead of visiting the table from one of the indexes. This reduces the
amount of bytes processed and CPU cost for processing query.
Optimization via Parameters #(912):
(9) Dynamic Sampling: This lets the optimizer to take sample rows of table to
compute missing statistics. This behavior is driven by the parameter
OPTIMIZER_DYNAMIC_SAMPLING or DYNAMIC_SAMPLING
hint.
This practice is benecial when there are frequently executed n-way joins.
By sampling a small sub-set of data, the DBMS Engine Cost-Based Opti-
mizer looks at faster join order for Tables
(10) Optimization Level: It denotes level of optimization used for compiling
PL/SQL library unit(s). Higher the value, higher the compiler effort. (Range:
0 to 2). Nevertheless, higher value will comparatively slow down the
processing.
(11) Performing Optimization contingent on Data Retrieval: Optimizer_Mode
parameter is very benecial and important to tune queries, especially if rows
are retrieved in chucks (ex: rst 10 or 100 rows). By default, it is set to
ALL_ROWS mode. If chunks of data are to be retrieved, then setting this
parameter to FIRST_ROWS_n will increase performance.
(12) Determining Wait Events: Wait event may take form of sundry kinds of
internal contention. Enabling timed statistics and wait interfaces allows us to
see what various components are doing by looking at where they are
spending their time waiting.
Ratio-Based Tuning #(1319):
(13) Ensuring In-Memory Sort: If memory is insufcient for sorting, disk memory
is used, which slows down a query. This situation can be recognized using v
$sysstat view. To resolve this issue, main memory limit has to be raised via
the parameter sort_area_size.
(14) Disk I/O Tweaking: Time taken to R/W database les or external les should
be minimized. This can be resolved by defragment les at OS level.
(15) User Wait Events: Users experiencing higher wait time (or logged in for long
time) can be terminated to release resources held by them and can be
1 Wait Event Tuning in Database Engine 5
rescheduled later. So that the other current users can make use of resources
released.
(16) Enhancing Parse Once Execute Many: Ideally hard parsing [2] should be
eliminated. If the ratio of hard parsing is found much higher than soft
parsing, enforce cursor sharing [4] and bind variables [4].
(17) Ascertaining Hard Parses: Higher amount of hard parsing designates that
repeatedly executed query is getting out of shared pool (shared memory).
Increasing the size of shared pool [5] or pinning object (ex: table/procedure)
to pool or caching table [2] will minimize hard parse.
(18) Minimizing Paging and Swapping: Effectively modifying (mostly increas-
ing) the size of (SGA) shared global area (Shared Memory) will reduce the
rate of paging and swapping.
(19) Tuning PGA for Optimal Memory Usage: Size of program global area
(user-specic session memory) affects cache hit ratio [7]. The parameter
PGA_AGGREGATE_TARGET (species target aggregate memory) should
be altered to enhance cache hit ratio.
Mending Database Objects #(2021):
(20) Cleaning Invalid Object(s): Often during application maintenance, there may
be code mutations, which might affect (ex: dependencies) other objects
making them invalid. This leads to build failure or functional failure at run
time. To prevent this, we should heed the status of dependent objects.
(21) Disabled Triggers: These are deleterious than invalid object(s) as they simply
does not execute. They will lead to failure of the business logic of an
application.
PL/SQL Tuning #(2225):
(22) Reducing calls to SYSDATE: Calls to SYSDATE escalate overhead for a
query. If multiple calls to SYSDATE are inevitable (prevalently in a loop),
then the technique of code motion should be preferred.
(23) Limiting Dynamic SQL: Dynamic SQL is benecial to application from
functional perspective. However, it will open windows for performance
degradation and SQL injection [6]. If dynamic SQL is integrated with static
SQL, optimizing generated SQL statements would be highly challenging.
(24) Implicit versus Explicit Cursor: Deem implicit cursor(s) worthy over explicit
cursor(s) for faster code processing, since implicit cursors are pre-optimized.
Explicit cursors must pass through expensive Declare, Open, Fetch and
Close phases.
(25) Using Associative Arrays: They ensure fast reference table lookups. When
the reference tables are searched via key, then query performance over ref-
erence tables can be drastically enhanced by loading reference tables into
associative arrays also referred as Index-by tables.
6 V.K. Myalapalli
In this section, the methodologies specied in the earlier section will be explained
via queries or code snippets that serve in the form of exemplars.
1 Wait Event Tuning in Database Engine 7
8 V.K. Myalapalli
1 Wait Event Tuning in Database Engine 9
10 V.K. Myalapalli
1 Wait Event Tuning in Database Engine 11
12 V.K. Myalapalli
1 Wait Event Tuning in Database Engine 13
After tuning queries, CPU cost and response time were enhanced and wait events
were signicantly minimized. Some of the results are represented beneath in the
form of output screens.
In order to get statistics for each query, we red some queries on database prior
to tuning, for evaluating statistics and execution plans.
SQL > set timing on //Displays elapsed time
SQL > set feedback on //Metadata of output
SQL > set autotrace on explain //Traces every query
Figure 1.1 shows the member count in each department retrieved through
Grouping method.
Figure 1.2 shows that data retrieved in Fig. 1.1 are retrieved via pivoting method
with less CPU cost and elapsed time. Also the method of retrieving data does not
include hashing as involved in formal method.
Figure 1.3 shows the statistics (CPU Cost-31, Bytes Accessed-31458942, and
Rows Processed-1023029) of a query, which retrieves rows from tables in chunks
(50 at a time). Here, optimizer mode is set to ALL_ROWS (default mode).
Figure 1.4 shows the statistics (CPU Cost-9, Bytes Accessed-31457406, Rows
Processed-1022981) after changing optimizer mode is to FIRST_ROWS_1. Here,
though the operations performed as per execution are same, more than 70 % of the
CPU cost is reduced.
Figure 1.5 designates that after increasing the size of sort area, disk sorts are
diminished to 0.
After enforcing the proposed explicated approaches, tuning is ensured at data-
base engine level as well as at the query level.
1.6 Conclusion
Database tuning necessitates each bit of our ingenuity to evade the harmful way.
Query rewrite can alter the way the data are accessed or processed. Efcient query
will always minimize the impact on underlying database.
16 V.K. Myalapalli
References
1. Myalapalli VK, Savarapu PR (2014) High performance SQL. In: 11th IEEE international
conference on emerging trends in innovation and technology, Pune, India, Dec 2014
2. Myalapalli VK, Shiva MB (2015) An appraisal to optimize SQL queries. In: IEEE international
conference on pervasive computing, Pune, India, Jan 2015
3. Myalapalli VK, Totakura TP, Geloth S (2015) Augmenting database performance via SQL
tuning. In: IEEE international conference on energy systems and application, Pune, India, Oct
2015
4. Myalapalli VK, Teja BLR (2015) High performance PL/SQL programming. In: IEEE
International conference on pervasive computing, Pune, India, Jan 2015
5. Myalapalli VK, Chakravarthy ASN, Reddy KP (2015) Accelerating SQL queries by
unravelling performance bottlenecks in DBMS Engine. In: IEEE international conference on
energy systems and application, Pune, India, Oct 2015
6. Myalapalli VK, Chakravarthy ASN (2014) A unied model for cherishing privacy in database
system. In: IEEE international conference on networks and soft computing, Andhra Pradesh,
India
7. Burleson DK, Celko J, Cook JP, Gulutzan P (2003) Advanced SQL database programmer
handbook, August 20031st edn. Rampant Tech press
8. Mishra S, Beaulieu A (2002) Mastering oracle SQL April 2002 1st edn. OReilly Publishing
Chapter 2
Machine Learning Using K-Nearest
Neighbor for Library Resources
Classication in Agent-Based Library
Recommender System
2.1 Introduction
Recommender system can help user to obtain relevant books and journal articles. It
lters library resources according to users interest. The agent-based architecture is
used to design the framework of recommender system. The recommendation per-
formance of the agents is improved by machine learning. Library recommender
agent performs the main tasks of ltering and providing recommendations. Rec-
ommender agent makes use of user proles to lter the library resources. User
proles and library resources are represented as vectors containing term frequencies
for every signicant keyword. In library resources, most of the book records have
table of contents, and journal articles include abstract and keywords. This provides
rich set of keywords available while computing similarity. More correct recom-
mendations are possible by the addition of semantically equivalent keywords into
the vectors. ACM computing classication system (ACM CCS) 2012 is used as
static ontology [1]. ACM CCS is in simple knowledge organization system (SKOS)
format which has 14 categories. Machine learning is useful to identify the category
of the book. The results of the experiments performed using k-Nearest Neighbor are
discussed in this paper.
The literature study is performed to learn the present state of the work in the eld of
development of recommender system for library resources. The overall study take
into consideration the various views, viz., approach used, technique applied for the
tasks such similarity computation and classication, and method for relevance
feedback and performance of the system. This paper describes the library resources
classication using k-Nearest Neighbor which is the important stage of library
recommender agent in the complete library recommender system. Therefore, the
part of literature study focusing on agent-based systems performing classication or
similar tasks to achieve the similar objectives is given in this paper.
The survey of recommender systems for libraries is performed by Gottwald and
Koch in 2011. BibTip, ExLibris bX, Techlens, Foxtrot, Fab, and LIBRA are listed
as existing solutions to the eld [2]. The workshop on new trends in content-based
recommender systems (CBRecSys2014) is identied. The goal of CBRecSys2014
was to provide the platform for the papers dedicated to all aspects and new trends of
content-based recommendations. There are many recommendation domains and
applications where content and metadata play a key role. In domains such as
movies, the relationship between content and usage data has seen thorough
investigation done already but for many other domains such as books, news, sci-
entic articles, and Web pages, it is not still known if and how these data sources
should be combined to provide the best recommendation performance. This is
2 Machine Learning Using K-Nearest Neighbor for Library 19
user prole, knowledgebase, and library resources. The properties of task environ-
ment are partially observable, deterministic, sequential, static, discrete, and single
agent. Actuators have ability to follow the links that are user interface and display the
information. Sensors have ability to parse documents and Web pages.
City block the so-called Manhattan metric is combined for numerical attributes
in City-SVD. The absolute value of difference between attribute values ax, ay
is the distance between those two numerical values. jax, ayj is divided by
standard the length of the range of the attribute a in training data as the normal-
ization parameter is set to the value range. The length is the difference between
largest and smallest values of the attribute in training data. The number of neighbors
(k) is kept 1, 2, 3, 4, and 5. Finally, the option for number of neighbors is set to
search optimal between 1 and 100 to identify the right value of k. Attribute
weighting/scaling is done by two methods: distance-based and accuracy-based. The
distance-based method works iteratively to choose the weights to optimize the
distance to correctly identied training library resource Li. Accuracy-based attribute
scaling method also works iteratively to choose the weights to optimize the accu-
racy of decision prediction for training library resource Li. Twenty iterations are
performed. In accuracy-based method, every iteration increases the weight of
attributes with high accuracy of classier having value k = 1. The second stage of
k-NN classier is determining the category using the recognized neighbors. The
distance-weighted voting method is selected to perform this. The votes are weighted
using the distance between the neighbor and the library resource for which category
needs to identify. Finally, the category is selected having the largest total of
weighed votes. If a neighbor generates a local rule inconsistent with other members
of neighborhood, then the classier excludes it from voting process as the option
lter neighbors using rules are set to true.
Figure 2.2 shows the design of k-NN classier for one sample experiment.
Testing and training library resources records are in tab le format. Eleven cate-
gories among fourteen are taken for classication from ACM CCS 2012. Three are
ignored because they are general. Experiments are performed using hold out, ten-
fold, leave one out cross-validation techniques. The values of k are 1,2,3,4,5 and in
24 S.B. Shirude and S.R. Kolhe
range 1 to 100 for all techniques. Figure 2.3 gives the confusion matrix generated
for one sample experiment.
For correctly classied library resources, classication accuracy is computed for
experiments. For evaluation of experiments, precision, recall, and f1 values [18] are
computed for each of the folds and LOOCV technique.
2 Precision Recall
F1 =
Precision + Recall
From the above table, it is clear that average classication accuracy is better
when the value of k lies between 1 and 100. The precision, recall, and f1 values are
computed for these experiments and are given in Table 2.5.
The graph is plotted for the above values to compare the approaches as given in
Fig. 2.4.
k-NN classier in experimented approaches takes decision for 231 library
resources. In tenfold approach, rst ninefold takes 22 records of library resources
while in the last tenfold 33 remaining records are added. In LOOCV technique, the
single record of library resource from all testing records is selected at a time. The
coverage for all experiments using k-NN is 100 % because the classier works
nonlinearly and classifying all testing library resources. It tries to nd nearest
category that may result in misclassication. Figure 2.4 shows the graphical
2 Machine Learning Using K-Nearest Neighbor for Library 27
2.6 Conclusion
References
1. http://www.acm.org/about/class/2012
2. Gottwald S, Koch T (2011) Recommender systems for libraries. In: ACM recommender
systems 2011. Chicago
3. Bogers T, Koolen M, Cantador I (2014) Workshop on new trends in content-based
recommender systems:(CBRecSys 2014). In: Proceedings of the 8th ACM conference on
recommender systems. ACM, pp 379380
4. Jennings N, Wooldridge M (1996) Software agents. IEE Rev 42(1):1720
5. Prakash N (2004) Intelligent search in digital libraries. In: 2nd convention PLANNER.
Manipur University, Imphal Copyright INFLIBNET Centre, Ahmedabad, pp 8390
6. Mnnich M, Spiering M (2008) Adding value to the library catalog by implementing a
recommendation system. D-Lib Mag 14(5):4
7. Prakasam S (2010) An agent-based intelligent system to enhance e-learning through mining
techniques. Int J Comput Sci Eng
8. Morales-del-Castillo JM, Peis E, Herrera-Viedma E (2010) A ltering and recommender
system for e-scholars. Int J Technol Enhanc Learn 2(3):227240
9. Pakdeetrakulwong U, Wongthongtham P (2013) State of the art of a multi-agent based
recommender system for active software engineering ontology. Int J Digit Inf Wirel Commun
(IJDIWC) 3(4):2942
10. Bedi P, Vashisth P (2015) Argumentation-enabled interest-based personalised recommender
system. J Exp Theor Artif Intell 27(2):199226
11. Lau SB-Y, C-S Lee, YP Singh (2015) A folksonomy-based lightweight resource annotation
metadata schema for personalized hypermedia learning resource delivery. Interact Learn
Environ 23(1):79105
12. Huang C-L, Yeh P-H, Lin C-W, Den-Cing W (2014) Utilizing user tag-based interests in
recommender systems for social resource sharing websites. Knowl-Based Syst 56:8696
13. Ghazanfar MA, Prgel-Bennett A (2014) Leveraging clustering approaches to solve the
gray-sheep users problem in recommender systems. Expert Syst Appl 41(7):32613275
14. Chen C-M (2013) An intelligent mobile location-aware book recommendation system that
enhances problem-based learning in libraries. Interact Learn Environ 5:469495
15. Hulseberg A, Monson S (2012) Investigating student driven taxonomy for library website
design. J Electron Res Librariansh 361378
16. Liao I-E, Hsu WC, Cheng MS, Chen LP (2010) A library recommender system based on
PORE and collaborative ltering technique for English collections. The Electron Libr 28:386
400
17. Bazan JG, Szczuka MS, Wroblewski J (2002) A new version of rough set exploration system.
In: Rough sets and current trends in computing. Springer, Berlin, pp 397404
18. Gunawardana A, Shani G (2009) A survey of accuracy evaluation metrics of recommendation
tasks. J Mach Learn Res 10:29352962
Chapter 3
An Efcient Dynamic Scheduling of Tasks
for Multicore Real-Time Systems
3.1 Introduction
K. Baital ()
National Institute of Electronics and Information Technology,
Kolkata, India
e-mail: kalyan_baital@yahoo.co.in
A. Chakrabarti
A. K. Choudhury School of Information Technology,
University of Calcutta, Kolkata, India
e-mail: amlanc@ieee.org
results are produced [1]. In non real-time system, throughput is the indicator of
measurement of the performance of the system, that is more and more tasks are
required to be executed in a certain time period for increasing the performance.
Performance in real-time system is measured based on the following criterion:
many tasks should be processed as much as possible, so that they will produce the
desired results before their deadline. Hence, the real-time system must be pre-
dictable in nature. Therefore, in strict real-time systems, a delay in result is not just
a delay but useless. In a real-time system, the system time (internal time) is mea-
sured with same timescale as the controlled environment (external time). The
deadline parameter classies the main difference between real time and
non-real-time systems. In real-time system, deadline has to be met under all even as
well as worst circumstances. Real-time system is divided into three types depending
upon the processing of tasks of different strictness levels, which are described as
follows:
(a) A real-time task is called hard if its miss in deadline may cause disastrous
consequences on the environment under control.
(b) A real-time task is soft if meeting its deadline is desirable but missing does not
cause serious damage.
(c) A real-time task is called rm if its miss in deadline makes the result useless,
but missing does not cause serious damage.
Two types of real-time task are there in respect of periodicity, namely periodic
and aperiodic. In a periodic task, a sequence of instances is generated with a xed
period, and in the case of aperiodic task, no period is present, that is next instance of
task is not known.
High-speed processors are essential for real-time system, and reduced cost of
high-speed processors has shown the way for solving real-time system demand
more efciently [25]. Today, multicore system is integrated into real-time system.
A multicore system integrates two or more processors in an integrated circuit for
performance enhancement and optimum power consumption. It executes multiple
tasks simultaneously with efciency. Unlike a single core, it performs concurrent
processing of tasks involving more than one core. The switching takes place either
at regular intervals or when currently executing task releases the control of the
processor. Multicore is an emerging trend currently due to its speed and perfor-
mance [6].
There are different ways of processing a multicore system which are summarized
below [7]:
(a) Symmetric multiprocessing (SMP)A single instance or image of the
real-time operating system managing the cores and shared resources of the
system. All the resources are available for all the cores and tasks; hence, no
external communication is required between the cores.
3 An Efcient Dynamic Scheduling of Tasks 33
a Worst Case Execution Time (WCET) estimation per task, which is the maximum
value for all execution scenarios of the task. [23] presents an approach that consider
not a single WCET estimation but a set of WCET estimation per task. This allows
the algorithm to reduce the number of resources for a given task set and hence
enhance the performance of the system. The authors of [24] implement a staggered
model for Proportionate-fair (Pfair) scheduling on a symmetric multiprocessor for
performance improvement and [25] uses Pfair to reduce the scheduling overhead,
and also supporting task-interrupt co-scheduling for embedded multicore systems.
One can visit [2637, 38] to have more exhaustive study on scheduling considering
different parameters, namely processor resources allocation, sequence of task,
dependent task, parallel multithread, power optimization, temperature and memory
access.
From all the previous work related to real-time scheduling done by the
researchers, it may be concluded that there is a need to have a flexible dynamic
scheduling model for real-time task execution in a multicore system under all
conditions, preserving the deadline among the tasks and also optimizing the
throughput of the system.
(C) Claim for novelty
This paper proposes a simulation model to nd the best possible way to schedule
the given set of tasks efciently to the available processing cores so that all the task
deadlines are met and also the throughput of the system is optimum for the set of
tasks.
The model maintains two-level queuesglobal and local [39, 40]. It also uses
the concept of two algorithms, namely EDF and Pfair as per follows:
(a) EDF at the global queue of new tasks and
(b) Pfair for context switching and forwarding the job to local queue.
Time complexity of both EDF and Pfair algorithm is O (logn); therefore, time
complexity of the proposed model is also O (logn), but the model has some nov-
elties in respect to efcient utilization of the processors as well as tting almost all
the tasks efciently meeting the deadline condition, which is given below:
The model is better than EDF scheme because EDF works well in underloaded
condition and when there is only one processor. But our proposed model is in
overloaded condition and has multiple processors. The same model can also be used
in underloaded condition having a single processor.
Further we can allocate almost all the new tasks into the cores if
1. Periodicity is big for the new task
2. Execution time of new task is small
3. Task is divided into huge number of jobs thus decreasing their execution time so
that jobs can be tted before next instance of the existing task starts.
3 An Efcient Dynamic Scheduling of Tasks 35
The architecture of the algorithm model with two-level queues is shown below
(Fig. 3.3):
The working principle of the proposed model is described below:
1. New tasks are rst stored in global queue (say T5, T6, T7) and are sorted in a
priority queue according to EDF or least time to go, i.e. task with earliest
deadline is given the highest priority. The rearrange queue is (T7, T6, T5).
Whenever a scheduling event occurs (task nishes, new task released, etc.), the
queue will be searched for the task with earliest deadline.
2. Using Pfair concept, front task (say T5) is divided into a number of jobs z with
execution time, Et ms (where Et = E/z), i.e. execution time is divided among the
jobs where Et is the execution time of divided jobs and the jobs have the same
periodicity P of the arrival task As per Pfair algorithm, each job is associated
with pseudo-release time and pseudo-deadline. Pseudo-release time and
pseudo-deadline of the ith job are as follows:
r = i 1 wtT
d = i wtT 1
Where wt T = E P and i 1
where r1, r2rz are the release times and d1, d2 dz are the deadlines of the
respective jobs.
3. Next, based on the pseudo-release time (say 5 ms) of new job of task T5 in the
global queue, search on TimeCore Map table is done and accordingly for-
warded to local queue of either C1 or C3 depending upon the lesser CPU
load/utilization of C1 or C3.
4. Based on the release time and execution time of that of the new job, we estimate
whether the job can be allocated to core of C1 or C3 before starting of next
instance of the existing task at core C1 or C3, honouring the deadline constraints.
If the condition is satised, we assign the job from local queue of C1 or C3.
5. The newly assigned job to core is preempted when another job from global
queue with highest priority (earliest deadline) releases or next period of existing
task of that core starts next instance (chance of which is near to zero as we break
the new task into number of jobs to decrease the execution time of jobs which
can easily t before next instance of existing task start). Preempted newly
assigned job switched to the global queue, and the global queue is rearranged
based on the EDF scheme.
6. Then again mapping from global queue to TimeCore Map table occurs for the
next job to nd a suitable core.
7. The global queue is also rearranged when (a) new task arrives at global queue
and (b) also new instance of task at global queue starts.
8. The process continues till all the tasks get executed.
9. Each core has private rst-level (L1) cache and a second-level (L2) cache, which
is shared by all the cores. The two-level caches (L1 and L2) are available in
many existing multicore architectures [20, 42].
Algorithm Dynamic_Schedule
Input: (T, C, P, I, e, N, Gq, J, z, r, Et, cn, d, p, l)
T = Initial Task, C = Core executing task T, P = Period of task T, I = Idle time
of core, e = execution time of task T, N = No of new tasks, Gq = Global queue,
J = new job, z = no of jobs for each new task, r = release time of job J,
Et = Execution Time of job J, cn = no of core, d = Deadline of job J, p = no of
slot in global queue Gq for storing p no of task, l = no of unit of local queue for
storing l no of jobs
Output: Assign job J into the system within deadline
3 An Efcient Dynamic Scheduling of Tasks 39
Step 1: /* Assign all the initial tasks (T ) to all the cores (C ) where no of initial
tasks = no of cores */
Step 4: /* N no of new tasks are arrived and stored in global queue Gq based on
the availability of slot*/
Step 7: /* Job is assigned from global queue to local queue of selected lesser CPU
utilization */
Step 8: /* Assign jobs from local queue to core of lesser CPU utilization */
(a) Initial_assignment(T, C, cn )
for(i=0;i<p; i++){
if(Gq [i] == null)
Gq [i] == N[i] //storing new task
}
(e) time_core_release(I, p, z, r )
for(c1=0;c1< z;c1++)
for(r2=0;r2< z;r2++)
for(c2=0;c2<p;c2++)
if(I[r2][c2]==r[r1][c1]{
Display I[r2][c2] //display idle time matches with release time
Display r2+1 //display idle core corresponding to idle time
}
(f) less_cpu(l, C )
for(k=0;k<l;k++){
if(C[k]!=null)
count++ }
Display count // no of unit full in local queue
//Repeat the process for all the selected idle cores
//minimum count core selected as lesser CPU utilization
for(k=0;k<l;k++) {
if(C[k]==null && di>=current time)
C[k]=Ji // ith job is Ji and deadline of ith job di
//assigned job to lesser utilized queue of core
}
We can draw similar graphs for all the new tasks with different period range and
different execution times maintaining the w value in such a way that almost all the
jobs of new tasks are accommodated into the system as drawn in the Fig. 3.4.
Hence, from the experimental results, we can conclude that approximately all the
random new tasks with different periods and execution times are accommodated in
the system meeting the deadline criteria.
We also achieve the result of CPU utilization of the system based on the
formula:
U = C i Ti
where Ci is the computation time of the task and Ti is the period of the task. We nd
that at a xed time period, the CPU utilization of the proposed system is greater
than 95 %. The CPU utilization can be increased further towards 100 % if we take
the periodicity and execution time in such a way that release time and deadline of
all the jobs are tted into the system.
(b) We can allocate almost all the new tasks into the cores if Execution Time is
reduced for the new task and hence increasing the utilization of the system
even more
In our experimental result, 4 out of 5 jobs found the idle time at different cores
when the execution time of new task is 4 ms (keeping periodicity constant) but only
one could be allocated at 10 ms execution time because others could not be tted
within their deadline or their release time could not be tted into the system. We
draw a graph with the experimental results, illustrating the variation in number of
jobs getting accommodated with the reduction in the execution time of tasks
(Fig. 3.5).
Experimental result shows that more and more jobs get accommodated into the
system if we reduce the execution time of the new task. Using the formula, we
calculate that the CPU utilization at a xed time interval approaches towards
100 %.
From experimental results, we nd that our model achieves the maximum CPU
utilization when we combine the above two scenario, i.e. increasing the periodicity
along with decreased execution time.
44 K. Baital and A. Chakrabarti
Fig. 3.6 The sample job release time is 15 ms and execution time is 0.4 ms as the task is divided
into ve jobs (execution time of task is 2 ms and execution time of each divided job is 0.4 ms).
The job is easily tted into the core C1 as it gets completed before next instance 16 ms of C1. If
the task is divided into four jobs, the job will take 15.5 ms to complete and if it is divided into 2
jobs the job will not be tted into the core as it will take 16 ms to complete and at that time the
next instance of C1 will be started. Therefore, it can be concluded that the job will be tted easily
to the core if new task is divided into more number of jobs
(c) We can allocate almost all the new tasks into the cores if Task is divided into
a huge number of jobs thus decreasing their execution time so that the jobs
can easily be tted before next instance of the existing task.
The result achieved through simulation is shown in Fig. 3.6
3.6 Conclusion
proposed algorithm has been veried through simulation, and the results show that
our algorithm performs excellently in all the cases. The other important issues such
as load balancing among the processing cores, task migration, memory overheads
and energy parameter, resource sharing and other performance parameters will also
be considered in the different phases of this model development.
References
1. Kopetz H (1997) Real-time systems: design principles for distributed embedded applications.
Kluwer Academic, Norwell
2. Nakate MSS, Meshram BB, Chavan MJP (2012) New trends in real time operating systems
IOSR J Eng 2(4):883892
3. Rutzig MB, Madruga F, Alves MA, Cota H, Beck ACS, Maillard N, Navaux POA, Carro L
(2010) TLP and ILP exploitation through a recongurable multiprocessor system. IEEE
978-1-4244-6534-7/10
4. Luque C, Moreto M, Cazorla FJ, Gioiosa R, Buyuktosunoglu A, Valero M (2012) CPU
accounting for multicore processors. In: IEEE transaction on computers, vol 61, No 2, Feb
2012
5. Ju M, Jung H, Che H (2014) A performance analysis methodology for multicore,
multithreaded processors. In: IEEE transaction on computers, vol 63, No 2, Feb 2014
6. Singh AK, Shaque M, Kumar A, Henkel J (2013) Mapping on multi/ many-core systems:
survey of current and emerging trends. In: DAC 13 May 29 June 07 2013, Austin, TX
7. Vaidehi M, Gopalakrishnan Nair TR (2008) Multicore applications in real time systems. J Res
Ind 1(1):3035
8. Amalarethinam DIG, Mary GJJ (2011) A new DAG based dynamic task scheduling algorithm
(DYTAS) for multiprocessor systems. Int J Comput Appl (09758887) 9(8), April 2011
9. Rammig F, Ditze M, Janacik P, Heimfarth T, Kerstan T, Oberthuer S, Stahl K (2009) Basic
concept of real time operating systems. Springer Science + Business Media B.V.
10. Mohammadi A, Akl SG (2005) Technical report no. 2005499 scheduling algorithms for
real-time systems. School of computing, Queens University, Kingston, Ontario, 15 July 2005
11. Lehoczky J, Sha L, Ding Y (1989) The rate monotonic scheduling algorithm: exact
characterization and average case behavior. IEEE. CH2803-5/89/0000/0166
12. Buttazzo GC (2005) Rate monotonic versus EDF: judgment day. Real time systems, vol 29.
Springer Science + Business Media, Inc. Manufactured in The Netherlands, pp 526
13. Kato S, Takeda A, Yamasaki N (2008) Global rate-monotonic scheduling with priority
promotion. IPSJ Trans Adv Comput Syst (ACS) 2(1):6474
14. Zhang J, Fang X, Qi L (2014) LCM cycle based optimal scheduling in robotic cell with
parallel workstations In: 2014 IEEE international conference on robotics and automation
(ICRA), 31 May 20147 June 2014
15. Wang G, Gong W, Kastner R (2008) Operation scheduling: algorithms and applications In:
Coussy P, Morawiec A (eds) High-Level synthesis, Springer Science + Business Media B.V.
16. Mathiyalagan P, Dhepthie UR, Sivanandam SN (2010) Enhanced hybrid PSOACO
algorithm for gri scheduling. ICTACT J Soft Comput (01), July 2010
17. Bertogna M, Baruah S (2009) Limited preemption EDF scheduling of sporadic task systems.
In: IEEE transactions on industrial informatics, vol. X, No X, Nov 2009
18. Herman JL, Kenna CJ, Mollison MS, Anderson JH, Johnson DM (2012) RTOS support for
multicore mixed-criticality systems. In: RTAS 12 Proceedings of the 2012 IEEE 18th real
time and embedded technology and applications symposium
46 K. Baital and A. Chakrabarti
19. Mollison MS, Anderson JH (2013) Bringing theory into practice: a user space library for
multicore real-time scheduling. In: IEEE real-time and embedded technology and applications
symposium 2013, pp 283292
20. Li Y, Suhendra V, Liang Y, Mitra T, Roychoudhury A (2009) Timing analysis of concurrent
programs running on shared cache multi-cores. In: 2009 30th IEEE Real-time systems
symposium
21. Pacher M, Brinkschulte U (2011) Ordering of time-dependent tasks for self-organized
resource scheduling. In: 2011 14th IEEE international symposium on
object/component/service-oriented real-time distributed computing workshops
22. Holenderski M, Bril RJ, Lukkien JJ (2012) Parallel-task scheduling on multiple resources. In:
2012 24th IEEE euromicro conference on real-time systems
23. Paolieri M, Quinones E, Cazorla FJ, Davis RI, Valero M (2011) IA3: an interference aware
allocation algorithm for multicore hard real-time systems. In: 2011 17th IEEE real-time and
embedded technology and applications symposium
24. Holman P, Anderson JH Implementing pfairness on a symmetric multiprocessor. In:
Proceedings of the 10th IEEE real time and embedded technology and applications
symposium (RTAS 04)
25. Park S (2014) Task-I/O co-scheduling for pfair real-time scheduler in embedded multicore
systems. In: 2014 IEEE international conference on embedded and ubiquitous computing
26. Jayaseelan R, Mitra T (2009) Temperature aware scheduling for embedded processors. In:
2009 IEEE 22nd international conference on VLSI design
27. Chen L, Boichat N, Mitra T (2011). Customized MPSoC synthesis for task sequence. In: 2011
IEEE 9th symposium on application specic processors (SASP)
28. Chen L, Marconi T, Mitra T (2012) Online scheduling for multi-core shared recongurable
fabric 978-3-9810801-8-6/DATE12 2012 EDAA
29. Ding H, Liang Y, Mitra T (2012) WCET-centric partial instruction cache locking. In: DAC
2012, Jun 37 San Francisco, California, USA
30. Pricopi M, Mitra T (2014) Task scheduling on adaptive multi-core. IEEE Trans Comput 63
(10):25902603
31. De Giusti L, Chichizola F, Naiouf M, De Giusti A, Luque E (2010) Automatic mapping tasks
to coresevaluating AMTHA algorithm in multicore architectures. IJCSI Int J Comput Sci
Issues 7(2)(1), Mar 2010
32. Aghazarian V, Ghorbannia A, Motlagh NG, Naeini MK (2011) RQSG-I: an optimized real
time scheduling algorithm for tasks allocation in grid environments. IEEE
978-1-61284-486-2/11
33. Bamakhrama M, Stefanov T (2011) Hard-real-time scheduling of data-dependent tasks in
embedded streaming applications. In: EMSOFT11, Taipei, Taiwan, 914 Oct 2011
34. Kim J, Shin T, Ha S, Oh H (2011) Resource minimized static mapping and dynamic
scheduling of SDF Graphs. IEEE 978-1-4577-2122-9/11
35. Wu G, Li Y, Ren J, Lin C (2013) Partitioned xed-priority real-time scheduling based on
dependent task-split on multicore platform. In: 2013 12th IEEE international conference on
trust, security and privacy in computing and communications
36. Qu WX, Fan XY, Liu Y, Yang H, Chen L (2010) Memory system prefetching for multi-core
and multi-threading architecture. In: 2010 3rd international conference on advanced computer
theory and engineering (ICACTE). IEEE
37. Peternier A, Ansaloni D, Bonetta D, Pautasso C, Binder W (2012) Hardware-aware thread
scheduling: the case of asymmetric multicore processors. In: 2012 IEEE 18th international
conference on parallel and distributed systems
38. Anane A, Aboulhamid EM, Savaria Y (2012) System modeling and multicore simulation
using transactions. IEEE 978-1-4673-2297-3/12
39. Pericas M, Cristal A, Cazorla FJ, Gonzalez R, Veidenbaum A, Jimenez DA, Valero M (2008)
A two-level load/store queue based on execution locality In: International symposium on
computer architecture IEEE
3 An Efcient Dynamic Scheduling of Tasks 47
40. Jayaseelan R, Mitra T (2009) A hybrid local-global approach for multi-core thermal
management. In: ICCAD 09, 25 Nov 2009, San Jose, California, USA Copyright 2009 ACM
978-1-60558-800-1/09/11
41. Lehoczky J, Sha L, Ding Y (1989) The rate monotonic scheduling algorithm: exact
characterization and average case behavior. In: 1989 IEEE real-time system symposium,
pp 166171
42. Suhendra V, Mitra T (2008) Exploring locking & partitioning for predictable shared caches on
multi-cores. In: DAC 2008, 813 June 2008, Anaheim, California, USA Copyright 2008
ACM 978-1-60558-115-6/08/0006
43. Levin G, Funk S, Sadowski C, Pye I, Brandt S (2010) DP-Fair: a simple model for
understanding optimal multiprocessor scheduling In: 2010 IEEE 22nd euromicro conference
on real-time systems
44. Anderson JH, Srinivasan A (2000) Pfair scheduling: beyond periodic task systems. IEEE
1530-1427/00
Chapter 4
Model-Based Approach for Shadow
Detection of Static Images
Abstract The existing computer vision systems accept computer-based images for
processing. These images can be processed differently depending upon the types of
applications. One such application that maps to the domain of our work is tracking
in which the image has a number of objects of which only the object of interest is
mapped for tracking. Due to the light sources and existence of surrounding light
conditions, the object of interest, if opaque, creates a hard cast shadow on the
background while if the object is semi-opaque, it creates soft cast shadows on the
background. These shadows resemble very closely to the shape of the object. Thus,
there are chances that they may be mistaken for the object leading to the problem of
false tracking. To avoid this problem, the shadows have to be detected in the
computer vision system before they enter the processing unit. To add to this, the
computer vision systems cannot differentiate between an object and its shadow and
thus passes both of them. So it was required as a necessary and important stage to
incorporate the shadow detection and elimination stage between the computer
vision and processing stage. This stage which is a complete system on its own has
its own input stage, processing, output and evaluation stage. The scope of the work
presented in the paper was for two methods, i.e. colour based and texture based
along with justied evaluation of the results obtained by them.
M.K. Sabnis ()
CS&IT Department, JVWU, Jharna, Jaipur, Rajasthan, India
e-mail: manojsab67@yahoo.co.in
M.K. Shukla
Amity School of Engineering, Amity University Noida, Noida, India
e-mail: mkshukla001@gmail.com
4.1 Introduction
As computer vision systems are not capable of separating the shadows form the
images, these shadows in general have to be separated from the objects before
entering the processing unit. This is done through the shadow detection and
elimination stage [14].
Shadow detection is a compulsory stage but shadow elimination is optional
which depends on image type and its use.
Figures 4.1: Image 1 and 4.2: Image 2 represent objects along with their shadows
having similar features in terms of shape and colour. In case of tracking, this may
confuse the tracking system to select the shadow instead of the object. This is called as
false tracking. This can also lead to miscounting of objects. In such type of applications
along with shadow detection, its elimination is also important [5, 6, 7].
For Figs. 4.3: Image 3 and 4.4: Image 4, the shadows can be used as an
advantage after their detection. For counting applications, the shadows of Fig. 4.3
can be used as the objects are not clearly detectable. For applications like object
identication, the shadow of Fig. 4.4 can be used as the object is not visible.
The shadow detection and elimination system has four stages. Stage one, the
input stage gives an idea of the type of images required for colour- and
texture-based methods. It is a collection of these types of images.
The processing stage rst evaluates the existing techniques in colour-based and
texture-based methods and further suggests how these techniques can be used to
develop an improved version of the algorithm.
The output stage gives the output image which detects the shadow. The per-
centage of the detected shadows by a particular algorithm is then evaluated by using
qualitative and quantitative parameters in the evaluation stage.
The research work done in the domain of image processing for shadow detection
using these two methods is presented as the remaining part of the paper inform of
detail explanation of each stage.
4.1.1 Shadows
As seen in Fig. 4.5 named shadow formation, the light coming from a single light
source reaches the background partially due to occlusion of light by the object. This
dark region on the background is called as the cast shadow. It has two regions, the
central dark region without any light from the light source called as umbra region and
the region between umbra and the outer boundary where there is a soft transition from
dark to a bright light due to some ambient diffused light or by the effect of other light
source is called as the penumbra region. In other words, umbra corresponds to back-
ground area where the direct light is totally blocked by the foreground object and the
penumbra region where the light is partially blocked [811].
There are large numbers of shadow types, depending upon their type of formation,
the properties they exhibit or the view point of their usage. But for shadow
detection as the working domain, shadow types have been classied to a limited
number of levels which are represented in a hierarchical form [9].
52 M.K. Sabnis and M.K. Shukla
The shadow type classication follows two levels of hierarchy: the rst level
classies the image as either moving or stationary. This is referred as image type or
coarse-level classication [12].
The next level, ne tunes the selected object shadows into a number of different
shadow classes based on their prominent properties observed, i.e. sharp and com-
plete, partial or attached. This type of classication is called as property-based
shadow classication [12].
In case of videos, the objects and their shadows are on the move. Such types of
shadows are called as dynamic shadows. In case of images, the object and their
shadow are stationary. Such types of shadows are called as static shadows [8].
This work mainly focuses on static images because lot of research is already
done on dynamic images in the domain of object tracking and detection [7, 8].
The object, i.e. the occluder, comes in the direction of light and casts its shadow on the
background. Depending on the properties such as position, intensity and edges which
the shadow exhibits, the shadows are classied as cast, self or attached [13, 14].
Occlusion of the light coming from the light source, due to the foreground
object, causes cast shadows on the back ground. This light source can have either
entire or partial occlusion of light. This then reduces the background region inci-
dent. Therefore, shadow points have lower luminance but similar chromaticity [15,
16].
4 Model-Based Approach for Shadow Detection of Static Images 53
The cast shadows, as show in Fig. 4.6, are divided into two regions, the umbra
and the penumbra. The complete blockage of direct incident light is referred as
umbra region and the one with partial blockage is referred as the penumbra region
[9, 10, 17].
Umbra being the darkest part of the shadow is easily seen and detected due to its
distinct edges. These edges, however, if not detected, leads to the detection of
shadows as the object instead the object of interest itself [18]. It is also referred as
projected shadows, since the cast shadow is the projection of the object on the
surface background [19].
When light falls on the object and if the object is not of a regular shape, then the
light does not reach some parts of the object due to occlusion of light by other parts
of the object. This creates a dark region on the object itself which is called as
self-shadow [12]. This is shown in Fig. 4.7.
Self shadows do not create a problem because the objects outer boundaries are
clearly detectable, which is the basic requirement of shadow detection [14, 17].
Compared to cast shadows, they have a higher brightness level as they receive more
secondary lighting from surrounding illumination objects [14]. Figure 4.8 repre-
sents attached shadow conditions [11, 17].
There are two views in case of attached shadows, one in which the shadows are
attached to the object and second is the one in which the shadows are attached to
another shadows. In the rst case, it does matter where the shadow is attached to the
object. If it is attached only at one place and at the bottom side, then such type of
images can be used to separate the shadows and its objects by geometrical methods.
Now for case number two, in a scene with multiple objects casting multiple
shadows, depending upon the object nearness to each other, or due to multiple
un-oriented light sources, shadow of one object may fall on shadow of other object.
Such types of shadows are called as attached shadows.
There are many approaches proposed for shadow detection and its elimination. This
is so because these approaches depend upon shadow models which vary depending
on the conditions applied, i.e. from an elementary model with assumptions to a
more practical one with realistic environmental conditions and minimal assump-
tions [10].
Also the shadow model for one environmental condition becomes nonapplicable
when the environmental conditions change making the algorithm unstable.
In 2003, Prati et al. reviewed all the shadow detection approaches and proposed
a standard shadow detection method having limited algorithms. These algorithms,
based on two-level taxonomy, were known as algorithm-based shadow detection
taxonomy [10].
The rst or the primary level considers the decision-making ability of the
algorithm to be certain or uncertain, thus dening two approaches: the statistical
and deterministic [10]
The second layer further classied the statistical methods as parametric and
nonparametric based while the deterministic methods were classied as
model-based approach or nonmodel-based approach. This was called as the Primary
Level Classication [10].
This model nalized the area of working for actual algorithmic development.
However, for this, the information required is image type and prominent areas to be
targeted within the image, respectively.
Another classication was also specied by Parti et al. called as the secondary
level classication which was also hierarchical but having only a single level. This
classication basically was feature based. These methods used features for
4 Model-Based Approach for Shadow Detection of Static Images 55
algorithm development and had been broadly classied into three working classes:
spectral, spatial and temporal [19, 20].
Thus, the algorithms based on properties such as intensity and chromaticity
mapped into the spectral domain. Geometry- and texture-based features mapped
into the spatial domain and its working areas were conned to frame based, region
based in the selected frame and pixel based in the selected region. The redundant
information was reflected in the temporal domain which could be used for further
improvement [4].
(1) Cast shadows cause loss of information of the surface under the shadow, as
this surface can be another shadow, object itself or another object or back-
ground. This presents difculties in image interpretation, image matching,
detection and others applications [8, 21].
(2) Attached shadows cause object merging, object shape distortion and object
loss.
(3) Further shadows usually degrade the visual quality of the image leading to
serious problems in segmentation and object extraction [13, 22].
Large numbers of different types of images are available. These images are rich in
feature such as intensity, colour, texture geometry and orientation. The algorithm
are developed based on the features that are prominent in an image so as to give
maximum output results. To limit the number of algorithms, four standard classes
of methods are developed for four prominent feature observable in an image [19].
This leads to the formation of four standard image classes which are method
based. A standard demarcation is not present between these image classes as some
images can belong to more than one class also. The image selection for a class also
56 M.K. Sabnis and M.K. Shukla
This is the main stage which represents the bulk of the work done in the domain of
shadow detection for colour- and texture-based methods.
As shown in Fig. 4.9, the processing module has two stages. The rst is the
preProcessing stage which is common for both these methods, for image enhancement
and ltering. The second stage is the processing stage which is divided into number of
blocks where each block corresponds to the model they represent [5].
The basic shadow model is acted upon by two types of light sources: the direct light
and ambient light. Direct light is from a point source and ambient light is from
environmental light, from the reflections of surrounding surfaces. This source is
regarded as the area light. For the formation of shadow area, the direct light is
occluded totally or partially [7, 1113].
The shadow model is represented as follows:
If ti is 0 < ti < 1, then the pixel points are in penumbra region. where
For more realistic shadow model, the image is captured by the camera and the
shadow model is represented in a more realistic way as follows: [8]
Fx y = ix y rx y 4:6
where
F (x y) Intensity of the pixel (x y)
r (x y) Reflectance component of object surface
i (x y) Illumination component of object surface
i (x y) is computed as the amount of light power received by the object, per
surface area and it is further expressed as follows:
i x y = Ca + Cp.cos( 4:7
58 M.K. Sabnis and M.K. Shukla
Equation 4.6 is for illuminated area, where Ca: intensity of ambient light, Cp:
intensity of light source and : Angle enclosed by light source direction and surface
normal.
ix y = Ca + tx y.cp.cos 4:8
ix y = Ca 4:9
The four models represent four types of image subclasses which can be evaluated
under colour-based methods. These models are brightness model, chromaticity
model, brightness chromaticity model and nonlinearity model.
Brightness Model: It is also called as illumination model. It works for greyscale
images, which are indoor and having uniform light source. This model represents
intensity invariant colour space where only the intensity changes, whereas the
colour information remains unchanged [14].
Chromaticity Model: It works on colour images (RGB) which are indoor, and
only intensity change is observed. It is not suitable for complex scenes [14].
Brightness Chromaticity Model: In case of outdoor scenes, both intensity and
colour information change. RGB images are not able to follow these changes in a
linear manner. HSV, HSI and C1C2C3 colour models are used [14]. This model
works under a number of assumptions such as shadow is casted on flat nontextured
surface, objects are uniformly coloured, only one light source illuminates the scene,
light source is strong and well illuminated, static and accurate camera, and no
sudden changes in the scene. Environmental changes and noise level are well with
the accepted limits [14].
Brightness Chromaticity Distortion Model: The image class is of outdoor
complicated scenes where no assumptions are applicable to give a practical con-
ditions scenario. To model such a scene, lots of correction have to be applied which
makes the modelling complicated.
4 Model-Based Approach for Shadow Detection of Static Images 59
4.3.1.2 Algorithms
The existing algorithm used thresholding as initial approach for static shadow
detection [13]. Then, the method of automatic thresholding was used for dynamic
images [11, 16, 18, 21]. The limitation of these methods was they assumed equal
distribution of intensity levels within the image which was not practical.
O Gorman proposed a method which was a technique for thresholding based on
image connectivity. The image was threshold at multiple intensities levels
depending upon the calculated connected values. From this stable set, connectivity
values were selected to give multiple intensities with this selected range [15, 17].
An improvement suggested over connectivity intensities was regions. The
region-based method had the advantage that the Euler number could be locally
countable and could be determined in a single raster scan of the image [15, 19]
The method which leads to the idea of multiple intensity-based thresholding was
basically edge-based using canny edge detection. The edges obtained from the
image were put in three predened classes: H (higher level threshold), L (lower
level threshold) and M (medium level threshold) [15, 20].
The edge-based method is further modied as region based. By thresholding
method in which two levels of threshold values are dened as H and L, the region
above H is retained as objects and the region below L or connected to L is rejected
as shadows [15].
The hysteresis methodology suggested is further used to combine local and
global information where the global thresholding is image-wise and local thresh-
olding is region-wise [23]. This is further used as the basis of the proposed
algorithm.
Two algorithms are proposed: one with single intensity for greyscale images and
another for colour images using the thresholding concept.
Intensity-Based Algorithm
(1) The greyscale image is read with image dimensions. The window size is
dened as per the output window requirements.
(2) The average intensity of the image is calculated to get a single value of the
threshold called as the global threshold.
By using the image dimensions, all pixels are added and its average is
obtained. This average, if taken as global threshold, will be lower than lighted
objects but the darker objects will fall below it and will not be detected. Thus,
to separate the darker objects from shadows, sixty per cent of average is taken
and added to average to get the global threshold.
(3) Binary image of the input image is now obtained.
Logic 1 is entered for those pixels whose intensity is greater than the reference
global threshold value, and logic 0 is entered for those pixels whose intensity
value is less than the threshold value.
60 M.K. Sabnis and M.K. Shukla
(4) Morphological operations such as closing, thickening and dilation are per-
formed [24, 25].
Closing operation is done for noise removal. It expands the boundaries and
closes on the background. The thickening is performed for selective growing
of foreground. Then, foreground boundaries are expanded by dilation.
(5) The binary image so formed has logic 1 for objects and logic 0 for shadows
and dark objects. To work with shadows, invert the image, so logic 1 now
represents shadows and dark objects and logic 0 represents the objects.
(6) Region labelling is performed
The image has object and their shadows. With the shadow size so selected,
regions are formed by four connected methods and labelled. Binary labelled
image with the number of labels is returned.
(7) Eliminate smaller regions
For the binary labelled image, nd threshold two. This threshold value is
image size divided by shadow size. Find labelled areas smaller than these
thresholds. These small regions are made logic 0 and thus put in background.
(8) Mask formation
The binary image has only larger regions. The mask so formed has logic 1 for
shadows and logic 0 for objects.
(9) Shadow region selection
This binary mask is multiplied element-wise with the original image. For
better resolution, the variable taken is double float. With elimination of image
contents under mask with logic 1, a shadow-free image is presented.
(10) Image Representation
For shadow representation, during mask comparison, shadow regions are lled
with green colour and objects darker than threshold are represented with red
colour.
The input image of Fig. 4.10 named as input being greyscale has a similar colour
range for object and its shadow. So the shadow in the output image is presented in
different colour as shown in Fig. 4.11 named as output.
Colour-Based Algorithm
The algorithm which is used as reference for coming up with proposed algorithm
has the basic RGB colour model transformed into invariant colour models such as
YCbCr, HCV, YIQ, HSI and HSV. Where Y is the luma, Cb is blue difference
chroma, Cr is red difference chroma, H is the hue content, C is the chroma and V is
the intensity value.
HSI model is selected due to higher H in shadow region and less blue colour
value with smaller difference between green and blue colour values.
The algorithm proposed depends on ratio map technique using thresholding
techniques.
The proposed algorithm is as follows:
(1) Selected input image
RGB is converted into HSV as RGB gives output under limited environmental
conditions, whereas HSV has a better adaptation capacity for environmental
variations.
(2) Channel separation
From the converted image, separate the hue, saturation and intensity value
channels
(3) Construct the ratio map for all pixels as follows:
(5) The ratio map is then converted to modied ratio map by applying exponential
function of ratio map. This stretches the gap between the ratio value of shadow
and nonshadow pixels.
He(x, y) is modied to He(x, y) + 1, and Ie(x, y) is sealed in the range of [0, 1].
This works satisfactory in small and medium range but gives unsatisfactory
results of shadow detection in large range.
(6) Then, calculate the global threshold value T for the images by using
Otsu thresholding method [23, 26].
where
Thus, it can be concluded that the output mainly depends on the type of input
image and the selection of the threshold. Though algorithm for greyscale is simple,
it has lot of limitations in its output representation.
Texture is a very general idea that can attribute to almost everything in nature. For a
human, texture analysis initially was dependent on look and feel concept where
texture related mostly to a specic spatially repetitive structure of surfaces points
formed by repeating a particular element or several elements in different relative
spatial positions [9].
4 Model-Based Approach for Shadow Detection of Static Images 65
The differences seen and felt by the human are difcult to be dened in a
quantitative manner. This leads to the dening of texture in the form of features
which can be measurable [9].
This leads to the denition of informal qualitative structural features such as
neness, coarseness, smoothness, granularity, lineation, directionality, roughness,
regularity and randomness [9].
Further, it is difcult to use human classication as a basis for formal denition
of image textures, because there is no obvious way of associating these features
which can be easily perceived by human vision with computational models that has
the goal to describe the texture [9].
After several decades of research and development on texture analysis and
synthesis, a variety of computational characteristics and property have been dened
for images [9].
Image texture can be dened as natural textured surface and articially created
visual patterns. For an image, the texture is considered as a texture pattern which
covers the entire image or a texture pattern with the region of the image which
repeats itself in a denite sequence covering the entire image [9].
The texture pixels have some denite property variations described as local
arrangement of the image signal in the spatial domain or the domain of Fourier or
other spectral transform. This is further used to model the texture image into
computational model.
The pixels have a relationship with their immediate neighbourhood pixels which
help them to dene a texture region with boundary pixels. This pixel set of region
as a whole has relationship with pixel sets of other region to identify identical
repeated patterns.
For modelling texture patterns in an image, at pixel level that texture-related
property of the pixel is selected which is more prominent.
There are number of such properties but CBIR standard has dened and accepted
six texture-based properties such as coarseness, contrast, degree of directionality,
line likeness, regularity, roughness and Markov random. These features are called
as Tamuras texture features [2729].
4.3.2.1 Algorithm
The existing algorithms examined on which the proposed algorithm is based are as
follows:
K Emily and Esther Rani suggested a solution for dynamic images by initially
using change detection mask and then canny edge map to separate shadow and
shaded areas [7].
A Leone and C Distante proposed a texture-based solution by feature extraction
method where texture-wise characterization is done of neighbourhood pixels by
projecting them on a set of Gabor functions [12].
The proposed algorithm is as follows:
66 M.K. Sabnis and M.K. Shukla
(1) Read the input image and convert the image into greyscale. The image usually
selected is HSV
(2) Find the entropy of the image. In an image, many pixel changes take place.
Entropy is the measure of this uncertainty of random variables, where Entropy
k is,
k = X Y Z 4:14
where
X = Log A + Log B.
Y = Log C log A.
Z = Log D Log B.
(3) If calculated entropy value is less than zero, then entropy(k) = 0, For K = 1 to
K = 256, calculate Entropy (k) and select the greatest value of entropy (k) as
the entropy of the image and then perform histogram equalization on the
image to increase its contrasts.
(4) Select a mask size m, compatible with the image and take the position of the
two least values in entropy and set those positions as 1 in the mask and set the
remaining mask position as 0. Shadow pixels are those pixels with minimum
variation, thus less entropy, i.e. 1 in the mask
(5) Replace the pixels of the original image wherever the mask is 1 so as to
remove shadow pixels
Figure 4.18 is the input image. This cannot be used for colour as the colour of
object and background is almost same. Similarly, it cannot be used for
geometry-based method as the object and its shadow are not connected to have the
angle of orientation. Thus, we restore for texture-based method. All shadows are
obtained depending upon the texture algorithm though our interest is only in the
shadow of the dog. This is seen in Fig. 4.19.
As shown in Fig. 4.20, the object has illumination almost similar to the back-
ground, and even though the object and its shadow are connected but still not
having a single orientation angle, we have to restore to texture-based techniques.
This output is shown in Fig. 4.21.
As the shadow is more prominent in the input image (Fig. 4.22), its advantage is
taken. The basic principal of texture analysis is used in this. Texture of the
4 Model-Based Approach for Shadow Detection of Static Images 67
background does not change due to shadow falling on the background, so back-
ground with object gives texture change and background with shadow no texture
change, so this is detected as shadow region as represented in Fig. 4.23.
The visual representation of the output of the image is available. This output image
is then evaluated on qualitative and quantitative bases for its justication.
Both the shadow methods give results under different conditions and types of
images. Thus, they cannot be compared on a direct basis. The types of errors that
are found common in all these methods are shadow detection failure (SDF) and
object detection failure (ODF).
At the shadow edges, the pixel information closely maps to the object pixels;
therefore, these pixels are misclassied as shadow pixels. This is called as shadow
detection failure. Similarly, in case of some dark regions of the objects, the pixels
are detected as shadows. This is called as object detection failure.
Now to measure and minimize these failures, some evaluation techniques are
suggested at the qualitative and quantitative levels.
4 Model-Based Approach for Shadow Detection of Static Images 69
The metrics dened for evaluation under qualitative are as follows: [30] robustness to
noise, object independence scene independence, computational load, flexibility to
shadow strength, width and shape, detection of indirect cast shadows and penumbra.
Table 4.1 explains what should be the state of these metrics for faithful shadow
detection.
DR = TP TP + FN 4:15
FAR = FP TP + FP 4:16
where FN: false negative where shadow points are classied as background or
foreground. TP: true positive where shadow points are classied as shadows, FP:
false positive where foreground or background points are detected as shadow
points.
Table 4.2 represents the detection rate and false alarm rate of different images
taken.
For greyscale with similar colour object and background, the detection rate is not
that high and more false pixels are detected as shadow pixels (Figure 4.11). Due to
large difference in colour of shadow and image, the detection rate is high
(Figs. 4.13, 4.15 and 4.17).
Table 4.3 represents the detection rate and false alarm rate for different images
tested under this method.
Table 4.2 Evaluation of Output Image Detection rate False alarm rate
colour-based method
0.751 0.14826
Figure 4.11
0.58033 0.0595
Figure 4.13
0.8696 0.18142
Figure 4.15
0.94698 0.91152
Figure 4.17
4 Model-Based Approach for Shadow Detection of Static Images 71
Table 4.3 Evaluation of Output image Detection rate False alarm rate
texture-based method
0.95 0.2472
Figure 4.19
1 0.2864
Figure 4.21
1 0.14
Figure 4.23
4.6 Conclusion
It can be concluded that no single method can be used for getting accurate results.
Hierarchical methods are recommended so that the accuracy can be improved. The
selection of the method and their sequence of implementation again depend on
individual type of image.
Intensity method can be used for greyscale or colour (RGB) for simple images
with uniform and single light source.
In case of multiple objects even with single light source, it gives multiple
reflections. The colour model gives both intensity and colour variations. In this
case, thresholding is preferred as threshold value selection totally depends on the
algorithm developer within the domain of image types.
When no conditions met in an image texture method, it is the only solution
available.
As this eld is very vast, the scope of this work is only limited to two methods. The
scope can be extended beyond these two methods to include intensity and
geometry-based methods so as to cover images of all types.
72 M.K. Sabnis and M.K. Shukla
In case of thresholding, maximum two thresholds with xed values are rec-
ommended. The value of these two thresholds can be made adaptive according to
the image requirements within a specic range.
In case of textures, the direct analysis of images using CBIR texture properties
makes the computation very complicated and very limited image processing
methods are available related to textures.
References
1. Chen T-M, Wang W-X (2006) Image shadow detection and classication based on rough sets.
In: Proceedings of fth international conference on machine learning and cybernetics, Dalian,
1-4244-0060-0/60/, IEEE, pp 38913896, 1316 Aug 2006
2. Blajovici C, Kiss PJ, Bonus Z, varga L (2011) Shadow detection and removal from a single
image. In: 19th summer school on Image Processing, Szeged, Hungary, pp 16, 7th-16th July
3. Gudmundsson K, Benediktsson JA, Cagatin FS (2008) Shadow extraction, Seventh
Euro-American workshop on information optics. J phys Conf ser 139(012032):17. IOP
publication ltd
4. Lakshmi S, Sankaranarayanan V (2010) Cast shadow detection and removal in a realtime
environment. IEEE, pp 245247. ISDN: 978-1-4244-9008-0/10/
5. Xu L, Qi F, Jiang R, Hao Y, Wu G Shadow detection and removal in real images: a survey,
CVLAB, Shaighai Jiao Tong University P.R China
6. Sanin A, Sanderson C, Lovell BC (2012) Shadow detection: a survey and comparative
evaluation of recent methods. Pattern Recognition, Elsevier, vol 45, No 4. pp 16841695.
ISSN: 0031-3203
7. Withagen PJ, Groen FCA, Schutte K (2007) Shadow detection using physical basis,
intelligent autonomous system Technical report, pp 114
8. Horprasert T, Harwood D, Davis L (2008) A statistical approach for real time robust
background subtraction and shadow detection, computer vision lab. University of Maryland,
pp 119
9. Madsen CB, Moeslund TB, Pal A, Balasubramanian S (2009) Shadow detection in dynamic
scenes using dense stereo information and an outdoor illumination model, computer vision
and media technology lab, Aalborg university Denmark, pp 110125
10. Rani EE, Jemilda G (2011) Shadow detection in single still image using tam based multi-step
algorithm, IJCST, vol 2, Issue 4. pp 496500, OctDec 2011. ISSN(Online): 0976-8491 ISSN
(Print): 2229-4333
11. Moving cast shadow detection, vision systems: segmentation and pattern recognition, pp 47
58 (2011)
12. Onoguchi K (1998) Shadow elimination method for moving object detection. Proce Fourteen
Int Conf Pattern Recogn 1:583587
13. https://www.cs.auckland.ac.nz/courses/compsci708s1c/lectures/Glect-html/topic4c708FSC.
htm
14. Howorth P, Ruger S Evaluation of texture feature for content based image retrieval.
Department of Computing, Imperial College, London, South Kensington Campus, London,
SW7 2AZ
15. Guo R, Dai Q, Hoiem D Single image shadow detection and removal using paired regions,
supported in part by the National Science Foundation under IIS-0904209, pp 20332038
16. Hierarchical static shadow Detection methods, US 7, 970, 168 B1
17. Rosin PL, Ellis T Image difference threshold strategies and shadow detection, Institute of
remote sensing applications, Joint Research Center, Italy
4 Model-Based Approach for Shadow Detection of Static Images 73
18. Jyothirmai MSV, Srinivas K, Srinivasa Rao V (2012) Enhancing shadow area using RGB
colour space. IOSR J Comput Eng 2(1):2428, July-Aug 2012. ISSN: 2278-0661
19. Bichsel M (1994) Segmenting simple connected moving objects in a static scene. IEEE
Trans PAMI 16:11381142
20. Gorman LO (1994) Binarization and multi thresholding of document images using
connectivity, Symposium on document analysis and Information retrieval, pp 237252
21. Leone A, Distante C (2006) Shadow detection for moving objects based on texture analysis.
J Pattern Recogn, pattern recognition, Elsevier 40(2007):12221233, Sept 2006. ISSN:
0031-3203
22. Otsu N (1979) A threshold selection method from grey level histogram. IEEE Trans Syst Man
Cybern SMC-9(1):6269, Jan 1979
23. Grey SB (1971) Local properties of binary images in two dimensions. IEEE Trans Comput
20:551561
24. Morphological image processing. https://www.cs.auckland.ac.nz/courses/compsci773s1c/
lectures/ImageProcessing-html/topic4.htm
25. Morphology fundamentals: dilation and erosion. http://in.mathworks.com/help/images/
morphology-fundamentals-dilation-and-erosion.html
26. Canny J (1986) A computational approach to edge detection. IEEE Trans PAMI 8:679698
27. Tuceryan M, Jain AK (1998) Texture analysis, the handbook of pattern recognition and
computer vision(2nd edn). World scientic Publication Co, pp 207248
28. Srinivasan GN, Shobha G (2008) Statistical texture analysis. In: Proceeding of world academy
of science, engineering and technlolgy, vol 36. pp 12641269, Dec 2008. ISSN: 2070-3747
29. Materka A, Strzelecki M (1998) Texture analysis methods-A review, Technical university of
Lodz, institute of electronics, COST B11 report, Brussels
30. wang SKE, Qin BO, Fan Z-H, Ma Z-S (2007) Fast shadow detection according to the moving
region. In: Proceedings of 6th international conference on machine learning and cybernetics,
Hong-Kong, IEEE 1-4244-0973-X/07, pp 15901595
31. Chung KL, Lin YR, Huang YH (2009) Efcient shadow detection of colour ariel image based
on successive thresholding scheme, transaction on Geoscience and remote sensing,
0196-2892, vol 42, No2. IEEE, pp 671682
Chapter 5
Light Fidelity (Li-Fi): In Mobile
Communication and Ubiquitous
Computing Applications
Abstract As the increasing nature of big data during the last few years, the
problem of data trafc arises and also the pervasive nature of smart phones the
number of end user smart phones was overtaken the total entire population of
planet. On the other hand, the emerging technology of Li-Fi has found to be the best
solution on these problems. This works on the concept of visible light communi-
cation (VLC) and also offering many solutions to reduce the cellular infrastructure
trafc and its needs. Not only the mobile communication but also the ubiquitous
computing adding the more overloads to the RF spectrum. In this paper, we have
covered the concepts of Li-Fi, how the Li-Fi technology can be enhanced in the
mobile communication, how it works, the Li-Fi cellular network, some ubiquitous
computing applications, common misconceptions about Li-Fi, Li-Fi in solar cell
and Internet of Things (IoT).
5.1 Introduction
According to the statistics of ITU, the use of mobile and Internet will reached per
month around 24 Exabyte of data which was more than the 30 times the size of the
entire global Internet in 2000 [1]. Also the different computing devices like laptops,
tablets and sensors are adding more load to the existing infrastructure making it
Li-Fi works on the principle of visible light communication which provides a much
wider license-free visible light spectrum to the service provider. The continuous
data streams are generated by flickering LED at transmitter side on higher rate. This
flickering is imperceptible to human eye. The PCB controls the electrical inputs and
outputs of the lamp and houses the microcontroller used to manage different lamp
functions [6].
The voltage regulator and level shifter circuits are used on both the side to regulate
the flickering [7]. Finally at receiver side photo detectors are used to sense the light
and to convert it into the respected pulses. These pulses then amplied and processed
to achieve the original data stream. The merging of illumination with wireless
communication provides a measurable reduction in both infrastructure complexity
and energy consumption [8]. As lighting used in indoor environments even during
day time, the energy used for communication would practically be zero. Fig. 5.1
It is estimated that in 2014, LEDs will secure 24 % of the lighting market. The
invention of blue LED had changed the view of the light sources. In Li-Fi LED is
used as a data transmitter that generates thousands of data streams as compare to
IR LED [9]. The data are transmitted from RS-232 cable to the power line and then
to the LED [10]. This transmission is not limited to line-of-sight direction. We can
use a single or an array of LED to transmit the data. The potential of these LEDs
can be controlled by using microcontroller and can be adjusted by some Luminaries
5 Light Fidelity (Li-Fi): In Mobile Communication 77
Design Optimization technique. We can use different colored LEDs for different
intensity channels. The experiments with a various dimmed intensities are observed
at Li-Fi R&D centre [9] showing the satisfactory reports for energy saving. The
innovations like nobel prize winner Blue LED add more advantages for Li-Fi by
increasing LED lifetime to 100,000 h and cheaper in cost [11].
The Avalanche photodiodes have found to be the better receiver in Li-Fi com-
munication. They read the flickering patterns of LED and interpret them as data.
Prof. Harald Haas has shown the rst receiver chip for Li-Fi with integrated
Avalanche photodiodes on CMOS [9] which is a 7.8 mm2 IC that houses 49
photodiodes [12]. When compared to Wi-Fi antenna it is just a 13 % of its length.
To make a full duplex communication, the mobile terminals uses infrared
(IR) transmission to talk with nearest optical AP. Various experiments of com-
bining both emitters and photodiode properties are in progress [10]. Some
78 N.V. Swami et al.
The visible light area is a cell in Li-Fi cellular network. This can be smallest in size
depending on the potential of LED source used. These cells are called as optical
attocell. [14] The decrease in cell size can signicantly increase cellular capacity
and user data rates [15]. The optical attocells can be covering an area of 1-10 m2
and distances of about 3 m [14].
The size of attocell is very small, and walls prevent the system from suffering
from co-channel interference between rooms [4]. The type of security is also
maintained as compare to Wi-Fi. The only thing is that the rate of hand-off will be
increased due to the small size attocells. As the Li-Fi works on VLC, there is no
heavy setup like RF network is required, so we can use Internet and make mobile
phone calls without any additional international roaming rates.
Li-Fi attocells can be deployed as part of a heterogeneous VLC-RF network.
They do not cause any additional interference to RF macro and picocells allowing
the system to hand-off users between the RF and Li-Fi subnetwork [4]. In Li-Fi
cellular network, the indoor attocells were surrounded by a microcellular network
[14]. At glance, data coverage in a room contains many attocells forming a very
dense cellular attocell network.
The Li-Fi attocell and RF macro and picocells together form a wide and unregulated
hybrid network in which the mobile users are served by both technologies. Due to
this hybrid nature, the hand-off falls in four categories: Li-Fi to RF, Li-Fi to Li-Fi,
RF to RF and RF to Li-Fi [16]. As studied in [16] both mobile users and xed users
are served by hybrid network. As attocell in indoor navigation has small size so
bandwidth reuse is possible which results in high spatial spectral efciency pro-
vided to the users [17]. On handover, the channel state information (CSI) must be
provided to the central unit (CU) which is monitoring the system continuously. The
continuous handover takes place in case of mobile users, and the signaling infor-
mation is exchanged between users and CU. This process takes an average time
ranging from around 30 ms to 3000 ms depending on the algorithm used. Some-
times the light beams are blocked due to some obstacle at that time the RF network
serves the coverage that is users having low optical CSI are served by RF APs.
A data rate threshold is used to identify whether the user is served by a Li-Fi AP or
a RF AP [16].
5 Light Fidelity (Li-Fi): In Mobile Communication 79
The ubiquitous with smart computing makes things live in environment. These
things can talk, interact, and operate with each other to make the environment more
sustainable and smarter. In this section, we are going to see how effectively Li-Fi
can be deployed in ubiquitous computing.
(a) Smart Homes:
Each home appliance can be enhanced with Li-Fi LED. It can be your micro-
wave oven cooking food for you, can be a vacuum cleaner cleaning room carpet by
checking the status of dust collected, can be a washing machine, freeze, thermos,
clocks, chairs, fans with LED, smart LED TVs playing favorite channels by
identifying your mood. It all is making together the social web of things in your
home.
Many companies like Ericson, Luminous, Revolve, Philips Hue, Staples connect
and Apple are participating to become a top of the developments in smart home
appliances. [11].
(b) Dense Area:
One of the achievements of this technology is that it works there where most of
the Wi-Fi technology fails. As we know in congested area that is in dense area
where more disturbances are present like construction materials, the Wi-Fi is not a
suitable option. Instead, a Li-Fi serves in a better way by smart lighting and
frequency reuse. This application can be possible in hotels, congested residency
areas, etc.
(c) Ubiquitous Healthcare:
It is the delivery of health-related services and information via ubiquitous
computing technologies to your family doctor [18]. This can be achieved by using
wearable Li-Fi transmitters. These can be rings, ear rings, wrist watches, jewelry,
etc., allowing monitoring not only your health issues but also adding a more beauty
to your personality. For example, your shirt buttons measuring your heart beat rate.
The information is send over to the smart applications of family doctor through
Internet. A xed rate of data packets can be decided for these data transfer like 30
packets per hour.
(d) Learning Environment:
Allowing each classroom light bulb to work as Li-Fi AP, we can make a smart
classroom. In which all students have access to a variety of digital devices like
PDAs, laptops, cell phones, headsets, wireless modems, webcams and services,
whenever and wherever they need them. Both teachers and students can be an
active participants in learning process both collaboratively and individually by
one-to-one or one-to-many or many-to-many communication facility in learning
environment.
80 N.V. Swami et al.
The latest invention of Professor Harald Haas team had made possible not only to
transfer a high-speed stream of data with the help of solar cell but also providing a
power to survive the system independently. This made a best scenario of power
5 Light Fidelity (Li-Fi): In Mobile Communication 83
saving by implanting each electronic object with self-powered Li-Fi solar cell. This
solar cell can act as Li-Fi receiver and solar cell at the same time. The demon-
stration of this experiment was done by Professor Haas at TED Global 2015 event
in London. He also stated that over a four billion people in the world are still do not
have access to the Internet but now with the little energy infrastructure using solar
energy this situation can be changed in developing countries.
They are focusing on the integration of power gathering and data reception at
solar panels and turning them into communication devices. In effect, solar cells
within the panel become communications nodes that receive high-bandwidth data
while also providing electrical power for the nodes operation. These solar panels
can be used on the roof of houses or vehicles to act as a broadband receiver from a
nearby Li-Fi transmitter. The translucent solar cells can be integrated into windows,
doors and in other glass furnitures. These can be also integrated into street fur-
nitures and billions of such devices forming the Internet of Things (IoT) [25].
The Google and Facebook like companies are participating to provide an
Internet to the entire population of the earth by their efforts like Google loon project
based on Internet balloons. The worldwide information and communication tech-
nologies will be requiring more than 100 nuclear power plants to serve for world
population. These self-powered nodes will remove a major barrier to data com-
munication growth. In conventional optical wireless communications, the steady
background component of the received optical signal is usually discarded, but can
instead be used to directly power to the receiving terminal.[26].
The Internet of things (IoT) is the one of the growing industry in market connecting
the network of physical objects, devices, vehicles, buildings and other items which
are embedded with electronics, software, sensors and network connectivity, which
enables these objects to collect and exchange data. Depending on the industry, the
sensor data can be related to temperature, humidity, pressure, machine vibration,
leakage and many other things. Now, of course, modern automation systems are
Internet enabled and these systems can now be called Internet of Things (IoT) ap-
plications. The pervasive nature of Internet of Things will make pervasive con-
nections between each unconnected services, machines, businesses and individuals.
Professor Harald Haas stated in his speech that there are 50 billion devices are
going to be connected by 2020 [26] which will required a huge amount of energy to
serve these devices. Li-Fi solar cell can be enhanced in this Internet of Things
concept for self-powered battery free devices. It is easy to attach Li-Fi tags than
RFID tags. The Li-Fi tags will be communicating to each other through visible light
communication.
84 N.V. Swami et al.
5.11 Conclusion
From above discussion, we can say that Li-Fi will serve many upcoming genera-
tions of mobile communication and other pervasive computing. It is perfectly tted
for mobile communication infrastructure in many better ways by providing a high
degree of mobility, saving energy, saving space, saving cost for deployment and
maintenance. It can be also used to make environment greener, safer and smarter.
No doubt it will fulll and shape the future technologically by making more sus-
tainable lifestyle for mankind.
References
1. Cisco visual networking index, Global mobile data trafc forecast Update 20142019, White
paper, CISCO
2. Li-Fi. Accessed https://www.en.m.wikipedia.org/wiki/Li-
3. Rani J, Chauhan P, Tripathi R (2012) Li-Fi (Light Fidelity)-The future technology. In:
Wireless communication, in International journal of applied engineering research. http://
www.ripublication.com/ijaer.htm
4. Tsonev D, Videv S, Haas H, Light delity (Li-Fi): towards all-optical networking. In: Institute
for digital communications, Li-Fi R&D Centre, Edinburgh, UK
5. Newton T (2013, Oct). British scientists shove 10 gbps through micro light bulbs in LiFi
experiment. https://www.recombu.com/digital/article
6. How light emitting plasma works. Accessed http://www.luxim.com
7. Bhut JH, Parmar DN, Mehta KV (2014, Jan) LI-FI technologya visible light communi-
cation. Int J Eng Dev Res
8. http://www.pureli.com/li-re/li-ame/
9. Haas H (2014, Apr) My Li-Fi revolution, Presented at Tam Dalyell prize lecture. http://www.
m.youtube.com/
10. Pujapanda KP (2013, Apr) LiFi integrated to power-line for smart illumination cum
communication. In: International conference on communication system and network
technologies. http://www.ieeexplore.ieee.org/
11. Wolf M (2014, Dec) 5 smart home companies to watch in 2014. www.forbes.com/sites/
michaelwolf/2013/12/31/5-smart-home-companies-to-watch-in-2014/
12. Savage N (2014, Nov) Li-Fi gets ready to compete with Wi-Fi. http://www.spectrum.ieee.org/
tlecom/internet/
13. Piltch A (2014, Jan). Bright idea: smartphone sensor receives data via light, Laptop. https://
www.m.blog.laptopmag.com/wysip-connect-li-downloads-data
14. Haas H (2013) High-speed wireless networking using visible light. Accessed https://www.
spie.org/x93593.xml
15. Uedinburgh NI, Partner on Li-Fi (2013, Nov). www.photonics.com/m/Article
16. Wang Y, Videv S, Haas H (2014) Dynamic load balancing with handover in hybrid Li-Fi and
Wi-Fi networks. In: IEEE 25th international symposium on personal, indoor and mobile radio
communications
17. Stefan I, Burchardt H, Haas H (2013) Area spectral efciency performance comparison
between VLC and RF femtocell networks. In: 2013 IEEE international conference on
communications (ICC), June 2013, pp. 38253829
18. Neethu MS (2013, Mar) Ubiquitous computing. https://www.lbsitbytes2010.wordpress.com/
2013/03/19/ubiquitous-computing-2/
5 Light Fidelity (Li-Fi): In Mobile Communication 85
19. Jin W-L (2012) SPIVC: a smartphone-based inter-vehicle communication system. Proc Trans
Res Board Ann Meet
20. Boukerche A et al (2008) Vehicular Ad Hoc networks: a new challenge for localization-based
systems. Comput Commun Sci Direct 112
21. Husain Fidvi NM Car to car communication system. Source: car communication system.
http://www.engineersgarage.com/contribution/car-to-car-commuincation-system?page=1
22. Swami NV (2015, Mar) Li-Fi (LIGHT FIDELITY)THE changing scenario of wireless
communication. Int J Res Eng Technol. http://www.ijret.org
23. Tower Infrastructure Solutions. Accessed https://www.bharti-infratel.com/cps-portal/web/
passive-infrastructure-solutions.html
24. Curran K, Furey E, Lunney T, Santos J, Woods D, Mc Caughey A (2011) An evaluation of
indoor location determination technologies. J Location Based Serv 5(2):6178, June 2011.
ISSN: 1748-9725, doi:10.1080/17489725.2011.562927, Taylor & Francis
25. http://www.li-centre.com/the-connected-solar-panel/
26. http://www.ted.com/talks/harald_haas_a_breakthrough_new_kind_of_wireless_internet?utm_
campaign=social&utm_medium=referral&utm_source=t.co&utm_content=talk&utm_term=
technology
27. http://visiblelightcomm.com/top-10-visible-light-communications-applications/automotive/
Chapter 6
Performance Analysis of Denoising Filters
for MR Images
Keywords MRI
Denoising Nonlinear lters NLM Bilateral Image
quality assessment
S.D. Oza ()
E & Tc Department, Army Institute of Technology, Pune, India
e-mail: sdoza@aitpune.edu.in
K.R. Joshi
E & Tc Department, PES Modern College of Engineering, Pune, India
e-mail: krjpune@gmail.com
6.1 Introduction
2n 2n
where I0 is the modied zeroth-order Bessel function of the rst kind, (.) is the
Heaviside step function, 2n is the noise variance, and A is the noise-free signal level.
For high SNR, the Rician distribution tends to be Gaussian. As the SNR falls to low
level (<2), it tends to be Rayleigh distributed.
The Rician noise is signal dependent and thus very difcult to remove. It is
necessary to denoise the MR image and support accurate diagnosis.
The denoising lters are broadly classied as linear and nonlinear. The linear lters
are the ones which obey the principles of superposition and shift invariance.
Gaussian lter is a classic example of linear lter. The linear lters can be
implemented in spatial domain or temporal (frequency) domain. The spatial domain
ltering techniques directly manipulate the intensity of pixels in the image [4],
while the performance of frequency-domain lter is governed by the selection of
frequency response of the lter. The linear lter though simple to implement tends
to give blurring effect. The nonlinear lter response does not obey the linearity
principles. Its output can vary in an intuitive way and tends to respect the edges in
the image. Non-local means (NLM), bilateral, and anisotropic diffusion lters are
few examples of this category [2, 5, 1316].
90 S.D. Oza and K.R. Joshi
Here,
0 wr, s 1 and wr.s = 1
where r is the point being ltered and s represents any other pixel in the same
image. If Nr and Ns are neighborhoods of the pixels r and s, respectively, then
weight w(r, s) indicates similarity between Nr and Ns.
Thus, this method averages all pixels in the image with similar neighborhood,
while weight w(r, s) judges the similarity between the neighborhoods. The com-
putation of weight leads to a huge time overhead. This can be optimized by
restricting the search to a region with radius R in the whole image.
b. Bilateral Filter
The bilateral lter is a neighborhood lter which is nonlinear. It was rst
introduced as nonlinear Gaussian lter in 1995 by Aurich et al. In 1998, Tomasi
and Manduchi rediscovered it and gave the name bilateral [6].
The bilateral lter replaces the current pixel intensity, with a weighted average of
geometric (spatial) and photometric (intensity) distances between neighboring
pixels in the search region. The weight assigned to each neighbor decreases with the
distance in the image plane (the spatial domain S) as well as the distance on the
pixel intensity axis (the range domain R). This helps in preserving edges along with
noise removal. The expression for the ltered value of pixel r is given as follows:
ks r k jIs Irj
2 2
1
Ir = s Nr e 22 d e 22 r Is 6:3
c
where d and r are standard deviations in spatial and intensity domains, respec-
tively [17, 18].
c. Gaussian Filter
Gaussian lter computes weighted average of the intensity of the adjacent positions
wherein weight decreases with the spatial distance to the center position p (origin):
6 Performance Analysis of Denoising Filters for MR Images 91
1 r2 +2 s2
gr, s = e 2 6:4
2 2
where r is the distance from the origin in the horizontal axis, s is the distance from the
origin in the vertical axis, and is the standard deviation of the Gaussian distribution.
The three lters, viz. NLM, bilateral, and linear lters, were implemented using
MATLAB (ver 2010) on Intel i5 machine with 4-GB RAM working at 2.2 GHz. The
image database includes ten MR images of bone and joint (658 495) and brain slice
(512 512). To each image, Rician noise was added with varying standard deviation
starting from S = 3 to S = 24 in ve steps. For each noise level, lter performance
was tested using three parameters, namely PSNR, MSE, and SSIM index.
In the case of each lter, as the noise level was increased, MSE increased
(Fig. 6.1), while PSNR (Fig. 6.2) and SSIM index reduced as was expected
(Fig. 6.3).
On application of NLM lter to brain MRI, the PSNR was 32.748 for S = 3
which was reduced to 23.4484 as S was increased to 24. When bilateral lter was
applied to the brain MRI, it outperformed the rest giving the PSNR index of
82.0602 for noise level S = 3 and 64.4933 for S = 24. In the case of linear lter,
PSNR was observed to be 24.7394 for minimum noise level of 3. This was the
lowest result as compared to the other two lters (Fig. 6.2).
The SSIM index was noted to be 0.9356 for minimum noisy image, in the case
of NLM lter. It was found to be reduced to 0.7871 as noise level was increased to
24. Here too, bilateral lter proved to be better with minimum drop, i.e., with
92 S.D. Oza and K.R. Joshi
300
MSE
200
100
0
3 6 12 18 24
STANDARD DEVIATION
60
PSNR in dB
50
40
30
20
10
0
3 6 12 18 24
STANDARD DEVIATION
0.8
SSIM
0.6
0.4
0.2
0
3 6 12 18 24
STANDARD DEVIATION
6 Performance Analysis of Denoising Filters for MR Images 93
maximum index of 0.99990.9319 being the least for even the noisiest image with
S = 24 (Fig. 6.4).
The average computation time, in the case of each lter, was noted down. For
the brain MR image, NLM lter with the search width of 15 was the slowest with
average computation time of 180 s, while linear Gaussian lter was proved to be
the fastest with average computation time of 3.8 s. The bilateral lter took optimum
average time of 4.5 s and provided best results (Fig. 6.5).
The experimentation results imply that bilateral lter performs better than NLM
lter in terms of PSNR as well as SSIM index. As shown in Figs. 6.6, 6.7, and 6.8,
SSIM index is close to the human perception of the quality of the denoised image.
94 S.D. Oza and K.R. Joshi
PSNR =64.4933
SSIM =0.9313
PSNR =23.4484
SSIM =0.7871
6 Performance Analysis of Denoising Filters for MR Images 95
PSNR =20.9520
SSIM =0.4070
It can also be seen that NLM and bilateral lter try to retain edges, while linear lter
blurs the image. In future, the performance may further be analyzed using edge
quality measure. In the present work, the computation time of the NLM lter was
noted to be too large as compared to the rest (Fig. 6.4). It may be reduced by
implementing the lter on parallel architectures [20]. The NLM lter being a
neighborhood lter can easily be mapped to parallel architectures. Also, variants of
NLM [21] and improvised bilateral, i.e., trilateral lter [22], can be analyzed for
performance.
References
1. Buades et al (2005) A non local algorithm for image denoising. IEEE Int Conf Comput Vision
Pattern Recogn CVPR 2(2005a):6065
2. Jose M et al (2008) MRI de noising using non local means. Med Image Anal 12:514523
Science Direct, Elsevier
3. Mohan J et al (2014) A survey on the magnetic resonance image denoising methods. Biomed
Signal Process Control 9:5669
4. Godbjartsson H, Patz S (1995) The Rician distribution of noisy MRI data. NIH Public Access
Author Manuscript
5. Kundu R et al (2014) De-noising image lters for bio-medical image processing. CSI
Commun
6. Tomasi C, Manduchi R (1998) Bilateral ltering for gray and color images. In: Proceedings of
the 1998 IEEE International Conference on Computer Vision, Bombay
96 S.D. Oza and K.R. Joshi
7. Perona P, Malik J (1990) Scale-space and edge detection using anisotropic diffusion. IEEE
Trans Pattern Anal Mach Intell 12:629639
8. Wu ZQ et al (2003) Wavelet-based Rayleigh background removal in MRI. IEEE Electron.
Lett. 39:603605
9. Avcibas I, et al (2002) Statistical evaluation of image quality measures. J Electron Imaging 11
(2):207
10. https://www.cs.sfu.ca/stella/papers/blairthesis/main/node11.html
11. http://www.drcmr.dk/mriinbrief
12. P Coupe et al (2010) Robust Rician noise estimation for MR Images. Med Image Anal
14:483493 Elsevier
13. Dolui S et al (2013) A new similarity measure for non local means ltering of MRI
images. 10401054
14. Teuber T et al (2012) A new similarity measure for non local ltering in the presence of
multiplicative noise. Comput Stat Data Anal 56:38213842 Elsevier
15. Manjon J et al (2010) Adaptive non local Means De noising of MR images with spatially
varying Noise levels. J Magn Reson Imaging 31:192203 Wiley-Liss 2009
16. Rajan J et al (2012) An adaptive non local maximum likelihood estimation method for
denoising magnetic resonance images. IEEE
17. Sylvain Paris et al (2008) Bilateral ltering: theory and applications. Comput Graph Vision 4
(1):173
18. Zhang M (2009) Bilateral ltering for image processing. Thesis, Beijing University
19. Wang Z et al (2004) Image quality assessment: from error measurement to structural
similarity. IEEE Trans Image Process 13(1)
20. Eklund A, et al (2013) Medical image processing on the GPU: past, present and future. Med
Image Anal http://www.dx.doi.org/10.1016/j.media.2013.05.008. Linkping University Post
Print
21. Kim DW et al (2011) Rician non local means de noising for MR images using nonparametric
principal component analysis. EURASIP J Image Video Process 2011:15
22. Wilbur CK, Wong et al (2004) Trilateral ltering for biomedical images. 0-7803-8388-5/04
IEEE
Chapter 7
A Detailed View on SecureString 3.0
Gnter Fahrnberger
Abstract These days, vicious globally playing enterprises and various culprits try
to exploit individuals sensitive data whenever opportunities arise. Public clouds and
private users with low (or even no) security awareness facilitate such malefactions.
On account of this, a rational data owners trust in public clouds should reach the
level semi-honest or semi-honest-but-curious at best due to their uncertainty of the
location(s) of their data respectively of undesired access on them. The homomorphic
cryptosystem SecureString 3.0 remedies to recapture the cloud users faith in secure
cloud computing on encrypted character strings by combining the paradigms blind
computing and secret sharing. While existing literature already covers the principles
of SecureString 3.0, herein, the adduced implementation details of this cryptosystem
given in pseudocode allow researchers to realize their own prototypes and practition-
ers to integrate SecureString 3.0 in their own security solutions. Decent security and
performance analyses prove the applicability of this cryptosystem in the eld, e.g.,
for secure instant messaging sifters.
7.1 Introduction
Nowadays, it is a common practice for private users to have their les and elec-
tronic mail accounts outsourced in public clouds due to comfortable accessibility
from nearly all over the world. Daily news about exploitations of sensitive data let
the private users security awareness upswing awhile before it reverts to its original
level, or even worse, people do not fear for the security of their data because they
suppose that their data do not arouse anyones interest. Even the latter would imme-
G. Fahrnberger ()
University of Hagen, North Rhine-Westphalia, Germany
e-mail: guenter.fahrnberger@studium.fernuni-hagen.de
diately change their opinion if somebody successfully looted their bank account due
to their stored plaintext credentials in an online folder.
Companies usually act much more carefully because lost or exploited business
data could entail the complete ruin of reputation or even existence of a rm. An
organization will not make use of a public cloud as long as its personnel assess
uncertainty for its valuable data there. If the security stas of an enterprise have
condence in a cryptosystem, then they will not hesitate to employ it for the stor-
age of crucial information in public clouds. Evenhandedly, the responsible sta will
only incorporate a viable homomorphic encryption scheme for blind computations
on ciphered data in public clouds if they rely on it.
Unfortunately, wrongdoing corporations take advantage of their prudent back-
ground about IT security and the users ignorance of equivalent ken to exploit user
data for entrepreneurial purposes. Ironically, the rms characterize such data abuse
as marketing action because the illegally observed user habits permit them to proer
their products more purposively.
Altogether, a decision maker has two options available. Either they entirely con-
de in a public cloud, or they completely suspect it. On account of this, basically,
trust represents a dichotomous property between two objects. Some authors allude
to semi-honest [22] or semi-honest-but-curious [5] parties which are only partially
trusted. Albeit such shady actors do not behave viciously, they can easily cause cru-
cial security breaches if they ignore or negate any prophylactic measures or respon-
sive countermeasures against genuine villains. The treatise about SIMS (Secure
Instant Messaging Sifter) characterized transmission or IM (Instant Messaging) plat-
form providers with such a misbehavior as neutralists [10]. The airy way is to eschew
such neutral resources, but in spite of security risks they often promise luring com-
mercial advantages. Cloud computing exemplies a classical neutralist. SLAs (Ser-
vice Level Agreements) and nondisclosure agreements bode well, but they cannot
change the fact that cloud capabilities opaquely reside and even roam in any com-
puting centers on our planet without control possibilities for the data possessor.
For this reason, a policy maker can consider to involve a public cloud in such a
way that an interloper cannot gure out what the cloud stores or computes. While
topical cryptosystems applied for end-to-end-encryption assure the safe detention of
sensitive data in cloud storages, SSE (Searchable Symmetric Encryption) schemes
abet keyword searches in ciphertexts. Reasonable modications of encrypted data
require the employment of homomorphic cryptosystems. While the famous encryp-
tion schemes of Rivest, Shamir and Adleman (RSA) [25], Paillier [23] and Gentry
[16] ensure homomorphic operations on ciphered numerical values, the SecureString-
family with the cryptosystems SecureString 1.0 (see Sect. 7.2.1), SecureString 2.0
(see Sect. 7.2.2), and SecureString 3.0 (see Sect. 7.2.3) proers various homomor-
phic functions on encrypted character strings. SecureString 2.0 improved the privacy
of SecureString 1.0, and SecureString 3.0 hardened SecureString 2.0 with better
authenticity, integrity, and privacy. The creator of these encryption schemes ear-
marked them for the secure sifting of chat and instant messages [10, 13] which does
not exclude them from utilization for other assignments.
7 A Detailed View on SecureString 3.0 99
The rst and sole paper about SecureString 1.0 was released under the title Com-
puting on Encrypted Character Strings in Clouds [7]. It treats of a homomorphic
cryptosystem with the following benets:
Underlying replaceable symmetric cryptosystem: SecureString 1.0 performs
each ciphering by a topical underlying exchangeable symmetric cryptosystem that
becomes inevitably exchanged if its vulnerability exceeds any observable modern
security norms.
Polygraphic encryption: SecureString 1.0 splits plaintext words into its disjoint
polygrams of length n in order to encrypt them individually.
Intra-word polyalphabetism [3]: SecureString 1.0 enciphers each polygram of a
plaintext word with a separate encryption transformation.
Position-based encryption transformation: SecureString 1.0 achieves unique
encryption transformations by ciphering each polygram together with the begin-
ning position within its encircling word.
Polygram order scrambling: SecureString 1.0 brings the ciphertext polygrams
out of sequence without losing the possibility to operate on them because their
100 G. Fahrnberger
order can be reconstructed after their decryption through their enveloped position
information.
The developer of SecureString 1.0 himself detected some drawbacks in his cryp-
tosystem as follows:
Finite length support: SecureString 1.0 can only fully support character string
operations on ciphertext objects up to a predened length.
Overt word boundaries: Due to the nite support of ciphertext object lengths,
the intentional use of an own SecureString 1.0 object for each word suggests itself.
Disadvantageously, this approach discloses the character string boundaries of texts
and abets eective repetition pattern attacks on them.
Repetitive ciphertexts: If polygrams are encrypted together with their beginning
positions but without any randomness (such as salts or nonces), then ciphertext
repetitions likely promote eectual statistical distribution attacks.
Due to these recognized aws, the inventor of SecureString 1.0 felt obliged to launch
the enhanced successor SecureString 2.0.
The foremost essay about SecureString 2.0 came up with the undermentioned advan-
tages [8]:
Underlying replaceable symmetric cryptosystem: SecureString 2.0 performs
each ciphering by a topical underlying exchangeable symmetric cryptosystem that
becomes inevitably exchanged if its vulnerability exceeds any observable modern
security norms.
Inter-word polyalphabetism [3]: SecureString 2.0 enciphers each plaintext word
with a unique encryption information. Therefrom, the arbitrary number of con-
catenated words per SecureString 2.0 object conceals their boundaries.
Salting: Morris and Thompson came along with the idea to salt plaintext pass-
words (i.e., to prolong them with a random character string) before their encipher-
ment to obtain dierent ciphertext versions of them by adhering various salts [20].
SecureString 2.0 attains a distinct encryption transformation for every plaintext
word by exclusively enciphering each character of the same plaintext character
string together with an identical salt.
Automatic salt updating [2, 10]: Each salt serves as input for a hash function that
outputs the salt for the encryption transformation of the successive plaintext word.
The so-called TTPG (Trusted Third Party Generator) can prepare the homomor-
phic computations for an untrustworthy (external) node just with a starting salt
(see Sect. 7.3.2).
Hereinafter, a list itemizes the features of SecureString 2.0 which represent pros and
cons at the same time:
7 A Detailed View on SecureString 3.0 101
That is why a prudent IM solution must inevitably build on a centralized sifter that
does not have the chance of getting to know and exploiting any instant messages.
Designs based on SecureString 2.0 were steps in the right direction, but too weak
achievements of the security goals authenticity, integrity, privacy, and resilience led
to the development of SecureString 3.0.
As its predecessors SecureString 1.0 and SecureString 2.0, SecureString 3.0 [12]
poses a non-size-preserving cryptosystem. The encryption scheme does nothing else
but to salt each individual plaintext character of a string, i.e., to append an arbitrary
character sequence, and to encipher the amalgamation of plaintext character and salt
with a contemporary high-performing cryptosystem in ECB (Electronic CodeBook)
mode [18].
The encryption algorithm may prolong as many characters with the same salt as
no ciphertext repetition appears. Therefore, before it scrambles a dedicated charac-
ter with an equal salt twice, it must switch to another salt and thereby defuses the
7 A Detailed View on SecureString 3.0 103
hazardousness of repetitions despite the usage of ECB mode. Automatic salt updat-
ing by inputting the recent salt in a (trapdoor) hash function denotes an easy and
secure alternative to obtain a fresh salt from an RNG (Random Number Genera-
tor) [2]. Blum and Micali pioneered with their general algorithmic scheme for con-
structing polynomial-time deterministic algorithms which stretch a short real ran-
dom input (also called seed) into a long sequence of unpredictable pseudo-random
bits [4]. Stream ciphers and re-keying approaches [1] capitalize on these algorithms
and draw pseudo-random material from their PRNGs (Pseudo-Random Number
Generators). In contrary to the PRNG of Petit et al. [24], SecureString 3.0 cannot
include payload content in salt derivations because the TTPG must not encounter
any message texts due to secret sharing with the untrusted host.
This kind of salt production does not only perpetuate backward security, i.e., no
salt can be concluded from its successors, but also lets a TTPG prepare homomorphic
computations for an untrustworthy node just with a starting salt per cohesive text
body. Every TTPG and every trusted sender (see Sect. 7.3.1) must possess the ability
to fabricate arbitrary salts. Especially mobile devices in the role of trusted senders
without a swift on-board RRNG (Real Random Number Generator) yearn for an
acceptable alternative.
If a modern block cipher combined with a safe mode of operation [18] (e.g., CBC
(Cipher Block Chaining), CFB (Cipher FeedBack), or OFB (Output FeedBack)) was
utilized instead of salting and ECB mode (as done in the SecureString-series), then it
would also warrant backward security, but its output ciphertext blocks would depend
on their antecedent blocks and a TTPG could not derive ciphertext alternatives for
any plaintext characters. It goes without saying that stream ciphering also does not
permit a TTPG to compute ciphertext alternatives for plaintext characters.
To make a long story short, SecureString 3.0 achieves non-repetitive ciphertexts
with controlled randomness attained through hashed salts in lieu of haphazardly
compassed ciphertexts through the contents of foregoing blocks.
7.3 Algorithms
Definition 1 Let A be the trusted sender, let Acert be A s certicate with A s public
key Ae , let AR be an ecient data structure in A, let ARNG be A s random num-
ber generator, let B be the untrusted host, let BR be B s public repository, let C
be the trusted receiver, let CA be the Certication Authority, let TTPG be the
Trusted Third Party Generator, let TTPGR be an ecient data structure in TTPG,
let sACTTPG be a securely stipulated session key between A, C and TTPG, let sAB
be a securely stipulated session key between A and B, let sBC be a securely stipu-
lated session key between B and C, let sBTTPG be a securely stipulated session key
between B and TTPG, let s XY be transport encryption between two parties X and
Y based on a topical symmetrical cryptosystem and a securely stipulated session
key sXY , let be an alphabet, let Dss (ciphertext) m be a decryption
and Ess (plaintext) m be an encryption function of a contemporary
symmetrical cryptosystem based on a session key s m , let H m1 be a
hash function, let v lv be plaintext of length lv and {vj |1 j lv } be
its ordered set of plaintext characters, let Jv be the ordered set of the character
positions of the needed salt changes for v, let w lv m be the SecureString 3.0-
object for v and {wj m |1 j lv } be its ordered set of ciphertext characters,
let R be a plaintext repository with search strings and their replacement
strings, let (ui , ti ) R|1 i |R| be the ith element of R, let Jui be the ordered
set of the character positions of the needed salt changes for ui , let lui the length of
ui , let d be a beginning index, let e be an ending index, and let salt m1
be a random word.
7 A Detailed View on SecureString 3.0 105
Algorithm 1 encrypt
Require: Acert with Ae , AR , Jv , sAB , sACTTPG , sBTTPG , v
Ensure: w or error
Algorithm 2 createPublicRepository
Require: Jv , R, TTPGR , lv , sACTTPG , salt
Ensure: BR
1: TTPG BR {} {Initialization}
2: for i 1 to |R| do {Iteration over search strings}
3: if lui lv then {Search string not too long}
4: TTPG Jui {}; TTPGR {} {Initialization}
5: for k 1 to lui do {Iteration over search string characters}
6: if uik TTPGR then {Salt change incrementation}
7: TTPG Jui (Jui k); TTPGR {};
8: end if
9: TTPG TTPGR (TTPGR {uik })
10: end for
11: if |Jui | |Jv | then {Proper number of salt changes in search string}
12: TTPG salt salt {Salt initialization}
13: for j 1 to lv lui + 1 do {Iteration over starting string character}
14: TTPG TTPGR {}; qi {}; salt salt {Initialization}
15: for k 1 to lui do {Iteration over search string characters}
16: if (j + k 1) Jv then {Salt change}
17: TTPG TTPGR (TTPGR k)
18: end if
19: end for
20: for k 1 to lui do {Iteration over search string characters}
21: if k TTPGR then {Salt change}
22: TTPG salt H(salt)
23: end if
24: if not (|Jui | > 0 and |TTPGR | > 0 and min(Jui ) = min(TTPGR ) and Jui
TTPGR ) then {Useful search string}
7 A Detailed View on SecureString 3.0 107
Once a faithless host B has got a BR from its assigned TTPG, it can properly exe-
cute the querying Algorithm 3 or the replacing Algorithm 4 on the accordant input
ciphertext w. Without an acquired BR , B can still produce reasonable substrings of
w by pruning w with the picking Algorithm 5. While a TTPG as the producer of BR
becomes aware of the meaning of all elements in BR , it neither gets in touch with v
nor with w nor with querying results. In contrast, B as the consumer of BR can read
and even write on w, but cannot learn anything about the content of v. Thus, this
division of labor between a TTPG and B enforces secret sharing.
Algorithm 3 query
Require: BR , m, sAB , w
Ensure: true or false
Algorithm 4 replace
Require: Acert with Ae , BR , m, sAB , sACTTPG , sBC , w
Ensure: v or error
Algorithm 5 pick
Require: Acert with Ae , d, e, sAB , sBC , w
Ensure: v or error
Algorithm 6 decrypt
Require: m, sACTTPG , w
Ensure: v
l
1: C lv mw ; w {} {Calculation of plaintext word length; Initialization}
2: for j 1 to lv do {Iteration over ciphertext characters}
3: C vj (rst character of DssACTTPG (wj )) {Character decryption}
4: C v (v {vj }) {Plaintext reassembly}
5: end for
6: return v
All components of the architectural model conde in the used certication author-
ity. The CA generally behaves passively and becomes merely active on demand to
approve/reject authentication requests for valid/invalid certicates (see Algorithms
1, 4, and 5), to issue new certicates, and to blacklist compromised certicates.
The next section describes how SecureString 3.0 withstands various typical oenses
against cryptosystems.
Menezes et al. itemized six attack classes whose objective is to systematically recover
plaintext from ciphertext, or even more drastically, to deduce the decryption key [19].
This section investigates the implications of probable examples of each class on
SecureString 3.0-objects in the descending order of their diculty respectively in
the ascending order of their dangerousness.
All subsections below assume the encryption scheme of SecureString 3.0 in Def-
inition 2.
Definition 2 Let denote an alphabet, then SecureString 3.0 extends each sin-
gle character vj |1 j |v| of a new plaintext message v with a salt sj
m1 in a vein so that the lowest number of salt changes adheres to (j )(i
)|vj sj = vi si . Thereafter, it lets the underlying cryptosystem scramble each plain-
text character-salt-combination vj sj m |1 j |v| with a secret key k l to the
ciphertext E l m m , k, vj sj E(k, vj sj ) = wj .
7 A Detailed View on SecureString 3.0 111
Proof The timely salt renewals of the encryption scheme of SecureString 3.0 war-
rant the singularity of all enriched plaintext characters within the plaintext v, i.e.,
(j )(i )|vj sj = vi si . The bijective characteristic of the encryption function
E leaves the ciphertext characters in w the uniqueness of the plaintext character-salt-
combinations in v.
|Sigmam | E(k, vj sj ) = wj } are known, and the secret key k l is unknown, then k
can be solely recovered with a brute-force-attack.
An expedient comparison at least requires a similar experiment setup for both cryp-
tosystems. On this account, the test setting of SecureString 2.0 [9] has become reused
for SecureString 3.0, complemented with a TTPG and a common RMI (Remote
Method Invocation) registry for all remote objects. Figure 7.2 illustrates the logi-
cal components of the elaborated experimental environment as well as a rudimen-
tary overview of the exchanged messages between them. The test program initially
executes the black-colored steps to interconnect the elements as needed. It solely
7 A Detailed View on SecureString 3.0 115
reruns them for recovery purposes after detected functional outages. The red-colored
sequence becomes oftentimes iterated to accomplish a statistically signicant sample
size for each evaluated data value.
The below-mentioned lines recapitulate the role of each unit in the test design.
Trusted Sender: This machine sets out to automatically fabricate and encrypt ran-
dom messages when all constituents have become interlinked through the (black-
colored) dialogs. The sender coercively pauses if at least one part of the testbed
malfunctions and resumes once everything is up and running again. This entails
that the present lab conguration does not support the examination of a store-and-
116 G. Fahrnberger
forward mode which would be necessary in a real use case to retry failed deliveries
towards an absent trusted receiver. The condential sender generates and conveys
a haphazard initial salt for each message to the TTPG before it pushes the enci-
phered message to the untrustworthy host. Together with the salt, the consigner
lets the TTPG know about the positions of salt transitions for each message to help
the TTPG minimizing the magnitude of the prepared cloud repository. In addition,
the sending process takes the responsibility to gage the turnaround time of every
spawned message through the red-colored ow in order to compile nal perfor-
mance statistics.
Trusted Third Party Generator (TTPG): This trusty entity is responsible to
quickly convert one or multiple sought plaintext keywords into a cloud reposi-
tory with all possible ciphertext alternatives in compliance with the assigned salt
change positions. Applications are conceivable in which such a cloud repository
embraces a ciphertext replacement string for each ciphertext search string. Just to
give an example, the TTPG of SIMS [10] fetches a bunch of deprecated vocables
from an online plaintext bad word dictionary, and for each of them it opposes an
equally long chain of asterisks as replacement string, alternatively the empty string
acts as replacement string. In the performance analysis of SecureString 2.0 [9],
the trustworthy sender generated one random replacement string for each search
string. The will for a convincing performance analysis of SecureString 3.0 calls
for dealing in the same way. The job of the TTPG nishes with the handover of
distilled cloud repositories to the untrusted host.
RMI Registry: The Java implementation of SecureString 3.0 works with the RMI
mechanism and, hence, demands a suitable reference library, also called an RMI
registry. While each of the two launched RMI servers in the SecureString 2.0
architecture (untrusted host, trusted receiver) sparked o an own RMI registry,
all RMI servers in the architectural layout of SecureString 3.0 (TTPG, untrusted
host, trusted receiver) share a collective RMI registry. The black-colored arrows
in Fig. 7.2 represent the registrations of the RMI servers in the RMI registry as
well as the lookups of the RMI clients (trusted sender, TTPG, untrusted host) for
RMI servers in the RMI registry.
Untrusted Host: This computer simulates a cloud resource that does not possess
full condence of the TTPG and of the trusted senders and receivers. Neverthe-
less, it shall take over considerable workload due to tendered unrivaled benets.
The workload incloses two simple assignments. First, the untrusted host queries
incoming ciphertext messages regarding ciphertext keywords out of the associ-
ated cloud repository and passes such messages without any found items on to
the trusted receiver. Second, it replaces each nding by the appendant ciphertext
replacement string out of the cloud repository and relays every processed mes-
sage to the trusted receiver. Third, it truncates the inbound messages and remits
the resulting substrings to the trusted receiver. Contrary to the recommendation in
[12], the untrusted host in the experimental buildup does not let the trusted sender
double-check and sign/reject modied ciphertexts to comply with the experimen-
tal makeup of SecureString 2.0 [9] and chalk up meaningful comparative perfor-
mance analysis.
7 A Detailed View on SecureString 3.0 117
Trusted Receiver: This node just deciphers each arriving SecureString 3.0-object
and noties the untrusted host of its accomplished reception. The untrusty node is
in charge of forwarding all these notications to the trusted sender.
Every reader of the introductive work [12] will notice that the experimental instal-
lation for SecureString 3.0 misses a CA. The rationale is the same as for the lack
of sender signaturesmaintaining comparability with the performance of Secure-
String 2.0.
By same token, both SecureString-editions even resort to the consistent, below-
quoted test details:
Hardware: HP DL380 G5 with 8 2.5 GHz cores and 32 GB main memory
Operating system: Fedora Core 64 bit Linux
Programming language/virtual machine: Java 8 Standard Edition
Underlying high-performing cryptosystem: AES (Advanced Encryption Stan-
dard) 128 bits
Sample size: 1,000,000 random messages for each pictured data point
The querying performance becomes combinatorially measured for the mean process-
ing time of each plaintext message length |v| {1, , 64} and each plaintext search
string length |u| {1, , |v|}, i.e., for 64
|v|=1
= 64(64+1)
2
= 2, 016 data points. A
Java RNG in the trusted sender issues one arbitrary search string per data point
that becomes enciphered in the TTPG and sought in all 1,000,000 samples by the
untrusted host. Figure 7.3 comparatively shows that SecureString 3.0 approximately
consumes the 4.vefold processing time of SecureString 2.0. The throughput time
rises from 0.1 ms to 0.45 ms for |u| = |v| = 1 with the utilization of SecureString 3.0
instead of SecureString 2.0. Evenhandedly, the diagram peak increases from 1 ms to
4.5 ms.
Fig. 7.3 Querying performance of SecureString 2.0 [9] (left) and SecureString 3.0 (right)
118 G. Fahrnberger
Fig. 7.4 Replacing performance of SecureString 2.0 [9] (left) and SecureString 3.0 (right)
The replacing performance becomes also clocked as the average cycle time of replac-
ing operations for 2,016 data values. For all 1,000,000 samples of each particular |v|-
|u|-combination, the identical random search string is employed as for the equivalent
data point of the querying performance evaluation. In this test run, the untrusted host
exchanges all occurrences of search strings in messages for the empty string, i.e., it
shortens aected messages. In contrast to the querying performance, Fig. 7.4 evi-
dences a lower rate of lead time increase for replacing operations with the practice
of SecureString 3.0 in place of SecureString 2.0. The raise from 0.3 to 0.6 ms for
|u| = |v| = 1 implies just doubling. The heightening of the top from 1.4 to 5.24 ms
roughly implicates a factor of 3.7.
The picking performance again becomes quantied as the mean runtime for 2,016
diagram points. A Java RNG in the trusted sender randomizes a beginning index d
Fig. 7.5 Picking performance of SecureString 2.0 [9] (left) and SecureString 3.0 (right)
7 A Detailed View on SecureString 3.0 119
|v| and an ending index e |v| for all 1,000,000 samples of each particular |v|-|u|-
combination that fullls the condition e d = |u|. Figure 7.5 attests for |u| = |v| = 1
that the run duration of 0.29 ms for SecureString 2.0 imperceptibly grows to 0.32 ms
for SecureString 3.0. Furthermore, it merely substantiates a marginal growth of the
summit from 0.51 to 0.70 ms.
7.6 Conclusion
Seemingly, even huge data scandals sensitize most private online users only for a
short time or bother them not at all. Thereby, they carelessly expose intimate infor-
mation about themselves in/through the Internet without being aware that malecent
parties are just waiting to exploit their behavior. Homomorphic cryptosystems thwart
villains with blind computations on enciphered data.
This scholarly piece continues previous introductive work about the homomor-
phic cryptosystem SecureString 3.0. After an introductory section and a further one
about related work, it unveils its implementation mechanisms with the help of a rudi-
mentary architecture model. The remaining section concerns the description of the
necessary constituents in this model and of six (pseudocode) algorithms spanning
them.
The security analysis examines the impact of Menezes six attack classes on
SecureString 3.0-objects. It arrives at the conclusion that the concealment of Secure-
String 3.0 mainly bases itself on an underlying high-performing cryptosystem with
strong robustness (e.g., against dierential cryptanalyses) and on a collision-free
(trapdoor) hash function in use for the fabrication of nonrecurring secret salts. More-
over, SecureString 3.0 wangles integrity by obliging message originators to sign
accepted manipulated messages.
A performance analysis of SecureString 3.0 constitutes another merit of this doc-
ument. It demonstrates that designers who count on SecureString 3.0 rather than on
SecureString 2.0 to eliminate perilous security lapses (like repetition pattern attacks)
must pay the price of worse but still agreeable time performance. In other words,
secure design philosophy forbids to cut a security edge in the name of eciency,
because there exist too many fast, insecure systems [14]. Under the adopted labora-
tory conditions of SecureString 2.0, querying operations on SecureString 3.0-objects
elongate around 4.5 times, replacing operations length up to 3.7 times, and picking
operations last just slightly longer.
This disquisition prompts the authoring of successional treatises that cope with
the embedment of the demonstrated SecureString 3.0 algorithms in live multi-user
applications.
Acknowledgements Many thanks to Bettina Baumgartner from the University of Vienna for proof-
reading this contribution!
120 G. Fahrnberger
References
1. Abdalla M, Bellare M (2000) Increasing the lifetime of a key: a comparative analysis of the
security of re-keying techniques. In: OkamotoT (ed) Advances in Cryptology ASIACRYPT
2000. Lecture notes in computer science, vol. 1976. Springer, Berlin, Heidelberg, pp 546559.
https://dx.doi.org/10.1007/3-540-44448-3_42
2. Anderson RJ (2008) Security engineeringa guide to building dependable distributed sys-
tems, 2nd ed. Wiley. https://www.cl.cam.ac.uk/~rja14/book.html
3. Bauer FL (2010) Decrypted Secrets: methods and maxims of cryptology, 4th edn. Springer
Publishing Company. https://dx.doi.org/10.1007/978-3-540-48121-8
4. Blum M, Micali S (1984) How to generate cryptographically strong sequences of pseudoran-
dom bits. SIAM J Comput 13(4):850864. https://dx.doi.org/10.1137/0213053
5. Chai Q, Gong G (2012) Veriable symmetric searchable encryption for semi-honest-but-
curious cloud servers. In: 2012 IEEE International Conference on Communications (ICC),
pp 917922, jun 2012. https://dx.doi.org/10.1109/ICC.2012.6364125
6. Elgaby M (2013) Computing on condential character strings in an enhanced SIP-framework.
Masters thesis, University of Hagen. https://dx.doi.org/10.13140/2.1.2059.4241
7. Fahrnberger G (2013) Computing on encrypted character strings in clouds. In: Hota, C., Sri-
mani, P.K. (eds) Distributed computing and internet technology. Lecture notes in computer sci-
ence, vol 7753. Springer, Berlin, pp 244254. https://dx.doi.org/10.1007/978-3-642-36071-8_
19
8. Fahrnberger G (2013) Securestring 2.0 a cryptosystem for computing on encrypted character
strings in clouds. In: Eichler G, Gumzej R (eds) Networked information systems. Fortschritt-
Berichte Reihe 10, vol 826, pp. 226240. VDI Dsseldorf, Jun 2013. https://dx.doi.org/10.
13140/RG.2.1.4846.7521/3
9. Fahrnberger G (2014) A second view on securestring 2.0. In: Natarajan R (ed) Distributed
computing and internet technology. Lecture notes in computer science, vol 8337. Springer
International Publishing, pp 239250. https://dx.doi.org/10.1007/978-3-319-04483-5_25
10. Fahrnberger G (2014) SIMS: a comprehensive approach for a secure instant messaging sifter.
In: 2014 IEEE 13th international conference on Trust, Security and Privacy in Computing and
Communications (TrustCom), Sep 2014, pp 164173. https://dx.doi.org/10.1109/TrustCom.
2014.25
11. Fahrnberger G (2015) Repetition pattern attack on multi-word-containing securestring 2.0
objects. In: Natarajan R, Barua G, Patra MR (eds) Distributed computing and internet tech-
nology. Lecture notes in computer science, vol 8956. Springer International Publishing, pp
265277. https://dx.doi.org/10.1007/978-3-319-14977-6_26
12. Fahrnberger G, Heneis K (2015) Securestring 3.0a cryptosystem for blind computing on
encrypted character strings. In: Natarajan R, Barua G, Patra MR (eds) Distributed computing
and internet technology. Lecture notes in computer science, vol 8956. Springer International
Publishing, pp 331334. https://dx.doi.org/10.1007/978-3-319-14977-6_33
13. Fahrnberger G, Nayak D, Martha VS, Ramaswamy S (2014) Safechat: a tool to shield childrens
communication from explicit messages. In: 2014 14th International Conference on Innovations
for Community Services (I4CS), Jan 2014, pp 8086. https://dx.doi.org/10.1109/I4CS.2014.
6860557
14. Ferguson N, Schneier B (2003) Practical cryptography. Wiley
15. Freitag D (2013) Erweiterung eines SMPP-Frameworks zur sicheren Verarbeitung ver-
traulicher Zeichenketten. Masters thesis, University of Hagen. https://dx.doi.org/10.13140/2.
1.4680.8641
16. Gentry C (2009) A fully homomorphic encryption scheme. Ph.D. thesis, Stanford University,
Stanford, CA, USA. http://crypto.stanford.edu/craig/craig-thesis.pdf
17. Halang WA, Komkhao M, Sodsee S (2014) Secure cloud computing. In: Boonkrong S, Unger
H, Meesad P (eds) Recent advances in information and communication technology. Advances
in intelligent systems and computing, vol 265, May 2014. Springer International Publishing,
pp 305314. https://dx.doi.org/10.1007/978-3-319-06538-0_30
7 A Detailed View on SecureString 3.0 121
Abstract The environment around us is very rich in acoustic information, with its
scope extending beyond mere speech signals. Perception of acoustic information
plays a signicant role in allowing human beings to comprehend sounds that they
hear in their surroundings. This paper focuses on the development of feature
extraction and accurate classication of variety of acoustic sounds in unstructured
environments, where adverse effects such as noise, distortion are likely to dominate.
This work attempts to classify ten different unstructured real-world acoustic envi-
ronments using empirical mode decomposition (EMD) which considers inherent
non-stationarity of acoustic signals by decomposing the signal into intrinsic mode
functions (IMFs). These IMFs are used for feature extraction. This work suggests
utility of composite feature set for classication and proposes an optimized, robust,
best-suitable feature set for classication of the diverse acoustic environments. For
the classication task, Gaussian mixture model (GMM) and k-nearest neighbor
(k-NN) classiers are used. Utilization of this optimized best-suitable feature set
yields the maximum classication accuracy of 100 % with GMM classier and an
average accuracy of 95 % for k-NN (k = 1, 3, 5). Lastly, this study presents the use
and comparison of various performance metrics to evaluate classication tech-
niques used.
S.D. Mainkar ()
Finolex Academy of Management and Technolgy, Ratnagiri, Maharashtra, India
e-mail: smart.extc@gmail.com
S.P. Mahajan
College of Engineering, Pune, Maharashtra, India
e-mail: spm.extc@coep.ac.in
8.1 Introduction
The rapid development in the elds of digital networks and multimedia technolo-
gies during last few years has opened the new paradigm of huge expansion in audio
signal processing. In particular, for a mobile device to be smart, it requires
knowledge about context of the surrounding environment. The introduction of
mobility in the system implies, in turn, that the user may access the system in an
adverse environment where acoustic disturbances may severely degrade the system
performance. Such diversity in acoustic environment arising due to user mobility
can provide the rich source of information about the current context. In this con-
cern, this work suggests the use of surrounding acoustic environment as a hint for
contextual identication. So, this paper focuses on most accurate recognition of
acoustic events, in challenging real-world, unstructured environments.
There are different factors which set this work of acoustic environment recog-
nition distinguishable from the topic of speech recognition. Firstly, the properties of
acoustic environments differ from those of speech, as the frequency content,
duration and prole of the sounds have a much wider variety than those of speech
alone. Secondly, no subword dictionary exists for real-world acoustic sounds in
contrast to the possibility of decomposition of words into their constituent pho-
nemes. Moreover, factors such as noise, reverberation and multiple sources are
possible in real-world unstructured scenarios, whereas research in the domain of
speech recognition has traditionally ignored these, by assuming the use of single
speakers using close-talking microphones.
Practical applications of such a system include smart sensing of acoustic envi-
ronment through intelligent wearable gadgets and hearing aids. Such knowledge
about the surroundings enables the device to provide better service as per users
needs. A mode of operation of a device can be adjusted in accordance with the
context; for instance, a smart phone may automatically switch to silent mode when
the user is at a lecture or in a conference meet, and in contrast it may ring louder
when the person is in noisy place like market, station. Further, a device may also
adjust processing parameters by appropriate judging of the surrounding environ-
ment. For example, some recent hearing aids have multiple equalization lters for
different scenarios but require manual switching between lters to be done by the
user. As an improvement to this, implementation of a proposed system could
smartly identify the surrounding acoustic environment so as to automatically apply
desired settings to the hearing aid.
In the past, most of the work have been done on audio categorization and framing
by using different features and methods. The use of feature extraction matrix is
elaborated in [1] to classify ve different audio categories, namely speech, music,
8 Performance Comparison for EMD Based Classication 125
noisy speech, speech with music and silence. The discrimination of vehicular trafc
noise sources including cars, motorbikes and heavy trucks is done in [2] by using
different spectral features and temporal feature. Three feature vectors for repre-
senting pitch content, rhythmic content and timbral texture of music signals are
suggested and assessed using statistical pattern recognition classiers with 61 %
accuracy in [3] for classifying ten musical genres. Categorization of environmental
sounds is carried out in [4] with the help of Chirplet, curvelet, and Hilbert trans-
forms. In [5], authors have made use of discrete wavelet transform (DWT) to
differentiate between speech and music. In [6], authors have put forth an algorithm
for segmentation and discrimination of audio clips into male speech, female speech,
silence, music and noise. They also have put forward best-suited features for
multiclass classication yielding accuracy of 96.34 % in audio classication.
Background noise sources of four different types are classied using EMD with
discrimination success rate of 77 to 85 % in [7]. In [8], authors have attempted
noise event recognition (car, truck, airplane, etc.) using three classiers. Classi-
cation of ve everyday noise types was done in [9], by comparing multiple
approaches, of which lter bank features proved to have edge over others. In [10],
authors have designed noise classication algorithms by using four pattern recog-
nition frameworks. In [11], authors have done a review on latest trends in recog-
nition of environmental sounds. A context awareness system on the basis of
acoustic signal is suggested in [12] for detection of sound events in ve different
real-world environmental sound categories.
Concerning machine learning methodologies, they have been utilized in various
audio processing tasks, e.g., music classication [13], generalized audio signal
categorization (news, music, sports, etc.) [14] and quantication of speech intelli-
gibility [15]. This literature review identied a need of system design for multiclass
acoustic environment discrimination with the prime target of improvement in
accuracy using EMD to deal with the intrinsic non-stationarity of acoustic signal.
The work presented in this paper constructs on the basis of abovementioned
related works with the following chief contributions. First of all, in this paper, we
put forward unique, optimized, most appropriate composite feature set to make a
distinction between multiple classes of 10 different real-world unstructured acoustic
environments from Diverse Environments Multi-channel Acoustic Noise Database
(DEMAND), achieving maximum classication accuracy of 100 % using GMM
classier and an average accuracy of 95 % by k-NN (k = 1, 3, 5) classier with
Euclidean distance and of 96 % with city-block distance. Another novel work
contributed in this paper is the use and comparison of performance criteria to
evaluate different multiclass classiers used, with the help of confusion matrix.
The further outline of this paper includes overview of EMD in Sect. 8.3. The
basic idea behind proposed methodology is presented in Sect. 8.4. Section 8.5
describes feature extraction and feature selection. The classication methodologies
employed are explained in Sect. 8.6. Experimental results along with database
description are covered in Sect. 8.7 followed by conclusion and future scope in
Sect. 8.8.
126 S.D. Mainkar and S.P. Mahajan
where envelope amplitude ak(t) and phase k(t) are time-varying quantities.
Physical interpretation and mathematical meaning can be extracted by means of the
amplitude and phase. In practice, most of the signals consist of more than one
oscillatory mode and so cannot be treated as IMFs. That means, EMD is equivalent
to numerical sifting process which empirically splits a signal into nite number of
IMFs that are basically hidden fundamental intrinsic oscillatory modes.
The mechanism of sifting technique for extraction of these modes from a given
acoustic time series s(t) can be briefly explained as follows [16]:
1. make out all local extrema (maxima and minima) of s(t);
2. interpolate between maxima (resp. minima) to obtain two envelopes sup(t) and
slow(t);
3. estimate mean envelope e(t) = [sup(t) + slow(t)]/2 and extract the difference
d1 t = st e1 t 8:1
4. check whether r1(t) is an IMF. If it is not an IMF then above steps should be
repeated till we obtain IMF. So, consider d1(t) as new data and iterate on it.
5. Suppose that after completion of such k iterations d1k becomes an IMF. That
means
8 Performance Comparison for EMD Based Classication 127
C1 t = d1k t 8:4
The rst IMF C1(t) comprises of the highest frequency component of the signal.
The residual signal r1(t) is then given by
r1 t = st C1 t 8:5
6. Treat r1(t) as new data and repeat above steps so as to extract all the IMFs. The
sifting procedure is ceased until remainder of signal becomes zero mean or of
monotonic nature as per predecided stoppage criterion given by
T d1k 1 t d1k t2
SDk = 8:6
t=0 d1k
2 t
Fig. 8.1 Generalized outline for EMD-based classication of diverse environmental sounds using
multiclass GMM/k-NN classiers
frame decomposition is done so as to split these frames into number of IMFs. From
these IMFs, different temporal and spectral features are extracted to form composite
feature vectors as described in Sect. 8.5. Finally, classication is done using
Gaussian mixture model (GMM) and k-nearest neighbor (k-NN) classiers, fol-
lowed by performance evaluation, so as to conclude with EMD-based unique
optimized feature set which is best suitable for discrimination of various real-world
acoustic environments. The generalized outline is shown in Fig. 8.1.
For obtaining the characteristics of audio data, the feature extraction is the vital
processing step in multiclass classication of environmental sounds. The primary
goal is to acquire that set of features from the real-world audio clips of interest
which are capable of expressing utmost information regarding preferred charac-
teristics of the actual signal. The analysis of the environmental sound audio clip is
done by using feature extraction. Two major sorts of feature extraction methods
involve: temporal analysis and spectral analysis approaches. Temporal analysis
makes use of time-domain waveform of the audio signal itself, whereas frequency
domain representation of the audio signal is utilized for analysis while dealing with
spectral analysis. IMFs generated by splitting the input audio signals into a series of
analysis windows or frames are used for feature extraction, and nally single value
corresponding to every feature per frame is computed.
In this work, we have selected short-time energy (STE) and zero-crossing rate
(ZCR) as temporal features and spectral flux (SF), spectral roll-off (SR) and spectral
8 Performance Comparison for EMD Based Classication 129
centroid (SC) as spectral features [3]. Further, we also have selected mel-frequency
cepstral coefcients (MFCCs) as one more feature. Further, here, we do experi-
mentations by using composite feature vectors formulated by aggregating different
features for all frames of rst IMF.
8.6 Classication
set to three (k = 3) and classied star belongs to blue polygons class. Second
circle represents ve neighbors (k = 5) considered in classication task. In the
second case, the classication result is an opposite and unknown star belongs to
green squares class [17].
(B) Gaussian Mixture Model (GMM) Classier
In GMM classication, Gaussian mixture model is used for statistical repre-
sentation of environmental audio recordings. The distribution of feature vectors
extracted from these audio clips is modeled by a mixture of Gaussian density
functions (Fig. 8.3).
Complete GMM is dened by mean vector, covariance matrix and mixture
weights. Every identied environment type has its own model which is then used as
its characteristic representation instead of speakers and their utterances to identify
the neighboring atmosphere of the speaker [17]. The identication assignment is
maximum likelihood classier. The main task of the system is to make a decision
whether input acoustic recording belongs to one of the set of environmental sounds,
which are represented by its models. This decision is based on computation of
maximum posterior probability for input feature vector [17].
Fig. 8.4 Performance comparison of GMM and k-NN classiers using ZCR-STE-MFCC as
composite feature set
8.8 Conclusion
References
6. Mahajan SP, Sahu J, Sutaone MS, Kokate VK (2010) Improving performance of multiclass
audio classication using SVM. CIIT Int J Data mining Knowl Eng 2(5):95103. ISSN
0974-9683
7. Jhanwar D, Sharma KK, Modani SG (2013) Classication of environmental background noise
sources using Hilbert-Huang transform. Int J Signal Process Syst 1(1), 6873
8. Couvreur C (1997) Environmental sound recognition: a statistical approach, Ph.D. thesis,
Faculte Polytechnique de Mons, Belgium, June 1997
9. Sawhney N (1997) Situational awareness from environmental sounds
10. El-Maleh K. Samouelian A, Kabal P (1999) Frame level noise classication in mobile
environments. In: Proceedings of the 1999 IEEE international conference on acoustics,
speech, and signal processing, pp 237240
11. Chachada S, Jay Kuo C-C (2013) Environmental sound recognition: a survey, In: Signal and
Information Processing Association Annual Summit and Conference (APSIPA), 2013
Asia-Pacic, pp 19, Oct.-Nov 2013
12. Sivaprakasam T, Dhanalakshmi P (2013) A robust environmental sound recognition system
using frequency domain features. Int J Comput Appl 09758887, 80(9) 510 Oct 2013
13. Bergstra J, Casagrande N, Erhan D, Eck D, Kegl B (2006) Aggregate features and
ADABOOST for music classication. Mach Learn 65(23):473484
14. Dhanalakshmi P, Palanivel S, Ramalingam V (2011) Classication of audio signals using
AANN and GMM. Appl Soft Comput 11(1):716723
15. Li FF, Cox TJ (2007) A neural network model for speech intelligibility quantication. Appl
Soft Comput 7(1):145155
16. Huang NE, Shen Z, Long SR, Wu ML, Shih HH, Zheng Q, Yen NC, Tung CC, Liu HH
(1998) The empirical mode decomposition and Hilbert spectrum for nonlinear and
non-stationary time series analysis. Proc Roy Soc London A 454:903995
17. Wu X, Vipin K (2009) The Top 10 algorithms in data mining, Chapman & Hall/CRC
18. Thiemann J, Ito N, Vincent E (2013) The Diverse environments multi-channel acoustic noise
database (DEMAND): a database of multichannel environmental noise recordings. In: 21st
International Congress on Acoustics, Jun 2013, Montreal, Canada
19. http://parole.loria.fr/DEMAND/
Chapter 9
Performance of Multimodal Biometric
System Based on Level and Method
of Fusion
M. Pathak ()
AISSMS Institute of Information Technology, Savitribai Phule Pune University,
Pune, India
e-mail: mrunalkpathak@gmail.com
N. Srinivasu
K. L. University, Guntur, India
e-mail: srinivasu28@kluniversity.in
9.1 Introduction
For todays scenario, security is major concern. A high-level industry uses bio-
metric authentication systems established on evidence of individual source of
information called unimodal systems [1], which make use of physiological char-
acteristics such as iris, ear, ngerprint, face, teeth, retina, palm print, veins or
signature, voice and gait. as behavioral characteristics [2, 3]. Each biometric has its
own strength and weakness in terms of accuracy, user acceptance and applicability,
and accordingly each biometrics is used in authentication application. The advan-
tage of the present biometric system is that it does not change, gets stolen and
misplaced. In real-world application, none of the individual biometric system is
anticipated to effectively meet all requirements after deploying. Variety of problems
occurs in unimodal biometric system like [4].
(a) Noisy sensor data: for example voice sample altered by cold or ngerprint
with a scar. Due to ambient condition, defective or improperly maintained
sensor or noisy data lead to inaccurate matching or false rejection.
(b) Non-universality: subset of user biometric system may not be able to provide
meaningful biometric data due to users illness or disabilities.
(c) Intra-class variation: This variation is caused by user when sensor character-
istics are changed during authentication or if user incorrectly interacts with
sensor. For example wrong facial pose. False rejection rate (FRR) in biometric
system increases due to large intra-class variation.
(d) Inter-class similarities: It means overlapping of feature space matching to
several users. False acceptance rate (FAR) in biometric system increases if
inter-class similarities increase. For example ideal twins face recognition.
(e) Failure-to-enroll: Unsuccessful attempts to create a template from low-quality
inputs.
(f) Spoof attacks: Submitting fake biometric traits such as face mask, mimicry of
voice or signature causes spoong attack. Due to imitation of data, unimodal
biometric system is vulnerable to spoong.
(g) Restricted degree of freedom: in unimodal biometric system, we are using
features from any single biometric traits such as face, iris and palm which will
restrict the performance of recognition.
(h) Unacceptable error rate: equal error rate (EER) have been decided based on
same value for FRR and FAR. Lower the EER, system will be considered as
more accurate. Otherwise error rate will be unacceptable.
Multimodal biometric system overcomes some of the limitations of unimodal
biometric systems. Multibiometric system utilizes information from one or more
modalities or multiple processing techniques or both. Therefore, multimodal bio-
metric systems are those which integrates more than one physiological or/and
behavioral characteristics for enrollment, identication or verication to improve
the performance and reliability [5, 6]. Some common multimodal biometrics are
9 Performance of Multimodal Biometric System Based 139
face and iris, iris and ngerprints, face and ngerprints, face and voice, face,
ngerprints and iris, face, ngerprint and signature.
In this paper, details of multimodal biometric systems are reviewed in different
sections. Section 9.1 is about introduction. In Sect. 9.2, general biometric system
will be discussed. Section 9.3 will be an introductory section on multimodal bio-
metric systems. This section gives an overview of a selection of well-known
multimodal biometric systems and set-ups that are used by researchers world wide.
Subsequently, Sects. 9.4 and 9.5 will be addressed on overview of level and
methods of multimodal fusion. Operation modes are discussed in Sect. 9.6, where
Sect. 9.7 focuses on issues in multimodal biometrics with its application in
Sect. 9.8. The paper is concluded in Sects. 9.9 and 9.10 with experimental results,
conclusion and discussion on the future directions in multimodal biometrics.
Test
Pre- Feature Template Test
Matcher
Processing Extraction Generator
Application
Sensor Device
Unimodal biometrics system have some of the limitations such as inter-class sim-
ilarities, intra-class variation, noisy data, spoong and non-universality which can
be overcome by multimodal biometric system by synthesizing information from
multiple biometric trait to claim the identity of person [9]. Multiple samples for a
single biometric trait are captured known as multisample biometrics and/or samples
for more than one biometric traits are captured known as multisource or multimodal
biometrics. Identity of claimed person is based on input taken from single or
multiple sensors measuring two or more different modalities of biometric charac-
teristics. To forge single biometric characteristics is more easier than to forge
multiple biometric characteristics; therefore, multimodal biometric systems are
generally much more vital to fraudulent technologies. Hence, these systems are
more reliable, highly accurate and secure in biometric identication as compared to
individual biometric modalities [6]. One of major advantages of this system is that
failure to enroll (FTE) rate is signicantly reduced in multimodal evaluation.
Nowadays multimodal biometric system has been widely deployed and adapted in
various civilian applications such as banking security to check online credit card
transactions and ATM security by providing login privileges. A multimodal bio-
metric system made decision either a genuine individual type of decision or an
142
imposter type of decision. Equal error rate [ERR], false acceptance rate [FAR] and
false rejection rate [FRR] are used to measure the accuracy of system [10].
Several biometric systems are developed in the literature which can be accom-
plished by fusing single trait with multiple representations and multiple matchers,
more than one traits of an individuals, or multiple future extractions and matching
algorithm accomplish on same biometric. Each representation has its own classier
in single trait with multiple representation fusion, and fusion takes place after
reporting score for class by each classier at matching stage. Performance degra-
dation can be avoided by selecting classiers based on their goodness [11]. In the
matching module, multiple matching strategies of biometric system are incorpo-
rated with the help of single biometric multiple matchers, and the performance is
improved by combining scores generated by these strategies [12]. To improve
speed and reliability of biometric systems, matching scores obtained from multiple
biometrics sources are integrated [13].
Various levels of fusion, possible scenarios, modes of operation, integration
strategies and design issues for multimodal biometric system have been proposed
by Ross and Jain (2003). Serial, parallel or hierarchical modes are the operational
modes of a multimodal system. Serial mode forces user to use the modalities one
after another. Therefore, there is no need to acquire information about multiple
sources (e.g. multiple traits) simultaneously. It reduces overall recognition time
because decision could be made before acquiring all the traits. Information from
multiple modalities is used simultaneously in order to perform recognition in the
parallel mode of operation. Strength of multimodal biometric system can be
enhanced by fusioning measurements from different biometric traits [14]. General
multimodal biometrics system can be represented as shown in Fig. 9.3.
Feature
Matching
Fingerprint
, palm System score or
print, iris, database Decision level
face, ear, etc.
Generally there are two phases, i.e. enrollment phase and authentication phase
on which multimodal biometric system can operate. The two phases are described
as follows [15].
Enrollment phase: Biometric traits of a user for two or more modalities are
captured and are stored as a template for that user in the system database which is
further used for authentication phase.
Authentication phase: Biometric traits of a user for different modalities are
captured which are for later use either to identify or verify a person by comparing
templates stored corresponding to all users in database [16].
Multimodal biometric system can be described with the help of following four
modules:
1. Sensor module
2. Feature extraction module
3. Matching module
4. Decision-making module
Sensor module: Biometric traits for different modalities are captured and used as
an input for feature extraction module after performing preprocessing.
Feature extraction module: Compact representation of different biometric traits
or modalities known as Features extracted from different modalities which are
then given to matching module for comparison and also stored as an template in
database.
Matching module: Templates in a database are compared with the extracted
features.
Decision-making module: Decision is taken for user either accepted or rejected
against template stored in database.
Features from various biometric traits are fused to perform multimodal biometric
fusion to enhance the strength and reduce the weakness of each measurement. The
purpose of multimodal biometric fusion is to take out content from a group of input
biometric traits. Multimodal fusion in biometric system is classied into two broad
types [17]: pre-classication and post-classication. In pre-classication, fusion
information is combined before applying any classication method or matching
algorithm. Information is combined after decision of classiers in post-classication
method. Pre-classication fusion uses raw input data from different biometric traits
[18, 19], so that biometric fusion takes place either at feature level (early fusion) or
at data level (sensor level). The post-classication fusion is divided into dynamic
9 Performance of Multimodal Biometric System Based 145
classier selection, abstract level fusion, rank-level fusion and matching score level
fusion [17, 20, 21].
(A) Data Level Fusion
It is the process of integration of multiple data and knowledge representing
more than one signals from a set of same type of modality source (e.g. same
view recorded by two webcams from various viewpoints) without loss of
information into a consistent, accurate and useful representation. It is highly
vulnerable to noise and failures due to the absence of preprocessing.
Data-level fusion is not commonly used because data required for fusion
should be compatible which is rare in biometric sensors, Fig. 9.4.
select classier to prove the correct decision for the precise input pattern [22,
23]. Dynamic selection requires the large data sets for estimating local clas-
sier accuracy. Future-level fusion produces best result when modalities are
related than when they are unrelated [24, 25].
9.8 Applications
Today biometrics has been used in wide variety of applications to provide security,
convenience, privacy enhancement in much more commercial, criminal and civil
applications, for e.g. to provide secure, economical and user-friendly system to
access personal information and perform business transaction against fraud or
information theft.
Multimodal and multiclassier are the different levels used to design multimodal
biometric system. Multiple algorithms or classier are used together to estimate
better results in multiclassier level. Previous research shows different biometric
150 M. Pathak and N. Srinivasu
Table 9.2 Previous research-based comparison of different fusion techniques and its FAR, FRR
and accuracy rate by various authors
Author Fusion Data set FAR FRR Accuracy
(%) (%) (%)
Muhammed et al. Score Face, nger, vein 0.05 0.23 95
Mohammed et al. Score Face, speech 0.087 0.67 96
Dhanashree et al. Feature Palmprint, palm, vein 0.029 1.0 99
Rattani et al. Feature Face, ngerprint 1.98 3.18 98
Bhagat et al. Feature Palmvein, face 0.5 1.0 98
Feifei et al. Score Fingerprint, F.vein 1.2 0.75 95
Krishneswari et al. Feature Palm and ngerprint 1.02 0.9 98
Nazmeen et al. Decision Face, ear 0 4 96
Nageshkumar et al. Score Palmprint, face 2.4 0.8 97
Krzyszof et al. Decision Face, speech 1.1 3.0 87
Mohamed et al. Decision Fingerprint, iris 2 2 98
Lin hong et al. Decision Palmprint, face 1 1.8 92
Gayatri et al. Feature Face, palmprint. 0.5 1.2 95
Mitul et al. Feature Palm and ngerprint 0.2 1.1 87
Jegadeesan et al. Feature Fingerprint, iris 10 5.3 91
and multibiometric systems developed and tested against FAR, FRR and accuracy
to check their performance. Table 9.2 shows FAR, FRR and accuracy of these
systems [41].
The performance of multibiometric system is evaluated on basis of score,
decision and feature fusion. Feature fusion techniques are better choice as compared
to score and decision fusion techniques [41].
References
1. Jain AK, Ross A, Prabhakar S (2004) An introduction to biometric recognition. IEEE Trans
Circuit Syst Video Technol 14(1):420
2. Ross A, Flynn P, Jain AK (2008) Handbook of biometrics. Springer, New York, USA
3. Aggithaya V, Zhang D, Luo N (2012) A multimodal biometric authentication system based
on 2D and 3D palmprint features. Proc SPIE, 6944:69440C-1
4. Ross A, Nandkumar K (2008) A. Jain: handbook of multibiometrics, Springer international
edition pp 271292
5. Telgad RL, Siddiqui AMN, Deshmukh PD (2013) Automated biometric verication: a survey
on multimodal biometrics, Int J Comput Sci Business Inform 6(1) 18. ISSN 1694-2108|
6. Ahuja MS, Chabbra S (2011) A survey of multimodal biometrics. Int J Comput Sci Appl
1:157160
7. Ahuja MS, Chabbra S (XXX) A survey of multimodal biometrics. Int J Comput Sci Appl 157
160. ISSN 2250-3765.
8. Hamming RW (XXX) Error detecting and error correcting codes. Bell Syst Tech J 26(2):147
160, 195
9. Sanjekar PS, Patil JB (2013) An overview of multimodal biometrics. Signal Image Process
Int J (SIPIJ), 4(1)
10. Mishra A (2010) Multimodal biometrics it is: need for future systems. Int J Comput Appl 3
(4):09758887
11. Prabhakar Salil, Jain Anil K (2002) Decision-level fusion in ngerprint verication. Pattern
Recogn 35:861874
12. Jain AK, Prabhakar S, Chen S (1999) Combining multiple matchers for a high security
ngerprint verication system. Pattern Recogn Lett 20:13711379 [24]
13. Ross Arun, Jain Anil (2003) Information fusion in biometrics. Pattern Recogn Lett 24:2115
2125
14. Rodriguez LP, Crespo AG, Lara M, Mezcua M (2008) Study of different fusion technique for
multimodal biometric authentication, in networking and communications. In: IEEE Interna-
tional Conference on Wireless and Mobile Computing
15. Chaudhary S, Rajender Nath (2015) A new multimodal biometric recognition system
integrating Iris, face and voice. Int J Adv Res Comput Sci Softw Eng 5(4):45150. ISSN 2277
128X
16. Golfarelli M, Maio D, Maltoni D(1997) On the errorreject tradeoff in biometric erication
systems. IEE Trans Pattern Anal Mach Intell 19(7):786796
17. Atrey P, Houssain A, Saddik A, Kanakanhalli M (2010) Mutimodal fusion for multimedia
analysis: a survey. Springer Trans Multimedia Syst 16: 345379
18. Sharma R, Pavlovic VI, Huang TS (1998) Towards multimodal Human computer interface.
In: Proc IEEE, 86(5):853860
19. Ross A (2003) A. Jain: information fusion in biometrics. J Pattern Recogn Lett 24(13):2115
2125
20. Zheng Y, Elmaghraby A (2011) A brief survey on multi spectra l face recognition and
multimodal score fusion. In: Proceedings of IEEE International Symposium on Signal
Processing and Information Technology (ISSPIT), Bilbao, pp 543550, 1417 Dec 2011
21. Kaur D, Kaur G (2013) Level of fusion in multimodal biometrics: a review. Int J Adv Res
Comput Sci Softw Eng 3(2)
152 M. Pathak and N. Srinivasu
22. Woods K, Kegelmeyer WP, Jr, Bowyer K (1997) Combination of multiple classiers using
local accuracy estimates. IEEE Trans Pattern Anal Mach Intell 19 (4):405410
23. Giacinto G, Roli F (2001) Dynamic classier selection based on multiple classier behaviour.
Pattern Recogn 34:179181
24. Nadheen F, Poornima S (2013) Feature level fusion in multimodal biometric authentication
system. Int J Comput Appl 69(18):09758887
25. Ross A, Govindarajan R (2005) Feature level fusion in biometric systems. SPIE conference
on Biometric technics for human identication II, vol 5779, Orlando, USA, March 2005,
pp 196204
26. Abaza A, Ross A (2009) Quality based rank-level fusion in multibiometric systems. In:
Proceedings of 3rd IEEE International Conference on Biometrics: Theory, Applications and
Systems (BTAS), (Washington DC, USA), September 2009
27. Siddiqui AMN, Telgad R, Deshmukh P (2014) Multimodal biometric systems: study to
improve accuracy and performance, Int J Current Eng Technol 4(1) ISSN 2277-4106
28. Meva DT, Kumbharana CT (2013) Comparative study of different fusion techniques in
multimodal biometric authentication. Int J Comput Appl 66(19):09758887
29. Garje1 PD, Agrawal SS (2012) Multibiometric identication system based on score level
fusion. IOSR J Electron Commun Eng (IOSRJECE), 2(6):0711. ISSN 22782834
30. Kazi M, Rode Y (2012) Multimodal biometric system using face and signature: a score level
fusion approach. Adv Comput Res 4(1)
31. Ross A, Govindrajan R (2005) Feature level fusion using hand and face biometrics. Proc SPIE
5779:196204
32. Hong L, Jain A (1998) Integrating face and ngerprints for personal identication. IEEE
Trans Pattern Anal. Match. Intell 20(1):12951307
33. Fierrez-Aguilar J, Ortega-Garcia J, Garcia-Romero D, Gonzalez Rodriguez J (2003) A
comparative evaluation of fusion strategies for multimodal biometric verication. In: Kittler j,
Nixon m (eds) Proceedings of 4th International Conference on Audio-video-based Biometric
Person Authentication, vol. LNCS 2688, pp 830837
34. Aguilar JF, Garcia JO, Romero DG, Rodriguez JG (2003) A cooperative evaluation of fusion
strategies for multimodal biometric verication. In: International conference on video based
biometric person authentication, pp 830837, Guildford
35. Razzak MI, Alghathbar MKK, Yusof R (2011) Multimodal biometric recognition based on
fusion of low resolution face and nger veins. Int J Innov Comput Inf Control ICIC Int. 7
(8):46794689. ISSN 1349-4198
36. Indovina M, Uludag U, Snelick R, Mink A, Jain A (XXX) Multimodal biometric
authentication methods: A COTS approach
37. Kisku D, Rattani A, Gupta P, Sing J (2009) Biometric sensor image fusion for identity
verication: a case study with wavelet-based fusion rules graph matching. In: Proceedings of
IEEE Conference on Technologies for Homeland Security, HST 09, Boston, pp 433439, 11
12 May 2009
38. Marcialis GL, Mastinu P, Roli F (2010) Serial fusion of multimodal biometric systems. In
Proceedings of BIOMS, Taranto, Italy, Sept 2010, pp 17
39. Anzar SM, Sathidevi PS (2012) Optimal score level fusion using modalities reliability and
separability measures. Int J Comput Appl 51(16):09758887
40. Asha S, Chellappan C (2008) Authentication of e-learners using multimodal biometric
technology. Presented at the International Symposium on Biometrics and Security
Technologies, ISBAST 2008. Islamabad, pp 16
41. Soruba Sree SR, Radha N (2014) A survey on fusion techniques for multimodal biometric
identication. Int J Innov Res Comput Commun Eng 2(12). ISSN(Online) 2320-9801, ISSN
(Print): 2320-9798
Chapter 10
DSK-Based Authentication Technique
for Secure Smart Grid Wireless
Communication
Abstract The evolution of the smart grid involves development and integration of
proceed information and communication technology model (ICT) into all aspects to
improve and optimize power generation, distribution, delivery, pricing, consump-
tion, and other effectiveness operations. The advanced functionality joined together
with the integration in information technology also comes with many safety and
privacy issues of user-related data in Smart Grid Communication. In proposed
technique, concept of dynamic secret key (DSK) is used to generate symmetric
cryptography keys to design an authentication and encryption scheme in smart grid
wireless communication. In this scheme, recently added device (e.g., smart meter)
is authenticated by a randomly chosen authenticated device. This scheme enables
mutual authentication between a control center situated in local management ofce
and the randomly chosen device as an authenticator to generate proper dynamic
secret-based dynamic encryption key (DEK) for consequent secure data commu-
nications. As stated in [1], a single misjudgment or missing retransmission
sequence (RS) in communication will prevent an adversary from obtaining the
dynamic encryption keys and hence prevent from unauthenticated access. The
transmission failure and packet loss in wireless communication are unpreventable
and cannot be predicted, and so it is not possible for an adversary to trace the
updating of dynamic encryption key (DEK).
M.S. Khedkar ()
Department of Computer Engineering, SESs GOIFE,
University of Mumbai, Mumbai, India
e-mail: mohan_khedkar@hotmail.com
V. Shelake
Department of Computer Engineering, SESs YTCEM,
University of Mumbai, Mumbai, India
e-mail: vijay.shelake@tasgaonkartech.com
10.1 Introduction
As mentioned above, there are various security solutions available for information
networks; however, in this section we summarize some existing and standard
security approaches and their limitations. An encryption technique is essential for
data integrity and condentiality in wireless smart grid communication. Thus, in
this subsection, we will summarize the various existing approaches for Secure
Wireless Smart Grid Communication in Table 10.1.
Since there are various cryptography and security needs of each of different
communication protocols and technologies used in wireless smart grid communi-
cation, employing interoperability between these systems is a critical task. A whole
solution based on this idea is needed. Hence, we took this as a challenge and
proposed a method that is introduced in the next section.
Table 10.1 Various existing approaches for secure smart grid wireless communication
Sr. Title Authors Advantages Disadvantages
no
1 Fault-Tolerant and Scalable Wu and Zhou (1) Proposed key management (1) In case of cross realm secure
Key Management for Smart technique consists of strong security, access, data aggregator may send
Grid, 2012 accessibility, scalability, request to local trust anchor for a
fault-tolerance, and efciency session key for communicating to a
(2) The key management at trust remote data aggregator, and trust
anchors is notably simplied anchor in remote will instead issue the
(3) It does not need to maintain shared actual session key
symmetric keys (2) It cannot be applied for wireless
smart grid communication because of
heavy computation
2 Security Technology for Metke and Ekl (1) It suggests Public Key (1) All suggested techniques in this
Smart Grid Networks, Infrastructure for Security paper increases the complexity in SG
IEEE Trans. Smart Grid, (2) It provides data encryption and (2) It also increases security issues as
vol. 1, pp. 99107, 2010 authentication third party is associated with it
(3) It uses the Standard cryptographic (3) It requires more resources and
technique computation power
3 EPPA: An Efcient and Rongxing Lu, Xiaohui Liang (1) Data aggregation is used as (1) EPPA does not improve the
Privacy-Preserving Xiaodong Lin, Xuemin (Sherman) multidimensional instead of existing efciency of the GW-to-OA
Aggregation Scheme for Shen one-dimensional communication
DSK-Based Authentication Technique for Secure Smart Grid
Secure Smart Grid E.g., It considers amount of energy (2) Is also does not address technique
Communications, 2013 consumed, at what time and for what for multidimensional data aggregation
purpose the consumption was, and so
on
(2) It also considers that power usage
information is often small in size, as
compared to the plain text space of the
encryption algorithm used
(continued)
157
Table 10.1 (continued)
158
Fig. 10.1 Block diagram of proposed method CC: control center, RS: retransmission sequence,
DS: dynamic sequence, TM: transceiver modules, SM: smart meter
where DS(k) is the kth new dynamic secret, Fhash is the hash function, and L_RS is
RS with length L.
C. Dynamic Secret Key (DSK):
In this module, generated dynamic secret (DS) is applied to generate and update
symmetric key at both sides of communication. This DS based updated dynamic
secret key (DSK) is known as Dynamic Secret Encryption Key (DEK). The XOR
function is used for encryption and decryption as a low computational function. The
sender encrypts and the receiver decrypts cipher using DEK(k) as follows [1]:
Authentication technique is the process to achieve data integrity [15]. Any new
device must be veried before added in the smart grid network by remote
authentication server located at the control center (local management ofce). So, in
proposed system, authentication and verication modules play an important role.
The authenticated device such as smart meter can play role of authenticator to
interchange messages between new device, called as supplicant, and authentication
server. Figure 10.2 illustrates this process in detail.
A. Verication:
As shown in Fig. 10.1, verication module veries devices in smart grid at
control center or local management ofce; if device is new it invokes registration
and authentication process. Next section describes the proposed authentication
protocol.
B. Registration and Authentication:
In registration and authentication process, new device, i.e., supplicant and the
authentication server, has a same key K0 which was precongured, not disclosed to
anyone else, even authenticator also not aware of this key K0. Both supplicant and
authentication servers identify each with key K0 and for the consequent data,
encryption/decryption between supplicant and authentication server each updates
K0 by applying concept of DSK as described in Sect. 10.3.1. We denote this
162 M.S. Khedkar and V. Shelake
Dynamic secret key-based demo system is designed and implemented for secure
smart grid wireless communication and authentication using Java ME embedded
emulator. Implemented system secures wireless smart grid communication by
encryption based on dynamic encryption key. It also supports addition of new
device, authentication, and conguration of devices in smart grid.
164 M.S. Khedkar and V. Shelake
10.5.1 Outcomes
In this section from Sect. 10.5.1.1 to Sect. 10.5.1.5, different results of implemented
system on Windows platform are discussed.
10 DSK-Based Authentication Technique for Secure Smart Grid 165
Above Fig. 10.4 screen shows receiver (CC) opens passive connection on a port
and waits for the senders (SM) request. Once connection is established, receiver
receives messages from various smart meters present in the smart grid network.
Another embedded smart device, i.e., receiver (CC or local management ofce),
accepts connection from sender (SM), starts packet transmission, and performs
analysis of each packet, and same RS is generated at receiver. Figures 10.5
and 10.6 prove that same RS is generated at both sender and receiver.
166 M.S. Khedkar and V. Shelake
10.5.1.5 Encryption/Decryption
10.5.2 Analysis
1.5
0.5
0
120 250 350 1000
170 M.S. Khedkar and V. Shelake
60
50
40
30
20
10
0
Trial1 Trial 2 Trial 3 Trial 4
Password Based Authencaon and
Authencaon Encrypon
provided dictionary of terms and trying each word until password matches. To
avoid Brute force, we need to create complex password which not feasible for SG
due to its limitations and constraints. (2) Also in password-based system, we must
store password in database to authenticate users; again we must provide strong
security to this database. It is concluded that authentication with encryption tech-
nique remains stable and superior as it is depending on dynamic and randomness of
the proposed technique. Implemented technique is more secure than
password-based because every time secret key is updated based on retransmitted
packets. Each trial of the percent of the DSK-based technique increases as shown in
Fig. 10.11.
10.6 Conclusion
Acknowledgements The authors would like to thank Dr. S.K. Ukarande, Principal, SESs
YTCEM and former Dean, Faculty of Technology, University of Mumbai and Dr. N.M Shekokar,
H.O.D. (Computer Engineering, DJSCOE) for their suggestions and help for system evaluation.
References
1. Liu T, Liu Y, Mao Y, Sun Y, Guan, IEEE Members (2014) A dynamic secret-based
encryption scheme for smart grid wireless communication. IEEE Trans Smart Grid 5(3)
2. Smart grid cyber security potential threats, vulnerabilities and risks (2009) California Energy
Commission, PIER EnergyRelated Environmental Research Program
3. Yan Y, Hu Q (2013) An efcient security protocol for advanced metering infrastructure in
smart grid. IEEE Netw
4. Li F, Xiong P (2013) Practical secure communication for integrating wireless sensor networks
into the Internet of Things. IEEE Sensors J 13
5. Metke AR, Ekl RL (2010) Security technology for smart grid networks. IEEE Trans Smart
Grid 1(1)
6. Jose EF, Hopkinson KM (2014) A trust-management toolkit for smart grid protection systems.
IEEE Trans Power Deliv
7. Grilo A, Sarmento H, Nunes M, Gonalves J, Pereira P (XX) A wireless sensors suite for
smart grid applications
8. Wu D, Zhou C (2011) Fault-tolerant and scalable key management smart grid. IEEE
9. Lu R, Liang X (2013) EPPA: an efcient and privacy-preserving aggregation scheme for
secure smart grid communications. Trans Parallel Distrib Syst
10. Rongxing MW, Zhang LK (2013) PaRQ: a privacy-preserving range query scheme over
encrypted metering data for smart grid. IEEE Trans Emerg Topics Comput 2013
11. Saranya SK, Karthikeyan R (2014) Security for smart distribution grid by using wireless
communication. IEEE
12. Zahurula S, Mariuna N, Grozescub V (XX) Development of a prototype for remote current
measurements of PV panel using WSN
13. Al-Anbagi I, Hussein T (2014) A reliable IEEE 802.15.4 model for cyber physical power grid
monitoring systems. Emerg Topics Comput Mouftah University of Ottawa
14. Fouda MM, Md. Fadlullah Z (2011) A lightweight message authentication scheme for smart
grid communications. IEEE Trans Smart Grid
15. Nicanfar H, Jokar P (2014) Efcient authentication and key management mechanisms for
smart grid communications. IEEE Syst J 8
Chapter 11
A Smart Security Framework
for High-Risk Locations Using Wireless
Authentication by Smartphone
Abstract The development and urbanization occurred in past decades led to the
concept of smart cities. These smart cities consist of various high-risk locations, so
there is a need of smart authentication technique which is more reliable and secured.
We proposed smart security framework with the integration of enhanced existing
technologies. This authentication framework provides security for the high-risk
location, such as government ofce buildings, airport, military bases, and space
stations. The area consists of various Wi- access points creating the geo-fence. As
soon as the person enters into the geo-fence with his smartphone, biometric
authentication is carried out using smartphone and accordingly the person is
allowed to enter inside the geo-fence. At the same time, Wi- positioning system
locates the path of person in the geo-fence. Biometric authentication uses the local
binary pattern matching algorithm having accuracy of 80 %. The location of the
person inside the geo-fence is tracked with the help of Trilateration method. The
Wi- positioning system based on Trilateration method nds the location of the
person with an error rate of 5 m, which is relatively very lower than the Global
Positioning System. This wireless authentication framework provides the better
solution for security of smart cities which in turn will be benecial for maintaining
the peaceful and healthy social relations.
A. Chandavale ()
SMIEEE, Department of Information Technology, MIT College of Engineering, Pune, India
e-mail: c.anjali38@gmail.com
V. Dorlikar
IBM, Pune, India
e-mail: v.vyenkatesh@gmail.com
11.1 Introduction
With the aim of providing peoples better place to live, the concept of the city was
born. Cities are developing quickly in all over the world; this urbanization intro-
duces the concept of smart cities and simultaneously makes them a necessity. Smart
cities consist of various communication centers, research centers, and space centers.
With the aim of providing them a smart security solution, we proposed this
framework. The proposed smart security framework is the novel approach based
upon the concept of geo-fence and Wi- positioning of the person. The proposed
system consists of several nodes such as a base station, sensor node, and the cluster
member. Smartphone nodes are generic wireless devices having smart capabilities.
Sensor nodes in a cluster communicate among each other and also communicate
with the cluster head. Cluster heads are more resource rich than cluster members.
In this framework, we use smartphones for the authentication, basically the
camera of the smartphones. The smartphone consists of an application installed in
it, which is automatically responding when the smartphone enters into the specic
area (military area/space stations/or research) (Fig. 11.1).
Those geographical areas covered by the Wi- range creating the virtual
boundaries called geo-fence. Those Wi- nodes also refer as sensor nodes, and the
base station is the server maintaining the authenticating activity. One intranet Web
site used to maintain authentication activity should be there which is managed by
administrator having various access rights. Based upon the access control policy,
person is either granted or denied to use the area.
Section 11.2 presents the related work in the area of authentication system.
Section 11.3 presents a methodology of the proposed framework. Section 11.4
Fig. 11.2 Different antenna orientation approaches (a) Omni-directional approach (b) Angle of
arrival approach (c) Min overlap approach
176 A. Chandavale and V. Dorlikar
packet reception rate threshold of <70 % and >30 % loss outside the region and a
threshold of >90 % and <10 % loss within the region [2].
Limitations to create a geo-fence with GPS mainly are: It cannot restrict the GPS
for the smaller area within the few square meters and steerable antennas increase the
overall cost. To overcome these limitations, framework proposed by us uses the
omni-directional approach to create a virtual boundary with the help of Wi- access
points (AP). Wi- APs are restricted according to areas, which also works great in
indoor locations where GPS fails. In omni-directional approach, steerable antennas
are not required so it automatically reduces the cost.
(B) Biometric Capture and Recognition
The second module is based upon the authentication. Using smartphone as an
authentication device is a challenging task because there are no biometric input
devices available with these systems. The main challenge is to capture the biometric
data from the smartphone by minimizing the surrounding noise, collect it, and
compare it with the large databases. A multilevel verication scheme using face,
ngerprint, hand geometry, and other biometric signatures which scans through
camera of smartphones increases the reliability of the system [3]
Face recognition is unique and reliable authentication system which is used with
smartphones. The face recognition module can integrate with 3D face geometry with
the skin texture analysis making it one of the most accurate biometric systems [4]
The available Eigenface and Fisherface approaches are scattered are maximized
within the classes as well as between the classes [5]. This reduces the accuracy of
Fisherface and Eigenface algorithms. Most of in-class scattering occurred due to
light variation, and it can be minimized by discarding the three most signicant
principal components [6, 7].
(C) Wi- Positioning
Extensive research has happened in Wi- positioning in the decade, and applica-
tions have been developed for the same. Lot of scope for accurate Wi- location
using Internet is available [8].
Wi- positioning algorithms are categorized into two types: RF ngerprint based
and access point position based. Some algorithms use both schemes. In an RF
ngerprint-based algorithm, overall space is divided into grids. Dedicated operators
collect radio scans at every grid. The data gathered from the entire grid construct an
RF signal map in the space [9]. When a smartphone scans radio signals at a
position, the radio scan is searched in the stored signal map. Position of the best
matching scans on the signal map is determined as the device position. This
technique consumes more processing power and time. Approximation and Trilat-
eration methods come into the category of access point-based algorithm. Approx-
imation algorithm is less accurate than among these algorithms [10].
Trilateration-based algorithm consumes less processing power, but less accurate
than RF ngerprint-based algorithm. Wi- Trilateration accuracy is generally
between 10 and 20 m; radio nger printing positioning shows an accuracy of at
least 510 m in the investigation and depends on the strength of Wi- in geo-fence
zone [11, 12].
11 A Smart Security Framework for High-Risk Locations 177
End
178 A. Chandavale and V. Dorlikar
searching time at server side, IMEI number is searched rst; then calculated binary
patterns are compared with the existing patterns. Face recognition is done with the
help of LBP algorithm, where face pattern is calculated in binary format.
If the face authentication fails 3 times, then smartphone application issues the
notication for PIN (Personal Identication Number) authentication. PIN is gen-
erated at the time of persons registration by using random number generation
algorithm (Figure 11.4).
After successful authentication, the next module for Wi- positioning started for
monitoring the persons activity inside the geo-fence. The values of RSS are
converted into the real-time distance in meter. Minimum three access points are
required to execute the Trilateration algorithm. The location values are calculated in
(X, Y) coordinate with respect to 4th quadrant. The data (name, IMEI, email id,
PIN, and location coordinates) related to person are stored in the Google App
Engine database. This database works on local host.
The extensive approach using various means such as person identication, face
recognition, and location tracking in geo-fence is a secured and reliable approach
for implementation in the eld of computer science.
(A) Creating Geo-fence
We are creating geo-fence using an omni-directional approach instead of a tradi-
tional directional antenna approach to minimize the implementation cost. The
proposed method uses overlapping coverage to create geo-fence [1]. Physical SSID
of all the APs inside the geo-fence is registered in application code.
Figure 11.5 shows the overview of creating geo-fence. Four Wi- APs of same
specication are used to create geo-fence around specic geographic area. All these
access points are registered while programming with the help of their SSID. When the
Wi- coverage of registered APs is found, the authentication procedure starts [13].
The algorithm for creating geo-fence is given below:
Step 1: While Wi- state is enabled.
Step 2: Search for available Wi- access points.
Step 3: For count = 0, set the initializing flag = false and increment the counter
according to number of Wi- SSIDs say n.
Step 4: Check the available Wi- SSID, if Wi-Fi SSID is present inside the code is
same as available SSID then set flag = true and run the wireless authen-
tication application.
Step 5: If expected SSID not present till count = n. //Expected SSID is already
present inside the code. n is the total number of available access points
Step 6: set count = 0. Go to step 2.
(B) Biometric Signature Capture and Recognition
In biometric signature recognition, minimum 1.3 megapixel camera is required to
capture the facial image. The framework supports face recognition and PIN for an
authentication. IMEI number is used for indexing of various face binary patterns.
Local binary pattern histogram algorithm is used for face recognition. The local
binary pattern methodology has its roots in 2D texture analysis. The basic idea
behind local binary patterns is to summarize the local structure in an image by
comparing each pixel with its neighborhood. Input image of size 200 200 pixels
is divided into 7 7 matrix zones. In each zone, intensity of the central pixel is
measured. If the intensity of the central pixel is greater than or equal to its neighbor,
then assign the value as 1; otherwise, assign value as 0. That ends up with a
binary number for each pixel. This process continues for all zones and it gener-
ates local binary patterns or sometimes referred to as LBP codes [14].
p1
LBPXc, Yc = 2p SIp Ic
p=0
Step 3: Align the number of neighbors on a circle with a variable radius in pixel,
which enables to capture the following neighborhoods:
Step 4: For a given point (Xc, Yc), the position of the neighbor (Xp, Yp), p P is
calculated by:
2p 2p
Xp = Xc + Rcos Yp = Yc Rsin
p p
where R is the radius of the circle and p is the number of sample points.
Step 5: If a points coordinate on the circle does not correspond to image coor-
dinates, then the point gets interpolated using bilinear interpolation [15]:
f 0, 0 f 0, 1 1y
f x, y = 1 x x
f 1, 0 f 1, 1 y
Here, face of the person is captured by the application in PNG format. This
image is captured in RGB32 format that means each and every pixel size is 32 bits
and then it is converted to the grayscale8. The dimension of captured facial image is
200 200 pixel. The binary code is calculated for each image and tagged to the
image at the time of storing the image into database.
11 A Smart Security Framework for High-Risk Locations 181
The local boundary pattern histogram algorithm is used for face recognition.
This algorithm is a supervised learning based, where the LBP code of facial image
is compared with the existing LBP code in database. The captured color image is
converted into a grayscale image before calculating binary code. LBP code is
calculated by dividing the facial image into the grid of 7 7 and accordingly 49
zones are created. For every zone, the LBP code or binary code is calculated using a
circular neighboring algorithm. The set of LBP code for single image is further
referred as a pattern or byte code.
(C) Wi- Positioning
To implement a Wi- positioning system, dedicated access points are respon-
sible for gathering radio scans with ground truth data. Off-the-shelf smartphones
have the capability to scan radio signals from Wi- access points (APs) [4]. We
propose to use Trilateration-based scheme to construct Wi- positioning system
(Fig. 11.6).
Trilateration algorithm:
Step 1: Let P1(X0, y0, r0), P2(X1, y1, r1), and P3(x2, y2, r2) be three access points
where (x, y) are the known position of access points and r be the radius.
Step 2: Calculate dx and dy are vertical and horizontal distances between the AP
centers.
dx = x1 x0
dy = y1 y0
11.4 Implementation
For implementation of this task, the imported packages are given below:
In Wi- positioning, static access points are used to calculate position of the
person inside the geo-fence. The standards taken into consideration are as follows:
When distance and frequency are in centimeters and kilohertz, the constant
becomes 87.55;
184 A. Chandavale and V. Dorlikar
When distance and frequency are in meters and megahertz, the constant
becomes 27.55;
When distance and frequency are in kilometers and megahertz, the constant
becomes 32.45;
These are the constant values to convert RSS value to centimeters, meters, and
kilometers. There should be minimum 3 access points to execute the Wi- posi-
tioning. The access points are registered during the code with their distance in any
of the quadrant (distance given below is in the 4th quadrant).
Source code of module 3: Wi- positioning.
In the android code given below, known static routers are registered while
coding with their distances in 4th quadrant.
The code given below shows how communication server is registered. The
address 192.168.0.15 is the IP address of access point through which server can
communicate with other access points and smart phone devices.
Figure 11.8 shows the working of backend module. After running the server
successfully in Android Studio, the local host with port number 8888 starts working
as backend Web page. In the same Web page, admin can see the registered devices
and their locations.
image. The smartphone is connected to server through the set of Wi- access points.
Server computer consists of facial image database and persons information which is
accessible with the help of Web browser.
(A) Geo-fence
Figure 11.10 below delineates distance as a function of received signal strength.
The received signal strength value goes on decreasing as the distance increases.
The RSS value is recorded with the help of Wi- analyzer testing tool. The results
differ from access point to access point. The experimental access point taken into
consideration is D-Link Dir 605L (2.4 GHz).
Table 11.1 LBP zones Zones Accuracy (%) Time (ms) Space (bytes)
3X3 60 257 198
5X5 70 325 329
7X7 80 423 524
9X9 96 736 780
11 A Smart Security Framework for High-Risk Locations 189
the face matching time done at server end. And space given in the table shows the
size required for each face block.
Comparison between accuracy of different face recognition algorithm:
Figure 11.14 shows accuracy in percentage of various face recognition algo-
rithms with respect to increasing number of face training set. Among all the
algorithms, LBP 7 7 gives the best result [14].
Recognition Versus Face Quality
As per results shown in Fig. 11.14, we executed LBP, Eigenface, and Fisherface
algorithm and compared the results of Eigenface and Fisherface with LBP7 7
algorithm. In Fig. 11.15, the accuracy of LBP7 7 is 1020 % more than
Eigenface and Fisherface algorithms (Fig. 11.16 and Table 11.2).
0 1 2 3 4 5 6 7 8 9 10 11
Table 11.2 Comparison result for Eigenface, Fisherface, and LBP algorithm
Algorithm Total no. Total no. of Total no. of Total no. of facial
of facial unrecognized incorrectly images correctly
images facial images identied facial recognized
images
Eigenface 100 14 21 65
Fisherface 100 13 17 70
LBP 100 10 10 80
11.6 Conclusion
The smart security concept is gaining popularity among various cities in the world.
In this paper, novel approach is proposed for wireless authentication with the help
of smartphone and biometric features of person. The accuracy of face recognition
algorithm based on binary pattern histogram technique is found to be 80 %. The
192 A. Chandavale and V. Dorlikar
Wi- positioning Trilateration method works with the approximate error rate of
5 m, which is relatively very smaller than Global Positioning System. With the
advancement in technology and widespread use of smartphones, it is found that
various techniques are developed for authentication. Here integration of two or
more techniques provides more secure and reliable framework. The proposed
approach is independent of android versions and applied on various smartphones
without requiring dedicated biometric scanners.
References
Abstract Medical image processing can be done through various methods, but it
may take long time for processing it. As the processing time is increased, it delays
the report generation which can affect patients life. So to scale down the medical
image processing time and without getting arousing effect on the quality of image,
high computational medical image processing methods are being used. Here in the
proposed system, to deal with the medical image, different theorems are used
(image registration, image segmentation, image de-noising). Image registration
regulates the image, and the image is then segmented into identical structures using
image segmentation. The image may have different types of noise in it. So, to expel
this noise image de-noising will be used. As the processing will be done in indi-
vidual environments (CPU and GPU), the efciency in both the situations will be
analyzed. Reasonably, GPU will be the capable environment for medical image
processing. This proposed system can be used to process different techniques like
MRI and CTScan.
12.1 Introduction
the eld of medical image processing [1, 2, 4, 5], full computation power is needed.
To achieve high accuracy and real-time performance, shared memory is used and
the parallel architecture like CUDA [13] (compute unied device architecture) is
also used for it. Notable anomalies include reviews centralized on specic medical
imaging algorithms such as image registration [4, 12], segmenting [1, 3] the image
and de-noising [5, 8] the image.
Here, we consider an extent of medical image methods and types of theorems
and basic functions. The aim of this proposed system is to give analysis of medical
image processing on GPU and CPU for introducing parallel programming
approaches [6] to many algorithms.
Focus of proposed system is to use GPU [13] and shared memory environment
to accelerate image processing. The front end of proposed system has a medical
image (MRI-medical resonance imaging) which will be processed through the
different image processing algorithms to obtain high-quality processed image.
While processing the image it will undergo different process like aligning the image
(i.e., image registration), segmenting the image into similar structures (i.e., image
segmentation) and image de-noising. The efciency of this proposed system will be
evaluated separately using high computational languages (GPU and CPU envi-
ronment [6]). Further considering the most efcient system, regarding
high-performance system environment will be concluded.
Medical image [9, 10] data that are collected with different medical imaging devices
such as MRI [16], CT scan (computed tomography scan) used for image diagnosing
by different practitioners. Processing this medical image by using CPU may take
large amount of time, which results delay in report generation, due to this may
cause adverse effect on patients life. So there is need to increase the efciency and
performance of medical image processing without affecting the quality of image.
Medical image analysis involves interleaving numerous methods, consisting of
different algorithms for processing. Processing the algorithms over a high com-
putational processor and without compromising the quality of image, the execution
time will be speedup. To check the efciency of algorithm, it has been analyzed on
two different architectures (i.e., CPU and GPU).
12.4 Methodology
In recent few years, image processing is playing major role in medical eld. But,
the overall process that is done during medical image processing [7, 15] is very time
consuming. So, to reduce the time consumption, the medical image processing is
done with the help of GPU and the shared memory.
Image processing consists of many algorithms through which the image is pro-
cessed, but in medical image processing the basic algorithms are used that are image
pre-processing (image acquisition), image registration, image segmentation and
image de-noising. These algorithms are basic algorithms of image processing. All
these algorithms consist of some methods through which the image processing is
done. The methods and the basic algorithms [11] will be discussed further in detail.
nothing but retrieving the data from the RF signal parameters. Performing image
acquisition is the rst step in image processing (this is because without an image no
process can be done). The magnetic eld is being changed during the acquisition
process as the distinct physical nature of tissue, or fluid or spectroscopic posses-
sions related to molecular structure, that is visible in the image. An image consists
of an acquisition cycle that is repeated many times. The tissue magnetization is
forced through a series of changes during each cycle. All the tissues and fluids do
not progress through these changes at the same rate. At the end of each cycle that
determines the intensity of the resulting tissue brightness in the image. It is the level
of magnetization on that is present at a special picture snapping time. The basic
types of image in MRI are proton density (PD), T1- and T2-weighted images. In
proton density, it produces strong signals or a bright appearance where the density
of protons is more. T1 and T2 describes or characterize the image by Magnetic
Relaxation time in this the tissue is flipped into the unstable state and the time
taken for the tissue to get back into the normal state is calculated and hence
differentiate into normal and unstable tissues or a part of the body. All these types
are used in a combined way and not only one type to gain more accuracy.
12.4.3 Algorithm
There are many different algorithms used for medical image processing. This
algorithms are image pre-processing in which image acquisition is done, image
registration which is done to align two images that are taken from two different
locations or at two different times. Image segmentation through which a particular
part may be a tissue or a membrane in body can be extracted from the MRI image.
Image de-noising is a noise reduction algorithm in which the noise like pepper noise
and salt noise, shot noise, impulse noise is removed using different methods in
image de-noising (Fig. 12.1).
The basic algorithms are as follows
Image acquisition
Image registration
Image segmentation
Image de-noising
The rst process that takes place in image processing is image pre-processing,
i.e., image acquisition.
Image acquisition is the rst process in the image processing in which the image is
retrieved from the source or hardware and stored temporarily in the processing.
Carrying out image acquisition is always the rst step in the work-flow cycle
12 High Performance Computation Analysis for Medical Images 197
Image registration, a process of aligning two images taken at two different times or
angles. To register two images means to line up them, so that common features
coincide and dissimilarities between them should be emphasized and readily visible
to the naked eyes. Image registration species to the process of considering two
images taken at different times.
While registering images, we are arranging a geometric conversion which
adjusts one image to t another. Basic image subtraction does not work for a
number of reasons. MRI images are assembled according one slice at a time.
Chances are that the slices from the two volumes are not parallel when correlated
with a 6-month-old MR volume with one captured yesterday. As a result, the
context would be different. Consider a right cylindrical cone. A plane slicing
through the cone, parallel to its base, forms a circle. If the slice is slightly off
parallel, an ellipse results. So for image registration different algorithms are used.
One of the algorithm which is efcient and suitable to medical image processing is
Fourier-based correlation using FFT. This algorithm is integral part of the
frequency domain method.
198 A. Shewale et al.
3. By, taking whole square (Norms) and power spectrum translation will be
removed.
Image segmentation algorithms are designed for segmenting the image to similar
structures or having similar features. Image segmentation is to arrange an image
into assorted sectors according to the feature of image, for example, the frequency
response. Up till now, lots of image segmentation algorithms exist and be abun-
dantly applied in day to day life. According to image segmentation method, we can
classify an image into region-based segmentation, region growing, region splitting
and merging, data clustering and edge-based segmentation.
In the survey of medical image processing, various image segmentation algo-
rithms are found to be good or efcient in their working. We cannot process the
whole image directly for the reason that it is incompetent and idealistic, thus image
segmentation algorithms are used for some applications, such as image recognition
or compression. It can identify the regions of interest in a scene.
For image segmentation, clustering method is most benecial for segmenting the
image. In clustering method, K-means clustering algorithm is most efcient
algorithm.
The K-means clustering algorithm includes following steps:
(1) Composite the N number of clusters required in the nal segregated result and
set them a number, and unconcerned select N patterns in the whole database as
N centroids for N clusters.
(2) Analyze the each pattern to nd the closest cluster centroid. The adjacent
centroid usually checks the pixel values similarity, but still it may also
consider the other features also.
12 High Performance Computation Analysis for Medical Images 199
(3) Associate the cluster centroids and then they all will have N centroids for N
clusters as we do in the step 2.
(4) Repeat the steps 2 and 3 to match the concurrency criterion.
(5) The typical convergence criteria are nothing but no realignment of any pattern
from one cluster into another, or the minimal decrease in squared error.
Usually noise is introduced in the image during image transmission. The added
noise can be in various forms such as shot noise, impulse noise, structured noise,
salt and pepper noise, Gaussian noise. Depending on the type of the noise, the
degradation of the image will differ. The conventional methods of noise removal
include clustering, NLMeans lter, Gaussian ltering, Anisotropic PDE (Partial
Differential Equation), Rudin-Osher-Fatemi TV, shrinkage models and different
transforms. Wavelet thresholding, isotropic diffusion wavelet multi-frame,
de-noising anisotropic diffusion, Bayesian estimation de-noising, and curvelet
transform and wave atom transform are the efcient ways for image noise reduction
algorithms. The noise removal algorithm must be picked according to the per-
centage of image quality reduction.
Most of the noise removal techniques suggested till now depend on what type of
noise is introduced. The required noise reduction algorithm is used to which the
image and video are based on. Gaussian ltering is considered as a perfect blur
for many applications, which provides kernel support that is large enough to t the
essential part of the Gaussian. In case of 2D ltering, it can be fragmented into a
series of 1D ltering for rows and columns. When the lter radius is narrow then
the direct 1D convolution is the fastest and efcient way to calculate the ltering
result. First of all, it is observed that the result of convolution has a length
N + M 1, where N is the signal size and M is a lter kernel size (equal to 2r + 1),
i.e., the output signal is longer than the input signal. Secondly, calculating FFT of
the complete image row is not superlative, as the complexity of FFT is ON log N.
By breaking the kernel, the complexity of FFT (fast Fourier transform) can be
reduced into sections with an approximate length M and performing overlap-add
convolution section-wise. When the size of FFT is considered as F and is selected
as the smallest power of 2 larger than 2M, and signal section size is selected as
FM + 1 for full utilization of FFT block then the superlative performance is
achieved. This reduces the overall complexity of 1D convolution to ONorm, thus
complexity of Gaussian blur per-pixel becomes Olog r, so for many practical
purposes Gaussian blur can be successfully implemented with simpler lters as the
value of constant is quite large.
200 A. Shewale et al.
12.4.4.1 GPU
CUDA stands for compute unied device architecture, and it is a new hardware and
software architecture for subjecting and conducting calculations on the GPU as a
data-parallel computing device without the constraint of mapping them to a
graphics application programmable interface. The operating systems multi-tasking
mechanism is bounded for managing the access to the GPU by several CUDA and
graphics applications running synchronously. Here by using CUDA, the process of
medical image processing will be done faster (Fig. 12.2).
The proposed system consists of MR image as an input image. The images will be
processed with the help of core image processing algorithms that are image reg-
istration, image segmentation, image de-noising. The whole processing will be done
on two different environments that are shared memory (CPU) and GPU (CUDA).
The analysis of both the environment will make more pre-sized decision regarding
the efciency of system. In the proposed system architecture, the system takes the
input from MRI machine. The input image undergoes three image processing
algorithms. Image registration, image segmentation and image de-noising. Image
registration is the most common algorithm in medical imaging. One reason for this
is GPUs hardware support for linear interpolation, which makes it possible to
transform image and volumes very efciently. Whereas image segmentation is often
used for segmentation of brain structures, blood vessels, organs, tumors and bones.
It immediately compares multiple candidate segmentation algorithms, once the
segmentation algorithm has been well established it works on massive datasets. It
also performs interactive segmentation and visualization.
In case of image de-noising, the main aim is to improve overall image quality.
Image registration and segmentation algorithms can also asset from a decreased
noise level. After ionizing radiation in MRI image, we get processed high-quality
image (Fig. 12.3).
202 A. Shewale et al.
Start
Obtain MRI image from the web by image acquisition.
Study the characteristics of MRI.
Implement image registration using Fourier-based correlation algorithm.
Implement segmentation by K-means clustering algorithm.
De-noise the image using Gaussian lter theorem algorithm.
Analyze the performance for the system.
Stop
12.5.2 Flowchart
The mentioned flowchart gives the brief idea of the whole process that takes place
in the medical image processing. In the gure, the process starts from the basic
stage of image processing that is image acquisition that is image retrieval and then
after studying its characteristics the main image processing begins with the image
registration, image segmentation and image de-noising algorithms used for effective
image processing. The overall process takes place in two different environments,
and then the performance is calculated in both of them. Whichever will be more
efcient will be concluded as the best system in both of the environments.
As shown in gure, we can show how the whole sequence of the system works
over a time.
12 High Performance Computation Analysis for Medical Images 203
Here for describing the overall medical image processing, the class diagram is
drawn in which the basic three algorithms are shown as the three different classes.
All the three classes perform the medical image processing algorithms in which
different methods are being used like Fourier-based coefcient method for image
registration, K-means clustering method segmenting the image into different
structures and for de-noising the image, Gaussian ltering method is used.
204 A. Shewale et al.
The class diagram for the proposed system is given above which gives the
description as follows:
Here for describing the overall medical image processing, the class diagram is
drawn in which the basic three algorithms are shown as the three different
classes.
All the three classes perform the medical image processing algorithms in which
different methods are being used.
The rst class in class diagram is Fourier-based coefcient method for image
registration which consists of Fourier transform, translations and rotations and
properties of Fourier transform.
Second class is K-means clustering method used segmenting the image into
different structures, which consist of clusters in it.
Then the third class is for de-noising the image, i.e., Gaussian ltering method
which consists of FFT and Gaussian lter (Fig. 12.5).
One view of analysis modeling, called structural analysis, considers data and the
processes that transform the data as separate entities. Data objects are modeled in a
way that de nes attributes and relationships. Processes that manipulate data objects
12 High Performance Computation Analysis for Medical Images 205
are in a manner that shows how they transform data as data objects ow through the
system. Let I be the set of input containing image and its properties.
I = fI1 , I2 , I3 , . . . , In g 12:5
O = f o1 , o 2 , o 3 , . . . , o 4 g 12:6
Consider Venn diagram for the given proposed system (Fig. 12.6).
Where
where
G = GPU Environment
C = CPU environment
F = set of functions
F = ff1 , f2 , f3 , f4 g
Where G C = F
Let f1 = Image Acquisition
f2 = Image Registration
f3 = Image Segmentation
f4 = Image Denoising
f1 = Image Acquisition
{
Taking Fourier transform of given images
206 A. Shewale et al.
}
f3 = Image Segmentation
f3 = Image Denoising
{
The normal distributions density function is called as Gaussian function. The
Gaussian function format for one dimension is given as follows:
1
f x = p e x2 2 2
2
Gx, y = 1 2 2 e x + y2 2 2
2
With the above two-dimensional function, we can calculate the weight of the
point.
Where is the standard deviation for the given distribution. And the mean for
the distribution is assumed to be 0.
The Gaussian function is used in different research areas:
It denotes a probability distribution for noise or data.
It is said to be a smoothing operator.
It is also used in mathematics.
There are some important properties of Gaussian function which are then veri-
ed by its integral function as which is given as follows:
p
I = exp x2 dx =
O = fo1 , o2 , o3 , . . . , on g
Processed image
12.7 Conclusion
Hence the high-performance computational analysis will be done with the help of
high computational languages. Here for analysis, the three algorithms are used that
are image registration, image segmentation and image de-noising. These algorithms
are the basic algorithms in the medical image processing. In image registration,
Fourier-based coefcient method is used for aligning the images. In image seg-
mentation, K-means clustering algorithm is used for segmenting the image into the
similar structures. For de-noising, the image Gaussian theorem is used. Using all
these methods and algorithms, the system performance is analyzed. After the per-
formance analysis, which system performs better will be concluded.
208 A. Shewale et al.
References
1. Abramov A, Kulvicius TF, Dellen B (2011) Real-time image segmentation on a GPU. In:
Facing the multicore challenge. Lecture notes in computer science, vol 6310, pp 131142
2. Bouman CA (2015) Magnetic resonance imaging (MRI). In: Digital image processing, 12 Jan
2015
3. Brady M, FRS FREng Hilary (2005) Image segmentation and classication. University of
Oxford
4. Buades A, Coll B, Morel JM (2003) A review of image denoising algorithms, with a new one.
Kostelec PJ, Periaswamy S (2003) Image registration for MRI. In: Modern signal processing,
vol 46. MSRI Publications
5. Buades A, Coll B, Morel J-M (2005) A review of image denoising algorithms, with a new
one. SIAM J Multiscale Model Simul A SIAM Interdiscip J
6. Chandra R, Dagum L, Kohr D, Maydan D, McDonald J, Menon R (2001) Parallel
programming in OpenMP. Morgan Kaufmann publishers
7. Eklund A, Dufort P, Forsberg D, LaConte SM (2013) Medical image processing on the
GPU-Past, present and future 17 (Elsevier B.V)
8. Gayathri R, Sabeenian RS (2012) A survey on image denoising algorithms (IDA), 5 Nov
2012. Int J Adv Res Electr Electron Instrum Eng
9. Goswami P (2015) Biomedical image processing, 19 Apr 2015. Published on Slideshare
10. Hanson LG (2009) Introduction to magnetic resonance imaging techniques. Copenhagen
University
11. Kaur A, Singh N (2012) Region growing and object extraction techniques. Int J Sci Res
12. Kostelec PJ, Periswamy S (2003) Image registration for MRI. In: Modern signal processing,
vol 46. MSRI Publications
13. NVIDIA CUDA (2007) Compute unied device architecture-programming guide. NVIDIA
Corporation
14. Smistad E, Falch TL, Bozorgi MM, Elster AC, Lindseth F (2015) Medical image
segmentation on GPUs-a comprehensive review 20 (Elsevier)
15. Zhu H Medical image processing overview. University of Calgary
16. Sprawls P. www.sprawls.org/mripmt/MRI01/index.html. Sprawls Educational Foundation
Chapter 13
Terrorist Scanner Radar and Multiple
Object Detection System
S. Kunjir R. Autee ()
Electronics and Communication Department, Deogiri Institute of Engineering
and Management Studies, Aurangabad, Maharashtra, India
e-mail: rmautee@gmail.com
S. Kunjir
e-mail: supriyakunjir@gmail.com
13.1 Introduction
In todays world, security system is becoming more essential and as per the need it
becomes more advanced. In this rapidly growing world, as technology changing
each and every eld with very rapid speed, simultaneously the requirement of
security also becomes an important need at every place of the world. And changing
technology has done great revolution in the eld of security also. In response to the
recent incidents of terrorism in the nation, it becomes necessary to develop an
advance and fast security system against terrorism for national border. Therefore,
here is an attempt made to have security system for national border against ter-
rorism using ultrasonic sensor network, which helps to protect the nation against
terrorism to a great extent. Proposed system goal is to build a more advance and
sensitive security system which can control terrorism. The development of the
system can be achieved by using ultrasonic radar sensor network. Here by con-
sidering the purpose of security system requires for the national border, the system
is designed such that it continuously scans the national border. A pair of ultrasonic
transmitter and receiver sensors with camera is placed on motor assembly and being
rotated by rotating assembly, i.e., motor in 180 angle using limit switches at
entry-restricted area of national border. This sensor network and camera are then
coupled to counter and display section, which will give us the count of obstacles
found by sensor network and also give photographs which are taken by camera
when any obstacle coming into their scanning path. Further the counter and display
system is coupled with recording and playback system to get voice announcement
at military base camp to get continuous updates of position of the obstacles from
border area. Due to the advantage of image provision, we are able to take suitable
control action against terrorist as per the situation. Figure 13.1 shows the overall
system conguration.
A typical radar technology includes emitting radio waves, receiving their reflection
and using this information to generate data. For imaging radar, the returning waves
are used to create an image. When the radio waves reflect off objects, this will make
some changes in the radio waves and can provide data about the objects, including
how far the waves travelled and what kind of objects they encountered. Using the
acquired data, a computer can create a 3D or 2D image of the target. Radar imaging
for combating terrorism (Grifths and Chirs J. Baker) [1] presented the review
application of imaging radar systems to counter terrorism. In his case, the imaging
radar is used to distinguish targets from the background, by exploiting differences in
signature, and wherever possible making use of prior knowledge. Further infor-
mation may be obtainable through the use of radar polarimetry and interferometer.
There are two distinct stages to this:
(i) The production of high-quality, artifact-free imagery and
(ii) The extraction of information from imagery.
This is followed by a discussion of four specic applications to counterterrorism:
(i) The detection of buried targets,
(ii) Through wall radar imaging,
(iii) Radar tomography and detection of concealed weapon and
(iv) Passive bistatic radar.
Radar is widely used in the application of weapon detection against terrorism.
Inmarsat Global Limited (Bulletin Global Government, November 2013) [2] has
identied the Blighter family of surveillance radars as the key component of a
solution that will enable border agencies to deliver very signicant improvements to
the security of their borders at an affordable cost.
The combination of BGAN with Blighter generates a highly capable, versatile
and mobile capability which fully meets the requirement and offers excellent value
for money. Blighter is the name of a family of modern state-of-the-art electronic
scanning radars, which are designed and built to provide continuous persistent
surveillance at borders, boundaries and perimeters. They detect moving targets over
both land and water, covering a wide area. Both mobile and man-portable variants
are available.
Blighter radars employ electronic rather than mechanical scanning. The elimi-
nation of mechanical scanning provides the highest possible levels of system
availability and reliability. Blighter can detect very slow moving targets, down to
0.4 km/h. This ensures that targets moving almost tangentially to the radar can still
be detected. Advances in sensor technology and computer processing power can
signicantly enhance border security. It is also a highly effective backup system that
can be guaranteed to provide resilient communications in the event of the accidental
or deliberate disruption of a primary terrestrial communications network.
212 S. Kunjir and R. Autee
Ajay Kumar Shrivastava et al. (Nov. 2009) [6] presented the system which gives
the effect of variation in separation between the ultrasonic transmitter and receiver
on accuracy of distance measurement. Distance measurement of an object in front
or by the side of a moving or stationary entity is required in a large number of
devices. To maintain the accuracy of measured distance, the separation between
transmitter and receiver is very important. In this system, the transmitter and
receiver are separated and readings are taken with the distance in the interval of
5 cm. When the distance increases, the error becomes constant and very less.
A correction may be applied to calculate the correct distance.
Ankita Sharma et al. (April 2010) [7] presented a report on ultrasonic 3D locator
which aims to locate the 3D coordinates of a point in space. Ultrasonic signals have
been used for the purpose. The device has two basic modules: location indicator and
position calculator. The location indicator is the transmitter module which is
independent of the calculation unit. The transmitter module sends out a modulated
pulse with carrier in ultrasound range. The receiver module has 5 receivers located
at pre-calculated positions so as to calculate the position of transmitter accurately.
The 3D coordinate of the object is detected based on the time difference of arrival at
the receivers. The transmitter sends pulses to the receivers, and the central unit
calculates the delay between the receivers. Using these measurements, it then
calculates the coordinate of the object (the transmitter). Continuous-wave radar is a
type of radar system where known stable frequency continuous-wave radio energy
is transmitted and then received from any reflecting objects.
A microwave measurement system for metallic object detection using swept
frequency radar (Yong Li et al. 2008) [8] described a microwave measurement
system for metallic object (gun, knives) detection using swept frequency radar. This
system reports on a 114 GHz swept frequency radar system for metallic object
detection using reflection conguration. The swept frequency response and resonant
frequency behavior of a number of metallic objects in terms of position, object
shape, rotation and multiple objects have been tested and analyzed. The system
working from 1 to 14 GHz has been set up to implement sensing of metal items at a
standoff distance of more than 1 m.
Helmut Essen et al. [9] presented two different approaches of concealed weapon
detection with active and passive millimeter wave sensors. The design of a passive
radiometric sensor in the W-band is presented. On the active side, an FMCW radar
system at 94 GHz is introduced for the scanning of persons. Sensors used for security
purposes have to cover the noninvasive control of humans, baggage and letters with
the aim to detect weapons, explosives and chemical or biological threat material.
Those sensors have to cope with different environmental conditions. Preferably, the
control of people has to be done over a longer distance. In times of increasing threat
by terrorist attacks, the control of passengers at airports and stations is one of the
major items. People carrying concealed weapons or explosives or those who have
other terroristic attacks in mind have to be detected under all circumstances.
Image segmentation of concealed objects detected by terahertz imaging (Sheeja
Agustin et al. 2010) [10] presents a method for image segmentation of concealed
objects detection by terahertz imaging. Terahertz radiation is emitted as part of the
214 S. Kunjir and R. Autee
black body radiation from anything with temperatures greater than about 10 K.
A reasonable penetration depth in certain common materials such as cloth, plastic,
wood, sand and soil can be performed by terahertz waves. Therefore, THz radiation
can detect concealed weapons since many nonmetallic, nonpolar materials are
transparent to this type of radiation (and are not transparent to visible radiation). In
this system, the multi-level thresholding method is applied to get the initial seg-
mentation of concealed objects in terahertz images. Then Gonzalez method and
Gonzalez Improved methods are proposed to detect and segment concealed objects
in terahertz images more correctly with specic shape.
Clive M. Alabaster et al. (2013) [11] proposed a virtual target radar system for
small arms re training. It provides the miss distance of a bullet from an aim point
in two axes as the bullet passes through the target plane. The VTR consists of two
radars at ground level separated by a baseline distance, d, which is approximately
2 m. Each radar bore sight is inclined upwards at 45 so that the bore sights
intersect at right angles. The region in which the beams overlap in the plane dened
by the two bore sight vectors denes the targeting plane. The shooter is presented
with an aim point by the intersection of two visible laser beams from a pair of diode
lasers aligned to the bore sight of each radar. A bullet passing through the beams
between set velocity and amplitude limits triggers the recording of the data by each
radar which is processed using inverse synthetic aperture radar (ISAR)-like tech-
niques to extract the range to the bullet. Each radar measures the range to the bullet
as it passes through the targeting plane. This yields the x and y coordinates of the
bullet and hence its miss distance in each plane from the target center.
Naidu (March 2012) [12] develops a system for 3D target tracking named fusion
of radar and IRST sensor measurements for 3D target tracking using extended
Kalman lter. Filters and multi-sensors are used to enhance the target-tracking
capabilities. Radar can measure azimuth, elevation and range of the target. It can
measure range with good resolution, but the angular measurements are not with
good resolution. Radar provides sufcient information to track the target since it
measures both angular position and range of the target. An infrared search and track
(IRST) sensor (sometimes known as infrared sighting and tracking) can measure
azimuth and elevation of a target with good resolution. It can provide only the
direction of the target but not its location because it does not provide the range. The
uncertainty associated with IRST might be represented as a square whose dimen-
sions are comparatively small perpendicular to the measured line of sight. With the
fusion of measurements from radar and IRST, the resultant uncertainty of the
estimated position of the target is smaller than the uncertainty of the estimates with
either measurement alone.
Technology Focus (Bulletin of Defence Research and Development Organisation,
April 2013) [13] is the bulletin of describes different radars with their technology and
applications for defense area. Since the radar is used to detect the target, it is necessary
to get the image of the target so that we can nd out exact scene of the target and able
to take suitable action against it. The radar imaging system gives the images of the
target, but it should not be expected that radar images should look like photographs.
Radar is also used for detection of metallic object and concealed weapon, but other
13 Terrorist Scanner Radar and Multiple Object Detection System 215
than metallic objects it will not detect any other weapons. An application of the radar
also includes determination of the position, distance and speed of the target. Further
effect of variation in separation between transmitter and receiver on the accuracy of
the distance measurement is also studied. To overcome the drawbacks, a specic
target detection arrangement is done in proposed system. In the proposed system,
camera is placed with the radar to capture the images of the target. Thus, it reduces the
problem of photographs of the target. The system provides an image of the target;
however, it needs a specic camera arrangement to be mounted on the radar.
The proposed system terrorist scanner radar and multiple object detection system
which consists of ultrasonic transmitter/receiver pair which is main part of system
which is use to detect the object which comes into its path. Once the object is
detected by ultrasonic transmitter/receiver, its output further sends to counter and
display unit. Counter counts the number of objects which detected by ultrasonic
transmitter, and display unit displays that count. Simultaneously the system gener-
ates a voice signal by using message driver unit and sends that voice signal to the
military base camp using FM transmitter and plays via FM receiver. The system
designed here divides into three main parts: ultrasonic sensor section which is
located at national border area, counter, display and message driver section which
are located at military ofce and voice announcement section located at military base
camp. The proposed system consists of the following blocks shown in Fig. 13.2.
When rotating assembly, i.e., motor starts rotates in 180 angle in forward and
backward direction, the ultrasonic transmitter and receiver and camera which are
placed on motor will also start to rotate. While rotating, it scans the area which
comes into its range and while scanning if it found any obstacle in its scanning path,
at the same time motor stops its rotation and the frequency which is transmitted by
transmitter gets reflected by obstacle and received by its receiver. When receiver
receives echo signal from obstacle, it then send one high pulse to the counter.
Counter gets incremented by one for one high pulse, incremented by two for two
high pulses and goes on increasing for every successive high pulses received from
ultrasonic receiver. One high pulse corresponds to one obstacle. Simultaneously
camera placed on motor will take the photographs of obstacle detected by US
sensors and sends images to the display unit. The display unit consists of
seven-segment display and PC or laptop as display devices.
Seven-segment displays are used to display the count of counter section, and
PC/laptop is used to display the images captured by camera. This total section is
then coupled with message driver unit, where one voice message generates after a
specic count reach by counter. This voice message is then passed to FM trans-
mitter to transmit it to military base camp. In military base camp, there will be FM
receiver and recording and playback assembly is located to play the voice
announcement in military base camp.
216 S. Kunjir and R. Autee
This terrorist scanner radar gives the simplest and more efcient method to detect,
count and identify the terrorist at nation border. As all the other existing system
does not provide exact identication of the terrorist, the proposed will help to
provide analysis of detection and identication of terrorist session.
The system analysis consists of output voltages at different test points before
detection of obstacle and after obstacle detection (Table 13.1).
13 Terrorist Scanner Radar and Multiple Object Detection System 217
Table 13.1 Voltages at different test points before and after obstacle detection
Sr. Test Test point description Voltage available Voltage available
no. point at TP before at TP after
obstacle detection obstacle detection
1 TP1 Across pin no. 4 and pin no. 6 of 0V 11.80 V
IC CA3140 which is output of
ultrasonic receiver
2 TP2 Across pin 1 and pin no. 3 of IC 0V 9.25 V
555 which output of motor
control circuit
3 TP3 Across pin 1 and pin no. 3 of IC 0V 7.53 V
555 which output of counter
circuit
4 TP8 Across pin 1 and pin no. 3 of IC Count below 8 Count above 8
555 which is output of message 0V 10.52 V
driver circuit
Figure 13.3 shows the waveform of output of ultrasonic transceiver when one
obstacle comes into scanning path. It shows that positively going high pulse when
detects an obstacle into its scanning path. Normally its output is at lower voltage
level approximately 6.87 V and when it detects some object into its path, the
voltage level increases positively up to 25.4 V which is given in Table 13.2.
Figure 13.4 shows the waveform of output of ultrasonic transceiver when two
obstacles come into its scanning path. It shows that positively going two high
pulses when detects two obstacles into its scanning path. Normally its output is at
lower voltage level approximately 6.87 V and when it detects some object into its
30
25
20
15
10
0
103
109
115
121
127
133
139
145
151
157
163
169
175
181
187
1
7
13
19
25
31
37
43
49
55
61
67
73
79
85
91
97
30
25
20
15
10
5
0
13
19
25
31
37
43
49
55
61
67
73
79
85
91
97
103
109
115
121
127
133
139
145
151
157
163
169
175
181
187
1
7
30
25
20
15
10
5
0
103
109
115
121
127
133
139
145
151
157
163
169
175
181
187
1
7
13
19
25
31
37
43
49
55
61
67
73
79
85
91
97
Figure 13.7 shows the output of inverter which is required to provide trigger as a
low pulse to the timer in mono stable mode.
13 Terrorist Scanner Radar and Multiple Object Detection System 223
30
25
20
15
10
5
0
103
109
115
121
127
133
139
145
151
157
163
169
175
181
187
13
19
25
31
37
43
49
55
61
67
73
79
85
91
97
1
7
0
1 3 5 7 9 11 13 15 17 19 21 23 25 27 29 31 33 35 37 39 41 43 45 47 49 51 53
-2
-4
-6
Time (ms)--->on horizontal axis and Voltage in (Volt)---> on vertical axis
1
F= 13:1
T
1
F =
11.40 10 6
1
F = 13:2
11.40 10 6
F = 87.719 KHz
13.5 Conclusion
Mainly for security of national border, government is taking a lot of efforts as well
as spending a lot of money. So by employing such type of system such as terrorist
scanner radar and multiple object detection system, we can achieve higher security
at the nation border with very nominal cost. The basic aim of the project is to
provide a security at nation border. This system is basically designed to provide
continuous scanning of entry-restricted area, to nd any obstacle which is coming
into that entry-restricted area and to detect exact kind of obstacle. While scanning
the entry-restricted area given to the system, if it found any obstacle into its path, it
rst detects exact type of obstacle; then, according to type of obstacle necessary
control action takes place. In control action, not only does it start ring but also
additional necessary control action can be taken by military as the FM transmitter
transmits message to military head ofce. The ring time varies according to the
obstacle it detects. Simultaneously FM transmitter transmits the message to FM
receiver which is located at military head quarter so that another necessary action
can be taken by military.
Acknowledgments Completion of this system is a task which would have not accomplished
without cooperation of my guide Prof. R. M. Autee. I would like to thank my parents for their
encouragement. At last, I am also thankful to my friends who have helped me in completion of this
system.
226 S. Kunjir and R. Autee
References
1. Grifths HD, Baker CJ (2006) Radar imaging for combating terrorism. Springer, Netherlands
2. Bulletin Global Government (2013) Border surveillance systems- blighter scanning radar and
sensor solutions
3. Kausadikar RB et al (2013) Terrorist scanner radar with military headquarters informing
system. Int J Comput Technol Electron Eng 3
4. Shrivastava P et al (2014) Project review on ultrasonic distance measurement. Int J Eng Tech
Res. ISSN 2321-0869
5. Dingley G, Alabaster C (2009) Radar based automatic target system. In: IEEE International
WD&D conference, pp. 9781-4244-2971
6. Shrivastava AK et al. (2009) Effect of variation of separation between the ultrasonic
transmitter and receiver on the accuracy of distance measurement. Int J Comput Sci Inf
Technol 1(2)
7. Sharma A et al (2011) Ultrasonic 3D locator. EE318 Electronic Lab Project Report, EE Dept,
IIT Bombay, April 2011
8. Li Y et al (2008) A microwave measurement system for metallic object detection using
swept-frequency radar. Proc SPIE 7117:71170K
9. Helmut E et al Concealed weapon detection with active and passive millimeter wave sensors
10. Agustin SA et al (2010) Image segmentation of concealed objects detected by terahertz
imaging. In: IEEE Conference on Computational Intelligence and Computing Research
11. Alabaster CM, Hughes EJ, Flores-Tapia D (2013) A virtual target radar system for small arms
re training. IEEE-2013
12. Naidu VPS (2009) Fusion of radar and ISRT sensor measurements for 3D Target tracking
using extended kalman lter. Defence Sci J 59(2):175182
13. Bulletin of Defence Research and Development Organisation (2013) Technol Focus 21(2)
Chapter 14
Inexact Implementation of Wavelet
Transform and Its Performance Evaluation
Through Bit Width Reduction
M. Acharya
C.V. Raman College of Engineering, Bhubaneswar 752054, Odisha, India
e-mail: acharya_moumita@yahoo.co.in
C. Pal (B) S. Maity A. Chakrabarti
A.K. Choudhury School of Information Technology, University of Calcutta,
92 APC Road, Kolkata, India
e-mail: palchandrajit@gmail.com
URL: http://www.caluniv.ac.in/
14.1 Introduction
from approximate computing as the input data and the human sense to perceive the
output data are inexact. In embedded systems, the most expensive resource is energy
and its efficient utilization is significant. So designers should be careful about the
tradeoff between the image accuracy and energy consumption.
Image processing is a process to translate an image into digital form and then analyze
and manipulate that digitized image including data compression and image enhance-
ment to improve its quality or to extract some convenient information from it. As
shown in Fig. 14.1 it shows a very simple way to compute the input image in fre-
quency domain (using wavelet transform) to extract the information by using different
processes like bit width reduction, filtering and perform the inverse transformation
to reconstruct the enhanced image. Image is usually treated as a two dimensional
function f (x, y); where x and y are spatial coordinates and the amplitude f at any
pair of coordinates (x, y) is its intensity or grayscale level of the image at that point.
An image is continuous with respect to x and y coordinates and in amplitude. When
x, y and amplitude of f are finite and discrete quantity, the image is known as
digital image. The purpose of image processing is visualization to human eyes,
image sharpening and restoration, multiscale signal analysis, image retrieval, etc.
For digital image processing some techniques that are generally used includes lin-
ear filtering, Wavelets, Pixilation etc. In natural scenes, an image may compose of
multiple resolutions because there is no particular scale or spatial frequency that has
a special status. Due to this reason any visual system may follow some uniformity
in the time of image representation and image processing over multiple scales [13,
14]. Image pyramid is the representation of an image at multiple levels.
Input Image
Input Image Wavelet
Matrix
Transform
Processing
Inverse
Output Image
Output Image Wavelet
Matrix
Transform
The subband coding can decompose a signal (image) into two or more components
that are known as subbands. Here the operation is performed with the combination of
two filter banks named as analysis filter bank and synthesis filter bank formed by one
low-pass filter h0(n) and one high-pass filter h1(n) to decompose the input sequence
f (n) and produces two half-length sequences: approximation subband f lp(n) and
details subband f hp(n) shown in Fig. 14.4. Similarity holds for the synthesis filter
14 Inexact Implementation of Wavelet Transform and Its Performance Evaluation 231
(b) (c)
(a) Approximaion 2 Approximation
Filter Downsampler (Level j-1)
2
Upsampler
(d) (e)
Interpolation
Filter
Prediction
Input -- Prediction
Image +
residual
(Level j) (Level j)
Fig. 14.3 a Image pyramid block diagram; b Original image; c Original image histogram; d
Pyramidal representation of original image; e Pyramidal image histogram
2 C3(n)
H(z)
C2(n)
2
H(z) 2
G(z) D3(n)
C1(n)
H(z) 2
2
G(z) D2(n)
f(n)=Co(n)
G(z) 2
D1(n)
C3(n)
H(z)
2
C2(n)
H(z)
2
G(z)
2 C1(n)
D3(n)
H(z)
G(z) 2
D2(n) 2
f(n)=Co(n)
2 G(z)
D1(n)
bank and produces two subbands f lp(n) and f hp(n) to form the reconstructed
signal f (n) as shown in Fig. 14.5. If the analysis filter bank is recursively applied to
the approximation subband, approximation pyramid is formed. The coefficients Ck
and Dk are produced by convolving the digital signal, with each filter, followed by
decimation of the output.
232 M. Acharya et al.
8X4 DWT
2 LH
G(z)
4X4 LL HL
Apply to Column
Apply to Column LH HH
2 HL
8X8 H(z) 8X 8
2 4X4
Convolution with LPF H(z) and
G(z) H then decimating the output
(b) (c)
Fig. 14.6 a Overall decomposition process for image; b Original image; c Resultant image after
subband coding
Subband coding for a 2D signal (image) may be applied consecutively along the
rows and columns. Figure 14.6 shows the hole decomposition process of subband
coding and its result. Initially the analysis filters are applied to the rows of an image.
It produces two new images. One is a set of common row coefficients and another
one is the set of detail row coefficients. The next analysis filters are applied to the
columns of each new image and produce four different images called subbands
which are the approximations A (LL), the vertical detail V (HL), the horizontal
detail H (LH), and the diagonal detail D (HH) [15]. Rows and columns which are
analyzed with a high-pass filter are designated with an H. Similarly, rows and columns
analyzed with low-pass filter are designated with an L. For an instance, if a subband
image is produced by using a low-pass filer on the rows and a high-pass filter on the
columns, it is called the LH subband. Each subband has different image information.
The LL subband consists of common approximation of the image and removes all
high-frequency information. Similarly the LH subband removes all high-frequency
14 Inexact Implementation of Wavelet Transform and Its Performance Evaluation 233
information along the rows and attenuates high-frequency information along the
columns. As a result the vertical edges of an image get highlighted. Likewise the HL
subband highlight horizontal edges and the HH subband diagonal edges of an image.
Wave is an oscillating function of time and space like sinusoid. The most popular
wave analysis is Fourier transform, which only deals with frequency resolution of
a signal but the temporal details are not present. To overcome this problem wavelet
analysis provides an alternative solution. Wavelet is a small wave that possesses
the energy concentration in time as well as frequency to provide a technique for
the analysis of nonstationary time varying phenomenon. As per the phenomena of
Heisenberg uncertainty principle it is impossible to know the exact time and fre-
quency of occurrence of this frequency in a signal. Either we have low-frequency
resolution with good temporal resolution or high-frequency resolution with poor
temporal resolution. Following this in wavelet transform the primary functions dif-
fers both in frequency and spatial range. This type of transformation is basically
intended to get good frequency resolution for low-frequency components that are
basically the average intensity values of the image and high chronological resolution
for high-frequency components forming the edges of the image. Wavelet analysis
are generally used to extract information from many different types of data including
audio signal, images and segregate that information in approximation and details sub
signal. There are two types of wavelet transformation; continuous wavelet transfor-
mation (CWT) and discrete wavelet transformation (DWT). To overcome the problem
of redundancy in CWT, discrete wavelet transform have been commenced. DWT are
not continuously translatable and scalable but can only be in discrete steps. When
DWT is used for continuous signal, it will generate a series of wavelet coefficients
and it is known as wavelet series decomposition.
The basic and simplest wavelet transformation is the Haar wavelet transformation,
introduced in the year of 1910 by an Hungarian mathematician Alfred Haar. The Haar
transform uses Haar function that is compact, dyadic and orthonormal, rectangular
pair. The Haar transform works as a model for the wavelet transformation and it is
closely related to the discrete Haar wavelet transform [16, 17]. For multiresolution
analysis, scaling function (x) create a series of approximations of a signal or
image and each varying by a factor of 2 from its nearest neighboring approximation.
Wavelet functions (x) are used to encode the difference (detail) in information
between adjacent approximations. So the Haar transform corresponding to the Haar
scaling function:
234 M. Acharya et al.
1 0x <1
(x) = (14.1)
0 otherwise
So after shifting and scaling the Haar wavelet functions are defined as
where k is the position of the functions and j is the width and its value define the layer
in the image pyramid. The wavelet function (x) relates to the difference between
the layer j and j + 1.
Now assume A is an N X N image where N is an even number. To compute
W N A and W N AW NT , D H W T is applied to each column of A and to each row of A,
respectively. W N is a transformation matrix.
a11 a12 a13 a14
a21 a22 a23 a24
A=
a31 a32
(14.5)
a33 a34
a41 a42 a43 a44
T
H H H AH T H AG T A V
W N AW NT = A = = (14.6)
G G G AH T G AG T H D
Partition A in 2 2 blocks as
a11 a12 a13 a14
a21 a22 a23 a24 A11 A12
=
= A21 A22
a31 a32 a33 a34
a41 a42 a43 a44
14 Inexact Implementation of Wavelet Transform and Its Performance Evaluation 235
Fig. 14.7 a Input image; b Level one HWT image; c Iterated HWT image
In this section we discuss about the proposed methodology for the bit width reduction
algorithm using discrete Haar wavelet transform to make a tradeoff between image
quality and computational energy. Generally the peak signal-to-noise ratio (P S N R)
is an evaluating parameter of image quality. P S N R is an approximate measure of
human perception on reconstruction quality. The P S N R is computed between the
reconstructed images and is compared with the original image. P S N R is defined
as the mean-squared error (M S E) of two m n size images f and z; where f is
original image and z is the reconstructed image.
m1 n1
M S E = 1/mn [ f (i, j) z(i, j)]2 (14.11)
i=0 j=0
P S N R = 10log10 (M AX 2I /M S E) (14.12)
P S N R = 20log10 (M AX I /M S E) (14.13)
Here M AX I is the maximum possible pixel value of the image under consider-
ation. In general, when samples are represented using linear PCM with B bits per
sample, M AX I is 2 B 1. In our experiment initially the original image pixels are
represented using 8 bits per sample, hence M AX I is 255 and after bit width reduction
the image pixels become 7 bits per sample, therefore M AX I is 128.
14 Inexact Implementation of Wavelet Transform and Its Performance Evaluation 237
Algorithm 14.1 Inexact computing of a grayscale image through LSB Bit slicing
using discrete Haar wavelet transform
Input: Load input image f (i) from the workspace. It is a grayscale image of resolution 256 256.
Output:Reconstructed image z(i).
Step 1. 1st level 2D D H W T is applied on the image and following the decomposition it will
generate the coefficients c A1, cH 1, cV 1, cD1.
Step 2. Then through single reconstruction A1, H 1, V 1, D1 is extracted from
c A1, cH 1, cV 1, cD1 respectively.
Step 3. Each A1, H 1, V 1, D1 is then converted into 8 bit binary number representation.
Step 4. Clipping the LSB bit from all 8 bit binary number and convert it into 7 bit equivalent
binary no.image representation.
Step 5. Rescaling the bit slicing image as coefficient matrix c A2, cH 2, cV 2, cD2.
Step 6.Finally reconstruct the image using I D H W T and compute P S N R and Entr opy
Return: Reconstructed image z(i).
Referring to Fig. 14.9 the From Workspace block designated as ImageSource variable,
which reads data from the MATLAB workspace where image data can be stored in
the form of a 2D array. A MATLAB preload scripting file reads the image and
Re-Scaling
Synthesis LSB Bit
Slicing
128x128 128x128
Synthesis Re-Scaling
CA2
unit
unit
delay delay
Gateway in unit
delay
unit
delay
adder
256
A<=B line buffer
2 CONSTANT Gateway out
DOWNSAMPLE
A
CONSTANT mux
2 fixed pt bit slicing
A<B
unit unit
delay delay
CH2
Gateway in unit
adder
256
..
..
CV2
...
...
...
CD2
...
serializes the pixels and temporarily stores in the workspace to make it in a format
suitable for hardware execution [18]. We have designed four parallel channels for
the computation of C A2, C H 2, C V 2 and C D2, respectively. The details of the
computation have been described in the Eqs. (14.5)(14.10). In the first phase c A1
has been computed. From the input matrix of the image, the total matrix is broken
down into 2 2 segments and as per the formula (Eq. 14.3) the summation of each
2 2 segment is first done followed by dividing the result by 2 is undertaken, which
produces a single value A1. This operation is computed for the entire input image
matrix say (M N ) and as a result a half resolution image is generated after the
DWT, i.e., (M/2 N /2). In the second phase cH 1 has been computed. From the
input matrix of the image, the total matrix is broken down into 2 2 segments and
as per the formula (in Eq. 14.4) the subtraction of each 2 2 segment is first done
followed by dividing the result by 2 is undertaken, which produces a single value
H 1. This operation is computed for the entire input image matrix say (M N ) and
as a result a half resolution image is generated after the DWT, i.e., (M/2 N /2).
Similarly the other two parallel sections are also computed as per the formulas in
Eqs. (14.5) and (14.6) which produces a single value V 1 and D1, respectively. This
operation is computed for the entire input image matrix say (M N ) and as a result
a half resolution image is generated after the DWT, i.e., (M/2 N /2).
T hr oughput: The design requires (512 512 + 1000 = 263144) clock cycles
to process one frame of resolution 512 512. The end blocks are working at a fre-
quency of 50 MHz. Therefore each clock period is 20ns. Therefore 263144 20 =
5262880 ns = 0.01 s is needed to process per frame. There the design can process
100 f rames per second for a frame of resolution 512 512. The algorithm has been
successfully implemented using FPGA hardware by using the Xilinx System Gen-
erator platform with Intel(R) Core(TM) 2 Duo CPU T6600 @ 3.2 GHz platform and
Xilinx Virtex-5 LX110T OpenSPARC Evaluation Platform (100 MHz).
Here we have implemented the bit width reduction algorithm using MATLAB for
four different images. Then we have performed the FPGA based hardware design and
its implementation for such inexact computing. We have compared the software and
the hardware implementations for all the four tested images as shown in Fig. 14.6. The
resource utilization and the power consumption data has been provided in Tables 14.1
and 14.2, respectively. The information regarding the resource and power reduction
percentages are also provided in Tables 14.1 and 14.2 respectively. The resource and
power reduction was achieved for the design through one bit slice reduction. We
compared the P S N R values between the original image with the DWT image and
that between the original image and the reconstructed IDWT image for both software
and hardware based approaches. Table 14.3 shows the comparative P S N R values
for the four 512 512 grayscale images. Figure 14.10 shows the supporting images
for the software and hardware validation.
240 M. Acharya et al.
Table 14.3 PSNR Calculation for four grayscale images measured in decibel(dB)
Sl. No. Image (res. 512 Tradeoff with hardware Tradeoff with software
* 512)
PSNR for PSNR for PSNR for PSNR for
DWT Image reconstructed DWT Image reconstructed
image after image after
IDWT IDWT
1. Leena.jpg 22.3167 20.0550 28.9009 28.1333
2. Cameraman.jpg 18.9383 17.2801 23.8522 22.8284
3. Einstein.jpg 24.0937 21.6293 26.7061 26.0342
4. Mandrill.jpg 19.7734 18.3522 24.2948 23.3573
14.8 Conclusion
We have performed a hardware design for DWT involving dynamic bit width reduc-
tion and hence an inexact computing approach. The prototype design has been imple-
mented on Virtex5 OpenSPARC FPGA. Implementation results show that compared
to the standard hardware implementation of DWT our approach gains in terms
of reduced resource utilization, power consumption while sacrificing a negligible
amount of image quality. To the best of our knowledge our work can be considered
as a first attempt in terms of inexact computing based hardware design of discrete
wavelet transform. We are in a process of developing a more robust algorithm for
resource and power adaptability.
Acknowledgements This work has been supported by the Department of Science and Technology,
Govt of India under grant No DST/INSPIRE FELLOWSHIP/2012/320 as well as grant from TEQIP
phase 2 (COE), University of Calcutta for the experimental equipments. We thank C.V. Raman
College of Engineering in Bhubaneswar, India for facilitating our work. We also thank Prof. (Dr.)
Kaushik Roy, School of Electrical and Computer Engineering, Purdue University, USA for the
encouragement and motivation provided to us.
References
1. Sinha A, Chandrakasan AP (1999) Energy efficient filtering using adaptive precision and vari-
able voltage. In: IEEE international conference on ASIC/SOC, 15 Sept 199918 Sept 1999, pp
327331. doi:10.1109/ASIC.1999.806528
2. Nawab SH et al (1997) Approximate signal processing. J VLSI Signal Process Syst Signal
Image Video Technol 15(1/2):177200
3. Keating M, Flynn D, Aitken R, Gibsons A, Shi K (2007) Low power methodology manual for
system on chip design. Springer Publications, New York
4. Banerjee N, Karakonstantis G, Roy K (2007) Process variation tolerant low power DCT archi-
tecture. In: Proceedings of the conference on design, automation and test in Europe, p 16
5. Gupta V, Mohapatra D, Raghunathan A, Roy K (2013) Low-power digital signal processing
using approximate adders. IEEE Trans CAD Integr Circ Syst 32(1):124137
6. Park J, Choi JH, Roy K (2010) Dynamic bit-width adaptation in DCT: image quality versus
computation energy trade-off. IEEE Trans Very Large Scale Integr (VLSI) Syst 18(5), May
2010
7. Allam MW (2000) New methodologies for low-power high performance digital VLSI design.
Ph.D. thesis, Waterloo, Ontario, Canada
8. Bellaouar A, Elmasry MI (1995) Low power digital VLSI design circuit and system. Kluwer
Academic Publication
9. Chippa VK, Roy K, Chakradhar ST, Raghunathan A (2013) Managing the quality vs. efficiency
trade-off using dynamic effort scaling. ACM Trans Embed Comput Syst 12(2s):90
10. Chippa VK, Venkataramani S, Chakradhar ST, Roy K, Raghunathan A () Approximate com-
puting: an integrated hardware approach. In: ACSSC 2013, pp 111117
11. Burt PJ, Adelson EH (1983) The laplacian pyramid as a compact image code. IEEE Trans
Commun 31(4)
12. Adelson EH, Simoncelli EP, Freeman WT (1990) Pyramids and multiscale representations. In:
Proceedings of European conference on visual perception
13. Adelson EH, Anderson CH, Bergen JR, Burt PJ, Ogden JM (1984) Pyramid methods in image
processing. RCA Eng 29(6):33-41
242 M. Acharya et al.
14. Niclas Brlin (2009) Image analysis wavelets and multi-resolution processing. Department of
Computing Science Ume University, 20 Feb 2009
15. Toufik B [1], Mokhtar N [2] (2012) The wavelet transform for image processing applica-
tions from [1] Automatic Department, Laboratory of Non Destructive Testing, Jijel University
Department of Electrical and Computer Engineering, Algeria. [2] Bristol Robotic Laboratory,
and University of the West of England, UK
16. Chao P-Y Haar transform and its applications. In: Time frequency analysis and wavelet trans-
form tutorial (D00945005)
17. Bnteau C (2011) University of South Florida Tampa, FL USA. Discrete haar wavelet transforms.
UNM - PNM state wide mathematics contest
18. https://www.mathworks.com/products/Image-analysis.html. using Discrete Wavelet Transfor-
mation
Chapter 15
A Vulnerability Analysis Mechanism
Utilizing Avalanche Attack Model
for Dependency-Based Systems
S. Hore S. Chatterjee
Department of CSE, Hooghly Engineering & Technology College,
Hooghly, West Bengal, India
e-mail: shirshendu_hore@yahoo.com
S. Chatterjee
e-mail: sankha.hetc@gmail.com
N. Dey ()
Department of Information Technology, Techno India College
of Technology, Kolkata, India
e-mail: neelanjan.dey@gmail.com
A.S. Ashour
Faculty of Engineering, Department of Electronics and Electrical
Communication Engineering, Tanta University, Tanta, Egypt
e-mail: amirasashour@yahoo.com
V.E. Balas
Faculty of Engineering, Aurel Vlaicu University of Arad, Arad, Romania
e-mail: balas@drbalas.ro
15.1 Introduction
15.2 Background
Since the proposed work tries to nd a generalized solution for nding the primary
cause behind avalanche therefore an extensive background analysis is done.
Korel [11] proposed a well-practiced method known as the Goal-Oriented
Approach. Through their work, the author uses data dependence analysis to guide
the process of test data generation. Data dependence analysis automatically iden-
ties statements that affect the execution of the selected statement, and this infor-
mation is used to guide the search process to nd the goal.
Ibrahim et al. [12] provided an analysis of the Feistel Networks (FN) structural
models known as Extended Feistel Network (EFN). The work inspected the models
with respect to its avalanche criterion to establish the optimal scheme proper for the
flexible block size cipher. The analysis showed that the F-function played a sig-
nicant role to achieve the avalanche effect. The experiential analysis demonstrated
that as more functions were used for each round, the smaller the threshold number
of rounds required by the model to gain avalanche probability of 0.5. Through the
same year, Newsome and Song [13] suggested a dynamic taint analysis for auto-
matic detection and analysis of exploits on commodity software. Taint Check is a
novel mechanism which used dynamic taint analysis to perceive a buffer overrun
vulnerability or format string vulnerability.
Kasthurirathna et al. studied the mechatronic software systems failure tolerance
under different software types and/or hardware failures by using a complex network
approach [14]. Subsequently, the constructed, imitated system architectures were
15 A Vulnerability Analysis Mechanism Utilizing Avalanche Attack 247
Typically, the optimization problems are concerned with determining the values for
one or some decision variables that convene the objective(s) the best without
violating the constraint(s). According to the objective function, optimization
problems may have multiple solutions some of which may be local optima.
Minimize expr subject to the constraints cons. All constraints are inequalities
and all variables are nonnegative.
15 A Vulnerability Analysis Mechanism Utilizing Avalanche Attack 249
where S is the solution space; N is the neighborhood structure; and the objective
function to be maximized is obj.
250 S. Hore et al.
Simulated Annealing
Simulated Annealing (SA) [33, 34] is a variation of the hill climbing, where it
includes a general survey of the scene to avoid climbing foot hills. Since, the whole
space has been explored initially and this avoids the being caught on a plateau or
ridges, by allowing for a probabilistic acceptance of poorer solution. It applies a
probabilistic rule to decide whether the innovative solution replaces the current one
or not. Through this algorithm, the probability of acceptance P of an inferior
solution changes as the search progresses and is calculated as follows:
P = exp E kT 15:3
Denition 4 Active node: If for any node i V, where V is the set of vertices in the
dependency-based system, there exists at least one direct dependency, it is called an
active node.
252 S. Hore et al.
Denition 5 Dead node: If for any node i V, where V is the set of vertices in the
dependency-based system, there does not exist any direct dependency or indirect
dependency, it is called a dead node.
Denition 6 Chain Reaction: A cumulative effect caused due to one active node
removal of a dependency-based system.
Denition 7 Avalanche Factor (AF): A heuristic which is used to measure the
effect of one active node removal from the dependency-based system. It also
measures the effect of the chain reaction (Denition 6) initiated by that active node.
Therefore, in Fig. 15.1 for node 1, the avalanche factor is 7. For node 5, the
avalanche factor is 10. Algorithm 1 has been devised to nd the avalanche factor of
any node in a dependency-based system.
Denition 8 Priority (Rank): is a heuristic which is used to measure the local
minim as in minimization problem and local maxima as in maximization problem.
The priority decreases with the increase in AF.
Since, it is quite obvious that an active node in a dependency-based system have
different direct dependencies and indirect dependencies, where the chain reaction
depends on them. Thus, different direct dependencies and indirect dependencies
would steer the chain reaction to different directions from the starting node and will
cover dissimilar set of nodes of the dependency-based system. Theoretically, there
may exist at least two such nodes for which the direct dependencies and indirect
dependencies may be same. Thus, it creates same pattern of the chain reaction
unless it nds different direct dependencies and indirect dependencies at some stage
of the chain reaction. Hence, the starting node does not have a strict and sole effect
on the chain reaction which it has initiated. Consequently, as a conclusion, the
pattern of the chain reaction may depend upon the active node which has initiated it.
Avalanche attack aims at nding the weakest or strongest point of a
dependency-based system by pointing out the node, the deletion of which will affect
the system to the lowest or highest extent. It is obvious to notice that in a
dependency-based system removal of one node will affect other nodes that are
directly or indirectly dependent on it. A node is said to be dead if it does not have
any type of connection with any other node of the system, whatever that is it neither
depends on any node nor any other node depends upon it. The existence of such
dead node is unrealistic in dependency-based systems. Hence, removal of one node
may create a dead node which in turn removes that freshly spawned dead node from
the system.
It is also noted that removal of one node from the system may partially affect
nodes having multiple dependency in the system which in turn may affect other
indirectly dependent nodes of the system. The chain reaction initiated by the
removal of one node will stop when it will not be possible to nd any further node
to affect. It can be observed that the overall effect on the system depends on the
starting node. Depending upon the dependencies, a different starting node may
result in different pattern of effects on the system.
15 A Vulnerability Analysis Mechanism Utilizing Avalanche Attack 253
To measure the level of the effect due to removal of one node is given by
avalanche factor (AF), a numeric value which gives an idea about consequences
that will take place after exclusion of one node from the system. A higher avalanche
factor indicates a higher vulnerability of that node in the system.
It is clear that the algorithm 1 returns a numeric value which reflects the effect
caused by removal of one particular node of a system. Figure 15.2 depicts the
dependency-based system depicted in Fig. 15.1, after applying avalanche factor
with starting node 5. The nodes lled with red depict the nodes which are
excluded due to the attack. Nodes with red boundaries depicts the affected ones but
not deleted. Edges in red are removed dependencies after the attack and the dark
blue ones carried the indirect effect of attack to other nodes. Nodes and edges in
light blue are unaffected ones. Figure 15.2 reveals that due to removal of node 5
254 S. Hore et al.
all the nodes are affected by the system, whereas Fig. 15.3 reveals the effect of the
attack started at node 3 and none, but 3 is the only node affected leaving others
untouched.
nodes. The average AF is compared with the average AF obtained from uninformed
search to observe the effect of deploying different search techniques. The informa-
tion regarding the heuristic search is obtained by using the Algorithm 1 which
gives a quite satisfactory approximation of the effect of individual members of a
dependency-based system. Simulated Annealing (SA) search technique has been
deployed to implement the model upon randomly generated dependency-based
systems with quite large variation of dependencies. The problem of nding out the
member having the highest effect on the system in its absence is found to be the
maximization problem which aims at nding out the member, termination of which
would have the highest effect on the system of interest. Similarly, the minimization
problem can be constructed as to nd out the member having lowest effect on the
system in its absence.
possible searches and has ability to maintain a reasonable difference with the
average obtained from the uninformed search.
Table 15.2 Search results of the maximization problem of the same system in Fig. 15.1
Starting node No of iteration Optimized result Avalanche factor
1 10 2 9
2 5 10
3 5 10
4 2 9
5 5 10
6 2 9
15 A Vulnerability Analysis Mechanism Utilizing Avalanche Attack 257
Figure 15.5 reveals that the results obtained using SA are always superior to the
results obtained using uninformed search.
Generally, in both the problems (minimization/maximization) the SA has pro-
vided superior result irrespective of the selection of starting node which clearly
indicates the ingenuity of the proposed model.
i. Different starting nodes number:
To examine the effect of using different starting nodes, Figs. 15.6, 15.7, 15.8, 15.9,
15.10, and 15.11 depict the results obtained by following the previous conguration
on dependency-based systems for both the minimization and maximization having
12, 25, and 200 nodes, respectively.
Fig. 15.7 Avalanche factor (AF) versus the starting node maximization of a typical
dependency-based system having 12 nodes
Fig. 15.8 Avalanche factor (AF) versus the starting node minimization of a typical
dependency-based system having 25 nodes
The gures reveal that in all the cases SA provides more accurate and optimized
results than that of the uninformed search.
It is obvious that using the information obtained by Algorithm 1 heuristic search
technique has been able to nd out a node having highest or lowest participation in
the post attack time in its absence irrespective of the selection of the starting node.
During the minimization problem, the local minim as having highest priority and it
15 A Vulnerability Analysis Mechanism Utilizing Avalanche Attack 259
Fig. 15.9 Avalanche factor (AF) versus the starting node maximization of a typical
dependency-based system having 25 nodes
Fig. 15.10 Avalanche factor (AF) versus the starting node minimization of a typical
dependency-based system having 200 nodes
260 S. Hore et al.
Fig. 15.11 Avalanche factor (AF) versus the starting node maximization of a typical
dependency-based system having 200 nodes
decreases with increasing value of the AF. While, in the maximization problem, the
local maxim have the highest priority and it decreases with increasing value of AF.
From the previous experimental results, it is clear that the proposed SA method
satised superior performance for either the minimization or the maximization
problems compared to the uniform search method. Also, it provides more stable
results compared to the AF method without search (traditional method).
15.6 Conclusion
This paper introduced a new vulnerability assessment analysis through the ava-
lanche attack model in a dependency-based system. The results show that the
proposed model has shown that how effectively avalanche attack model can safe-
guard the proposed system. Also, it helps to prevent any destruction that may ruin a
dependency-based system. The extensive experimental result shows that using
some heuristic technique allows an easy identication for the most valuable part or
parts of the proposed system and therefore either we can prevent attacks or can
destroy the enemy system.
15 A Vulnerability Analysis Mechanism Utilizing Avalanche Attack 261
References
26. Kriti Virmani J, Dey D, Kumar V (2015) PCA-PNN and PCA-SVM based CAD systems for
breast density classication. Appl Intell Optim Biol Med Curr Trends Open Probl
27. Dey N, Samanta S, Yang X-S, Chaudhri SS, Das A (2013) Optimisation of scaling factors in
electrocardiogram signal watermarking using cuckoo search. Int J Bio-Inspired Comput
(IJBIC) 5(5):315326
28. Samanta S, Acharjee S, Mukherjee A, Das D, Dey N (2013) Ant weight lifting algorithm for
image segmentation. In: 2013 IEEE international conference on computational intelligence
and computing research(ICCIC), Madurai
29. Chakraborty S, Samanta S, Mukherjee A, Dey N, Chaudhuri SS (2013) Particle swarm
optimization based parameter optimization technique in medical information hiding. In: 2013
IEEE international conference on computational intelligence and computing research
(ICCIC), Madurai
30. Rani J, Seth K (2012) Dependency analysis for component based systems using minimum
spanning tree. In: International conference on advances in computer applications (ICACA).
pp 1116
31. Russell S, Norvig P (2009) Articial intelligence: a modern approach. Prentice Hall Press
32. Sullivan KA, Jacobson SH (2001) A convergence analysis of generalized hill climbing
algorithms. IEEE Trans Autom Control 46:12881293
33. Granville V, Krivanek M, Rasson JP (1994) Simulated annealinga proof of convergence.
IEEE Trans Pattern Anal Mach Intell 16:652656
34. Fleischer MA (1995) Simulated annealing: past, present, and future. In: Alexopoulos C,
Kang K, Lilegdon WR, Goldsman D (eds) Proceedings of the 1995 winter simulation
conference. IEEE Press, Arlington, Virginia, pp 155161
35. Metropolis N, Rosenbluth A, Rosenbluth M, Teller A, Teller E (1953) Equation of state
calculations by fast computing machines. J Chem Phys 21(6):10871092
Chapter 16
Performance of Statistical and Neural
Network Method for Prediction
of Survival of Oral Cancer Patients
N. Sharma ()
Zeal Institute of Business Administration, Computer Application
and Research, Savitribai Phule Pune University, Ganesh Khind,
411007 Pune, Maharashtra, India
e-mail: nvsharma@rediffmail.com
H. Om
Computer Science and Engineering Department, Indian School of Mines,
Dhanbad, Jharkhand, India
e-mail: hari.om.cse@ismdhanbad.ac.in
16.1 Introduction
Recently, it has been observed that statistical methods and articial neural network
techniques are widely used in healthcare industry, as modern medical facility
generates huge amounts of heterogeneous data on daily basis [1]. A majority of
areas related to medical services that require data to be analyzed are mainly for
prediction of effectiveness of surgical procedures, medical tests, medication, and the
discovery of relationships among clinical and diagnosis data [2]. This study intends
to provide advantageous understanding into the capabilities and performance of
statistical methods and articial neural networks for prediction of survival of oral
cancer patients, as it is considered as a grave problem globally and in India as well.
The information extracted from the survey carried out by GLOBOCAN indicates
that oral cancer ranks among the top three types of cancer in India and has one of
the highest incidences in Indian subcontinent [3] as shown in Fig. 16.1.
Approximately, 83,000 new cases and 46,000 deaths are recorded each year
because of oral malignancy in the country [4]. Figure 16.2 presents the statistics of
oral cancer cases in India stratied by sex. The statistics informs that there is a high
rate of mortality in males than females [5]. If we do not take this disease seriously,
then the situation in future is going to be very alarming as projected by GLO-
BOCAN 2008 data (refer Fig. 16.3).
Incidence Mortality
Fig. 16.1 Incidence and mortality rates (age standardized) by cancer type in India (sexes
combined) data extracted from GLOBOCAN 2008: cancer incidence and mortality worldwide
Incidence Mortality
9.8
7.5
6.8
5.2 5.2
3.6
Fig. 16.2 Incidence and mortality rates (age standardized) lip and oral cancer in Indiaby sex
extracted from GLOBOCAN 2008: cancer incidence and mortality worldwide
16 Performance of Statistical and Neural Network Method 265
Fig. 16.3 Crude incidence projection for lip/oral cavity cancer (20082030). Data extracted from
GLOBOCAN 2008: cancer incidence and mortality worldwide
In this paper, we have attempted to build two models for predicting the sur-
vivability of oral malignant patients. First model is a linear regression model
(LR) which is a statistical method, and the second is multilayer perceptron model
(MLP) which is an articial neural network. Linear regression performs better for
extremely small sample size and when hypothesis or experience shows an under-
lying relationship between dependent and predictor variables [6], whereas the true
power of multilayer perceptron model lies in its ability to learn linear as well as
nonlinear relationships straight from the data being modeled, and its performance
enhances with specimen measure [7]. This paper is arranged as accompanies: The
following section reviews the related work done by different researchers; Sect. 16.3
gives the information about predictive models adopted in this research; the
experimental results are presented in Sects. 16.4 and 16.5 covers the conclusions.
Finally, acknowledgment and references is mentioned in the last section.
factors alone are used, then the articial neural networks (MLP and RBF) are
certainly more exact than the TNM staging system. Moreover, there are many
researchers who have studied oral tumor evolvement and progression using various
types of neural networks and data mining approaches, and found it useful in
identifying early potential relapses [1621].
where Yactual is actual target value and Ypredicted is observed or predicted target
value. The sum of the squared residuals (SSR) is calculated to yield the desired
goodness of t of the linear function that can be expressed in form of following
form:
N
SSR = Yactual Ypredicted 2 16:2
i=1
where N is number of instances. The smaller the value of SSR, the closer is the
predicted target value to the actual observed target value. Therefore, the goal of
regression analysis is to obtain the regression equation by adjusting the coefcients
a and b until the sum of the residuals errors that are associated with all the points
reaches its minimum. It is known as least-squares regression (LSR) t or ordinary
least-squares (OLS) regression. The coefcients a and b can be derived as follows:
N XY X Y
b= a = Y bX 16:3
N X 2 X 2
coefcient values. The t-statistics is a measure of the likelihood that the actual value
of the variable Y is not zero. The larger the absolute value of t, the less is the chance
that the actual value of the variable could be zero. The prob(|t|) is the probability of
obtaining the estimated value of the coefcient if the actual coefcient value is zero.
The smaller the value of prob(|t|), the more signicant the coefcient and the less
likely that the actual value is zero. Here absolute value of t is considered in prob(|t|).
The condence interval (CI) conrms the range of values for the calculated coef-
cient that covers the actual coefcient value with the specied condence. Linear
regression can be performed using several algorithms; however, in this case study,
singular value decomposition (SVD) algorithm is used to implement linear
regression as it is robust and less sensitive to predictor variables that are nearly
collinear [24, 25].
1
Output of Hidden layer = hi = f ui = 16:5
1 + e ui
The hidden layer of the MLP model uses nonlinear sigmoid function (). The
output from the hidden layer is distributed to each of the neurons in the output layer
[30]. The output layer also applies the summation function to compute the weighted
268 N. Sharma and H. Om
sum, denoted by vi, and feeds it to the activation function to compute the output of
the layer (yi) as follows:
n
Weighted Sum = vi = hi Wijy 16:6
i=1
1
Output of Output layer = yi = f vi = 16:7
1 + e vi
where hi is output of hidden layer and Wyij is weight from ith hidden layer node to jth
output layer node (i and j are indices of neurons). The back-propagation learning
algorithm is based on minimization of the error between the actual output (Yi) of
the network and the desired output (Yi(d)). The error E is calculated using Eucli-
dean function as follows:
1
E= Yi d Yi 2 16:8
2
where E shows error of ith neuron of the output layer. To normalize the error, the
weight of the neuron must be updated as follows:
E
WijK + 1 = WijK 16:9
Wij
where Wk+1
ij is updated weight, Wkij is old weight, and E Wij is rate of change of
error with respect to weight. The process of learning algorithm adopted by the
multilayer perceptron neural network is presented below:
Step1: Read inputs and desired/expected output of the network.
Step2: Initialize weight of inputs with random values.
Step3: Compute summation function.
Step4: Compute activation function.
Step5: Calculate prediction error E (=desired outputactual output).
Step6: Average the error over the complete set of training cases.
Step7: Transmit the error backward through the network and calculate the gra-
dient of change in error with respect to changes in weight.
Step8: Reduce the error by adjusting or updating the weight.
The table presents the number of rows with a particular category (A = Alive or
D = Dead) that were misclassied by the model, for training as well as validation
data. Along with the misclassied count, the table also presents misclassied
weight, misclassied percent, cost, and classication accuracy of the model.
Table 16.5 presents the misclassication table for MLP, and Table 16.6 presents
the misclassication table for linear regression model for training and validation
data, respectively.
270 N. Sharma and H. Om
Table 16.1 Linear regression parameter for survival = Dead: coefcient b values, standard
error, t-statistics, Prob(t), 95 % condence interval
Variable Coefcient Std. error t Prob(|t|) 95 % Condence
interval
Clinical symptom
Burning sensation 0.0248006 0.03886 0.64 0.52350 0.1011 0.05146
Loosening of tooth 0.0293704 0.06485 0.45 0.65070 0.1566 0.09788
Mass 0.459061 0.1678 2.74 0.00634 0.7884 0.1298
History of addiction
None 7.52112 e 2.424 e 0.31 0.75645 5.509 e 4.005 e
+012 +013 +013 +01
Tobacco chewing 7.52112 e 2.424 e 0.31 0.75645 5.509 e 4.005 e
+012 +013 +013 +01
History of addiction 1
None 6.25376 e 2.821 e 2.22 0.02689 1.179 e 7.17 e
+013 +013 +014 +012
Comorbid condition 1
None 1.32596 e 5.517 e 2.40 0.01643 2.433 e 2.409 e
+014 +013 +013 +014
Gross examination
Inltrative 0.0429256 0.1913 0.22 0.82246 0.4182 0.3324
Plaque like 0.246707 0.2641 0.93 0.35043 0.7649 0.2715
Polypoidal 0.184125 0.177 1.04 0.29848 0.5315 0.1632
Ulcero-proliferative 0.178139 0.1131 1.58 0.11546 0.4 0.04374
Site
BM 0.267512 0.1413 1.89 0.05867 0.5448 0.009821
LA 0.0985091 0.0981 1.00 0.31555 0.291 0.094
Palate 0.0596517 0.131 0.46 0.64907 0.1975 0.3168
RMT 0.0883389 0.2038 0.43 0.66471 0.3115 0.4882
Tongue 0.0658472 0.09115 0.72 0.47021 0.2447 0.113
Tumor size
24 cm 5.96114 e 4.236 e 1.41 0.15972 2.352 e 1.427 e
+013 +013 +013 +014
<2 cm 7.00587 e 3.193 e 2.19 0.02848 7.393 e 1.327 e
+013 +013 +012 +014
Neck nodes
Present 7.52112 e 2.424 e 0.31 0.75645 5.509 e 4.005 e
+012 +013 +013 +013
LFT
Normal 0.199452 0.2686 0.74 0.45789 0.7265 0.3276
FNAC of neck node
Yes 0.0277584 0.04156 0.67 0.50435 0.1093 0.0538
Diagonistic biopsy
None 4.87633 e 2.126 e 0.23 0.81859 4.659 e 3.683 e
+013 +014 +014 +014
(continued)
16 Performance of Statistical and Neural Network Method 271
Table 16.5 Misclassication table for multilayer perceptron (training and validation data)
Category Misclassied (training data) Misclassied (validation data)
Count Weight % Cost Count Weight % Cost
A 212 212 34.6 0.34 236 236 38.56 0.38
D 95 95 23.0 0.23 74 74 17.91 0.17
Total 307 307 29.9 0.30 310 310 30.24 0.30
Overall accuracy = 70.05 % Overall accuracy = 69.76 %
Table 16.6 Misclassication table for linear regression (training and validation data)
Category Misclassied (training data) Misclassied (validation data)
Count Weight % Cost Count Weight % Cost
A 10 10 1.63 0.01 196 96 32.02 0.32
D 399 399 96.6 0.96 201 201 48.66 0.48
Total 409 409 39.9 0.39 397 397 38.73 0.38
Overall accuracy = 60.10 % Overall accuracy = 61.27 %
The classication of data rows by the model is presented in the confusion matrix.
The matrix has a row and a column for each category (A = Alive and D = Dead) of
the target variable. The categories in the rst column are the actual categories of the
target variable. The categories shown across the top of the table are the predicted
categories. The gure in each cell is the value of the actual category of the row and
the predicted category of the column. The diagonal cells contain the values those
are correctly classied by the model, whereas off-diagonal cells have misclassied
data. Confusion matrix for training and validation data for MLP and LR is shown in
Table 16.7.
Table 16.7 Confusion matrix for multilayer perceptron and linear regression (training and
validation data)
Category Multilayer perceptron Linear regression
Training data Validation data Training data Validation data
A D A D A D A D
A 400 212 376 236 602 10 416 196
D 95 318 74 339 399 14 201 212
274 N. Sharma and H. Om
The sensitivity and specicity report is regarding positive category and the negative
category. Sensitivity represents the probability of an algorithm to correctly predict
malignancy, and it is computed as Sensitivity = TP/(TP + FN). Specicity repre-
sents the probability of an algorithm to correctly predict non-malignancy, and it is
computed as Specicity = TN/(FP + TN). Survival = D is considered as a positive
and Survival = A is considered as negative for both the models developed. The
report also provide details about negative predictive value (NPV), positive pre-
dictive value (PPV), geometric mean of sensitivity and specicity, geometric mean
of ppv and npv, recall, precision, f-measure and area under ROC curve for training,
and validation data for both the models in Table 16.8.
Area under ROC curve (AUC) plots the true positive rate against the false
positive rate for the different possible cut-points of a diagnostic test. It shows the
trade-off between sensitivity and specicity (any increase in sensitivity will be
accompanied by a decrease in specicity). The area under it ranges between 0 and
1. It is used as a criterion for measuring the prediction ability of a model. The
number nearer to 1 predicts better. Figures 16.4 and 16.5 present area under ROC
curve for Survival = Dead for MLP and Linear Regression models, respectively.
Lift is a ratio between expected response using predictive model and without any
model for specic population, and gain is a ratio between expected response using
predictive model and without any model for entire population. The lift and gain
chart used following terms:
Bin indexThe entire dataset is divided into equal number of bins. The rst bin
represents the data rows that have the highest predicted probability for the specied
target category.
Cutoff ProbabilityIt is the smallest predicted probability of data rows falling
in the bin.
Class % of binThe percentage of the cases in the bin that have the specied
category of the target variable.
Cumulative % populationThe cumulative percentage of the total cases (with
any category) falling in bins up to and including the current one.
Cumulative % of classThe cumulative percentage of the rows with the
specied category falling in bins up to and including the current one.
16 Performance of Statistical and Neural Network Method 275
Table 16.8 Sensitivity and specicity report for multilayer perceptron and linear regression
(training and validation data)
Category Description Multilayer perceptron Linear regression
Training Validation Training Validation
data (%) data (%) data (%) data (%)
True Predicting malignant 31.02 33.07 1.37 20.68
positive (TP) patients among the
malignant patients
True Predicting non-malignant 39.02 36.68 68.73 40.59
negative (TN) patients among
non-malignant patients
False positive Patients who are predicted 20.68 23.02 0.98 19.12
(FP) as malignant among
non-malignant patients
False Patients who are predicted 9.27 7.22 38.93 19.61
negative (FN) as non-malignant among
malignant patients
Sensitivity Probability to correctly 77.00 82.08 3.39 51.33
predict malignancy
specicity Probability to correctly 65.36 61.44 98.37 67.97
predict non-malignant
cases
Geometric Geometric mean of 70.94 71.01 18.26 59.07
mean of sensitivity and specicity
sensitivity
and
specicity
Positive Proportion of patients with 60.00 58.96 58.33 51.96
predictive the disease who are
value (PPV) correctly predicted to have
the disease
Negative Proportion of patients who 80.81 83.56 60.14 67.42
predictive do not have the disease
value (NPV) who are correctly
predicted as not having the
disease
Geometric Geometric mean of PPV 69.63 70.19 59.23 59.19
mean of PPV and NPV
and NPV
Precision Proportion of cases 60.00 58.96 58.33 51.96
selected by the model that
have the true value;
Precision is equal to PPV
Recall Proportion of the true 77.00 82.08 3.39 51.33
cases that are identied by
the model; recall is equal
to sensitivity
(continued)
276 N. Sharma and H. Om
Fig. 16.4 Area under ROC curve for multilayer perceptron model
Table 16.9 Calculation of lift and gain for linear regression for training data for
Survival = Dead
Bin Cutoff Class % Cum % Cum % Cum % of % of Lift
Index probability of bin Population of class Gain population class
1 0.79395 64.08 10.05 15.98 1.59 10.05 15.98 1.59
2 0.73145 63.11 20.10 31.72 1.58 10.05 15.74 1.57
3 0.66895 64.08 30.15 47.70 1.58 10.05 15.98 1.59
4 0.60645 59.22 40.20 62.47 1.55 10.05 14.77 1.47
5 0.57520 51.46 50.24 75.30 1.50 10.05 12.83 1.28
6 0.45020 51.46 60.29 88.14 1.46 10.05 12.83 1.28
7 0.15332 17.48 70.34 92.49 1.31 10.05 4.36 0.43
8 0.05957 13.59 80.39 95.88 1.19 10.05 3.39 0.34
9 0.00000 10.68 90.44 98.55 1.09 10.05 2.66 0.27
10 0.00000 6.12 100.00 100.00 1.00 9.56 1.45 0.15
Table 16.10 Lift and gain for linear regression for training and validation data
Lift/Gain Training data Validation data
A D A D
Average gain 1.259 1.4134 1.149 1.170
Maximum lift 1.54 1.59 1.51 1.59
Percent of cases with survival (%) 59.71 40.29 59.71 40.29
278 N. Sharma and H. Om
(a) Lift
(b) Gain
Fig. 16.6 Lift and gain for survival = Dead for linear regression model
Table 16.11 Calculation of lift and gain for multilayer perceptron for training data for
Survival = Dead
Bin Class % Cum % Cum % of Cum % of % of Lift
Index of bin population class gain population class
1 66.02 10.05 16.46 1.64 10.05 16.46 1.64
2 65.05 20.10 32.69 1.63 10.05 16.22 1.61
3 58.25 30.15 47.22 1.57 10.05 14.53 1.45
4 66.02 40.20 63.68 1.58 10.05 16.46 1.64
5 47.57 50.24 75.54 1.50 10.05 11.86 1.18
6 46.60 60.29 87.17 1.45 10.05 11.62 1.16
7 24.27 70.34 93.22 1.33 10.05 6.05 0.60
8 14.56 80.39 96.85 1.20 10.05 3.63 0.36
9 4.85 90.44 98.06 1.08 10.05 1.21 0.12
10 8.16 100.00 100.00 1.00 9.56 1.94 0.20
Table 16.12 Lift and gain for multilayer perceptron for training and validation data
Lift/Gain Training data Validation data
A D A D
Average gain 1.278 1.416 1.278 1.308
Maximum lift 1.61 1.64 1.53 1.57
Percent of cases with survival (%) 59.71 40.29 59.71 40.29
The attempt has been made to compare linear regression and MLP models on the
basis of various criteria to assess their performance, and the same is discussed
below:-
Misclassication Table: Classication accuracy of MLP model for training data
is 70.05 % and validation data is 69.76 %, whereas classication accuracy of linear
regression model for training data is 60.10 % and validation data is 61.27 %.
280 N. Sharma and H. Om
(a) Lift
(b) Gain
Fig. 16.7 Lift and gain for Survival = Dead for multilayer perceptron model
Confusion Matrix: The matrix shows that MLP model generates better true
positive and false positive results, whereas linear regression generates better true
negative and false negative results.
16 Performance of Statistical and Neural Network Method 281
Sensitivity and Specicity: It indicates that MLP model can predict the malig-
nant cases more efciently than that of Linear Regression.
Lift and Gain: Average gain and percentage of cases with Survival = A or D is
slightly better in the case of MLP model in comparison to Linear Regression
models.
Figures 16.8 and 16.9 present the graphical comparison of both the models for
training and validation dataset. MLP displays better results in terms of accuracy,
true negative, false negative, specicity, geometric mean of sensitivity and speci-
city, positive predictive value, geometric mean of PPV and NPV, precision,
f-measure, and area under ROC curve in comparison with linear regression. The
results clearly indicate that MLP is a better model to predict the survivability of the
oral cancer patients.
282 N. Sharma and H. Om
16.6 Conclusions
In this paper, we have built two models for the prediction of survival of oral cancer
patients, contemporary statistical model, i.e., linear regression model and a neural
network model, i.e., multilayer perceptron model. Both the models are different in
the way they are developed and the way they handle the dataset. The performance
of both the models is compared systematically. Classication accuracy of neural
network model is certainly better than statistical method, and denitely a better
model to predict the survival of oral cancer patients.
Acknowledgments The authors would like to place on record their sincere thank Dr. Vijay
Sharma, MS, ENT, for his valuable contribution in understanding the occurrence and diagnosis of
Oral Cancer. The authors offer their gratitude to the management and staff of Indian School of
Mines, for their constant support.
16 Performance of Statistical and Neural Network Method 283
References
1. Setyawati BR, Sahirman S, Creese RC (2002) Neural networks for cost estimation. AACE Int
Trans EST13:13.113.8
2. Smith K, Palaniswamy M, Krishnamoorthy M (1996) A hybrid neural approach to
combinatorial optimization. Comput Oper Res 23(6):597610
3. Buntine WL, Weigend AS (1991) Bayesian back-propagation. Complex Syst 5:603643
4. Warner B, Misra M (1996) Understanding neural networks as statistical tools. Am Stat 50
(4):284293
5. Lippmann RP (1987) An introduction to computing with neural nets. IEEE ASSP Mag:422
6. Freeman JA (1994) Simulating neural networks with mathematica reading. Addison-Wesley,
MA
7. Hertz J, Krogh A, Palmer RG (1991) Introduction to the theory of neural computation. Santa
Fe institute studies in the sciences of complexity, vol 1. Addision-Wesley, Redwood City, CA
8. Fausett L (1994) Fundamentals of neural networks: architecture, algorithms and applications.
Prentice Hall
9. Rumelhart DE, Hinton GE, Williams RJ (1986) Learning representations by back-propagating
error. Nature 323:533536
10. Smith M (1993) Neural networks for statistical modeling. Van Nostrand Reinhold, New York
11. Taylor JG (1999) Neural networks and their applications. Wiley
12. White H (1992) Articial neural networks: approximation and learning theory. Basil
Blackwell, Oxford
13. Langdon JD, Russel RC, Williams NS, Bulstrode CJK (2000) Arnold: oral and oropharyngeal
cancer practice of surgery. Hodder Headline Group, London
14. Dinshaw KA, Ganesh B (2008) Annual report 20022005, hospital based cancer registry,
Tata Memorial Hospital, 2008
15. Willium GS, Hine MK, Levy BM A text book of oral Pathol. 4th edn. W.B. Saunders
Company, Philadelphia
16. Baxt WG (1995) Application of articial neural networks to clinical medicine. Lancet
346:11351138
17. Bottaci L (1997) Articial neural networks applied to outcome prediction for colorectal cancer
patients in separate institutions. Lancet 350:469472
18. Fogel DB, Wasson EC, Boughton EM, Porto VW, Angeline PJ (1998) Linear and neural
models for classifying breast masses. IEEE Trans Med Imaging 17:485488
19. Guh JY, Yang CY, Yang JM, Chen LM, Lai YH (1998) Prediction of equilibrated
postdialysis BUN by an articial neural network in high-efciency hemodialysis. Am J
Kidney Dis 31:638646
20. Lapeer RJA, Dalton KJ, Prager RW, ForsstrUom JJ, Selbmann HK, Derom R (1995)
Application of neural networks to the ranking of perinatal variables influencing birth weight.
Scand J Clin Lab Invest 55:8393
21. Hornik K, Stinchcombe M, White H (1989) Multilayer feedforward networks are universal
approximators. Neural Netw 2:359366
22. Fujita H, Katafuchi T, Uehara T, Nishimura T (1992) Application of articial neural network
to computer-aided diagnosis of coronary artery disease in myocardial spect Bulls-eye images.
J Nucl Med 33(2):272276
23. Poli R, Cagnoni S, Livi R, Coppini G, Valli G (1991) A neural network expert system for
diagnosing and treating hypertension. Computer:6471
24. Shang JS, Lin YE, Goetz AM (2000) Diagnosis of MRSA with neural networks and logistic
regression approach. Health Care Manag Sci 3(4):287297
25. HariKumar R, Vasanthi NS, Balasubramani M (2012) Performance analysis of articial
neural networks and statistical methods in classication of oral and breast cancer stages. Int J
Soft Comput Eng (IJSCE) 2(3):263269
284 N. Sharma and H. Om
26. Exarchos KP, Rigas G, Goletsis Y, Fotiadis DI (2012) Modelling of oral cancer progression
using dynamic Bayesian networks. In: Data mining for biomarker discovery. Springer
optimization and its applications, pp 199212
27. Rosenblatt F (1961) Principles of neurodynamics: perceptrons and the theory of brain
mechanisms. Spartan Books, Washington DC
28. Kent S (1996) Diagnosis of oral cancer using genetic programminga technical report,
CSTR-96-14
29. Kaladhar DSVGK, Chandana B, Bharath Kumar P (2011) Predicting cancer survivability
using classication algorithms. Int J Res Rev Comput Sci (IJRRCS) 2(2):340343
30. Milovic B, Milovic M (2012) Prediction and decision making in health care using data
mining. Int J Public Health Sci 01(2):6978
31. www.dtreg.com
32. Sharma N, Om H (2012) Framework for early detection and prevention of oral cancer using
data mining. Int J Adv Eng Technol 4(2):302310
Author Index
A M
Acharya, Moumita, 227 Mahajan, S.P., 123
Ashour, Amira S., 243 Mainkar, Sujay D., 123
Autee, Rajesh, 209 Maity, Satyabrata, 227
Myalapalli, Vamsi Krishna, 1
B
Baital, Kalyan, 31 O
Balas, Valentina E., 243 Om, Hari, 263
Oza, Shraddha D., 87
C
Chakrabarti, Amlan, 31, 227 P
Chandavale, Anjali, 173 Pal, Chandrajit, 227
Chatterjee, Sankhadeep, 243 Pathak, Mrunal, 137
D S
Dey, Nilanjan, 243 Sabnis, Manoj K., 49
Dorlikar, Vyanktesh, 173 Sharma, Neha, 263
Shelake, Vijay, 153
F Shewale, Ashwini, 193
Fahrnberger, Gnter, 97 Shirude, Snehalata B., 17
Shukla, Manoj Kumar, 49
H Sirsat, Narayan Balaji, 75
Holambe, Prabhakar Ramesh, 75 Sonawane, Anuja, 193
Hore, Sirshendu, 243 Srinivasu, N., 137
Swami, Nitin Vijaykumar, 75
J
Joshi, K.R., 87 T
Teke, Utkarsha, 193
K
Khedkar, Mohan S., 153 W
Kolhe, Satish R., 17 Waghmare, Nayan, 193
Kumar, Santosh D., 193
Kunjir, Supriya, 209