Escolar Documentos
Profissional Documentos
Cultura Documentos
PROCEEDINGS
Gediz University Publications
www.gediz.edu.tr
EDITOR
Prof. Dr. Mustafa GNE
EDITORIAL BOARD
Asst. Prof. Dr. Ali Kemal INAR
Asst. Prof. Dr. brahim GRLER
PUBLICATION COMMITTEE
Asst. Prof. Dr. Ozan AKIR
Asst. Prof. Dr. Selim SOLMAZ
Asst. Prof. Dr. Uur TRKAN
Asst. Prof. Dr. Yavuz BAYAM
Instr. Dr. Gkhan AYLI
Instr.Yavuz NCE
Ress. Asst. erife TOZAN
Ress. Asst. Gkhan AKYOL
Ress. Asst. Mehmet APUTU
ii
MAIN SPONSORS
iii
PREFACE
It was only a year ago, in June 2010, when the Faculty of Engineering and Architecture of Gediz University inaugurated an
international symposium where 291 research outputs were presented with the participation of more than 400 academicians
and researchers from 106 universities around the globe. The high rate of interest and the success of this 1st Symposium
provided us enough motivation and encouragement to reorganize the event in the consecutive year and decide maintaining on
a yearly basis. Therefore, this year, The 2nd International Symposium on Computing in Science & Engineering (ISCSE
2011) was held at Kuadas (located on the western coast of Turkey) during 01 and 04 June 2011.
The 629 participants from a total of 27 countries, 60 of whom were postgraduate students, had opportunities for scientific
discussions and exchanges that greatly contributed to the success of the symposium. The participating countries included
Algeria, Albenia, Australia, Bahamas, Belgium, Bosnia Herzegovina, Burkina Faso, Cyprus, France, Gambia, Germany,
Ghana, Indonesia, Iran, Iraq, Ireland, Italy, Jordan, Northern Cyprus, Malaysia, Mali, Qatar, Saudi Arabia, Syria, Sweden,
Tunisia, Turkey and USA.
The main objective of ISCSE 2011 is to bring together national and international researchers and practitioners to discuss and
exchange scientific theories and engineering applications over a rather wide spectrum of topics. The symposium coverage
was intentionally kept as comprehensive as possible in order to help create platforms of cooperation for science people and
promote and facilitate future interdisciplinary researches. With these aims in mind, the Symposium Organizing Committee
determined xx topics including, for instance, Quantum Computation, Robotics and Automation, Artificial Intelligence, DNA
Computing, Fuzzy Logic and Discrete Optimization etc.
ISCSE 2011 received the total of 300 full-paper and 42 posters submissions. The full-paper submissions were sent to the
members of the Scientific Committee and additional reviewers for evaluation. Though the full list would exceed the space
allotted here, we would like to thank the 70 reviewers for their hard work and dedication in helping us to design such a highquality program of talks.
Besides the symposium talks and poster presentations, the program included 5 plenary lectures covering various fields of the
symposium, given by the invited speakers: Prof. Dr. Vesa A. NISKANEN from the University of Helsinki on Aspects of
Approximate Probability, Prof. Dr. Burhan TURKSEN from ETU University of Economics and Technology on Review of
Fuzzy System Model: Philosophical Grounding, Rule Bases and Fuzzy Function, Prof. Ibrahim A. Ahmad and Associate
Prof. Dr. Dursun DELEN from the Oklahoma State University on Reversed Hazard rate and Mean Idle Time: Two New
Notions in Quantifications Aging Derived from Stochastic Orderings and Applications, and Business Intelligence and
Advanced Analytics for Real World Problems and Prof. Dr. Mike NACHTEGAEL from Ghent University on The
Application of Fuzzy Logic for Noise Reduction in Image Processing. Their invaluable contribution and illuminating public
talks are acknowledged and greatly appreciated. Without their participation, the symposium would have been incomplete.
Knowing that the organization of such an international event needs great effort and time spent, we would like to extent our
appreciation to the Rector of the Gediz University, Prof. Dr. Seyfullah evik and the members of the organizing committee,
including Mustafa Akdag, Haluk Gumuskaya, Serdar Korukolu, Vedat Pazarlioglu, Mehmet Pakdemirli, Fuat Okumus,
Ozlem Erkarslan, Ibrahim Gurler, Aziz Kolkiran, Ozan Cakir, Fatih Mmtaz Duran, Hadi Zareie, Haidar Sharif, Seza Filiz,
Ali Kemal Cinar, Yavuz Bayam, Selim Solmaz, Ugur Turkan, Mehmet Aksarayli, Emin Ozyilmaz, Gokhan Cayli, Mumin
Ozcan, Salahattin Yildiz, M. Rdvan zel, Ozlem Kiren Gurler, Aysegul Gungor, Mehtap Ozdemir Koklu, Gulsen Senol,
Ibrahim Cakirlar, Mehmet aputu, Zerife Yildirim and Serife Tozan Ruzgar.
Lastly, but not least, The Symposium would not have been possible without the generous financial and in-kind support
provided by our sponsors. The contributions received from Tubitak, Turksat, BTK (Information and Communication
Technologies Authority of Turkey), Petkim, Kavuklar Group, Orkide Group, MEGA International Communication, Emarket
Turkey, Derici Group, Cisco and Hamle Digital are greatly appreciated.
Given the enourmous speed with which science and engineering applications has been advancing in the areas covered by
ISCSE, we expect that the next symposium will be as exciting and stimulating as this years one was, as seen through the
pages of this proceedings volume.
Please stay tuned to the web site of the Gediz University for updates and details.
Looking forward to seeing you all and new researchers in ISCSE 2012.
Prof. Dr. Mustafa GNE
Symposium Coordinator
iv
HONARY CHAIR
Prof. Dr. Seyfullah CEVIK, Gediz University, Rector
SYMPOSIUM CHAIR
Prof. Dr. Mustafa GUNES, Gediz University, Dean of Faculty of Engineering and Architecture
SYMPOSIUM ORGANIZING COMMITTEE
Prof. Dr. Mustafa GUNES
Prof. Dr. Mustafa AKDAG
Prof. Dr. Haluk GUMUSKAYA
Prof. Dr. Lotfi Askar ZADEH
Prof. Dr. Mehmet SEZER
Prof. Dr. Serdar KORUKOLU
Prof. Dr. M. Vedat PAZARLIOGLU
Prof. Dr. Mehmet PAKDEMIRLI
Prof. Dr. Serkan ERYILMAZ
Assc. Prof. Dr. Fuat OKUMUS
Assc. Prof. Dr. Ozlem ERKARSLAN
Asst. Prof. Dr. Ibrahim GURLER
Asst. Prof. Dr. Aziz KOLKIRAN
Asst. Prof. Dr. Ozan CAKIR
Asst. Prof. Dr. Fatih Mmtaz DURAN
Asst. Prof. Dr. Hadi ZAREIE
Asst. Prof. Dr. Md. Haidar SHARIF
Asst. Prof. Dr. Seza FILIZ
Asst. Prof. Dr. Ali Kemal CINAR
Asst. Prof. Dr. Yavuz BAYAM
Asst. Prof. Dr. Selim SOLMAZ
Asst. Prof. Dr. Ugur TURKAN
Asst. Prof. Dr. Mehmet AKSARAYLI
Asst. Prof. Dr. Emin OZYILMAZ
Instr. Dr. Gokhan CAYLI
Instr. Mumin OZCAN
Instr. Salahattin YILDIZ
Ress. Asst. Dr. Ozlem KIREN GURLER
Ress. Asst. Aysegul GUNGOR
Ress. Asst. Mehtap OZDEMIR KOKLU
Ress. Asst. Gulsen SENOL
Ress. Asst. Ibrahim CAKIRLAR
Ress. Asst. Zerife YILDIRIM
Lab. Tech. Serife TOZAN RUZGAR
SCIENTIFIC COMMITTEE
Prof. Dr. Ronald R. YAGER
Prof. Dr. Reza LANGARI
Prof. Dr. Yalcn CEBI
Prof. Dr. Mustafa GUDEN
Prof. Dr. Ali CALISKAN
Prof. Dr. Mustafa GUNES
Prof. Dr. Serdar KURT
Prof. Dr. Orhan GUVENEN
Prof. Dr. Mehmet SEZER
Prof. Dr. Daniel LANE
Prof. Dr. Ali SURMEN
Prof. Dr. Tuncer OREN
Prof. Dr. Salih OKUR
Prof. Dr. Robert SHORTEN
Prof. Dr. Mustafa AYTAC
Prof. Dr. Nihat BOZDA
Prof. Dr. Ahmet GOKEN
Prof. Dr. Etienne E. KERRE
Prof. Dr. Saban EREN
Prof. Dr. Osman BALCI
Prof. Dr. rfan ALAN
Prof. Dr. Martin CORLESS
Prof. Dr. Mehmet Cudi OKUR
Prof. Dr. Gultekim CETINER
Prof. Dr. J.N.K.RAO
Prof. Dr. Serkan ERYILMAZ
Prof. Dr. Said Ali HASAN
Prof. Dr. Senay UDOGRUK
Prof. Dr. Orhan TORKUL
Prof. Dr. Levent SENYAY
Prof. Dr. Gulhan ASLIM
Prof. Dr. Masoud NIKRAVESH
Prof. Dr. Cemali DINCER
Prof. Dr. Erdal CELIK
Prof. Dr. Ismihan BAYRAMOGLU
Prof. Dr. Jorg FLIEGE
Prof. Dr. Mike NACHTEGAEL
Prof. Dr. Talip ALP
Prof. Dr. Ibrahim A. AHMAD
Prof. Dr. Selim ZAIM
Prof. Dr. Yavuz AKBAS
Prof. Dr. Turan BATAR
Prof. Dr. Harun TASKIN
Prof. Dr. Efendi NASIBOGLU
Prof. Dr. Gulen CAGDAS
Prof. Dr. Ugur CAM
Prof. Dr. Gokmen TAYFUR
Prof. Dr. Murat SOYGENIS
Prof. Dr. Yunus CENGEL
Prof. Dr. rfan ALAN
Prof. Dr. Haluk GUMUKAYA
Assc. Prof. Dr. Nurullah UMARUSMAN
Assc.Prof. Osman I. TAYLAN
Assc.Prof. brahim DARRAB
Assc. Prof. Dr. Dursun DELEN
Assc. Prof. Dr. Mustafa TOPARLI
vi
SCIENTIFIC COMMITTEE
Assc. Prof. Dr. Musa ALI
Assc. Prof. Dr. Sinan SENER
Assc. Prof. Dr. Yusuf OYSAL
Assc. Prof. Dr. Arzu GONENC SORGUC
Assc. Prof. Dr. Can BAYKAN
Assc. Prof. Dr. Ali Ihsan NESLITURK
Assc. Prof. Dr. Gamze TANOGLU
Asst. Prof. Dr. Muhammed CINSDIKICI
Asst. Prof. Dr. Hadi ZAREIE
Asst. Prof. Dr. Aziz KOLKIRAN
Asst. Prof. Dr. Selim SOLMAZ
Asst. Prof. Dr. Murat CAKIRLAR
Asst. Prof. Dr. Ozdemir CETIN
Asst. Prof. Dr. A.Turan OZCERIT
Asst. Prof. Dr. Uur TURKAN
Asst. Prof. Dr. Yavuz BAYAM
Asst. Prof. Dr. Koray KORKMAZ
Asst. Prof. Dr. Nurcan BAYKUS
Asst. Prof. Dr. Haldun SARNEL
Asst. Prof. Dr. Cuneyt AKINLAR
Asst. Prof. Dr. Ahmed FREEWAN
Asst. Prof. Dr. Mustafa Emre LAL
Asst. Prof. Dr. Ahmet ZENGN
Asst. Prof. Dr. Kadir ERKAN
Asst. Prof. Dr. H. Secil ARTEM
Asst. Prof. Dr. Osman CaglarAKIN
Asst. Prof. Dr. Alpaslan DUYSAK
Asst. Prof. Dr. Jens ALLMER
Asst. Prof. Dr. Yenal AKGUN
Asst. Prof. Dr. Ayce DOSEMECILER
Asst. Prof. Dr. Sabri ALPER
Asst. Prof. Dr. MD Haidar SHARIF
Asst. Prof. Dr. Hurevren KILI
Asst. Prof. Dr. Aysegul ALAYBEYOLU
Asst. Prof. Dr. AhmetTuran OZCERIT
Asst. Prof. Dr. Ibrahim GURLER
Asst. Prof. Dr. Sahin UYAVER
Asst. Prof. Dr. Fahrettin ELDEMIR
Asst. Prof. Dr. Ozan CAKIR
Assc. Prof. Dr. Mustafa TOPARLI
Assc. Prof. Dr. Fuat OKUMUS
Asst. Prof. Dr. Sinan KOKSAL
Asst. Prof. Dr. Ahmet Afsin KULAKSIZ
Assc. Prof. Dr. Bulent EKICI
Assc. Prof. Dr. Allaberen ASHYRALYEV
Assc. Prof. Dr. Ipek DEVECI
Asst. Prof. Dr. Emel KURUOLU
Asst. Prof. Dr. Mehmet AKSARAYLI
Asst. Prof. Dr. Istem KESER
Asst. Prof. Dr. Nahit EMANET
Dr. Mehmet Emre GULER
Dr. Emre ERCAN
Dr. Gokhan CAYLI
Dr. Murat TANIK
Dr. Ozlem KIREN GURLER
Dr. Efe SARIBAY
vii
CONTENTS
Chapter 1 - COMPUTER ENGINEERING
16
26
33
40
46
54
60
67
74
79
86
viii
91
99
106
111
118
125
131
136
143
149
156
161
167
ix
173
180
188
193
200
206
212
218
224
229
237
244
250
256
263
268
274
280
286
293
300
306
315
321
327
334
339
347
357
358
362
366
373
379
385
391
xi
398
403
409
419
426
434
442
450
458
465
472
Proceeding Number:200/39
Design of CMOS Analog Operational Amplifiers using EKV Transistor Model and
Multi-Objective Optimization
Proceeding Number: 200/41
478
483
489
496
Proceeding Number:200/46
I-PD Controller Design for A 4-Pole Hybrid Electromagnet on the Basis of
Coefficient Diagram Method (CDM)
xii
504
511
512
518
525
532
539
546
552
560
567
574
581
588
xiii
595
601
607
611
618
625
631
641
642
648
655
662
672
680
687
xiv
693
704
712
720
728
735
742
756
763
772
779
786
793
xv
801
809
816
825
832
840
849
856
864
872
879
856
893
898
904
909
917
xvi
925
932
939
949
950
957
964
971
978
985
992
999
1003
1010
Chapter 6 - ARCHITECTURE
1017
1024
1035
1036
xvii
1042
1048
1056
1067
1068
1075
1081
1089
1097
1106
1112
1118
1125
1131
1137
1144
xviii
1151
1160
1167
1173
1178
1184
1189
1194
1200
1206
1212
1219
1225
1230
xix
1237
1244
1253
1260
1266
1272
1285
1286
1293
1299
1304
1310
1316
1322
1326
1334
xx
1340
1349
1356
1360
1369
1370
1375
1387
The Application of Fuzzy Logic for Image and Video Noise Reduction
Mike Nachtegael
xxi
1388
Chapter 1
COMPUTER
ENGINEERING
Keywords: image processing, image segmentation, distributed and parallel processing, high performance
computing.
1 Introduction
Nowadays, concurrency in software is an important way to manage the sharing of hardware resources used at
the same time. The multiprocessor systems allow true parallel execution; multiple threads or processes run
simultaneously on multiple processors. The trade-off made in the multiprocessor case is increasing the overall
system cost [1]. The multi-core processors are composed of two or more independent cores in a single physical
package [2]. There are certain important considerations when using multi-core processors for improvement of
single programs performance of color or gray-scale image filtering or segmentation (FoS) process: the optimal
application performance (OAP) and the decreasing of execution time. The OAP will be achieved by effectively
using threads to partition software workloads [1]. The Open Multi-Processing (OpenMP) is an application
programming interface (API) used to explicitly direct multi-threaded and shared memory parallelism [2]. It
supports much functionality required for parallel computation based on thread-level parallelism (TLP) [2], [3]. In
most modern Fortran and C/C++ compilers as well as modern operating systems (OS), OpenMP is available [2],
[4]. The computationally intensive processes such as image FoS can be executed more efficiently on the systems
based on multi-core processors than the systems based on single-core processors. There are some works on
multi-core image processing with Open Multi-Processing (OpenMP) standard in the literature [2]. The OpenCV
library is an Open source Computer Vision library which involves over 500 functions that span many research
areas in vision [5], [6], [7]. Image segmentation is a low-level computer vision task. It is a decomposition
process of a gray-scale or color image into homogeneous partitions. There are some clustering techniques for
image segmentation such as k-means [8] and mean-shift (MS) [9], [10] clustering. Thresholding is a simplest
segmentation process. Otsus method and Entropy method are optimal thresholding techniques [11], [12].
Similar benchmark systems and/or softwares are published in the literature. Venkata et al. [13] presented the San
Diego Vision Benchmark Suite (SD-VBS). In SD-VBS, they implemented an algorithm for image segmentation.
There are other systems/softwares such as MediaBench [14] and Spec2000 [15] benchmark suites which include
different algorithms from different research areas. In my study, the benchmark software is developed to evaluate
TLP performance of four well-known image segmentation techniques (i.e., clustering and thresholding) which
are implemented and tested at same platform via same way on high-definition (HD) images. The software
benchmarks single- and multi-thread performance of related FoS techniques using OpenMP and OpenCV
infrastructure, and it plots performance results via its built-in graphic/plot viewer. My software is called ISeeMP,
which is an acronym for Image Segmentation Extended via Multi-Processing. The software has an original
approach for mutual performance comparison of well-known image FoS techniques.
The remainder of this paper is organized as follows. Section 2 describes FoS methods used in my benchmark
study. Section 3 explains the parallelization approach of ISeeMP. In section 4, I introduce my benchmark
software with details of their modules and/or structural parts. In section 5, I give the experimental performance
results obtained via my software on hardwares/machines of my testbed. Section 6 concludes the paper, and the
obtained results are discussed.
The MS segmentation based on MS clustering is proposed in [10] that the details can be found in Comaniciu and
Meer [10] study.
0 0 11 T , 0 1 1
(1)
w2 B2 T2
(2)
For Eq. (1), (k ) and (k ) are zeroth- and first-order cumulative moments of the histogram up to the k th
level. Also, T is the total mean level of the original image [11]. For Eq. (2), it is noticed that W2 and B2 are
functions of threshold level k , but T2 is independent of k . Furthermore, W2 is based on the second-order
statistics (i.e., class variance), while B2 is based on the first-order statistics (i.e., class means) [11], [21]. The
optimal threshold k * is:
B2 (k*) max B2 (k )
(3)
1 k L
The details of optimal thresholding via Otsus method can be found in [11] and [21].
(4)
where P(x) is the probability that X is in the state x , and P log 2 P is defined as 0 if P 0 . Many entropybased thresholding methods use their histogram entropy (HE) as a criterion function. The threshold related to the
maximum entropy value is chosen as the correct among all thresholds [12], [23]. The HE is given below by Eq.
(5).
n1
H p g log p g 0
(5)
g 0
where, g 0, ..., n 1 shows the n number of discrete gray levels that a pixel can take. The details of Entropy
method can be found in [12], [23].
3 Parallelization Approach
In parallel execution of the programs, the multi-core processors offers support to execute threads in parallel.
Therefore, the cost of communication is very less [3]. The TLP is the parallelism way inherent in an application.
This application runs multiple threads at once. TLP is coarser-grained then instruction-level parallelism (ILP). In
my study, Microsoft Visual C/C++ compiler, OpenCV library, and OpenMP 2.0 standard API are used to
develop my software (i.e., ISeeMP [24]). OpenMP supports the thread-based programming model. Also, the
sequential codes of related four FoS algorithms in my study are converted to multithreaded codes. OpenMP
allows the coder/programmer to indicate the regions of entire programs code that should be executed in parallel.
In this way, it enables to state how the workload in those regions should be shared among the available resources
[25]. Also, OpenMP allows nested parallelism that I used frequently this property in parallelization of my
applications sequential code. The strategy, which could form the basis for an implementation of parallelization
of related image FoS algorithm, involves specific parallel regions and private variables for related nests of
parallelizated loops. The threads in thread pool of OS are distributed to work dynamically by OpenMP API, but
in programs design if necessary I reorganized my programs code for assignment of the number of thread to
relevant section of code.
4 Software
In design of my software (i.e., ISeeMP [24]), I considered the benchmark tools in the literature. In my study, I
aimed to develop an evaluation platform. An evaluation tool help you make the best decisions for your specific
situations. My software contains five independent structural parts (i.e., modules). The segmentation based four
modules are issued for related image FoS method. These modules contain subparts that they work as single- or
multi-thread manner for related methods algorithm. The performance plots (PP) module also assigned to
evaluate and plot the TLP performance results which are speedup and efficiency. For the benchmark, the
software repeatedly runs with user-desired times (i.e., count of runs) that the relevant algorithms sequential or
parallel performance is to be measured and evaluated with respect to user choices. The performance
measurement (PM) is made by related part of relevant segmentation module, and PM is sent to PP module.
Therefore, PM of all methods are obtained, evaluated and displayed by PP module. In Fig. 1, (a) shows the
softwares main graphical user interface (GUI) window which has Settings&Outputs panel for available
parameter settings of related image FoS method; and (b) shows the performance plot of speedup. The plotting
interface is based on NPlot [26] charting library.
Fig. 1.(a) ISeeMPs GUI main window, and (b) its performance plots window.
5 Experimental Results
My testbed hardware suite involves two different computer equipped different kind of CPUs (i.e., one
computer with single-core and one computer with dual-core) and OSs. They are a single-core Intel Pentium IV
processor (#531 with Hyper-Threading support) at 3.0 GHz, 3 Gbytes of RAM, with Microsoft Windows XP 64
bit OS, and a dual-core Intel Core 2 Duo Centrino processor (#P8600) at 2.40 GHz, 3 Gbytes of RAM, with
Microsoft Vista 32 bit OS, respectively. In this paper, these hardwares/machines are called in given order as
Machine #1 and Machine #2, respectively. The ISeeMP software benchmarks four different segmentation
techniques which are implemented and tested at same platform via same way. I used home-made image dataset
to test my software. This dataset involves 10 images which are 24 bit 2048 x 1536 pixels color images [24].
From five test executions, I computed sequential and parallel execution times for each image segmentation
method via my benchmark software, and then I compared these results in related plots. These plots cover simple
models for speedup and efficiency of related segmentation methods performance for the TLP. The speedup ( S pt
) is the ratio between sequential execution time and parallel execution time of related segmentation algorithm,
and it is given in Eq. (6).
S pt Speedup
(6)
T (1)
T (n pt )
Efficiency
(7)
T (1)
(n pt ) T (n pt )
The TLP performance benchmark results over five executions are given via related tables. The speedup and
efficiency results of TLP performance for each one of methods are shown in Table 1 for Machine #1 and in
Table 2 for Machine #2. In these tables, the first row shows the average speedup values, and second row shows
the average efficiency values. In addition, the better values are shown as bold and italic, and same values in both
tables are shown just as italic.
Table 1. Speedup and efficiency of TLP performance for each one of image FoS methods executed with two
threads on Machine #1.
S pt & values for Machine #1
Image no.
#1
#2
#3
#4
#5
#6
#7
#8
#9
#10
K-means
0.82 0.98 0.82 0.83 0.82 0.84 0.82 0.81 0.76 0.83
0.41 0.49 0.41 0.41 0.41 0.42 0.41 0.40 0.38 0.41
Mean-shift
1.53 1.60 1.62 2.25 1.22 1.48 2.53 3.04 3.57 2.06
0.76 0.80 0.81 1.12 0.61 0.74 1.26 1.52 1.78 1.03
Otsu
0.90 0.83 0.89 0.92 0.87 0.87 0.88 0.92 0.87 0.91
0.45 0.41 0.44 0.46 0.43 0.43 0.44 0.46 0.43 0.45
Entropy
1.05 1.00 0.99 1.04 1.03 0.99 1.00 1.00 1.00 1.00
0.52 0.50 0.49 0.52 0.51 0.49 0.50 0.50 0.50 0.50
Table 2. Speedup and efficiency of TLP performance for each one of image FoS methods executed with two
threads on Machine #2.
S pt & values for Machine #2
Image no.
#1
#2
#3
#4
#5
#6
#7
#8
#9
#10
K-means
1.01 0.74 0.95 0.98 0.80 0.81 0.97 0.78 0.97 1.08
0.50 0.37 0.47 0.49 0.40 0.40 0.48 0.39 0.48 0.54
Mean-shift
1.71 1.46 1.55 2.04 1.24 1.40 2.70 2.91 2.67 1.49
0.85 0.73 0.77 1.02 0.62 0.70 1.35 1.45 1.33 0.74
Otsu
1.04 0.79 0.78 0.91 0.91 0.85 0.88 0.78 0.89 0.87
0.52 0.39 0.39 0.45 0.45 0.42 0.44 0.39 0.44 0.43
Entropy
1.31 1.33 1.01 0.89 1.06 1.23 1.13 1.72 0.96 1.28
0.65 0.66 0.50 0.44 0.53 0.61 0.56 0.86 0.48 0.64
I see from these results in both tables that the MS method has better parallelization results than other methods.
The overall results can be seen on ISeeMPs website [24].
6 Conclusions
In this study, the main contribution is to develop a platform for benchmark and performance evaluation for
TLP performance for some well-known image FoS methods. For some images and/or methods, the test results
obtained via dual-core equipped hardware are better than the results of single-core hardware. I have come to the
conclusion that overall results are depended on intra-region homogeneity of pixel colors and parallelization level
of nested loops in programs code.
References
1.
Akther S. and Roberts J. (2006). Multi-Core Programing. Increasing Performance through Software Multithreading.Intel Press.USA.
2. Slabaugh, G.; Boyes, R.; and Yang X. (2010). Multicore Image Processing with OpenMP. IEEE Signal Proc. Mag.
[Applications Corner] 27(2): 134-138. Mar. 2010.
3. Packirisamy V. and Barathvajasankar H. (2008). OpenMP in multicore architectures. Tech. Report.Uni. of
Minnesota. http://www.cs.umn.edu/~harish/reports/openMP.pdf
4. OpenMP. (2011). The Open Multi-Processing. http://openmp.org/. Accessed on 01March 2011.
5. Bradski G. and Kaehler A. (2008). Learning OpenCV: computer vision with the OpenCV library. OReilly Media
Inc. Publication. Sebastopol, CA.
6. OpenCV.
(2011).
The
open
computer
vision
library.
http://sourceforge.net/
projects/opencvlibrary/. Accessed on 01 March 2011.
7. Willow Garage. (2011). Willow Garage Software website. http://www.willowgarage.com/pages/software/opencv/.
Accessed on 01March 2011.
8. Ng H.P., Ong S.H., Foong K.W.C., Goh P.S. and Nowinski W.L. (2006). Medical Image Segmentation Using KMeans Clustering and Improved Watershed Algorithm. In Proceedings of IEEE Southwest Symp. on Image Analy.
and Interpr. 61-65.
9. Sonka M., Hlavac V. and Boyle R. (2008). Image Processing, Analysis, and Machine Vision, Intl. Student Ed., 3rd
ed. Thomson Learn. Thomson Pub.Toronto.Canada.
10. Comaniciu, D. and Meer, P. (2002). Mean shift: A robust approach toward feature space analysis. IEEE Trans. on
Pattern Analy. and Mach. Intel. 24(5): 603619.
11. Otsu N. (1979). A threshold selection method from gray-level histograms. IEEE Trans. on Syst., Man and Cybern.
9(1): 62-66.
Biographies
Bahadir KARASULU He graduated from the Science and Arts Faculty-Physics Dept. at KocaeliUniversity, Kocaeli,
Turkey in 2003. Afterwards, in 2006, he completed a M.Sc. thesis study at MaltepeUniversity, Istanbul, Turkey. In 2010, he
obtained his Ph.D. degree in Computer Engineering from EgeUniversity, Izmir, Turkey. His research interests are artificial
intelligence, computer vision, distributed and parallel computing, simulation and optimization.
Department of Electronics and Computer Education, Technical Education Faculty, Karabk University, Karabk, Turkey
1
sgorgunoglu@karabuk.edu.tr,2safakaltay@karabuk.edu.tr, 3baha.sen@karabuk.edu.tr
Abstract. Motion estimation is widely used especially for observinghuman movements. Tracking human
movements is one of the important issues in security. Checking regularly singular movements of a person and
group movements is necessary in environments where providing security is important. In this study, a system
which is the estimation of dominant motion direction on video sequences based on k-means clustering
method was developed. Video sequences which are included within the scope of this study consist of human
movements.The aims of the system are determination of motion vectors accurately on video sequences and
utilisation of clustering process properly.Therefore, software for determining dominant motion direction
accurately was developed using C# visual programming language. User can select any video, start and stop
motion analysis and also see motion vectors and times of processes on interface of the software.
1 Introduction
Studies which are on image and video processing have been increasing nowadays. These studies usually
include observing moving objects. Motion estimation and motion analysis are used generally for this process.
A lot of academic studies were made on motion analysis and motion estimation. Some of them searched only
motion of people. One of them is nonparametric density estimation with adaptive, anisotropic kernels for human
motion observations [1]. In this paper, it is suggested to model priors on human motion by means of
nonparametric kernel densities.Some studies searched groups of people and evaluated motion of people in
groups. One of them istracking groups of people [2].
Moreover, there are studies which deal with estimation of traffic density by watching motion of vehicle in
traffic. They find regions where traffic density is much. One of them is estimation of vehicle velocity and traffic
intensity using rectified images [3]. Other one is vehicle detection and counting by using real time traffic flux
through differential technique and performance evaluation [4]. Automatic density estimation and vehicle
classification for traffic surveillance systems using neural networks is a study which used artificial intelligence
[5].
Motion estimation is widely used in medicine. There are a lot of studies on medicine. One of them is
correlation analysis of ultrasonic speckle signals for correlation based motion estimation [6].The other one is
motion estimation for abdominal MR[7].Studies were made on different block matching algorithms such as
spiral search algorithm, diamond search algorithm, adaptive search algorithm and orthogonal search algorithm
[8-12].
In this study, full search algorithm which is a block matching algorithm is used for finding motion vectors and
k-means clustering method is used for clustering motion vectors.
2 Motion Analysis
At first, video images are subjected to specific pre-processing operations in motion analysis process. Motion
vectors occurred in moving regions in video images are determined after these processes. Finally, motion vectors
are grouped.
(1)
MB1 denotes macro block which will be compared in the first frame and MB 2 denotes macro block at search
region in the second frame at the equation. (x,y) represent the coordinates of macro block pixels. Full Search
algorithm applied in this study is showed in Figure 1.
10
{ For each j
{ Calculate SAD value between MB2(i,j) in the second frame and MB1(x,y)
If i=x-N and j=y-N The lowest value =SAD
'
'
'
(2)
11
Start
Define k value of cluster number which will be formed
Define random cluster centers as much as k value (C1,C2, , Ck)
do
{ Store existing centers
(tempC1= C1, tempC2= C2, ... tempCk= Ck )
Calculate Euclidian space between each active block and cluster centers
Include active block to inlying center
Find the middle center of blocks included in each cluster (C1,C2, , Ck)
}
While (tempC1
End
3 Software
Estimation of dominant motion direction software was developed by C# which is a visual programming
language. The screenshot of the software developed is given in Figure 3. Video scene which was analysed,
motion vectors formed and processing times of algorithms can be seen in screen.
12
There are various menus provided to users in this software. Video which will be used in motion analysis can
be selected, played and stopped via Video menu. Additionally, characteristics of video can be displayed. Motion
analysis procedure can be initiated and stopped via Motion Analysis menu. Number of groups and whether the
motion vectors will be displayed in the screen can be selected via submenu Settings menu situated under Motion
Analysis, and frames which will be compared are selected via Pre-processing Operations submenu. Preprocessing procedures carried out are displayed via Pre-processing Operations submenu.
4 Conclusions
In this study, a system which is the estimation of dominant motion direction on video sequences based on kmeans clustering method was successfully developed. In this manner, motion which is out of accustomed and the
dominant direction of this motion can be determined. Therefore, some pre-processing procedures were applied to
video sequences at first. Afterwards, motion vectors were formed using Full Search algorithm. These motion
vectors were clustered with K-means algorithm. Screenshots, and detection and classification of the direction of
motion in these screenshots obtained in experimental studies are displayed in Figure 4a and Figure 4b
respectively. Video sequences utilised are in 320x240 pixels dimensions. The red circles in video sequences
display motion centres.
Durations of algorithms used in the study are given in Table 1. Durations of algorithms were obtained from a
notebook consists of hardware as follows: Intel(R) Pentium(R) Dual T3400 2.16 GHz microprocessor, 3GB Ram
with Windows 7 operating system.
13
a)
b)
Fig. 4.Screenshots displaying the centre of motion direction obtained and its classification.
a) Motion in video sequence was classified as bidirectional. b) Motion in video sequence was classified as three directional.
Time (ms)
Pre-processing
240
Operations
0-5
Converting Grayscale
0-12
40-140
15-60
15-60
0-300
K-Means Algorithm
0-1
0-30
Total Time
300-700
References
1.
2.
3.
4.
5.
6.
7.
8.
9.
Brox, T.; Rosenhahn, B.; Cremers, D.; and Seidel, H. P. (2007). Nonparametric density estimation with adaptive,
anisotropic kernels for human motion tracking. Workshop on Human Motion. Rio de Janeir. Brazil. 152-165.
McKenna, S. J.; Jabri, S.; Duric, Z.; Wechsler, H.; and Rosenfeld, A. (2000). Tracking groups of people. Computer
Vision and Image Understanding 80(1): 42-56.
Maduro, C.; Batista, K.; Peixoto, P.; and Batista, J. (2009). Estimation of vehicle velocity and traffic intensity using
rectifield images. IbPRIA 4th Iberian Conference on Pattern Recognition and Image Analysis. Povoa de Varzim.
Portugal. 64-71.
Mohana, H. S.; Ashwathakumar, M.; and Shivakumar, G. (2009). Vehicle detection and counting by using real time
traffic flux through differential technique and performance evaluation. ICACC International Conference on
Advanced Computer Control. Singapore. 791-795.
Ozkurt, C. and Camci, F. (2009). Automatic density estimation and vehicle classification for traffic surveillance
systems using neural networks. Mathematical and Computational Applications An International Journal 14(3):
187-196.
Bilge, H. . (1997). Yapay akla dayal ultrasonik grntleme iin hareket tahmini. Yksek Lisans Tezi.
Krkkale niversitesi Fen Bilimleri Enstits. Krkkale. 1-25.
imek, Yldrm, M. (2007). Abdominal MR grntlerinde hareket tahmini. Yksek Lisans Tezi. Gebze Yksek
Teknoloji Enstits Mhendislik ve Fen Bilimleri Enstits. Gebze. 15-31.
Kroupis, N.; Dasygenis, M.; Soudris, D.; and Thanailakis, A. (2005). A modified spiral search algorithm and its
embedded hardware implementation. IEC. Prague. 375-378.
Zhu, S. and Ma, K. K. (2000). A new diamond search algorithm for fast block matching motion estimation. IEEE
Transactions on Image Processing 9(2): 287-290.
14
Biographies
Salih GRGNOLU- was born in orum, 1970. He graduated from Gazi University in 1995. He received his M.Sc. and
Ph.D. degrees from the same university in 2000 and 2007 respectively. He worked as instructor at Gazi University between
the years of 1999-2007. His areas of interest include biometric systems, image processing, microprocessors and neural
network applications. Asst. Prof. Dr. Grgnolu has been still working for Karabk University, Department of Computer
Engineering.
afak ALTAY- was born in Erzurum, 1986. She graduated from Sleyman Demirel University in 2008. She is interested in
image and video processing. Research Assistant Altay has been still working for Karabk University, Department of
Electronics and Computer Education.
Baha EN- was born in Mersin, 1974. He received his B.Sc. in Computer Science Department from the Gazi University,
Ankara/Turkey in 1996. He received his M.Sc. degree from Institute of Science and Technology, Gazi University in 1999,
and his Ph.D. degree from same department. His research interests include graphics, vision, genetic algorithms, expert
systems, geographical information systems, 3d modeling and simulation systems. Asst. Prof. Dr. en has been still working
for Karabk University, Department of Computer Engineering.
15
Abstract.At the present day, high performance computing (HPC) architectures are designed to resolve
heterogeneous sophisticated scientific as well as engineering problems across an ever expanding number of
HPC and professional workloads. The computation of the fundamental functions
sin(x)
and
cos (x)
isfrequent in both scientific and engineering applications but somewhat time consuming task in highperformance numerical simulations. In this paper, we have addressed the problem of high-performance
numerical computing of sin(x) and cos(x) pair for specific processor IA-64 and optimized it for a vector of
input arguments
xi
processor microarchitecture and the manual optimization techniques improve the computing performance
significantly as compared to the standard math library functions with compiler optimizing options.
1 Introduction
The computation of the fundamental trigonometric functions sin(x) and cos (x) are frequent in scientific
and engineering applications but some what very time consuming task in numerical simulations. For examples,
to simulate scattering at the Coulomb potential at Rutherford model needs to calculate tens of millions of
fundamental functions e.g., sin(x), cos(x), etc. (see Fig. 1). This function can be computed accurately and
efficiently by calling math library routine which will also deal with exceptional case e.g., input argument x = 0 .
But standard routine is often incapable of achieving the demanding performance of high-performance computing
in simulation. Our current trend in constructing high-performance numerical computing for specific processor
IA-64 (Intel Architecture-64) has been optimized for
or less simultaneously.
In 2000, Strebel [1] optimized sin(x) and cos (x) functions along with others for the static scheduling
architecture e.g., Alpha 21164. The Alpha 21164 microprocessor was designed for in-order (static) scheduling
[1, 2]. Nevertheless, the optimization is not workable for the processors like Itanium or even Alpha 21264. Thus,
we have lined up our efforts towards Itanium and hence our work would be considered as an incremental
improvement of the optimization result presented by him.
We have optimized sin(x) and cos (x) functions for Itanium processor (IA-64) with an 1000MHz clock.
There are mainly 2 types of scheduling may observe in Itanium: (i) code scheduled to minimize the number of
16
bundles; and (ii) code scheduled to minimize the number of clock cycles. We have emphasized the first
scheduling and the most common bundles MMF (Template value 14 and 15 [3]) throughout our optimizations.
The Itanium processor core is capable of up to 6 issues per clock, with up to three branches and two memory
references. The memory hierarchy consists of a three-level cache. The first level uses split instruction and data
caches; floating-point data are not placed in the first-level cache. The second and third levels are unified caches,
with the third level being an off-chip cache placed in the same container as the Itanium die. However, the IA-64
is a unique combination of innovative features such as explicit parallelism, predication, speculation and more.
The architecture is designed to be highly scalable to fill the increasing performance requirements of various
server and workstation market segments. In the optimized implementations it has been used an eminent aspect
that one add and one multiply are issued each cycle and the result will be available after result latencies for both
add and multiply. It has been extensively used the standard manual optimization techniques e.g., loop unrolling
and software pipelining to compute sin(x) and cos (x) efficiently. The Loop unrolling technique combines
several iterations of a loop into one basic block. The technique of software pipelining is a way to overlap the
execution of different instances of loop body in a systematic way. Loop unrolling and software pipelining are
combined by first unrolling the loop and therefore increasing the size of the loop body to some extent and then
software pipelining the resulting codes which increases the potential for instruction level parallelism a bit more.
Fig. 1.Laboratory system: scattering at the Coulomb potential. If the case of numerical simulation of this
model, it is important to calculate sin(x) and cos(x) functions as quickly as possible because tens of millions of
particles are involved in such type of simulation.
2 Implementation Steps
sin( x) = k =0
k
(1) k 2 k 1
( 1)
x
cos
(
x
)
=
x 2k .
and
k =0
(2k 1)!
(2k )!
17
Instead of using these formulas for larger values x as well, we may use the addition theorems of
trigonometric functions and use table lookup for some
xt being only b bits wide, sin(x) and cos (x) are given by:
sin( x) = sin( xt xs ) = sin( xt )cos ( xs ) cos ( xt )sin( xs )
cos ( x) = cos ( xt xs ) = cos ( xt )cos ( xs ) sin( xt )sin( xs ) .
Since
xs is small we have cos ( xs ) 1, so it is a good idea not to compute cos ( xs ) directly, but work with
us = cos ( xs ) 1 instead. A schematic representation of the code which calculates pair of sin(x) and
cos (x) is given below. For simplicity, we assume that the number of bits for the table lookup is b = 7 .
y = x 1/2
y = y 3 244
yh = y 3 244
xs = ( x yh h ) yh s xs = x 2yh
t =< y > 6..0
table lookup
c = ct (ct ( t us ) st ss ) c = cos ( x)
s = st (st ( t us ) ct ss ) s = sin( x)
return(c, s)
In this algorithm, the way of computing
53 bits in the mantissa. Note that the lookup table does not have to consist of 2 2 entries, but we may exploit
identities like cos (2 x) = cos ( x) and sin(/2 x) = cos ( x) . Our implementation uses a lookup table
for
18
of vendors in July 2001. Many vendors now market systems with this processor e.g., NEC and Fujitsu Siemens,
while HP and IBM offer them as alternative processors in their high-end systems such as the HP SuperDome and
the IBM x-series. In computing, IA-64 is a 64-bit processor architecture developed in cooperation by Intel and
Hewlett-Packard, implemented by processors such as Itanium and Itanium 2. The goal of Itanium was to produce
a post-RISC era architecture, using a VLIW design. Unlike Intel x86 processors, the Itanium is not geared
toward high performance execution of the IA-32 (x86) instruction set. Three instructions are grouped together
into 128-bit sized and aligned containers called
instructio n slots and a 5-bit template field. The format of a bundle is depicted in Table 1.
Table 1. Bundle Format
slot 2
slot 1
slot 0
template
(41-bit)
(41-bit)
(41-bit)
(5-bit)
First 4 bits of the template explicitly indicate types of instructions, and the last bit (stop bit) indicates whether
the bundle can be executed in parallel with the next bundle[7]. Bundled instructions don't have to be in their
original program order, and they can even represent entirely different paths of a branch. Also, the compiler can
mix dependent and independent instructions together in a bundle, because the template keeps track of which is
which. During execution, architectural stops in the program indicate to the hardware that one or more
instructions before the stop may have certain kinds of resource dependencies with one or more instructions after
the stop [7]. IA-64 instructions are packed by compiler into 128 bit length bundle. The bundle contains three IA64 instructions along with template. The template indicates dependencies between instructions whether the
instructions in the bundle can be executed in parallel or if one or more must be executed serially due to operand
dependencies. The template also indicates whether the bundle can be executed in parallel with the neighbor
bundles.A single bundle containing three instructions corresponds to set of three functional units. IA-64
processors would contain different numbers of such sets. And these processors will be able to run the same code.
Indeed, assume an IA-64 processor has N sets of three functional units each; then using the template information
on dependencies between bundles it is possible to chain the bundles to create instruction word of 3 N
instructions (N bundles) in length. This is the way chosen by HP and Intel to provide scalability of IA-64.
Certainly the concept is beautiful. Unfortunately IA-64 is not absolutely perfect. IA-64 CPUs are extremely
efficient chips, which can process more instructions per clock cycle than any other CPU. That does not mean
there are no concerns. The quality of compilers will make or break the success of IA 64. Complexity of the
compiler increases drastically. Therefore the number of errors in compiler and compilation time increases badly
too.
19
Fig. 2.Four times unroll and instructions scheduling for the first phase. Each time a clock cycle will issue 2
bundles simultaneously - one from 1st bundle sequences and another from the corresponding 2nd bundle
sequences.
Fig. 3.Non-optimized first phase pre-loop. Each time a clock cycle will issue 2 bundles simultaneously - one
from 1st bundle sequences and another from the corresponding 2nd bundle sequences.
Fig. 4.Optimized first phase loop. Each time a clock cycle will issue 2 bundles simultaneously - one from 1st
bundle sequences and another from the corresponding 2nd bundle sequences.
Fig. 5.Non-optimized first phase post-loop. Each time a clock cycle will issue 2 bundles simultaneously - one
from 1st bundle sequences and another from the corresponding 2nd bundle sequences.
20
Fig. 6.Four times unroll and instructions scheduling for the second phase. Each time a clock cycle will issue 2
bundles simultaneously - one from 1st bundle sequences and another from the corresponding 2nd bundle
sequences.
Fig. 7.Non-optimized second phase pre-loop. Each time a clock cycle will issue 2 bundles simultaneously one from 1st bundle sequences and another from the corresponding 2nd bundle sequences.
Fig. 8.This structure depicts (n/4 - 2) times looping of 4 times unrolling and 3 times pipelining of instruction
scheduling to the resulting optimized code. The second phase loop is highly optimized. Each time a clock cycle
will issue 2 bundles simultaneously - one from 1st bundle sequences and another from the corresponding 2nd
bundle sequences.
Fig. 9. This structure indicates second phase non-optimized post-loop of 4 times unrolling and 3 times
pipelining of instruction scheduling to the resulting optimized code. Each time a clock cycle will issue 2 bundles
simultaneously - one from 1st bundle sequences and another from the corresponding 2nd bundle sequences.
21
phase with floating point operations. Such an issue structure is pertinent for software pipelining. We have
overlapped 3 such issue structures so that the resulting structure will be able to minimize loop overhead and to
increase the potential for instruction-level parallelism to more extent. The resulting optimized code delineates in
the Fig. 3, 4, and 5 concerning pre-loop, loop, and post-loop successively of the first phase. After 4 times
unrolling and scheduling the cycles become 42n/4 11n . It is easy to observe, from Fig. 2, only 20.23% of
the total instruction slots are filled with instructions. There are several Itanium-dependent restrictions that cause
a bundle to be split and issue a stop [2]. So, it is often good idea to make overlap 3 such issue structures resulting a structure with 4 times unrolling and 3 times pipelining. After fabricating software pipelining, there
are
t = t0 c1
where
parameter
c2
n
(1)
t 0 is the only barely adequate time required for compute a certain task with t0 = 16ns . The
c1 will be determined by the overhead which occurs for every iteration of the innermost loop such as
22
and return to its original or usable and functioning condition, and the penalty due to non-optimal pre-loop and
post-loop codes. For large vectors n , the overhead
t t0 c1 . Every piece of overhead can not be classified to either class c1 or c2 in all likelihoods e.g., load
misses due to table lookup can occur in every iteration for small n , but for larger n table entries will be reused
and the number of load failures to succeed will decrease.
To find the overheads
c1 and c2 for pair of sin(x) and cos(x), we have measured the execution time t per pair
ready those data to the simple model in the least squares sense, minimizing the relative errors between predicted
and measured execution times. The execution times t for the manually optimized codes are shown Table 2 for a
couple of vector lengths n without using any compiler optimized options. After fitting the data of Table 2 to the
simple model at Eq. 1 in the least squares sense, we have estimated
c2 = 247ns overhead per vector. In relation to the idealized execution time of 16ns the overall
3.9100
overhead drops to about
% 24% for n . Taking into account that the integer pipelines are
16
cos(x), and
highly utilized and that casual first level cache misses will lead to partial stalls, the resulting efficiency seems
satisfactory.
Table 2. With IA-64 (1.00 GHz), the execution time t in nanoseconds, per pair of sin(x) and cos(x)
computation for varying vector lengths n without using any compiler optimized option.
Vector length n
Nanoseconds
38.56
30.79
16
26.73
32
24.28
64
22.12
128
20.92
256
20.11
512
19.23
23
object file becomes larger, which can significantly hurt cache performance. To whatever extent, we can say that
the speed-up of our manually optimized pair of sin(x) and cos(x) depends on the quality of the compiler used.
Table 3. With IA-64 (1.00 GHz), the execution time t in nanoseconds, per pair of sin(x) and cos(x)
computation for varying vector length n using gcc and ecc following their optimized options -03 -funroll-loops
and -O2, respectively.
Vector length n
Compilers
gcc
ecc
t in nanoseconds
4
37.21
48.29
29.58
31.33
16
25.11
27.46
32
23.29
24.65
64
21.15
22.37
128
20.51
21.94
256
19.86
21.04
512
19.13
20.31
Built-in function
106.48
102.95
Table 4. Speed-up with respect to built-in function for vector length n deeming similar results and optimized
options from the Table 3.
Vector length n
Speed up
gcc
ecc
2.86
2.13
3.60
3.29
16
4.24
3.75
32
4.57
4.18
64
5.03
4.60
128
5.19
4.69
256
5.36
4.89
512
5.57
5.07
Table 3 depicts the time t in nanoseconds, for executables generated by gcc and ecc compilers. The best
results obtained by using the various optimizing options available for the different compilers have been shown.
The execution time of the optimized pair of sin(x) and cos(x) may vary to a little extent. As a result, the
execution time on the Table 3 has been computed as the geometric mean of 100 measurements with the same
optimized option for every vector length. However, in computing the optimized pair of of sin(x) and cos(x) for a
enough large vector length n with or without using compiler options in gcc and ecc give a satisfactory factor of
reduction as compared to the standard built-in approach.
24
We have defined the speed-up as the ratio of the execution time of the built-in function and the execution time
of the optimized code in varying vector lengths. The Table 4 depicts different speed-up with varying vector
lengths. It is remarkable that higher vector lengths are desirable. But for simulation, higher vector lengths require
higher number of dummy inputs. For examples, maximum 7 dummies would be important for a choice of vector
length 8 and similarly 15 for 16, 31 for 32, 63 for 64, 127 for 128, etc. Although the lower vector lengths expect
to calculate a smaller number of dummies in simulation, the degree of performance of them is significantly low
and consequently they are hardly use in high performance computing. Depends on simulation work, one would
be benefited from vector lengths like 32 or 64 or so.
4 Conclusions
We addressed a specific problem to compute and presented a detailed analysis of possible software
optimizations to improve its performance. The processor micro-architecture and the manual optimization
techniques improve the computing performance significantly. The optimized routine for computing pair of sin(x)
and cos(x) expects some smarter compilers like gcc and ecc. for high performance. The routine is highly
optimized manually. Consequently, compiler optimization options will not have a big effect on performance. The
routine is significantly faster than using standard library routine, provided that the vector length n is large
enough. Although the routine is much more efficient on IA-64 (Itanium), with or without some maintainance it
can be compiled and will execute correctly on any processor which conforms the IEEE 754 floating point
standard.
References
[1] R. Strebel. Pieces of software for the Coulombic m body problem. Diss. ETH No. 13504, 2000.
[2] C. C. Corporation. Alpha Architecture Handbook. Compaq Computer Corporation, 4 edition, 1998.
[3] D. Patterson and J. Hennessy. Computer Architecture : A Quantitative Approach. Morgan Kaufmann Publishers, Inc.,
3rd edition, 2003.
[4] M. H. Sharif, A. Basermann, C. Seidel, and A. Hunger. High-performance computing of 1/sqrt(xi) and exp(xi) for a
vector of inputs xi on Alpha and IA-64, Journal of Systems Architecture, 54, 2008, 638-650.
[5] I. Corporation. IA-64 System Architecture : IntelR IA- 64 Architecture Software Developers Manual, 2, Intel
Corporation, 1.1 edition, 2000.
[6] I. Corporation. IA-64 Application Architecture : IntelR IA-64 Architecture Software Developers Manual, 1, Intel
Corporation, 1.1 edition, 2000.
[7] I. Corporation. Intels IA-64 Application Developers Architecture Guide. Intel Corporation, 1999.
[8] A. Oram and M. Loukides. Programming with GNU Software. OReilly & Associates, Inc., 1 edition, 1997.
25
Abstract. Morphologic analysis is the study of word formation in a language. It describes how the minimal
meaningful units called morphemes come together to form a word. Morphological parsing and generation is
the basis for many natural language processing applications such as machine translation. Kyrgyz is one of the
major languages in Turkic Languages Family. In this paper we introduce Kyrgyz Morphology using
Koskemnienis two-level finite state method with an implementation on Dilma Machine Translation
Framework. The moprhology is given using two tools: (1) A set of orthographic rules describing the
phonological rules when affixing morphemes to roots or stems (ii) A finite state automaton describing
morphemes and the order in which the morphemes affixed to words.
Keywords: Kyrgyz, morphology, phonology, orthographic rules, two level morphology, finite state machine,
machine translation
1 Introduction
Kyrgyz is one of the major Turkic languages. It is the official language of the independent state of Kyrgyzstan.
It is spoken by more than 4 million people Kyrgyzstan, China, Afghanistan, Kazakhstan, Tajikistan, Turkey,
Uzbekistan, Pakistan and Russia.
In this paper we present an overview of finite state Kyrgyz morphology using Koskenniemis two level
morphology (Koskenniemi, 1983, Karttunen 1983) with a set of orthographic rules modeling the phonological
dynamics during word formation. Morphology is one of the fundamental building blocks of any natural language
application. This is more so, especially for agglutinative languages in which most words are formed by joining
morphemes. Accurate morphological parsing is required for a proper syntactic and semantic analysis of
sentences upon which many applied linguistics applications are built. Developing a comprehensive Kyrgyz
morphology is part of our ongoing Kyrgyz-Turkish Machine Translation Project.
The RUSLAN project is developed to translate documents from Czech to Russian (Hadic 1987) which are
both Slavonic languages. Later another translation project Cesilko was undertaken to translate from Czeck to
Slovac (Hajic et al. 2000).Ceilko was extended to other Slavonic languages, Polish and Lower Serbian (Dworak
et al. 2006), and the Baltic language Lithuanian (Hajic et al. 2003)
Corresponding Author
26
InterNOSTRUM project was another project for closely related languages. It was developed to translate
between Catalan and Spanish (Canals-Marote et al. 2000). More work for Romance Language pairs was
alsoimplemented. Portuguese-Spanish machine translation system was implemented in a similar manner.
(Garrido-Alenda et al. 2003)
The two-level Turkish morphology was described by Oflazer (Oflazer, 1994) using PC KIMMO (Antwort
1990) which is milestone project. In a way it paved the way for similar studies on Turkic languages, since
Turkish and other Turkic languages have lexical, morphological and syntactic similarities.
The machine translation between Turkic languages is limited to only a few projects. Lexicon-based
translation system was implemented between Azerbaijani and Turkish (Hamzaoglu 1993). Another early system
was implemented between Crimean Tatar and Turkish (Altintas, icekli 2001).
A Turkmen Turkish machine translation system was implemented at Fatih University (Shylov 2008) which
turned into morphological machine translation system for Turkic languages, called DLMA. Another Turkmen
Turkish system was implemented at ITU (Tantug 2006,2007). We use DLMA to implement Kyrgyz Turkish
translation system. Two-level morphology of Kazakh Language was described and implemented in DILMAC
(Zafer, 2011). In a similar study two-level morphology of Kazan Tatar language was developed using DILMA
(Gkgz, 2011)
This paper is organized as follows: We describe the Kyrgyz orthography using two level rules in Section 2.
Orthographic rules are formulated for the correct spelling during word formation. We then represent the Kyrgyz
morphotactics using finite state machines (FSM) in Section 3. The FSMs for nominal and verbal morphotactics
describe how the words of the language can be formed in detail. We present a number of nominal, verbal and
adverbial word formation examples to demonstrate the system. The paper concludes in Section 4.
2 Kyrgyz Orthography
Kyrgyz is written in a modified version of Cyrillic script.The Kyrgyz Cyrillic alphabet is used in Kyrgyzstan.
We used Kyrgyz Cyrillic to Latin converter [Hata! Bavuru kayna bulunamad.]. Comparison of Kyrgyz
Cyrillic - Latin alphabet is given in Table 3.
Table 3. Comparison of Kyrgyz Cyrillic - Latin Alphabet
Cyrillic
Latin
A
Bb
Vv
Gg
Dd
E e;Ye ye
Yo yo
C c; J j
Zz
i
Yy
Kk
Cyrillic
Latin
Ll
Mm
Nn
Oo
Pp
Rr
Ss
Tt
Uu
Cyrillic
Latin
Ff
Xx
TS ts
I
inceltme
kesme
Ee
Yu yu
Ya ya
Like other Turkic languages Kyrgyz is also agglutinative. It has a rich set of inflectional and derivational
suffixes (Kasapolu 2007). A word is formed by affixing inflectional and derivational morphemes to roots
successively. During the formation of words some phonological rules is to be followed. These orthographic rules
are expressed in certain formalism as shown below. Some subsets of letters of Kyrgyz language that used in rules
are defined in Table 4
27
Letters in Set
{ f,s,t,k,,,h,p }
{ b,d,c,g,v,z,m,n, ,l }
{ l,d,t }
{ a,e,o,, ,I,u, }
{ r,y }
In this section orthographic rules are described using two-level morphology. In this model word morphology
is represented in two levels: lexical and surface. Orthographic rules describe transformation form lexical level to
surface level. In the rule (a:b), the colon ( : ) character is represent transformation between two levels. Letter a is
in the left side of the colon, indicates that it refers to symbol of lexical level and letterb is in the right side of the
colon, indicates symbol of surface level. These rules are represented four different types:
a : b = > LC__RC ; meaning a lexical a, corresponds to a surface b only if it follows the left context
(LC) and precedes the right context (RC).
a : b < = LC__RC ; meaning a lexical a always corresponds to a surface b if it follows LC and
precedes RC.
a : b < = > LC__RC ; meaning a lexical a corresponds to a surface b always and only if it follows LC
and precedes RC.
a : b / < = LC__RC; meaning a lexical a, never corresponds to a surface b betweeb LC and RC.
i.
ii.
iii.
These rules,are example of consonant harmony similar to Turkish, represent changing from L to set letters
{l,t,d}. First rule states that the letters in lexical L turn into letter l if the last letter of the word is one of the
letters in V or R. An application of the rule is given in the following example
Lexical : cete+LA
Surface: cete0le cetele (Kyrgyz); yedekle (Turkish); back up (English)
Second rule states that the letters in lexical L turn into letter t if the last letter of the word is one of the letters in
Cs. An application of the rule is given in the following example.
Lexical: ti+LA
Surface: ti0te
Third rule states that the letters in lexical L turn into letter d if the last letter of the word is one of the letters in
Ct. An application of the rule is given in the following example. This rule is a consonant harmony like Turkish.
Lexical: kol+LA
Surface: kol0da kolda (Kyrgyz); kolla (Turkish); protect (English)
28
In these examples the surface words are respectivelycetele, tite,kolda are formed by affixing +LA morpheme
to the rootsti, cete, kol. L in the lexical level becomes respectively l,t,d in the surface obeying the rules above.
I:0 =>V+:0__@*
This rule, is a example of vowel elision, represents dropping vowels in subset of I. If last letter of the word is
one of the letters in V and affix begins with Ip, the letter I drops. In the following example the surface word is
apam is formed by affixing +Im (Possesive 1st Person Single for noun) to root apa. I in the lexical level drop in
the surface level obeying the rule above.
Lexical : apa+Im
Surface: apa00m apam(Kyrgyz); annem (Turkish);my mother (English)
The orthographic rules are specified to the system using an XML file. Part of XML file representing the rules
about changing from L to I, t, or d is shown in Figure 1.
29
3 Kyrgyz Morphotactics
Morphotactics describe the morphemes and the order in which these morphemes are affixed to roots or stems.
Morphotactics allows one to generate valid or meaningful words out of morphemes and a lexicon. It also allows
one to check the validity of a given word. We describe Kyrgyz morphotactics using finite state machines (FSM)
as this is the usual method. A finite state machine as shown below is in principal a directed graph. It consists of a
set of states and a set of transitions among these states. Possessive suffixes for nouns are shown in Table 5.
Definitions of possessive suffixes for nouns morphotactics of Kyrgyz language using the finite state model is
given Figure 2.
Table 5. Possessive Suffixes for Nouns
Possesive for Nouns
Possesive 1st Person Single
Possesive 2nd Person Single
Possesive 3rd Person Single
Possesive 1st Person Plural
Possesive 2nd Person Plural
Possesive 3rd Person Plural
Abbreviation
Poss1PS
Poss2PS
Poss3PS
Poss1PP
Poss2PP
Poss3PP
Suffix
+(I)m
+(I)
+(s)I
+(I)bIz
+(I)Ar,+(I)Iz
+LarI, +(s)I
Example
bala+m
bala+
bala+s
bala+bz
bala+z
bala+lar, bala+s
Table 6. Kyrgyz Morphotactics are implemented in a XML file. Part of xml file, for transition of states from
s-1(noun) to s-2 (plural), s-3(verb) and s-4;9 (possessive persons), is shown in Figure 3.
30
Root+Suffix
bala+(I)m
bala+(I)
bala+(s)I
bala+(I)bIz
bala+Iz
bala+LArI, bala+(s)I
bala+LAr+(I)m
Kan+sIrA
Az+Ay
Transition of States
S-1(Noun)s-4 (Poss1PS)
S-1(Noun) s-5 (Poss2PS)
S-1(Noun) s-6 (Poss3PS)
S-1(Noun) s-7 (Poss1PP)
S-1(Noun) s-8 (Poss2PP)
S-1(Noun) s-9 (Poss3PP)
S-1(Noun) s-2(Plural) s-4 (Poss1PS)
S-1 (Noun) s-3 (Verb)
S-1 (Noun) s-3 (Verb)
31
XML files. System is available as a web application. Lexicons and XML specifications are uploaded through the
web interface. Testing is also performed on the web interface.
This is to our knowledge the first study on Kyrgyz language which is the first precondition for a successful
machine translation between Kyrgyz and Turkish. As a result of this study we now have a morphological parser
and a morphological generator for Kyrgyz. Our ultimately goal in this project is to develop a Kyrgyz-Turkish
machine translation system using DILMAC.
References
1.
2.
3.
4.
5.
6.
7.
8.
9.
10.
11.
12.
13.
14.
15.
16.
Koskenniemi K. (1983). Two level morphology: A general computational model of word-form recognition and
production. Tech. rep. Publication, No. 11, Department of General Linguistics, University of Helsinki.
Karttunen L (1983). PC-KIMMO: A General Morphological Processor. In Texas Linguistics Forum 22, pp.165186.
Oflazer, K. (1994). Two-level description of Turkish morphology, Literary and Linguistic Computing, Literary and
Linguistic Computing Volume9, Issue2 pp. 137-148.
Shylov M.(2008).DLMA: Turkish and Turkmen Morphological Analyzer and Machine Translation Program.
Masters thesis, FatihUniversity, stanbul Turkey.
Gkgz E., et al (2011). Two-Level Qazan Tatar Morphology. 1st International Conference on Foreign Language
Teaching and Applied Linguistics.
Zafer H.R, et al (2011). Two-Level Description of Kazakh Morphology. 1st International Conference on Foreign
Language Teaching and Applied Linguistics.
Kasapolu engel H. (2007). Krgz Trkesi, in Trk Leheleri Grameri [Prof. Dr. Ahmet Ercilasun], 481-542,
Ankara, Aka.
Altinta K & icekli . (2001) A Morphological Analyser for Crimean Tarar. In: Procedings of the 10 th Turkish
Sym on AI and NN TAINN pp 180-189, North Cyprus.
Canals-Marote R., et al. (2000) interNOSTRUM: a Spanish-Catalan Machine Translation System. Machine
Translation Review, 11, 21-25
Tantu A.C., Adal E. & Oflazer K. (2006) Computer Analysis of the Turkmen Language Morphology. In:
FinTAL, Lecture Notes in Computer Science, pp. 186-193. Springer
Dvok B., Homola P. & Kubo V. (2006) Exploiting similarity in the MT into a minority language. In: LREC2006: Fifth International Conference on Language Resources and Evaluation. 5th SALTMIL Workshop on
Minority Languages,Italy
Garrido-Alenda A., et al. (2003) Shallow Parsing for Portuguese-Spanish Machine Translation In: TASHA 2003:
Workshop on Tagging and Shallow Processing of Portuguese, Lisbon, Portugal
Haji J. (1987) RUSLAN - An MT System Between Closely Related Languages. In: Third Conference of the
European Chapter of the Association for Computational Linguistics (EACL'87), Copenhagen, Denmark
Haji J., Homola P. & Kubo V. (2003) A simple multilingual machine translation system. In: MT Summit IX,
New Orleans, USA
Haji J., Hric J. & Kubo V. (2000) Machine translation of very close languages. In: Proceedings of the sixth
conference on Applied natural language processing pp. 7-12. Morgan Kaufmann Publishers Inc., Proceedings of the
sixth conference on Applied natural language processing
M.Corbi-Bellot A., et al. (2005) An open-source shallow-transfer machine translation engine for the Romance
languages of Spain. In: 10th EAMT conference "Practical applications of machine translation", Budapest, Hungary
32
1 Introduction
The major aim of search engines is actually a question of deciding which documents in a text collection should
be obtained from search algorithm to satisfy users need for information. They use any stemming techniques and
various different techniques to provide the best result first. For example, Google heavily relies on links when it
comes to determine the ranking of a web site. Moreover, the ranking process varies widely from one engine to
another. In addition to these, the stemming technique is used by search engines is not clear. Furthermore,
different stemming techniques dont support by commercial search engines. Hence, all possible cases cant be
examined in terms of the information retrieval for Turkish. Therefore, a search engine for Turkish (SET 1) is
developed . To the best of our knowledge, this is the first online project for researches on Turkish IR
(Information Retrieval). With developing a custom search engine, the effects of stemming techniques can be
examined for Turkish IR. In this search engine, no-stemming (NS) is selected to be the baseline; Turkish
stemmer (TS) and word truncation (WT) method are used as comparator.
Another matter is the test collection used by search engine. In Turkish, only a very limited number of test
collections are available for IR studies. We suffered from similar limitations in creating dataset at our previous
studies [1], [2]. Therefore, a Turkish sentence collector, which obtains data over internet, was developed. For
1Classes
of SET developed in C# programming language are open source and available via the web page
http://bilgmuh.nku.edu.tr/SET/. Moreover, you can try all modules by using this web page.
33
this study, the Turkish sentence collector has been upgraded and developed a crawler that collects news
specifically from Milliyet which is an online newspaper. Using this crawler, news of Milliyet stretching from
2004 to 2007 was able to be accessed and created.
The paper is organized as follows: Section 2 introduces properties of the SET projects and gives general
information such as information models, stemming techniques, crawling techniques and evaluating techniques
that is used in SET. Section 3 examines the results of obtained from an evaluator module. The paper ends with
some concluding remarks.
Description
(?<=href=\").*?(?=\")
openWindow\('(.*?)'\)
Links in Javascripts
popup\('(.*?)'
Before adding an URL into the URL Listing, URL normalization process and unique web page control process
are applied to URL for ensuring a duplicate record is not created. After obtaining the URLs, this crawler
generates a XML file for every web page. For generating this file, it takes string between comment tags of <!-PRINT BASLADI --> and <!-- PRINT BTT --> by using regular expression because the rest is made up of
banners, menus, advertisements and other unnecessary information. These comment tags didnt use in another
years so news between 2004 and 2007 are chosen to use. The generated XML file begin with an information
about news article (<DOCNO>, <DATE> and <CATEGORY> tags) and end with a text of a news article
(<TEXT> and<P> tags), as shown in Fig. 1. The first paragraph of texts is usually the title of the news articles.
34
<DOC>
<DOCNO> 1103 </DOCNO>
<DATE>2007/08/20</DATE>
<CATEGORY>SPOR</CATEGORY>
<TEXT>
<p>
12 Dev Adam'da rota Efes Cup
</p>
<p>
A Milli Basketbol Takmmz, 3 - 16 Eyll tarihlerinde spanya'da dzenlenecek
olan 2007 Avrupa Basketbol ampiyonas ncesi son ciddi provasn zmir'de
yapacak. Ay - yldzllar, 22 - 26 Austos'ta zmir Halkapnar Salonu'nda
gerekletirilecek 6. Efes Pilsen World Cup Mini Dnya Kupas'nda boy
gsterecek. Yarn sabah bu turnuvaya katlmak zere zmir'e gidecek olan
millilerimiz, bugn sabah kondisyon ve akam da taktik olmak zere iki
antrenmanla hazrlklarn srdrecek. Milli Takmmz, turnuvada Letonya ve
Hrvatistan ile A Grubu'nda yer alrken, B Grubu'nda ise in, Srbistan ve Polonya
mcadele verecek.
</p>
</TEXT>
</DOC>
2Zemberek
35
Turkic languages
the web using
n=5is selected in this study. While indexing the test collection with these stemming techniques, they change the
size of storage. (See below Table 2.)
Table 2. The indexing information of NS and stemming techniques
Information
NS
TS
WT5
864035
364786
156328
214
163
148
9.91
6.25
4.73
55
55
1068
906
998
29
11
49
49
49
990
846
804
Stemming techniques reduce the size of storage in memory. TS and WT5 cause 15% and 20% of decrease in
the storage efficiency when comparing with NS.
where t is the weight of a term and d is the document vector. After obtaining term weight, the similarity vector
is computed by using the formula
sim(q, d )
t ,d
2
t ,d
wt ,q
2
t ,q
36
adding these words to the search term, users need parentheses to show the order in which relationships should be
considered. In this case, keywords can be used ((Beikta OR (Kara AND Kartal)) and (Fenerbahe OR
Kanarya)) and (Basketball OR Pota).
3 Experiments
110 ad hoc queries are prepared to evaluate the effectiveness of stemming techniques. These query terms are
divided into three sub groups by using the MAP values of NS and TS.
NS
TS
WT5
0.91
0.67
0.64
ampiyonlar ligi
1.00
0.59
0.77
isizlik sorunu
1.00
0.21
1.00
1.00
0.61
0.57
As these examples indicate, stemming techniques are statistically significantly inefficient in analyzing query
terms derived with inflectional and derivational suffixes. In fourth example, the results gathered by TS and WT5
are different because the stem, i (job), is used in TS however WT5 uses the word, isiz (jobless). According
to this selection, WT5 is more effective because the word that WT5 uses is closer to the users original word,
isizlik (joblessness), in meaning than the word that TS uses. In terms of count of retrieved search results,
these words ( i, isiz, isizlik) appear in 59832, 3533 and 1809 documents, respectively. Number of
relevant documents gathered by using the stem, i, is more than the number of relevant documents gathered by
using the words, isiz and isizlik. But the relevancy percent of the results is lesser due to the fact that there
is 2058 words derived from the stem, i, in test collection and these 59832 documents include all possible
forms of i. In other words, using the stem in search is most effective, if all the words derived from it have
similar meanings. Now, the positive cases are examined by analyzing four example queries. (See below Table 4.)
37
NS
TS
WT5
film festivalleri
0.55
1.00
1.0
0.52
0.87
0.82
0.33
0.62
0.76
Kaltsal hastalklar
0.15
1.0
1.00
In negative examples taking plural suffix -lar out has a negative effect on results, despite the fact that in
positive examples taking plural suffix -ler has a positive effect on results. In last two examples, NS have no
any effect because of query term that has no suffix. NS is efficient when user use suffixes in his query term. In
four positive cases, stemming techniques are observed to be adequate to improve search engine results.
Table 5. Comparison of three different cases
NS
TS
Wt5
Case
Query
MAP
Cnt
MAP
Cnt
MAP
Cnt
Cnt
Positive Cases
0.25
46
0.52
394
0.50
270
48
Equal Cases
0.76
561
0.76
1522
0.77
1193
16
Negative Cases
0.56
233
0.46
1519
0.53
656
46
Overall Results
0.49
166
0.53
1048
0.56
493
110
While evaluating the average performance on 110 query terms, stemming techniques seems to be slightly
better than the other techniques. Improvements of TS, WT5 are 6.68%, 13.16% better than NS, respectively. The
average count of retrieved search results show that no-stemming can be used to narrow the search results when
comparing other methods. In the other hand, TS is an appropriate technique for obtaining more search results
than the others.
In Turkish, Ekmekcioglu and Willett [6] compare the effectiveness of information retrieval by employing
stemmed and non-stemmed query word terms using Turkish news articles of size 6,289 and 50 queries. Sever
and Bitirim [7] evaluate the effectiveness of a new stemming algorithm, namely FINDSTEM, which employs
inflectional and derivational stemmers. Their algorithm provides 25% retrieval precision improvement with
respect to no-stemming. Pembe and Say [8] investigate the question of whether NLP techniques can improve the
effectiveness of information retrieval. Can et al. [4] compare the effects of four different stemming options by
using a large-scale collection. They used 72 ad hoc queries. All these studies are local. To the best our
knowledge, this study is the first online project in Turkish IR.
4 Conclusion
In this study, an online Turkish search engine is described and evaluated the effects of the stemming
techniques on Turkish search results obtained from this search engine. Stemming techniques usually gives
satisfactory results in IR[9], [10]. Still, this study indicates that there are some cases that stemming is proved to
be inadequate. Suffixes that are used in users query term can be critical in some cases. However, stemming
techniques eliminates importance of these suffixes. Regular agglutinative languages such as Turkish encode
38
more information with suffixes than the other languages. Because of that, not only stems but also suffixes are
important in agglutinative language based IR.
Some future research possibilities are modifying this Crawler to obtain different Turkish texts. Another future
work would be to try different information retrieval models such as OKAPI, language modeling and mutual
information model. Moreover, synonyms of query terms are to be used in searching.
References
Uzun, E., Klaslan, Y., H.Agun, V. and Ucar, E., (2008). "Web-based Acquisition of Subcategorization Frames
for Turkish", ICAISC 2008, Editors: Rutkowski L. et.al., ISBN 978-83-60434-50-5, pp. 599-607.
2. Uzun, E., (2007). An internet-based automatic learning system supported by information retrieval [nternet tabanl
bilgi eriimi destekli bir otomatik renme sistemi], Ph.D., Department of Computer Engineering, Trakya
University, Edirne, Turkey
3. Sever,H. and Tonta,Y. (2006). Truncation of Content Terms for Turkish, CICLing 2006, Mexico City, Mexico.
4. Can, F., Kocberber, S., Balcik, E., Kaynak, C., Ocalan, H. C., Vursavas, O. M., (2008). "Information retrieval on
Turkish texts." Journal of the American Society for Information Science and Technology. Vol. 59, No. 3, pp. 407421.
5. Sanderson, M. andZobel J. Information retrieval system evaluation: Effort, sensitivity, and reliability, ACM SIGIR
05, 2005, pp. 162-169.
6. Ekmekcioglu, F.C. and Willett, P. (2000). Effectiveness of stemming for Turkish text retrieval. Program, 34(2), pp.
195-200.
7. Sever, H. and Bitirimi Y. (2003). FindStem: Analysis and evaluation of a Turkish stemming algorithm. Lecture
Notes in Computer Science, 2857, pp. 238-251.
8. Pembe F. C., and Say ACC, (2004). A linguistically motivated information retrieval system for Turkish. Lecture
Notes in Computer Science, 3280, pp. 741-750.
9. Krovetz,R. (1993). Viewing morphology as an inference process, ACM SIGIR93, Pittsburgh: ACM, 1993, pp. 191202.
10. Savoy, J. (2006). Light stemming approaches for the French, Portuguese, German and Hungarian languages, ACM
SAC 06, pp. 1031-1035.
1.
Biography
Erdin Uzun received his BSc (2001), MSc (2003) and PhD (2007) degrees from Computer Engineering Department of
Trakya University respectively. He was a research assistant in this university during this time. Then, he has begun to work in
Computer Engineering Department of Namik Kemal University as assistant professor. He was the vice dean between 2007
and 2010 Faculty of Corlu Engineering. Since 2010, he is the vice chairman of Computer Engineering Department of this
faculty. His research interests include information retrieval, machine learning, data mining, web mining and natural language
processing. He is the developer of SET (Search Engine for Turkish - bilgmuh.nku.edu.tr/SET/) and Teaching (Lesson
Content Management System - bilgmuh.nku.edu.tr/Teaching/)
39
Abstract. Artificial Neural Network (ANN) and Multiple Linear Regression (MLR) models based on
maximal and non-exercise variables are developed to predict maximal oxygen consumption (VO 2max) in
healthy volunteers ranging in age from 18 to 65 years. The dataset includes data of 100 subjects andthe input
variablesof the dataset aregender, age, body mass index (BMI), grade, self-reported rating of perceived
exertion from treadmill test (RPE), heart rate (HR), perceived functional ability (PFA) and physical activity
rating (PA-R). The performance of the models are evaluated by calculating their standard error of estimate
(SEE) and multiple correlation coefficient (R).The results suggest that the performance of VO2maxprediction
models based on maximal variables can be improved by including questionnaire variables (PFA and PA-R) in
the models.
Keywords: maximal oxygen consumption, artificial neural networks
1. Introduction
Cardiorespiratory fitness (CRF) is the ability to perform dynamic, moderate-to-high intensity exercise using
the large muscle groups for long periods of time [1]. CRF depends on the respiratory, cardiovascular, and
skeletal muscle systems and, therefore, is an important component of health and physical fitness [2].The standard
test for determining CRF is the direct measurement of maximal oxygen uptake (VO2max) during a maximal
graded exercise test [3]. Additionally, this method is useful when educating individuals about their overall fitness
status, stratifying cardiovascular risk and prescription and control of fitness programs in healthy individuals and
patient who could be treated using physical exercise [4].VO 2max is the most accurate way to assess CRF,
however maximal exercise tests that directly measure VO2max is impractical in most physical therapy clinics [5]
because maximal tests require expensive gas analysis and ventilation equipment. As a result, researchers have
sought to develop alternative methods that would allow clinicians to predict VO 2max without having to perform
direct measurement of oxygen uptake [6].VO2max could be predicted by using maximal exercise variables,
submaximal exercise variables, non-exercise variables or any combination of these variables. Maximal tests
require expensive equipment, trained staff and a highly motivated client [7]. Submaximal tests are most often
performed on a cycle ergometer, a treadmill or a track to indirectly estimate VO 2max. Although submaximal
tests are faster, safer, and cheaper to administer, they are less accurate than maximal tests [8]. Non-exercise
variables include various demographic (age, gender) , biometric (body mass, height, weight) and questionnaire
(PFA to walk, jog or run given distances and PA-R) information [9].
Froelicher et. al. [10] investigated the hypothesis that an individuals VO2max can be realistically predicted by
the maximal time achieved in the Balke treadmill protocol. The oxygen consumption in the final minute
40
ofexercise of 1025 healthy men who performed a maximal effort in the Balke protocol were linearly regressed on
their maximal treadmill time using a least-squares fit technique. Their results for standard error of estimate (SEE)
and multiple correlation coefficient (R) were 4.26 and 0.72, respectively.
Foster et. al. [11] had an investigation on 230 male patients. Individuals were categorized according to health
status on the basis of having or not having cardiovascular disease (documented myocardial infarction, typical
angina pectoris, or myocardial revascularization surgery). They used the time variable and tried to show
nonlinear regression could produce more accurate results than linear regression for VO 2max prediction. SEE and
R of their models were reported to be 3.35 and 0.97, respectively.
Patton et. al. [12] used a maximal predictive cycle ergometer test for estimating VO 2max. The dataset
contained 15 males and 12 females. The test consisted of pedaling a cycle ergometer (Monark) at 75 rev/min,
beginning at an intensity of 37.5 watts and increasing this amount each minute until the subject could no longer
maintain pedal rate. The highest work rate achieved was recorded as the end point of the test and used to
construct regression equations to predict VO2max. This was compared with two direct measures of VO 2max (an
interrupted treadmill run and an interrupted cycle ergometer procedure at 60 rev/min) and with the submaximal
predictive test. The R of their models was 0.85, however they did not report any results for SEE.
Storer et. al. [13] used 116 females and 115 males, aged 20-70 years, sedentary, nonsmoking and apparently
healthy subjects. They hypothesizedthat cycle ergometer VO 2max could be accurately predicted due to its more
direct relationship with watts. Therefore, they developed an equation including the variables watts, weight and
age. They reported that the R of their model was 0.94.
Jette et. al. [14] investigated if VO2max could be predicted from independent variables measured during the
administration of the Canadian Home Fitness Test. They used 59 subjects between the ages of 15 and 74years
underwent the fitness test and a progressive exercise treadmill test for the direct determination of volitional
VO2max. Average oxygen cost of the last completed exercise stage, body weight, post exercise heart rate and
age were the variables for their models. Their results indicated that VO2max could be adequately predicted. SEE
and R were reported to be 4.08 and 0.90, respectively.
Stickland et. al. [15] investigated the accuracy of the 20-m multi-stage shuttle run test to predict VO2max in 60
men and 62 women young adults. Their models included only one variable called the last half-stage of the shuttle
run completed. For men, the reported values of SEE and R were 4.07 and 0.88, respectively. For women, the
reported values of SEE and R were 3.64 and 0.81, respectively.
Flouris et. al. [16] used the maximal attained speed from the 20-m shuttle run test to predict VO2max of 110
healthy young males. SEE and R of their best model were reported to be 1.70 and 0.96, respectively.
Ruiz et. al. [17] developed an artificial neural network (ANN) equation to estimate VO2max from 20 m shuttle
run test performance (stage), sex, age, weight, and height in 122 boys and 71 girls adolescents aged 13-19 years.
The SEE and R of their model was reported to be 2.78 and 0.96, respectively.
Leger et. al. [18] validated a maximal multistage 20-m shuttle run test for the prediction of VO2max in 91
adults. Maximal speed was the only variable in their model. They hypothesized that the 20-m shuttle run
testcould be a valid and reliable test for the estimation of VO2max. The presented SEE and R values were 5.4 and
0.84, respectively.
There are several limitations of the studies in literature that focus on predicting VO 2max by using maximal
variables. First of all, almost all of these studies used multiple linear regression for developing VO2max
prediction equations. Intelligent data processing techniques such as ANNs have been used in one study
only[17]. Secondly, cross validation was not used in most of these studies, which raises a question about the
reliability of the presented results. Thirdly, all the studies used maximal variables and standard non-exercise
variables such as sex, age, weight and height. In order words, they did not consider including questionnaire type
non-exercise variables in their VO2max prediction models.
In this study, new VO2max models for fit adults based on both maximal exercise and non-exercise variables
are developed by using ANNs. The dataset involves measured VO 2max values of 100 individuals as well as the
maximal test variables; heart rate, grade and self-reported rating of perceived exertion from treadmill test and
also non-exercise variables; gender, age, body mass index (BMI), PFA and PA-R questionnaires. Six different
41
VO2max prediction models are developed using 10-fold cross validation on the dataset. SEE and R values of the
models are computed and it is shown that the performance of the prediction VO 2max model based on maximal
and standard non-exercise variables can be improved by adding questionnaire variables in the prediction model.
The reported values of SEE and R for the best model are 2.28 and 0.94, respectively.Section 2 gives dataset
generation. Section 3 presents VO2max prediction models. Section 4 gives results and discussion. Finally,
Section 5 concludes the paper.
2. Dataset Generation
The dataset contains 100 subjects (50 females and 50 males), that are healthy volunteers ranging in age from
18 to 65 years. All subjects were recruited from the Y-Be-Fit Wellness Program at Brigham Young University
(BYU) and employees from the LDSHospital in Salt Lake City, Utah. During the exercise testing, a physical
activity readiness questionnaire (PAR-Q) and an informed consent document that was approved by the BYU
Institutional Review Board for Human Subjects, were given all subjects to complete. Using a stadiometer and
balance beam scale, volunteers height and body mass were measured and recorded, respectively. Prior to
testing, instructions about the maximal graded exercise test were given to all subjects. Subjects performed a
maximal graded exercise test using a modified version of the Arizona State University (ASU) maximal protocol
[19], with only a slight change in the warm-up. The warm-up consisted of having subjects complete up to three
4-min stages and reach 75% of maximum heartrate (HRmax). During the first stage, subjects walked at a selfselected brisk pace, 4.86.4 km/h. If subjects progressed to the second stage, the speed was increased to a selfselected jogging pace between 4.8 and 6.4 km/h. If 75% of HR max was not achieved by the end of the second
stage, subjects continued to the third stage where the speed was increased to 9.6 km/h or greater. After the warmup, while the test administrator explained the test procedures for the maximal graded exercise test and fit
subjects with a mouthpiece and nose clip, subjects rested for approximately 5 min. As the necessary pace for the
maximal graded exercise test the ending speed of the warm-up served. Thereafter, the grade was increased 1.5%
each minute, while pace remained the same, and continued until subjects reached volitional fatigue and were no
longer able to continue.
Metabolic gases were collected using the TrueMax 2400 metabolic measurement system during the maximal
graded exercise test. Every 15s, the VO2max and respiratory exchange ratio (RER) were computed, averaged,
and printed by an online computer system. At the end of each stage, the subjects exercise heart rate (HR) and
RPE scores were recorded. If at least two of the following three criteria were met, VO2max was considered
valid:
1. HRmax within 15 beats of age predicted maximal heart rate.
2. RER equal to or greater than 1.10.
3. Plateau in VO2max despite an increase in work load.
Subjects who failed to meet at least two of these criteria were dropped from the study. Prior to exercise testing
subjects completed the PFA and a modified PA-R non-exercise questionnaire.
42
each model was determined by trial and error. Table 1 also shows the number of neurons in the hidden layer for
each model.
Table 1: VO2max prediction models and their variables
Model
No
1
Variables
Gender, Age, BMI, RPE
Number of Neurons
In the Hidden Layer
8
11
10
RPE, self-reported rating of perceived exertion from treadmill test; HR, heart rate; GRD, grade
In order to assess the performance of the ANN-based models, MLR-based models have also been developed.
(1)
(2)
In (3) and (4), Y is the measured VO2max value, Y is the predicted VO2max value, Y is the mean of the
measured values of VO2max and N is the number of samples in a test subset.For both ANN-based and MLRbased VO2max prediction models, the individual SEE and R values for each fold have been calculated and
average values for each model are shown in Table 2.As Table 2 shows, among models 1 through 3, which
contain maximal and standard non-exercise variables, model 3 has the lowest SEE and highest R. As is clearly
seen from Table 2,adding questionnaire variables (PFA and PA-R) to models 1 through 3 yields lower SEEs and
higher Rs both for ANN-based and MLR-based models. Table 3 shows the performance improvement of models
in percentage after including the questionnaire variables in the models.
43
Models
Model No
1
2
3
4
5
6
Variables
SEE
3.72
3.59
3.15
2.34
2.23
2.28
R
0.86
0.86
0.89
0.93
0.94
0.94
SEE
5.44
5.43
5.09
3.82
3.77
3.73
R
0.65
0.65
0.72
0.86
0.86
0.87
Regression Method
ANN
1
2
3
MLR
SEE
SEE
-37.19%
-37.78%
-27.55%
+8.4%
+15.12%
+6.09%
-29.86%
-30.44%
-26.75%
+32.18%
+31.96%
+20.37%
Model 2 has the highest decrement in SEE both for ANN-based and MLR-based models. Model 2 also has the
highest increment in R for the ANN-based model whereas model 1 has the highest increment in R for the MLRbased model. It is clear from the presented results that ANNs perform better than MLR for prediction of
VO2max. For instance the SEE of the best ANN model is 2.23 and the SEE of the best MLR model is 3.73. Also,
the R of the best ANN model is 0.94 and the R of the best MLR model is 0.87.
5. Conclusions
In this study, we use ANNs to develop VO2max prediction models by using maximal variables such as heart
rate, grade and self-reported rating of perceived exertion from treadmill test and non-exercise variables gender,
age and BMI. We show that the accuracy of the VO2max prediction models based on maximal and standard nonexercise variables can be improved by including questionnaire variables PFA and PA-R in the models. Models 1
through 3 yielded 37.1%, 37.88% and 27.62% decrements in their SEE values, after including the questionnaire
variables. The most accurate VO2max prediction model is the one that includes the variables gender, age, BMI,
heart rate, PFA and PA-R. According to the presented results, it is understandable that ANN performs better than
MLR for prediction of VO2max. The reported values of R for all ANN-based models are higher than those
reported for MLR-based models. Also, the reported values of SEE for all ANN-based models are lower than
those reported for MLR-based models.
44
References
1.
2.
3.
4.
5.
6.
7.
8.
9.
10.
11.
12.
13.
14.
15.
16.
17.
18.
19.
Wilmore, J. H., et. al. (2008). Physiology of Sport and Exercise. USA: Human Kinetics.
Williams, L.& Wilkins. (2000). AmericanCollege of Sports Medicine. Philadelphia: ACSMs guidelines for
exercise testing and prescription 6th ed.
Jackson, A.S., et. al. (1990). Prediction of functional aerobic capacity without exercise testing. Medicine and
Science in Sports and Exercise 22(6):863-870.
Duque, I. L., et. al. (2009). A new non exercise-based VO2max prediction equation for patients with chronic low
back pain. J Occup Rehabil 19:293-299.
Chatterjee, S., et al. (2006). Prediction of maximal oxygen consumption from body mass, height and body surface
area in young sedentary subjects. Indian J Physiol Pharmacol 50(2):181186.
Balmer, P. B. and OConnel, D. G. (2009). Research corner regression analysis for prediction: understanding the
process. Cardiopulmonary Physical Therapy Journal 20(3):23-26.
Waddoups, L., et. al. (2008). Validation of a single-stage submaximal treadmill walking test. Journal of Sports
Sciences 26(5):491-497.
Hunn, H. M., et al. (2002). The influence of pre-test anxiety, personality and exercise on VO2max estimation.
JEPonline 5(1):5-14.
George, J. D., et. al. (2009). Prediction of maximum oxygen uptake using both exercise and non-exercise data.
Measurement in Physical Education and Exercise Science 13(1):1-12.
Froelicher, V. F. and Lancaster, M. C. (1974). The prediction of maximal oxygen consumption from a continuous
exercise treadmill protocol. Am HeartJ 87:445-450.
Foster, C., et. al. (1984). Generalized equations for predicting functional capacity from treadmill performance. Am
Heart J 107:1229-2134.
Patton, J. F., et. al. (1982). Evaluation of a maximal predictive cycle ergometer test of aerobic power. Eur J Appl
Physiol 49:131-140.
Storer, T. W., et. al. (1990). Accurate prediction of VO2max in cycle ergometry. Med Sci Sports Exerc 22:704-712.
Jette, M., et. al. (1976). The canadian home fitness test as a predictor of aerobic capacity. CMA Journal 114:680692.
Stickland, M. K., et. al. (2003). Prediction of maximal aerobic power from the 20-m multi-stage shuttle run test.
Canadian Society for Exercise Physiology 28(2): 272-282.
Flouris, A. D., et. al. (2005). Enhancing the efficacy of the 20 m multistage shuttle run test.Br J Sports Med 39:166170.
Ruiz, J. R., et. al. (2008). Articial neural network-based equation for estimating VO2max from the 20 m shuttle
run test in adolescents. Articial Intelligence in Medicine 44: 233-245.
Leger, L. A. and Lambert, J. (1982). A maximal multistage 20-m shuttle run test to predict VO2max. Eur J Appl
Physiol 49:1-12.
George, J. D. (1996). Alternative approach to maximal exercise testing and VO2max prediction in college students.
Research Quarterly for Exercise and Sport 67(4):452457.
45
Computer Engineering Department, Iran University of Science and Technology, Tehran 16846-13114, Iran.
1
Abstract. IMS (IP Multimedia Subsystem) network is considered as an NGN (Next Generation Network)
core networks by ETSI. Decomposition of IMS core network has resulted in a rapid increase of control and
signaling message that makes security a required capability for IMS commercialization. The control messages
are transmitted using SIP (Session Initiation Protocol) which is an application layer protocol. IMS networks
are more secure than typical networks like VoIP according to mandatory of user authentication in registration
time and added SIP signaling headers. Also different vulnerabilities have been occurred that lead to SIP
servers attacks. This paper studies the main SIP servers of IMS (x-CSCF) based on ETSI Threat,
Vulnerability and Risk Analysis (TVRA) method. This method is used as a tool to identify potential risks to a
system based upon the likelihood of an attack and the impact that such an attack would have on the system.
After identifying the assets and weaknesses of IMS SIP servers and finding out the vulnerabilities of these
hardware and software components, some security hints that can be used for secure deployment of IMS SIP
servers are proposed. Modeling shows the effects of server weaknesses and threats that reduces availability.
Any designed system has some assets with weaknesses. When threats have accrued based on weaknesses, the
system will vulnerable. Vulnerability analysis optimizes costs and improves security.
1 Introduction
Importance of system security and assets has led different methods to risk assessment and threats analysis.
There are some differences between these methodologies in simulation and implementation standpoint, but goal
of all them appropriates response system requirements, vulnerabilities, threats and activities could prevent
destroying of system assets. Providing suggestions for protection of system privacy, integrity and availability are
the main outputs of vulnerability assessment. TVRA is standard by ETSI organization as a method of systematic
modeling to identify threats, specify protection requisites to reduce risks and enhance system security, also
weaknesses, threats, security objectives and the overall security arrangements will be obtained [1]. System assets
have Weaknesses that may be attacked by threats and may lead to an unwanted incident with breaking certain
pre-defined security objectives. Vulnerability, consistent with the definition given in ISO/IEC 13335, is modeled
as the combination of a weakness that can be exploited by threats. The goal of security design is to ensure
reduction of unwanted incidents and attacks. Unwanted incident is dependent upon the weakness in an asset.
Modeling is completed with the following steps [2]:
46
1)
2)
3)
4)
5)
Vulnerability assessment determines necessary countermeasure items in implementation. This paper has been
trying to use the behaviour methodology and identify system vulnerability. Fist, an IMS network modelled by
TVRA is explained, as a related similar work then, IMS SIP server modelled is given. The final provided table is
used to balance between designing of SIP servers with suitable countermeasures to reduce attack potential.
47
There are deferent methods to ensure the IMS SIP proxy message integrity [11]:
1.
2.
3.
IMS AKA
HTTP digests with integrity supporting
Any both Methods with IPSec or TLS.
User and network communication requires integrity security mechanism. We can implement HTTP digest
authentication with TLS. This mechanism seems also establish some network layer security like transport layer.
The integrity protection mechanism may have no encryption methods, but it must provide message integrity.
Security protection mechanism and algorithm types are agreed in TLS security layer. At user registration time by
AKA, the connection will be protected by TLS and the security level will be higher than the time just IMS AKA
authentication protocol is used.
Attacker cannot guess the produced session key in the TLS, even if she/he could influence AKA protocol [9].
The next important issue raised SIP server security is users accounting. Any user is associated with the SIP
proxy by definition of charging conditions. The network has to provide the desired service to all authorized
users. However, in case of single SIP proxy modeling, this article discussion may not have relevance, but it is
one of the considered implementation cases in modeling. I-CSCF server is billing responsible. User charging
information is added to the SIP message header by P-CSCF (p-charging-vector).
Another major issue that should be considered in modeling is SIP server availability. The most important
weaknesses of SIP proxy is the Vulnerability comes from flooding attacks that is the serious threat for SIP
infrastructure in IMS network [9]. Floods of invite, response and register messages with random source
addresses/ports, are sent to P-CSCF by attacker. It attempts to keep track of the large number of calls. When
legitimate users attempt to make a call, the SIP element is not able to process the messages due to lack of
resources (memory, CPU, processing time, etc). Reduction of the flooding attacks will increase availability of
SIP Proxy servers. To promote of availability, definition of comparison and controlling methods on transferred
message sequences are more practical. For example, a substantial imbalance between an accounting of SIP invite
and SIP 180 ringing messages indicates a flooding attack. Increasing of invite message number comparing with
Ok messages is shown the invite flooding attack [10].
The last part of security objectives is authentication. It is the initial necessary steps to IMS network
communication [6]. The authentication mechanisms are done by the user and the network agreed in
''authorization'' header. Available authentication methods are the following:
Early IMS authentication
Network Attachment Sub-System(NASS)
Http digest (challenging)
http digest over TLS/IPSec
IMS AKA : AV(Rand(IK,CK), Autn, Xres)
These methods are sequenced according to increasing of safety. Digest authentication is password-based
method that allows secure user identification without sending password in plaintext over the network. TLS is
based on Secure Sockets Layer (SSL) and provides data integrity and confidentiality for reliable connection
(TCP). AKA is not only mutual authentication. It has a confidentiality and integrity key negotiation for
connection security. After user authentication ''P-ASSERTED-IDENTITY'' header is added to SIP message by
P-CSCF for showing of IMS trust to a user [3]. This header identifies user has authenticated by network and
causes recently requests are not authenticated due to save the resources [11]. This header field identifies the
source of message even when amount of ''from'' field not provided properly [11]. Table 1 categorized the
discussed IMS SIP server security objectives using ideas contained in TVRA.
48
Server
P/S-CSCF
Accounting
Availability
P/S-CSCF
Logical Asset
Authentication parameters (ISIM, user credentials)
Integrity, cryptography & session keys(P/S-CSCF)
x-CSCF
IPSec, TLS parameters
SIP Methods (invite, Bye, Cancel, Register)
49
CSCF server to connect even with no encryption. With elimination of ''authorization'' field in ''Register'' message
header, server shall select ''Early IMS authentication'' method. Thus eavesdropping could be possible, so the
confidentiality is not supported. Session characteristics would be eavesdropped by attacker and an illegal ''bye''
or ''cancel'' message will be sent, so a legitimate dialog could be dropped. We suggest authenticating ''Bye'' and
''Cancel'' messages, when we have no encryption mechanismin communication.
50
selection of packets have to be discarded. If server CPU load remains on or crosses the threshold limit (80%), for
a period of time, server is under flooding attack [9].
The P-CSCF is the most important IMS SIP servers in security protection. It is suggested to implement PreParsing system on P-CSCF as a packet inspection mechanism. Pre-parsing system can implement out of x-CSCF
Server. Protection system makes delay in packet processing, when it implements out of the P-CSCF as a firewall.
The implementation of pre-parsing mechanism inside of P-CSCF is more practical. One of the prevention
mechanisms against the flooding attack is message processing with severity consideration. For example less
important SIP headers will be discarded. The following notes are suggestedin boundary conditions:
Unknown user ''REGISTER'' message should be blocked. These messages may be the cause of server
overloading.
The ''REGISTER/INVITE'' messages from known legitimate User Equipments (users authenticated
successfully) should be carried out.
TLS and IPSec mechanisms are useful in integrity and confidentiality protection.
Countermeasure must have severity in detail design stage to reduce the cost.
Although authentication of all user requests is a strong control mechanism to remove the user ID fraud,
forgery attacks and message manipulation, it will increase P-CSCF, S-CSCF processing time [13].
The registration process is tightly coupled with authentication process in IMS SIP [11].''REGISTER'' message
authentication is a force function. Choosing authentication mechanism without confidentiality is one of the IMS
SIP server vulnerabilities. It is recommended to restrict this condition. When no encryption mechanism is ready,
an attacker can due to terminate dialogue by sending illegal ''Bye'' and ''Cancel''. We can also control attack by
authentication of these methods.
To detect SQL Injection, pre-parsing system parses message and checks ''username'' field with (;) in
''REGISTER'' method. Protection system alerts: ''SQL-injection has been identified'' and further processing of
message will be stopped. User name is inserted in the Blacklist. Some other''REGISTER'' method fields like:
''Realm'' and ''URI" can be in parsing subjects, too.
Weaknesses
Position
(x-CSCF)
Threat& Attacks
Security
Objectives
Countermeasures
Cancel/bye message
mandatory authentication
when we have no
encryption.
Cancel/Bye
method
p-cscf
communication
without
cryptography
P/S/I
DOS
Authentication
Invite/ Register
method
No pre-parsing
P/S
DOS/Flooding
Confidentiality
Invite/ Register
method
Long
processing time
P/S
DOS/Flooding
Confidentiality
Availability
51
Limitation in using of
Early IMS
Auth.,NBA,Http basic
Auth.
Register method
SIP header
(Cseq, Call-Id,
Via)
SigCmp, TLS,
IPSec data
Weak
authentication/
no integrity
mechanism
SIP message
Heavy algrthm/
increasing of
processing
capacity
Availability
Integrity supporting in
HTTP digest
Transmission
with no.
encryption
Authenticati
on with no
encryption
SQL injection
Message content
processing {syntax (;) with
cut/copy/del world}
Sniffing
SIP message
manipulation
DOS
Confidentiality
Integrity
Availability
Availability
TLS Mandatory
encryption in registration
time
Using of TLS/IPSec
ESP in communication
security protection
Limitation in compression
algorithm choosing in high
level QOS
Conclusion
In this paper a vulnerability analysis was done on the IMS SIP servers based on the TVRA methodology. The
results of this analysis has shown that flooding attacks on the P-CSCF and S-CSCF servers are the most
important treats that target the availability of these servers. Using encrypted authentication mechanism is one of
the best proposed approaches that not only prevent against eavesdropping and message integration attacks, but
also make a preventive mechanism in encountering with denial of service attacks. Some consequences has been
shown in table 3.
References
1.
2.
3.
4.
5.
6.
7.
8.
9.
10.
11.
12.
13.
14.
ETSI TS 102 165-1, (2006), (TISPAN); Methods and protocols; Part 1: Method and proforma for Threat, Risk,
Vulnerability Analysis.
ETSI TR 187 002,(2010), (TISPAN); TISPAN NGN Security (NGN_SEC); Threat, Vulnerability and Risk Analysis
V3.0.2.
Wang, D., Liu, C.(2009), Model-based Vulnerability Analysis of IMS Network, Journal of Networks, VOL. 4, NO. 4.
ETSI TR 187 014,(2009),(TISPAN); eSecurity; User Guide to eTVRA web-database.
ETSI TR 187 011,(2008),(TISPAN); NGN Security; Application of ISO-15408-2 requirements to ETSI standards -guide,
method and application with examples.
3GPP TS 23.228, 3GPP Technical Specification Group Services and System Aspects; (IMS);age 2 (Release 10).
ITU-T, (2003), Security architecture for systems providing end-to-end communications, ITU_T X.805.
3GPP TS 33.203 V10.0.0 ,(2010), 3GPP Technical Specification Group Services and System Aspects; 3G Security;
Access security for IP-based services (releases 10).
Sisalem, D., Floroiu, J, Kuthan, J., Abend, U., Schulzrinne, H.,(2009), SIP Security, published by Jhon Wiley.
Rebahi, Y., Sher, M., Magedanz, T.,(2008),"Detecting flooding attacks against (IMS) networks," Computer Systems and
Applications, 2008. AICCSA 2008. IEEE/ACS International Conference on , vol., no., pp.848-851.
Russell, T., (2008), The IP Multimedia Subsystem (IMS); Session Control & Other Network Operations, published by Mc
Graw Hill.
3GPP TS 33.978, (2008), 3GPP Technical Specification Group Services and System Aspects; Security aspects of early
IMS; Release8.
Bremler-Bar, A., Halachmi-Bekel, R., Kangasharju, J., (2006), '' Unregister Attacks in SIP '', 2nd IEEE workshop on
Secure Network Protocols (NPSec), CA.
3GPP TS 24.229 V9.3.1, (2010), 3GPP Technical Specification Group Core Network and Terminals; IP multimedia call
control protocol based on (SIP) and (SDP)
52
Biographies
Afsane Madani Afsaneh Madani received her M.Sc in ICT (secure telecommunication) from Computer Engineering
Department of Iran University of Science and Technology (IUST) in 2011. She has received her B.S degree in electronic
engineering from Amirkabir university of Technology (AUT). She is working in network security department in Iran ITRC
institute. Her main research interests include network security, network signaling and protocols, 3G and NGN.
Hassan Asgharian Hassan Asgharian received his M.Sc. in Computer Engineering from Amirkabir University of
Technology (AUT) in 2009 and He also has got his B.Sc. from Computer Engineering Department of Iran University of
Science and Technology (IUST) in 2006. Currently he is a PhD candidate in computer engineering in IUST and working as a
research assistant in the Network Research Group (NRG) of the Research Center of Information Technology (RCIT) of
Computer Engineering Department in Iran University of Science and Technology.
Dr. Ahmad Akbari Ahmad Akbari received his Ph.D. in Electrical Engineering from the University of Rennes 1, Rennes,
France, in 1995. Dr. Akbari is currently an associate professor at Iran University of Science and Technology, Iran. Dr. Akbari
works on Speech Processing related research topics (especially speech enhancement) for more than 20 years. His current
research interests include Intrusion Detection and Response Systems, VoIP Communications, Next Generation Networks,
VoIP and SIP Security and also Data Communications Networks. Dr. Akbaris work has appeared in several peer-reviewed
journals and conference proceedings.
53
Abstract. In this paper, we develop artificial neural network (ANN) models for predicting the performance
measures of a message passing multiprocessor architecture interconnected by the Simultaneous Optical
Multiprocessor Exchange Bus (SOME-Bus), which is a fiber-optic interconnection network. OPNET Modeler
is used to simulate the SOME-Bus multiprocessor architecture and to create the training and testing datasets.
The performance of the ANN models is evaluated by calculating metrics such as mean absolute error (MAE),
root mean squared error (RMSE), relative absolute error (RAE), root relative squared error (RRSE) and
correlation coefficient (r). It is concluded that ANN models shortens the time quite a bit for obtaining the
performance measures of a message passing multiprocessor and can be used as an effective tool for this
purpose.
1. Introduction
The demand for even more computing power has never stopped. A number of important problems have been
identified in the areas of defense, aerospace, automotive applications and science, whose solutions require a
tremendous amount of computational power [1]. Parallel computers with multiple processors are opening the
door to teraflops computing performance to meet the increasing demand of computational power [2]. The
programmingmodel for multiprocessor systems is important because it affects the amount of operating system
overhead involved incommunication operations as well as the level of involvement required by the programmer
tospecify the processor interaction required by the application.Shared address programming and message passing
are examplesof the programming models that can be supported by multiprocessor systems.The message passing
paradigmrequires a higher level of programmer involvement and knowledge of the details of theunderlying
communication subsystem in order to explicitly direct the interprocessorcommunication. Shared memory
programming offers the application programmer amodel for using shared data that is identical with that used
when writing sequential programs [3].
The performance analysis of a multiprocessor architecture is an important factor in designing such
architectures. Statistical simulation is a powerful tool in evaluating the performance measures of a
multiprocessor architecture. In such simulation, the architecture is modeled by using specific probability
distributions [4, 5]. The simulation can be time consuming especially when the architecture to be modeled has
many parameters and these parameters have to be tested with different values or probability distributions.
54
There is only a single study in literature that shows artificial intelligence techniques could be used to predict
the performance measures of a multiprocessor architecture [4]. In that study, a broadcast-based multiprocessor
architecture called the SOME-Bus employing the distributed shared memory programming model was
considered. The statistical simulation of the architecture was carried out to generate the dataset. The dataset
contained the following variables: ratio of the mean message channel transfer time to the mean thread run time,
probability that a block can be found in modified state, probability that a data message is due to a write miss,
probability that a cache is full and probability of having an upgrade ownership request. Support vector regression
was used to build prediction models for estimating average network response time (i.e. the time interval between
the instant when a cache miss causes a message to be enqueued in the output channel until the instant when the
corresponding data or acknowledge message arrives at the input queue), average channel waiting time (i.e. the
time interval between the instant when a packet is enqueued in the output channel until the instant when the
packet goes under service) and average processor utilization (i.e. average fraction of time that threads are
executing). It was concluded that support vector regression model is a promising tool for predicting the
performance measures of a distributed shared memory multiprocessor.
In this paper,we develop artificial neural network (ANN) models to predict the performance measures of a
multiprocessor architecture employing the message passing protocol. The simulated architecture is the SOMEBus, which is a high performance optical multiprocessor architecture. ANN models are developed to predict
average channel waiting time, average channel utilization, average network response time, average processor
utilization and average input waiting time. The following five error metrics are used for evaluating the accuracy
of the prediction models: mean absolute error (MAE), root mean squared error (RMSE), relative absolute error
(RAE) and root relative squared error (RRSE) and correlation coefficient (r). Considering the results, it is clear
that except for the average network response time, ANN performs a good job in predicting the performance
measures of a multiprocessor architecture employing the message passing protocol.
This paper is organized as follows. Section 2 summarizes the SOME-Bus architecture. Section 3 presents
overview of ANNs. Section 4 gives the statistical simulation and dataset generation. Section 5 presents
results.Finally, Section 6 concludes the paper.
55
(1)
A wide range of algorithms have been developed for training the network to achieve the optimum model
performance, while ensuring generalization and computational efficiency. For feed-forward ANNs, the back
propagation algorithm is the most frequently used [11].
56
processor (ranging from 1 to 6), ratio of the mean thread run time to channel transfer time (ranging from 0.05 to
1), thread run time (exponentially distributed with a mean value of 100), and pattern of the destination node
selection (uniform and hot region). In the uniform traffic, the destination node is selected using uniform
distribution among the nodes. .In the hot region pattern, the destinations of the 25% of the packets are chosen
randomly within a small hot-region consisting of 12.5% of the nodes [13].The dataset obtained as a result of the
simulation contains four input and five output variables. The input variables are: T/R, node number, thread
number, traffic pattern. The output variables are: average processor utilization, average network response time,
average channel waiting time, average input waiting time, and average channel utilization. Table 1 shows
descriptive statistics of the output variables.
CWT
CU
NRT
PU
IWT
21.84
0.30
521.45
0.60
189.37
Minimum
0.02
0.02
47.72
0.24
15.87
Maximum
186.40
0.85
1027.36
0.95
356.91
31.66
0.20
209.31
0.19
94.23
Mean
Standard
deviation
CWT:Channel Waiting Time; CU:Channel Utilization; NRT:Network Response Time; PU:Processor Utilization; IWT:Input Waiting Time
A multistage feed forward ANN [8] is used for developing the prediction models. Levenberg-Marquardt back
propagation algorithm is used for training the networks. A tansigmoid transfer function is used in the hidden
layer and a pure linear transfer function is used in the in the output layer.
57
5. Results
The following five metrics are used for evaluating the accuracy of the prediction models:MAE, RMSE, RAE
and RRSE and r [14]:
MAE
RMSE
1
n
| Y
1
n
i 1
(2)
(Y Y ) 2
RAE
Y |
i 1
| Y
(3)
Y |
i 1
n
(4)
| Y Y |
i 1
RRSE
(Y
Y ) 2
i 1
n
(5)
(Y Y ) 2
i 1
Y Y Y ' Y '
1
n 1
Y
Y '
(6)
where n is the number of data points used for testing, Y is the measured value, Y is the predicted value,
the average of the measured values,
measured values and
Y is
Table 2 shows the performance of the prediction models. As is clearly seen from Table 2, except for the
average network response time, ANN performs a good job in predicting the performance measures of a
multiprocessor architecture employing the message passing protocol.
Table2. Summary of regression models to estimate message passing architecture performance
Performance Measures
RMSE
MAE
RAE
RRSE
0.99
1.63
1.04
0.20
0.05
0.99
0.01
0.00
0.16
0.03
0.99
14.63
10.98
0.26
0.08
0.99
0.09
0.05
0.24
0.06
0.99
8.26
5.80
0.27
0.09
58
6. Conclusion
This paper proposes to use ANN-based models to predict the performance measures of a message passing
multiprocessor architecture. The basic idea is to collect a small number of performance measures by using a
statistical simulation and predict the performance of the system for a large set of input parameters based on
these. Future research can be performed in a number of areas. The first area would be expanding the number of
input parameters in the dataset. The second area would be feature extraction on input variables. In this case, the
critical attributes that best predict performance measures can be selected from a candidate set of attributes
through feature selection algorithms combined with ANNs.
Acknowledgments
We would like to thank Dr. Constantine Katsinis for letting us include the material about the SOME-Bus
architecture in this paper. We also would like to thank OPNET Technologies, Inc. for letting us use the OPNET
Modeler under the University Program.
References
1.
2.
3.
4.
5.
6.
7.
8.
9.
10.
11.
12.
13.
14.
Duato, J., Yalamanchili,S., andNi. L. (2003). Interconnection Networks: An Engineering Approach. USA: Morgan
KaufmannPublishers.
Guller,D.E., SinghJ.P., and Gupta, A. (1999).Parallel Computer Architecture: A Hardware/Software
Approach.USA: Morgan Kaufmann Publishers.
Chan,F., CaoJ., and Sun, Y. (2003). High-level abstractions for message-passing parallel programming. Parallel
Computing29(11):1589-1621.
Akay, M.F. and Abaskele, . (2010). Predicting the performance measures of an optical distributed shared
memory multiprocessor by using support vector regression.Expert Systems with Applications37(9): 6293-6301.
Wenisch, T.F., et al. (2006). Statistical sampling of microarchitecture simulation. In Proceedings of the 20th
Parallel and distributed Processing Symposium. RhodesIsland, Greece.25 - 29 April 2006.
Katsinis, C. (2001). Performance analysis of the simultaneous optical multi-processor exchange bus.Parallel
Computing 27(8):1079-1115.
Katsinis, C. and Hecht, D. (2004). Fault-tolerant DSM on the SOME-Bus multiprocessor architecture with message
combining.In Proceedings of the 18th International Parallel and Distributed Processing Symposium (IPDPS04).
Santa Fe, New Mexico. 26-30 April.USA.
Alpaydn,E. (2010).Introduction to Machine Learning.London. England: MIT press.
Lavinea B.K. and Blank T.R. (2009). Comprehensive Chemometrics Chemical and Biochemical Data Analysis
3(18):571-586.
Chen M.-S. and Yen H.-W.(2011). Applications of machine learning approach on multi-queue message scheduling.
Expert Systems with Applications38(3):33233335.
Andries P. Engelbrecht (2002). Computational Intelligence: An Introduction, Wiley.
OPNET Modeler, OPNETUniversity program, http://www.opnet.com/university_program.
Aci, C.I. and Akay, M. F. (2010). A new congestion control algorithm for improving the performance of a
broadcast-based multiprocessor architecture.Journal of Parallel and Distributed Computing70(9):930-940.
Witten, I.H. and Frank, E. (2005).Data Mining: Practical Machine Learning Tools and Techniques. Canada:
Morgan KaufmannPublishers.
59
1 Introduction
Many real life problems can be modeled by differential equations, displaying the interrelations between
mathematics, physical and biological sciences or engineering [1, 2]. Therefore, studying the solution methods for
differential equations has always been an important research subject [3-5].
Perturbation methods [6] are widely used for computing approximate analytical solutions of differential
equations. Due to the small parameter limitation of perturbation methods which makes the solution valid for
weakly nonlinear systems, alternative methods such as the modified and multiple-scale Lindstedt-Poincar
method [7, 8], the linearized perturbation method [9], the homotopy perturbation method [10] and iterationperturbation methods [11-13] etc. are used for strongly nonlinear systems.
However, a general solution method valid for all types of equations which do not require special assumptions
and transformations is lacking in the literature.
As an extension of the perturbation-iteration algorithms (PIAs) for algebraic equations [14-16], the PIAs in [1,
17] which do not require special transformations or initial assumptions are applicable to a wide range of
differential equations.
Although the number of correction terms in the perturbation expansion is just one ( n 1) and the algorithms
are limited to at most third order in the Taylor series expansion ( m 3) , the three PIAs described in [1, 17]
involve lengthy algebraic computations which are hard to do by hand. Computer algebra systems can be very
useful for carrying out these computations.
60
With a generalization towards working for general n and m such that 1 n m , the PIAs in [1, 17] are
implemented in Mathematica, a leading computer algebra system. The Mathematica package
PerturbationIteration.m[18] automatically carries out all the steps of the algorithms.
The paper is organized as follows. Section 2 outlines the key steps of the PIA(2, 3) algorithm up to 2
iterations, i.e., n 2 and m 3 , on a Bratu type initial value problem. In Section 3, the steps of the PIA(1, 6)
algorithm up to 2 iterations are illustrated on the same initial value problem. Section 4 has comparison of our
results with earlier perturbation-iteration results [1]. In Section 5,Mathematica package
PerturbationIteration.mis briefly described.
2 PIA(2, 3)
The Bratus initial value problem [1]:
u
u 2 e 0,
(8)
0 x 1
u (0) u (0) 0
where u u (x) will be used as our leading example to illustrate the steps of the algorithm. The exact solution
for (1) is given as [1]: u 2 ln(cos x) .
First introduce an auxiliary perturbation parameter to the nonlinear term in (1):
u 2 e
(2)
0.
u n1 u n (u c,1 ) n (u c,2 ) n .
Upon substitution of (3) into (2), expanding in a Taylor series with up to third order, and re-arranging the
terms with respect to similar powers of yields:
2
(u n 2) (( u c,1 ) n 2 u n ) (( u c, 2 ) n 2 (u c,1 ) n u n )
(4)
1 3
3
( u n 2 u n (u c,1 ) n (u c, 2 ) n ) 0.
3
Equation (4) contains two unknowns (u c,1 ) n and (u c,2 ) n which require two equations to be solved. Hence
equation (4) is separated into two blocks as follows:
(5)
(u c,1 ) n 2 u n u n 2 0,
2
(u c, 2 ) n 2 (( u c,1 ) n (u c, 2 ) n ) 2 u n (u c,1 ) n u n
1 3
u n 0,
3
u0 0
and solving (5) with (6) yields
61
(7)
2
(u c,1 ) 0 x ,
(u c , 2 ) 0
1 2 x
2 x
2 2 x
2 x 2
e
(1 2 e
e
2e
x ).
2
u1
(8)
1 2 x
2 x
2
e
(e
1) .
2
Now substituting (8) in (5) leads to a system with variable coefficients. To overcome this, in (5), we replace
(u c,1 )1 0,
(u c,2 )1 2 (( u c,1 )1 (u c,2 )1 )
1
( 16 e
3 2 x
9e
2x
9e
2x
3 2x
) 0,
24
(u c,i )1 (0) 0, (u c,i )1 (0) 0, i 1, 2.
Solving (9) gives:
(10)
(u c,1 )1 0,
(u c , 2 )1
e
3 2 x
(e
2 x
1)( 1 e
2 x
64(e
2 2 x
3 2 x
384
4 2 x
5 2 x
36 2 x (e
2 2 x
3 2 x
) ).
Hence from (3), (8) and (10), the result of the second iteration is:
u2
3 2 x
(1 512 e
3 2 x
6 2 x
4 2 x
( 255 36 2 x ))
(11)
384
3e
2 2 x
(85 12 2 x ) ).
3 PIA(1, 6)
Still working on the Bratus initial value problem (1), we will now describe the algorithm with one correction
term and with terms up to sixth order in Taylor series expansion.
Only one correction term is taken in the perturbation expansion:
u n1 u n (u c ) n .
(12)
Substituting (12) into (2), then expanding in a Taylor series with up to sixth order, and re-arranging the terms
with respect to similar powers of yields:
62
(13)
(u n 2) (( u c ) n 2 u n ) ( 2 (u c ) n u n )
1 3 3
1 4
4
2
2
(u n 6 u n (u c ) n )
( u n 12 u n (u c ) n 12 (u c ) n )
3
12
1
60
1
(u n 20 u n (u c ) n 60 u n (u c ) n )
6
( u n 30 u n (u c ) n 180 u n (u c ) n 120 (u c ) n ) 0.
360
60
1
3
un
360
6
2
u n u n (u c ) n 2 (u c ) n (u c ) n
u n (1 (u c ) n )
(u n
1
2
12
1
3
(14)
(u c ) n
u n (1 (u c ) n )
u n ) ( 2 2 (u c ) n (u c ) n ) 0,
(u c ) n (0) 0, (u c ) n (0) 0.
Taking u 0 0 for the initially assumed function in (14) and omitting the nonlinear terms in (u c ) n leads to:
(15)
(u c ) 0 2 (u c ) 0 2 0,
(u c ) 0 (0) 0, (u c ) 0 (0) 0.
We leave better treatment of the nonlinear terms in (14) for future work and continue with solving (15), which
yields
(u c ) 0
(16)
1 2 x
2 x
2
e
(e
1) .
2
u1
(17)
1 2 x
2 x
2
e
(e
1) .
2
Now substituting (17) in (14) leads to a nonlinear system with variable coefficients. To overcome this
difficulty, in (14), we replace u n in the coefficient of (u c ) n by u 0 instead of u1 and omit the nonlinear terms
in (u c ) n . Doing so yields
(u c )1 2 (u c )1
604e
e
3 2x
8 2x
6 2 x
(e
2x
2x
2 2x
4
1) (1 4e
76 e
23040
4390e
4 2x
604e
5 2x
76 e
6 2x
4e
7 2x
) 0,
(u c )1 (0) 0, (u c )1 (0) 0.
63
(18)
(u c )1
6 2 x
2 x
(e
1)( 1 e
2 x
155 e
(19)
2 2 x
1612800
1555e
3 2 x
29730 e
29730e
7 2 x
4 2 x
379050 e
8 2 x
155 e
1555e
238560 2 x (e
5 2 x
6 2 x
5 2 x
9 2 x
379050 e
6 2 x
10 2 x
11 2 x
e
e
)).
Hence from (12), (17) and (19), the result of the second iteration is:
u2
6 2 x
(1 154 e
2 2 x
1400e
(20)
3 2 x
1612800
28175e
4 2 x
2370900 e
6 2 x
28175 e
8 2 x
1400 e
9 2 x
10 2 x 12 2 x
7 2 x
154e
e
120 e
(1988 2 x 9631)
120 e
5 2 x
(1988 2 x 9631)).
4 Numerical Comparisons
Comparisons of the different perturbation-iteration algorithms with numerical solutions of equation (1) are
given [1] in Table 1. Although all iteration algorithms rapidly converge to the numerical solutions, PIA(1, 6)
seems to perform slightly better than the others.
Table 8. Comparisons of perturbation-iteration algorithms and numerical results for Bratuss initial value
problem [1]
x
PIA(1, 3)
Numerical
PIA(2, 3)
PIA(1, 6)
Solution
u1
u2
u1
u2
u1
u2
0.1
0.01001
0.01001
0.01001
0.01001
0.01001
0.01001
0.01001
0.2
0.04026
0.04026
0.04026
0.04026
0.04026
0.04026
0.04026
0.3
0.09138
0.09135
0.09138
0.09135
0.09138
0.09135
0.09138
0.4
0.16445
0.16431
0.16445
0.16431
0.16445
0.16431
0.16445
0.5
0.26116
0.26059
0.26116
0.26059
0.26116
0.26059
0.26116
0.6
0.38393
0.38212
0.38391
0.38212
0.38391
0.38212
0.38392
0.7
0.53617
0.53134
0.53610
0.53134
0.53610
0.53134
0.53614
0.8
0.72278
0.71124
0.72249
0.71124
0.72249
0.71124
0.72264
0.9
0.95088
0.92542
0.94983
0.92542
0.94983
0.92542
0.95037
1.23125
1.17818
1.22772
1.17818
1.22772
1.17818
1.22947
64
(1)
as
an
example,
call
the
function
In[3]:=PerturbationIterationAlgorithm[
{''[]2*Exp[[]]==0},
{[0]==0,[0]==0},,,2,3,2
]
which will produce the results (8) and (11). Similarly, all results in this paper can be reproduced by using our
package.
References
1.
2.
3.
4.
5.
6.
7.
8.
9.
10.
11.
12.
13.
14.
15.
16.
17.
18.
Aksoy Y.; and PakdemirliM. (2010). New perturbation-iteration solutions for Bratu-type equations.Computers and
Mathematics with Applications 59: 2802-2808.
Hoppensteadt, F. C. and Peskin, C. S. 1992. Mathematics in Medicine and the Life Sciences, Springer-Verlag,
New-York.
Ince, E. L. 1944. Ordinary Differential Equations, Dover, New York.
Kamke, E. 1959. Differentialgleichungen: Lsungsmethoden und Lsungen, Akademische Verlagsgesellschaft,
Leipzig.
Murphy,G. M. 1960. Ordinary Differential Equations and Their Solutions, D.van Nostrand Company, Inc.,
Princeton.
Nayfeh,A. H. 1981. Introduction to Perturbation Techniques. John Wiley and Sons, New York.
Hu,H. (2004). A classical perturbation technique which is valid for large parameters.Journal of Sound and
Vibration 269: 409-412.
Pakdemirli, M.; Karahan M. M. F.; and BoyacH. (2009). A new perturbation algorithm with better convergence
properties: Multiple Scales Lindstedt Poincar method.Mathematical and Computational Applications 14: 31-44.
He, J. H. (2003). Linearized perturbation technique and its applications to strongly nonlinear oscillators.Computers
and Mathematics with Applications 45: 1-8.
He, J. H. (2003). Homotopy perturbation method: A new nonlinear analytical technique.Applied Mathematics and
Computation 135: 73-79.
He, J. H. (2006). Some asymptotic mehods for strongly nonlinear equations.International Journal of Modern
Physics B 20: 1141-1199.
He, J. H. (2001). Iteration perturbation method for strongly nonlinear oscillations.Journal of Vibration and Control
7: 631-642.
zi T.; and Yldrm, A. (2009). Generating the periodic solutions for forcing van der Pol oscillators by the
iteration perturbation method.Nonlinear Analysis: Real World Applications 10: 1984-1989.
Pakdemirli M.; and Boyac, H. (2007). Generation of root finding algorithms via perturbation theory and some
formulas.Applied Mathematics and Computation 184: 783-788.
Pakdemirli, M.; Boyac H.; and Yurtsever, H. A. (2007). Perturbative derivation and comparisons of root-finding
algorithms with fourth order derivatives.Mathematical and Computational Applications 12: 117-124.
Pakdemirli, M.; Boyac H.; and Yurtsever, H. A. (2008). A root finding algorithm with fifth order
derivatives.Mathematical and Computational Applications 13: 123-128.
Pakdemirli, M.; Aksoy, Y.; and Boyac, H. (2011). A new perturbation-iteration approach for first order differential
equations.Mathematical and Computational Applications 16: 890-899.
Gkta, . (2011). Software is available on http://www.turgutozal.edu.tr/ugoktas
65
Biographies
nal Gkta holds a BSc. degree (1993) in mathematics from Boazii University (Turkey), MSc. (1996) and Ph.D. (1998)
degrees in mathematical and computer sciences from the Colorado School of Mines (U.S.A.).
His research focuses on integrability tests of nonlinear partial differential equations and nonlinear semi-discrete lattices.
He also works on symbolic solution methods of differential and difference equations, symbolic summation methods and
various other aspects of symbolic computation.
Dr. Gkta was employed by Wolfram Research, Inc. (makers of Mathematica) as a senior kernel developer prior to
joining the Department of Computer Engineering at Turgut zal University (Turkey) in 2010.
Mehmet Pakdemirli was born in Ankara, Turkey in 1963. He received his BS. degree (1985) and MS. degree (1987) in
mechanical engineering from Boazii University (Turkey). He completed his Ph.D. studies in 1991 from stanbul Technical
University (Turkey). He worked as a postdoctoral research associate in Virginia Tech (U.S.A.) for two years. He became
assistant professor in 1993, promoted to associate professor in 1994 and finally to full professor in Celal Bayar University,
Manisa, Turkey. He visited the University of Michigan (U.S.A.) for six months in 1996 and King Fahd University of
Petroleum and Minerals (Saudi Arabia) in 2003-2004. He is currently professor and president of Celal Bayar University.
His major areas are applied mathematics, mechanical vibrations of continuous systems and non-Newtonian fluid
mechanics. He has authored, co-authored some hundred papers in internationally refereed journals.
Yiit Aksoy was born in zmir, Turkey in 1982. He received his BS. degree (2005) and MS. degree (2007) in mechanical
engineering from Celal Bayar University (Turkey). Currently, he is a Ph.D. candidate in mechanical engineering at Celal
Bayar University.
His major areas are fluid mechanics and approximate analytical methods of mathematical physics.
66
Abstract. In many web content extraction applications, parsing is crucial issue for obtaining the necessary
information from web pages. A DOM parser is preferably used in this task. However, major problems with
using a DOM parser are time and memory consumption. This parser is an inefficient solution for applications
which need web content extraction between tags. Therefore, we develop a lightweight parser, namely SET
Parser, which is utilized regular expressions and string functions for extracting these tags. In experiments,
2000 web pages are crawled from Sabah and Milliyet for experimental setup. The average size of a DOM tree
created from these web pages is as large as 9.58 times of the average size of the web pages. On the other
hand, SET Parser in parsing contents is less time consuming than DOM. These experiments indicate that SET
Parser is more useful for extracting contents from web pages.
1 Introduction
Document Object Model (DOM, www.w3.org/DOM) parsing is widely used a key operation for processing
web pages. Especially, a DOM parser is the preferred way of extracting web contents [1], [2], [3], [4]. Web
documents have a hierarchy of informational units called nodes; DOM is a way of describing those nodes and
the relationships between them. In DOM, everything is an object that consists of element names and contents of
them. A DOM parser presents a document as a tree-structure in memory. However, the size of DOM tree created
from a web document is larger than the size of the original web document. This means that memory and time
consumption will increase constantly. For exceptionally large documents, parsing and loading the entire
document can be slow and resource intensive.
Web pages often contain irrelevant contents such as pop up ads, advertisements, banners, unnecessary images,
extraneous links etc. around the body of an article that distracts users from actual content. There are lots of
applications about extraction of useful and relevant content from web pages including text summarization, cell
phone and pda browsing, speech rendering for the visually impaired [5], [6]. For extraction of relevant
content,unnecessary tags and contents in Web pages should be removed. In general, extracting content of text in
conventional techniques required knowledge on structure of the text [7]. Web pages are semi-structured
documents mostly written in HTML that defines tags. These tags can be structured with DOM parser. However,
Cai et al. [8] proposed that only some tags (<table>, <tbody>, <tr>, <td>, <p>, <ul> and <li>) are very important
and commonly used in web pages. Yerlikaya and Uzun [9] use the layout html tags, <div>, <table> and <ul>, for
67
obtaining relevant contents. In content extraction, because of A DOM parser takes all tags into account, it is not
efficient.
In search engine for Turkish (SET6) Project, we developed an intelligent crawler that automatically detects
irrelevant contents and extracts relevant contents. We realized that memory and time consumptions are important
issue in this task. Therefore, we developed a lightweight parser (namely SET Parser) for reducing these
consumptions. In this study, we describe this parser and examine the improvements in memory and time.
The paper is organized as follows: Section 2 gives general information about DOM, Regular Expressions and
the use of them C# programming language. Section 3 introduces the SET parser which is utilized regular
expressions and string functions for extracting contents. Section 4 examines memory and time consumptions in
DOM and SET Parser. The paper ends with some concluding remarks.
To access the DOM programmatically, mshtml library is used. write function creates a DOM object for a web
document. Nonetheless, while structuring a web document, a DOM object takes all tags into consideration. It is
clear at this point that memory consumption and time consumption will increase. Moreover, regular expressions
can be used for extracting links and images.
The origin of regular expressions lies in automata and formal language theory. Mathematician Stephen Cole
Kleene described these models of using his mathematical notation called regular sets in the 1950s. Regular
expression is a language that can be used to edit texts or obtain sub-texts that based on a set of defined rules.
6Classes
of SET developed in C# programming language are open source and available via the web page
http://bilgmuh.nku.edu.tr/SET/. Moreover, you can try all modules by using this web page.
68
Regular expressions can be used for obtaining relevant contents from web pages. For example, the following
table indicates patterns that are used for extracting links and images from web pages.
Table 1. Regular expression patterns for extracting links and images
Links
(?<=href=\").*?(?=\")
Images
<\\s*img [^\\>]*src\\s*=\\s*([\"\\'])(.*?)>
Regular expressions are supported by many programming languages such as C, C++, .NET, Java, Perl, Ruby,
Python to search and manipulate text based on patterns. For accessing .Net regular expression engine from C#,
we import the namespace
System.Text.RegularExpressions.
Regular expressions can save you time when considering DOM. Therefore, we write two functions which are
used regular expression for this task. However, regular expressions have been a bottleneck for the unknown
nested structure that is a layout stored within the structure of another layout. Now, we describe the main function
of SET Parser.
3 SET Parser
SET Parser has three functions. Two functions of SET Parser are used only regular expressions for extracting
links and images.Nevertheless, unfortunately regular expressions dont give accurate results in the nested
structure of tags. Therefore, we write codes for obtaining inner contents from layout tags.
69
For example, when using SET Parser in a web page for <div class="content"> tag, firstly regular expression
pattern <div.class="content"> (by replacing blanks with a point) is prepared for calculating the count of this
pattern in a web page. Then, start tag as <div and end tag </div are created as appropriate to the tag. Start
position and end positions are found by using these tags. indexof string function can be used for this task.
substring can be utilized for extracting a proposed content by using these positions. Afterwards, the count of
start tag and end tag are calculated whether the proposed content is correct or not. If the count of start and end
tags are equal, this content is added to the result array. While implementing this algorithm, it is clear that the last
control operation increases the response time in contents which have a lot of inner tags. Therefore, we examine
the effects of the count of inner tags in the experiments.
4 Experiments
In experiments, test collection obtained from Milliyet (www.milliyet.com.tr) and Sabah (www.sabah.com.tr)
contains 2000 web pages. 1000 pages are crawled for each newspaper. The total sizes of these collections are
around 160.47 MB in UTF-8 format. Table 2 gives more information about this collection.
Table 2. Information about Test Collection
Document Size
Information
Tag Count
Milliyet
Sabah
Milliyet
Sabah
56.32 MB
104.15 MB
Median a Document
57.67 KB
106.65 KB
1103.41
1320.07
Minimum a Document
30.04 KB
36.7 KB
587
496
Maximum a Document
112.66 KB
140.39 KB
3730
1998
This collection contains tags that are used to design web pages and contents that present between tags. A web
document in this collection may take approximately 82.16 KB. A crude observation on the text collections
reveals that text collection of Sabah contains more unnecessary tags and contents than Milliyet when considering
that the sizes of actual contents are approximately similar. While doing experiments, we also take into account
the effect of file size and the number of used tags. In the first analysis, we compare the parsing time of links and
images for parsers of DOM and SET. The following table gives more information about this comparison.
Table 3. Information about the average time and the average count of extracted links and images
DOM
SET Parser
Average Count
Links
Images
Links
Images
Milliyet
214.49
355.54
248.84
55.42
Sabah
741.48
1098.93
272.75
171.99
Milliyet
21.05
4.68
248.84
55.42
Sabah
29.91
7.83
272.75
171.99
This analysis indicates that DOM Parser needs longer time to parse when compared with SET Parser which is
used regular expressions. Moreover, the parsing time of a DOM object for Sabah is very slow when compared
with Milliyet because of the file sizes and the count of tags used in web pages. The use of Regular expressions is
70
more suitable solution for extracting links and images than the use of DOM. The analysis show that the pattern
used for extracting links is slow than the pattern of extracting images in terms of time efficiency since patterns
used for extraction have an effect on the results. The second analysis is on examining the parsing time of layout
tags which have nested structure. We select three layout tags for each newspaper. These tags for respectively
Milliyet and Sabah:
1. Main layouts: <div class="content"> and <div id="container_body">
2. Menu layouts: <div class="menu"> and <div id="altmenu" class="altmenu_style altmenu_font_style">
3. Related news Layouts: <div class="haber"> and <div class="pad10lrb print_hidden" style="">
In this analysis, the parsing time of DOM and SET parsers is examined for better understanding the
improvements. (See below Table 4)
Table 4. Analysis of the Parsing Time in the Nested Structure
DOM Parser
SET Parser
Milliyet
Sabah
Milliyet
Count of the
Sabah
Count of the
(ms)
(ms)
(ms)
Inner Tags
(ms)
Inner Tags
639.68
1729.51
3.15
17.88
228.02
424.24
636.15
1690.21
2.50
3.00
37.47
176.68
617.21
1637.98
1.79
1.00
5.54
3.00
Avg.
631.01
1685.90
2.48
7.29
90.34
201.31
Tags
Table 4 indicates similar results when compared with Table 2. However, the count of inner tags increases the
parsing time of SET Parser. For example; the count of the inner tags in Tag 3 of Sabah is approximately 3.00
and the parsing time is 5.54 milliseconds. Nevertheless, the count of the inner tags in Tag 1 of Sabah goes up to
424.24 and also the parsing time increases to 228.02 milliseconds. But still, the SET Parser is more suitable
solution for obtaining contents between tags. The average parsing time of DOM Parser is 254.4 and 18.7 times
shorter than the average parsing time of DOM parser, respectively in Milliyet and Sabah. Third analysis is
examined the impact of the size of web pages on parsing time. For this analysis, the main layouts of Milliyet and
Sabah are selected. The following figure indicates the parsing time of DOM parser for tag <div class="content">
of Milliyet and tag <div id="container_body"> of Sabah. (See Fig 1. (1))
500
450
400
350
300
250
200
150
100
50
0
ms
ms
2200
2000
1800
1600
1400
1200
1000
800
600
400
200
0
405060708090100110120130140
Milliyet KB
405060708090100110120130140
KB
Milliyet
Fig 1. The parsing time for DOM Parser (1) and SET Parser (2) in different file sizes
The parsing time of Sabah is longer than the parsing time of Milliyet in similar file sizes because of the count
of tags used in Sabah is more than the count of tags used in Milliyet. That is, the parsing time of DOM parser
depends on the count of tags used in web pages. Moreover, Fig 1. (1) shows that the increasing size of web pages
has a negative effect on parsing time.The Fig 1. (2) indicates the parsing time of SET Parser for same tags. The
71
parsing time of Milliyet is lesser than the parsing time of Sabah. The parsing time of SET parser depends on the
count of inner tags. In the other words, the Fig 1. (2) shows that Sabah contains more inner tags than Milliyet.
The last analysis is for investigating the memory consumption of DOM Parser and SET Parser. (See below Table
5.)
Table 5. Analysis of Memory Consumption
Average of File Sizes
(KB)
Usage (KB)
Usage (KB)
(%)
Milliyet
57.67
539.08
61.90
88.52
Sabah
106.66
1035.42
142.77
86.21
Avg.
82.16
787.25
102.34
87.00
DOM parser stores all contents and relations in memory. However, SET Parser only contains required part of
contents and file size in memory. Therefore, SET Parser improves the memory usage with %87.00. The size of a
DOM tree created from a document is as large as 9.58 times of the size of the original document. Wang et al.
[10] claim that the size of a DOM tree crated from a document may be 10 times larger than the original size.
However, this finding is not experimental but our study argues this claim for HTML DOM.
Conclusion
DOM tree constructed from a web page is widely used step for many content extraction algorithms. However,
the use of DOM is not efficient on algorithms that are focused on specific tags. Because of DOM takes all tags
into account in this task. In this study, the experiments show thatthe use of SET Parser is more suitable solution
for extracting contents than the use of DOM. That is, a lightweight parser like the one in this study can be
utilized for this task.
These analyze show that SET Parser was used for obtaining relevant contents of Turkish newspapers, namely
Milliyet and Sabah. This method provides less parsing time and memory than the use of DOM. Some future
research possibilities are adapting this parser to extract different web contents. Furthermore, the results of our
search engine (SET) may be improved by using texts obtained from this parser.
References
1.
2.
3.
4.
5.
6.
7.
8.
lvarez-Sabucedo L. M., Anido-Rifn L. E., Santos J. M. (2009). Reusing web contents: a DOM approach.
Software: Practice and Experience 2009, 39(3): 299314.
Kaasinen E., Aaltonen M., Kolari J., Melakoski S., Laakko T. (2000). Two Approaches to Bringing Internet
Services to WAP Devices. WWW9; 231-246.
Wong W. and Fu A. W. (2000). Finding Structure and Characteristics of Web Documents for Classification. In
ACM SIGMOD 2000; Dallas, TX., USA
Zheng Y., Cheng X., Chen K. (2008). Filtering noise in Web pages based on parsing tree. The Journal of China
Universities of Posts and Telecommunications; 15, 46-50.
BuyukkoktenO., Garcia-MolinaH., Paepcke.A. (2001). Accordion Summarization for End-Game Browsing on
PDAs and Cellular Phones. In Proc. of Conf. on Human Factors in Computing Systems (CHI'01)
Buyukkokten,O., Garcia-Molina, H., Paepcke.A. (2001). Seeing the Whole in Parts: Text Summarization for Web
Browsing on Handheld Devices, In Proc. of 10th Int. World-Wide Web Conf..
Gupta S., Kaiser G. E., Peter G. (2005). Chiang M F, Starren J. Automating Content Extraction of HTML
Documents. World Wide Web: Internet and Web Information Systems; 8, 179-224.
Cai D., Yu S., Wen J. R., Ma W. Y. (2003). Extracting content structure for web pages based on visual
representation. APWeb'03 Proceedings of the 5th Asia-Pacific web conference on Web technologies and
applications.
72
Biographies
Erdin Uzun received his BSc (2001), MSc (2003) and PhD (2007) degrees from Computer Engineering Department of
Trakya University respectively. He was a research assistant in this university during this time. Then, he has begun to work in
Computer Engineering Department of Namik Kemal University as assistant professor. He was the vice dean between 2007
and 2010 Faculty of Corlu Engineering. Since 2010, he is the vice chairman of Computer Engineering Department of this
faculty. His research interests include information retrieval, machine learning, data mining, web mining and natural language
processing. He is the developer of SET (Search Engine for Turkish - bilgmuh.nku.edu.tr/SET/) and Teaching (Lesson
Content Management System - bilgmuh.nku.edu.tr/Teaching/)
Tark Yerlikaya received his BSc degree in Electronics & Communications Engineering Department of Yldz Technical
University (1999), MSc (2002) and PhD (2006) degrees in Computer Engineering Department of Trakya University. He was
a research assistant between 1999 and 2006 in this university. Since 2006, he is the assistant professor in same university.
Between 2006 and 2008, he was the head of the Division of Computer Hardware in this department. Since 2009, he is the
chairman of Computer Technologies and Information Systems Department in Kean Yusuf apraz School of Applied
Sciences of Trakya University. His research interests are cryptology, security, web mining and intelligent systems.
Meltem Kurt received her BSc degree in Computer Engineering Department of Trakya University in 2010. Now, she is the
MSc student in same department. Her research interests are cryptology, web mining and information retrieval.
73
Iran University of science and technology, Computer Engineering Department, University Road, Hengam Street,
Resalat Square, Tehran, Iran
1
m_iranmanesh@comp.iust.ac.ir, 2akbari@iust.ac.ir
Abstract. In recent years, intrusion detection has emerged as an important technique for network security.
Most of intrusion detection systems use primary and raw features which are extracted from network
connection without any preprocessing on the features. Dimensionality reduction is crucial when data-mining
techniques are applied for intrusion detection to transform features. Many data mining algorithms have been
used for this purpose. Manifold learning is an emerging and promising approach in nonlinear dimension
reduction. Isometric feature mapping (Isomap) is one of the most prominent manifold learning methods that
have been investigated in this domain. Researchers successfully applied Isomap on intrusion detection
systems as a nonlinear dimension reduction method. But, manifold learning algorithms have some problems
such as operation on batch mode. Therefore they cannot be applied efficiently for a data stream. Incremental
Landmark Isomap which is an increment version of classical landmark Isomap can handle the problem of
new data points. This method is applied on NSLKDD dataset. The results demonstrate higher detection rate
for this method, comparing to linear methods such as PCA.
Keywords: intrusion detection system, dimensionality redution, manifold learning, isomap, incremental
landmark isomap, PCA.
1 Introduction
With rapidly development of communication, there is no restriction of real distance for people to contact each
other. It helps people in many areas, such as business, entertainment, education, etc. In particular, the Internet
has been used as an important component of business models[1] . Besides the improvement and efficiency of the
network, it is distinguished that unauthorized activities by external attacker or internal sources are increased
dramatically. So the intrusion detection system which was introduced by Anderson [2] is an important problem
nowadays.
The aim of the intrusion detection system is to classify benign users from intruders and attackers with high
accuracy. One of the important factors to achieve this aim is input data and features, which are extracted from
the raw input data. Whenever input features have more information about the benign and intruder users, the
intrusion detection system can better separate different type of users, and the security of the protected
system/network is improved. From this point of view, preprocessing on the raw input features has an important
role in intrusion detection systems. Feature preprocessing methods are divided into two main groups: feature
selection and feature transformation. We investigate feature transformation in this study.
74
Typical datasets for intrusion detection are generally very large and multidimensional. With the growth of
high speed networks and distributed network based data intensive applications storing, processing, transmitting,
visualizing and understanding the data is becoming more complex and expensive. To tackle the problem of highdimensional datasets, researchers have developed some methods such as PCA [3] and MDS [4]. These methods
fail to act properly on real-world datasets, because these datasets are non-linear. Therefore using non-linear
dimension reduction instead of linear dimension reduction for mapping manifolds to their intuitive dimensions is
more rational.
Non- linear manifold learning algorithms such as LLE[5] and Isomap[6] have been used for nonlinear
dimension reduction. These methods are divided in two main categories. First category includes methods that try
to maintain globally structure of dataset while second category maintains locally structures. For instance, Isomap
is based on maintaining globally geometry of dataset.
Landmark Isomap [7] is a variation of Isomap, which preserve all of attractive attributes, but is more efficient.
It selects some data points as landmarks to construct the map. Landmark points are chosen, randomly.
An important fact that exists in data mining domain is that sometimes the information should be collected
sequentially through dataflow. Manifold learning algorithms operate in batch mode. It means that all data
should be available during training and they cannot be applied on dataflow. Incremental manifold learning has
been invented for this purpose. Martin law et all in [8] proposed incremental Isomap and incremental L-Isomap
that can solve this problem. Their method can handle unseen data, and it can be applied on data stream too.
In this paper, we use the incremental version of L-Isomap. This method could handle the problem of data
streams. Furthermore, because of landmark part of this algorithm, it can solve bottlenecks of classical Isomap.
The rest of the paper is organized as follows. Section 2 provides a brief overview on Isomap, L-Isomap and
Incremental Isomap and then it describes our contributrion. The experimental results are described in section 3.
The conclusions are presented in section 4.
2 Methods
2.1 Isomap
Isomap is a generalization of MDS (Multi-Dimensional Scaling) in which the pairwise distance matrix is
replaced by the matrix of pair wise geodesic distances. The algorithm contains three steps:
1.
Construct graph that shows manifold of data points in high dimensional space. This graph is built on
input data according to certain rules which should reflect the structure in the neighborhood of data
points. It uses one of two prominent methods to build the graph. These two prominent methods are
KNN-neighborhood and -neighborhood. In KNN method would be the neighbor of if it was in the
k nearest neighborhood of , and in -method is the neighbor of if the distance between them is
less than .
2. Compute pair wise distance matrix D with the Floyd or Dijkstra algorithm.
3. Apply MDS on D.
MDS is a linear dimension reduction technique that embeds data points in low dimensional space. When input
data points has a low-dimensional Euclidean structure, classical MDS is guaranteed to find embedding which
exactly preserve Euclidean distance between data points. Given Euclidean distance matrix D, this algorithm
embeds input data points such as and to a low dimensional space and so that:
|
(9)
Here, the complexity is O(kN2), Where N is the number of data points, and k is the dimension of the
embedding [4].
75
2.
3.
Storing the distance matrix: classical Isomap needs O(N 2)storage while L-Isomap needs O(Nn). Here, N
is the number of data points and n is the number of Landmarks. When N is too large, landmark version
will need much less memory storage in comparison with classical Isomap.
Calculating distance matrix: Using Dijkstras algorithm this is O(knNlogN) where k is the size of
neighbourhood.
Finding the embedding: in classical Isomap it takes O(N3), but in L-Isomap it reduces to O(n2N).
2.3Incremental Isomap
Suppose that data points for
exist. Batch Isomap algorithm can calculate for
, so that
is embedded form of original data points in low dimensional space. Then, the new data point
arrives.
The goal of incremental Isomap [8] is to update the transformed data points
so that to best preserve the
updated geodesic distances. This is done in three stages:
3 Results
In this section, we evaluate methods using the NSL-KDD dataset[9]. NSL-KDD is a dataset suggested to solve
the inherent problems of the KDD99 dataset mentioned in[9]. However, this new version of the KDD data set
still suffers from the problems discussed by McHugh[10] and may not be a perfect representative of existing real
networks. Because of the lack of public datasets for network-based IDS, we believe it still can be considered as a
suitable benchmark allowing researchers to compare different intrusion detection methods. Furthermore, the
76
number of records in the NSL-KDD training and test sets is reasonable. This advantage makes it affordable to
run the experiments on the complete set without the need to select a small portion randomly. As such, evaluation
results of the different works will be consistent and comparable[11]. Table (1) shows the summary of the NSLKDD dataset.
Application
Number of Features
Number of Samples
Train +
Train
41
125973
Train 20% +
Train
41
25192
Test +
Test
41
22544
Test -
Test
41
11850
NSL-KDD has two different training datasets: Train+ and Train 20%+. The dataset Train 20%+ contains 20%
samples of the Train+, with almost all varieties of samples in Train+. We use Train20%+ as the training dataset.
For the test phase, we use two test datasets of NSL-KDD, Test+ and Test-. We ran the experiments on a Pentium
IV 2 GHz machine with 1 GB memory.
As mentioned before, using incremental manifold learning method instead of classical manifold learning can
handle problem of data streams. The number of neighbors to construct manifold, is a parameter that has to be set.
After feature reduction with incremental L-Isomap, we classified data points with decision tree classifier.
Another time, after feature reduction with PCA, data points were classified with decision tree. Table (2-3) shows
number of features and the accuracy of classification for each method. Each value in this table is the average of
10 independent runs of various methods.
Table(2):Accuracy of classification for NSl-KDD(+)
Number of neighbor s
for L-Isomap
40
NSL-KDD(+)
PCA
Classical L-Isomap
Original dataset
58.39
69.48
78.01
10
67.78
71.37
20
76.57
79.78
30
76.97
80.21
Number of embedding
PCA
Classical L-Isomap
Original dataset
for L-Isomap
dimension
40
46.67
57.34
58.19
10
51.67
59.65
20
55.34
62.7713
30
57.95
61.90
Results show higher accuracy for incremental L-Isomap, in comparison with PCA. As we can see in tables,
when the embedding dimension gets lower, the difference between accuracy of two methods becomes more. For
example, when the embedding dimension is 5, the difference between accuracy of two methods is more than
10%. This difference is less than 4%, when the embedding dimension is 30. This fact shows us that the accuracy
of incremental L-Isomap is better in comparison with PCA, when the number of features is very low.
77
Besides, original datasets without applying feature reduction algorithms are classified. It can be seen that
incremental L-Isomap improves accuracy of classifying, when it reduces the number of features for each data
point to its half. This shows that some features are irrelevant or noise and incremental L-Isomap can omit them.
4 Conclusion
In this paper we applied incremental L-Isomap for dimension reduction in IDS and compared it with PCA.
Experiment results demonstrate higher accuracy for incremental L-Isomap in comparison with PCA.
Furthermore, the difference between accuracy of two methods is more, when the number of reduced features is
low.
Additionally, we compared accuracy of original dataset without applying feature reduction method. The use of
Incremental L-Isomap improves the overall performance of IDS mainly due to three reasons. Firstly, it can
handle the problem of new data points. Secondly, it reduces the dimension, thereby making the computational
efforts less. Third reason is the reduction of noise in the data. By reducing the noise, we gained a better
classification of normal and abnormal data.
References
1.
Shon T, Moon J. "A hybrid machine learning approach to network anomaly detection", Information Sciences, 2007,
p. 3799-3821.
2. J.P. Anderson, Computer Security Threat Monitoring and Surveillance, James P. Anderson Co., Fort
Washington, PA, Tech. Rep. 79F296400, Apr. 1980, p. 1-56.
3. I.T. Jolliffe, "Principal Component Analysis", Springer-Verlag, New York, 1986.
4. T.F. Cox, and M.A.A. Cox., "Multidimensional Scaling", Chapman &Hall, London, 1994.
5. S.T. Roweis, L.K. Saul, Nonlinear Dimensionality Reduction by Locally Linear Embedding, Science, vol. 290,
December 2000, p. 2323-2326.
6. J.B. Tenenbaum, V. de Silva, and J.C. Langford, A Global Geometric Framework for Nonlinear Dimensionality
Reduction, Science, vol. 290, 2000, p. 2319-2323.
7. Vin De Silva and Joshua B. Tenenbaum. "Global Versus Local Methods in Nonlinear Dimensionality Reduction" ,
In Advances in Neural Information Processing Systems 15, MIT Press, 2003, p. 705712.
8. Law, M., Zhang, N., Jain, A.: "Nonlinear manifold learning for data stream", In Berry, M., Dayal, U., Kamath, C.,
Skillicorn, D., eds.: Proc. of the 4th SIAM International Conference on Data Mining, Lake Buena Vista, Florida,
USA. 2004, p. 3344.
9. Mahbod T, Ebrahim B, Wei L, Ali A.G, "A Detailed Analysis of the KDD CUP 99 Data Set", In proceeding of
Computational Intelligence in Security and Defense Application (CISDA 2009).
10. McHugh J. "Testing intrusion detection systems: a critique of the 1998 and 1999 darpa intrusion detection system
evaluations as performed by lincoln laboratory", ACM Transactions on Information and System Security, 2000, vol
3, p. 262294.
11. http://iscx.ca/NSL-KDD/
78
Abstract. This paper presents the results of our experimental study about primality testing algorithms and
their parallelization. Both of these topics are require for the implementation of security protocols. MillerRabin, Chinese Hypothesis and Lucas-Lehmer-Riesel primality test algorithms are used; it is aimed to assure
the more secure systems to prevent the data theft. After each algorithm is rendered to run on multicore system
parallel, they are run on distributed memory, so the instant key change will be provided faster for security.
Keywords: Computer Security, Distributed and Parallel Processing, Primality testing algorithms.
1 Introduction
In recent years need for security increased as data theft rose. New security systems become vital in our lives.
Big percentage of PC users started using antivirus softwares. Lots of people securely shops via internet with the
products of security protocols. These protocols use math to obtain security [1]. Generation and testing of prime
numbers are important for cryptosystems like RSA and Diffie-Hellman gain their security [2]. While prime
number gets bigger security becomes higher in these methods. However finding a big prime is not easy, specially
numbers bigger than 64bit due to todays computer architecture. Also producing only one prime and building up
the system with it may create security issues. So, also new primes should be generated. When finding a big
prime takes a long time, this methodology also becomes inefficient. If the time needed to get a big prime as
output can be shortens, this methodology can be used to create more secure and efficient security systems.
Using parallel systems may be the answer. With a library called OpenMP[3] or by the multithreading;
instructions that can run parallel can be executed simultaneously. We have been expected that this shortens
execution time dramatically[4]. Even with one more step, instructions may be executed in discrete memory with
an API called MPI. By the help of these two approaches, creating an efficient cryptosystem that works within
desired time interval may be possible.
2 General Info
In this project execution time is the issue. The methodology used is determined by terms of reducing running
time as much as we can. While deterministic primality testing algorithms returns absolute results they are
slow[5]. Hence, probabilistic approaches are chosen named Miller-Rabin Primality Testing Algorithm for
accuracy and Chinese Primality Test for faster execution. One deterministic approach named Lucas-LehmerRiesel also studied for execution time comparison.
OpenMP library is used to divide code on shared memory and MPI API is used to execute the code on discrete
memory.
79
Function millerRabin ( n, k )
n = 1 + 2**a * m;
foreach i <= k do
b Random( )
( 1, n);
z (b**m) mod n;
while not (( j = 0 and z = 1 ) or ( z = n 1 ))do
if j > 0 and z = 1 or ++j = a then
return false;
z (z**2) mod n;
return true;
Function chineseTest ( N )
80
The algorithm is very similar to the LucasLehmer test, but with a variable starting point depending on the
value of k.
If we define a sequence {ui} for all i > 0 by:
ui = u2i-2 2
Then N is prime if and only if it divides un2.
There are some ways to find the starting value of sequence ui -> (u0);
If k is equal to 1 and n is prime, then we are dealing with Mersenne numbers and can take u 0 = 4.
If n 3(mod 4), then we can take u0 = 3.
If k = 1 or 5 mod 6 and 3 does not divide N, then we take
Otherwise, we are in the case where k is a multiple of 3, and it is more difficult to select the right value of u 0.
Function lucasLehmerRiesel ( N )
If un-2 (mod N) = 0
N is prime
Else
N is composite
else
N is not suitable to test
3 Experimental Discussion
After algorithms are selected, they have been coded in Visual Studio in C language. To be able to observe
execution times and compare them, first they implemented without using any additional library. Due to todays
computer architecture, biggest number we could test was a 32 bit number. In each algorithm results are
compared. All tests are run in Intel Core i7 CPU Q740 @1.73GHz, on single core.
81
Run Time(sec)
32-bit Prime
Run Time(sec)
4534761
0.063
4534771
0.063
7568367
0.094
7568369
0.093
47483657
0.578
47483659
0.563
67543931
0.844
67543937
0.813
345234567
4.328
345234581
4.453
563459887
7.688
563459891
7.532
1432562571
19.392
1432562581
19.549
2147483641
30.283
2147483647
30.428
It is seen at Table 1 and Table 2 that Chinese Primality Test is faster than Miller-Rabin Primality test.
However Miller-Rabin is more accurate. We knew that if Chinese Test determines that number is composite, it is
a deterministic result. Then there is no point to test the number with Miller-Rabin which takes longer to check.
So two algorithms are combined and one hybrid code is implemented. With this code it takes slightly longer to
check a number, however it is more accurate. To comparison in terms of time a deterministic algorithm LucasLehmer-Riesel also tested.
Run Time(sec)
32-bit Prime
Run Time(sec)
4534761
0.000
4534771
0.032
7568367
0.000
7568369
0.094
47483657
0.64
47483659
0.328
67543931
0.484
67543937
0.984
345234567
2.625
345234581
3.672
563459887
3.516
563459891
4.048
1432562571
10.137
1432562581
13.407
2147483641
27.362
2147483647
15.126
After those implementations GMP library is used.[11][12] With help of this library, numbers bigger than 64bit
can be tested. Only two implementations Hybrid code and Lucas-Lehmer-Riesel- are converted to work with
big numbers and the experimental results are presented in Table 3 and Table 4.
82
Run Time(sec)
0.000
224.816
0.00002
712.523
0.00013
1928.12
0.00006
64-bit Prime
1432562581(32bit)
2147483647(32bit)
4323413483
5341265353
7516192771
15245325703
60245325709
Run Time(sec)
186.001
332.325
497.036
600.948
1140.503
2342.043
8767.142
As seen from the Figure 1, in hybrid primality testing approach, the part that makes the code not possible to
parallelize is SuccessiveSquare function. Due to it recursively calls itself, that part should be in critical region.
Due to the most time wasting part is that function, there is no meaning to parallelize the algorithm.
83
Figure 2 shows us, each output is an input for next loop. Due to second iteration must wait for the first one;
parallelization is also not suitable for this approach.
4 Conclusion
The main purpose of this project was to decrease the time needed to find a big prime in acceptable time
interval. By doing that, creating instant keys for example to use in RSA systems, would became much easier,
therefore crypto systems were expected to be more secure. If the time spent to create a key; that is the cost to
create a key, becomes lower, life-time of a key gets smaller. When and if a key is cracked, thanks to this system,
it would have been already changed. However due to selected algorithms were not suitable to parallelization, we
could not test and compare results. In spite of everything, this project still can continue by selecting different
primality testing algorithms.
5 Future Works
Our study is continuing by checking other primality testing approaches to find one which can be coded to run
parallel. After we seen that, algorithms we have chosen cannot be executed in terms of parallelism, next we will
try to execute Elliptic Curve Primality Testing algorithm on multicore architecture. The reason why this
algorithm selected is because it is more likely to be adaptable on parallel systems. With this primality approach,
it can be possible to test primality on multicore systems and achieve our purpose.
84
References
1.
2.
3.
4.
5.
6.
7.
8.
9.
10.
11.
12.
13.
85
didem.oktem@ege.edu.tr, 2sen@cse.deu.edu.tr
Abstract. Expert Systems (ES) are intelligent computer programs, which use knowledge and inference
procedures to solve problems in the way that a human-expert should do. Rule base component of the ES is
used to make decisions in an ES. But rule base component can also be used in other artificial intelligence
techniques like Genetic Algorithms (GA). In this study, a GA is designed to optimize the contents of an inservice training program of a corporation and the rule base mechanism of the ES is adopted in a GA in XML
format to be used when needed.
1 Introduction
Nowadays people us many different computer programs to achieve many different tasks in different sectors. In
some cases, the computers and the computer programs replace with human beings completely. In such cases,
Expert Systems (ES) are used to handle tasks without any employee. ES are intelligent computer programs,
which use knowledge and inference procedures to solve problems in the way that a human-expert should do [1].
These systems are composed of mainly three components mutually attached to each other. These components are
user interface (UI), inference engine and the rule base, which usually occur in the form of a set of IFTHEN rules [2]. The inference engine uses the IF-THEN statements of the rule base to make decisions. For some
ES shells, there exists some certain rule base syntax but in some cases, if the ES is prepared from scratch, the
software developer may feel himself free about representing rule base statements.
Correspondingly, an independent rule base representation can be used not only in an ES, but also in some
other problem solving methods like Genetic Algorithms (GA). GAs are introduced in 1970s by John Holland
[3]. They are considered as stochastic search algorithms and in GAs, the probable solutions to a problem are
represented as chromosome like strings. For example, a rule base component can easily be added to a GA if it
is used for solving a Constraint Satisfaction Problem (CSP) or an optimization problem with some hard and soft
constraints.
XML is a simple text-based format for representing structured information: documents, data, configuration,
books or transactions [14].In this study, Extensible Markup Language (XML) is used as a rule base
representation technique in a software, which finds optimum solutions to the curriculum planning problem of inservice education of corporations by using the genetic algorithm and expert system mechanisms together.
The training, which is applied to the employees of a company, is composed of different modules and each
module may have prerequisite modules. The rules of the system clarify the relations of the modules with each
other from the prerequisite notion point of view. That is, some modules have other modules as their
prerequisites. The trainee has to take a module after its prerequisites have been taken. This approach leads us to
prepare a list of rules among the modules. The list prepared is later used for optimization of the modules in
different phases of the study. For this reason, the rule list is stored in XML format.
86
2 Literature Review
Timetabling problems are one of the most popular optimization problems which the researchers have been
seeking for various methods to solve them in an efficient manner. Evolutionary algorithms are the most efficient
and convenient way to solve these kind of problems in short run time periods. Optimization can be done in many
application areas like education, flight traffic control in airports, nurse rostering or operating room timetable in
hospitals [4]. There are several studies on timetabling problem which seek for optimum solutions especially for
educational corporations. The researchers have studied on high school timetables and obtained quite effective
results when compared to the timetables prepared manually. Petrovic and Burke [15] and Qu et al. [16] are some
of the most popular studies on this application area. Pillay and Banzhaf [17] have studied on examination
timetabling problem and have implemented a heuristic system to handle the problem.
While handling optimization problems, there exist many studies in which the ES methods and GAs are used
together as in the studies of Tseng [5] for land-cover classification, Choi [6] for managing the auction
mechanisms and Nguyen [7] for compressor scheduling. The problem solving techniques of evolutionary
algorithms meet with the requirements of the expert system software in many different application areas like
product design [8], cost management [9], unmanned air vehicles [10], decision making tools in fashion [11],
composite laminate design [12] and problems like determining substation locations [13] can be mentioned as the
application areas that GAs and ESs are used together.
3 Methods
In the study, the IF-THEN statements of a rule base are converted to the chromosomes of the initial population
of the GA. In order to represent the IF THEN statements as permutation encoded chromosomes, the rule base
data is first converted to a sparse matrix, in which each row of the matrix corresponds to one of the rules. The 1s
in the ith row of the sparse matrix represent the prerequisite modules of the ith module. The rules of the modules
are inserted into the system via the UI of the software. The number of prerequisite modules depends on the
importance of the module in the in-service education package.
The software is prepared with C#.NET platform. In the program, there exists a user interface to insert the
relevant data to the system before the GA is run. The most important input for the GA to be run is the
prerequisite rules among the modules of a training program. These rules are inserted via the user screen by
picking the prerequisite modules of a particular module and defining them as prerequisite for each other. When
the rules about the modules are inserted into the system, an XML file, including the rules among the modules is
prepared as given in the program code given below:
XmlTextWriter textWriter=newXmlTextWriter("RulesXML.xml", null);
textWriter.WriteStartDocument();
textWriter.WriteStartElement("modules");
for (int i = 0; i < 39; i++)
{ textWriter.WriteStartElement("module");
textWriter.WriteAttributeString("mno", i.ToString());
textWriter.WriteAttributeString("hour", hours[i].ToString());
for (int j = 0; j < 39; j++)
{ if (prereqs[i, j] == 1)
{
textWriter.WriteStartElement("prereq");
textWriter.WriteString(j.ToString());
textWriter.WriteEndElement();
}
}
textWriter.WriteEndElement();
}
textWriter.WriteEndDocument();
textWriter.Close();
87
The XML file has some tags related to the rule information of each module. The prerequisite module numbers
for each module are represented between the tags <prereq> and </prereq>. In the optimization phase of the
software, the initial population of the GA is constructed from the XML file. XML file format provides a generic
and platform independent data storage environment for the rule base data. The sample XML file is given in Fig.
1.
Fig 1. The XML file with the prerequisite rules of the modules.
C# has the components to form (XmlTextWriter) and parse (XmlTextWriter, TreeNode) the XML documents.
The XML file for the rules is obtained with the program code for the chosen lesson. The prerequisite rules for the
different training programs are stored in separate XML files.
While converting the XML rules to the sparse matrix used in the GA, the following program code is used:
XmlTextReader reader = newXmlTextReader("RulesXML.xml");
TreeNode tnmodules = newTreeNode();
tnmodules.Text = "modules";
treeView1.Nodes.Add(tnmodules);
TreeNode tnmodule = null;
TreeNode tnprereq = null;
while (reader.Read())
{if (reader.NodeType == XmlNodeType.Element && reader.Name == "module")
{ tnmodule = newTreeNode();
tnmodule.Text = reader.GetAttribute("mno");
row = Convert.ToInt32(tnmodule.Text);
tnmodules.Nodes.Add(tnmodule);
}
elseif (reader.NodeType == XmlNodeType.Element && reader.Name == "prereq")
{ reader.Read();
tnprereq = newTreeNode();
tnprereq.Text = reader.Value;
col = Convert.ToInt32(tnprereq.Text);
tnmodule.Nodes.Add(tnprereq);
prereqs[row, col] = 1;
}
}
reader.Close();
88
89
References
1.
2.
3.
4.
5.
6.
7.
8.
9.
10.
11.
12.
13.
14.
15.
16.
17.
Giarratano J., Expert Systems: Principles and Programming, Riley, Fourth Edition, (2004).
Sutrisna M., Potts K. and Buckley K., 2003, An Expert System As a Potential Decision Making Tool In the
Valuation of Variations, Proceedings of the RICS Foundation Constructionand BuildingResearch Conference,
COBRA Press, ISBN: 1-84219-148-9, pp: 314-326.
Holland H.J., Adaptation in Natural and Artificial Systems, Cambridge, MIT Press, (1975).
Cardoen B. et al., Operating Room Planning and Scheduling: A Literature Review, European Journal of
Operational Research 201, Elsevier, pp: 921-932 (2010).
Tseng M-H. et al., A Genetic Algorithm Rule-Based Approach for Land-Cover Classification, ISPRS
Journal of Photogrammetry & Remote Sensing 63, Elsevier, pp: 202-212 (2008).
Choi J.H. et al., Utility Based Double Auction Mechanism Using Genetic Algorithms, Expert Systems with
Applications 34, Elsevier, pp: 150-158 (2008).
Nguyen H.H. et al., A Comparison of Automation Techniques for Optimization of Compressor Scheduling,
Advances in Engineering Software 39, Elsevier, pp: 178-188 (2008).
Chaoan L., The Expert System of Product Design Based on CBR and GA, International Conference on
Computational Intelligence and Security Workshops, IEEE, pp: 144-147 (2007).
Chou J-S., Generalised Linear Model-Based Expert System for Estimating the Cost of Transportation
Projects, Expert Systems with Applications 36, Elsevier, pp: 4253-4267 (2009).
Kuroki Y. et al., UAV Navigation by an Expert System for Contaminant Mapping with a Genetic Algoritm,
Expert System with Applications, Elsevier, (In press) (2010).
Wong W.K. et al., A Decision Support Tool for Apparel Coordination Through Integrating the KnowledgeBased Attribute Evaluation Expert System and the T-S Fuzzy Neural Network, Expert Systems with
Applications 36, Elsevier, pp: 2377-2390 (2009).
Kim J-S., Development of a User-Friendly Expert System for Composite Laminate Design, Composite
Structures 79, Elsevier, pp: 76-83 (2007).
Chakravorty S. ve Thukral M., Choosing Distribution Sub Station Location Using Soft Computing
Technique, International Conference on Advances in Computing, Communication and Control (ICAC309),
ACM, pp: 53-55 (2009).
Online: http://www.w3.org/standards/xml/core
Petrovic S. and Burke E.K., University Timetabling, Handbook of Scheduling: Algorithms, Models and
Performance Analysis, CRC Press, Boca Raton, Chapter 45 (2004).
Qu R. et al., A Survey of Search Methodologies and Automated Approaches for Examination Timetabling,
Journal of Scheduling 12 (1), pp: 55-89 (2009).
Pillay N. and Banzhaf W., An Informed Genetic Algorithm for the Examination Timetabling Problem,
Applied Soft Computing 10, Elsevier, pp: 457-467 (2010).
90
ysahin@fatih.edu.tr, 2emanetn@fatih.edu.tr
Abstract.In this study, three different pattern recognition methods, random forest (RF), artificial neural
networks (ANNs) and support vector machines (SVMs), are used to develop decision support system for the
diagnosis of Parkinsons disease (PD) by using voice records. Acoustic analysis of voice records are done by
Praat speech analysis and synthesis tool. The aim of our study is threefold. First, it investigates the
effectiveness ofnewly introduced attribute, standard deviation of fundamental frequency, for the diagnosis of
PD. Secondly, it measures the importance of each attribute in the classification. Finally, it evaluates the
performance of each classifier. All classifiers have been tested and their performance comparisons have been
done. The results show a high accuracy for all classifiers.
Keywords: Parkinson diagnosis, voice classifier, artificial neural networks, support vector machine, random
forest.
1 Introduction
Neurological diseases are very common all over the world. Parkinsons disease (PD) is one of the most
common neurological disease [1] which impairs motor skills, speech, and other functions such as mood,
behaviour, and talking [2], [3], [4]. Moreover, statistics show that each year, approximately 50,000 Americans
are diagnosed with Parkinsons disease. In fact, all studies show that age is the mainspring effect of PD, which
increases steeply after age 50 [5], [6]. Although there is currently no cure found for PD, medication is used to
increase the quality of life of the patients if it is diagnosed at the early stages of the disease [1], [7].
Most of the people with Parkinsons disease have vocal impairment such as dysphonia [8, 9]. Vocal
impairment may also be one of the earliest indicators for the onset of the illness. Since the measurement of voice
is noninvasive and simple to measure, it is often used to detect Parkinsons disease. The purpose of this study is
to develop a decision support system to diagnose PD by using patients voice records. We used pattern
recognition methods for classification based on Random Forest (RF), Artificial Neural Networks (ANN) and
Support Vector Machines (SVM) [19], [22], [26] which are widely used for pattern recognition systems. These
three classifiers have been tested and their performance comparisons have been done.
There are extensive studies about general voice disorders [10], [11], [12],[13],[15], [16] by using traditional
speech measurement methods which include measurement of fundamental frequency F0, absolute sound pressure
level, time variation of periodic signals known as jitter, and shimmer. All these measurements have been used
for discriminating healthy people from person with Parkinsons disease.
More recently, the studies have been concentrated on machine learning tools for acoustic measurement to
detect voice disorders [2], [10]. Many interdependent vocal features can be extracted by using both traditional
and novel methods such as linear discriminating analysis (LDA), Wavelet Decomposition, Fast Fourier
91
Transformations. All these features have been used to diagnose PD at the early stages by Artificial Neural
Networks (ANN) and Support Vector machines (SVM) [2], [17], [18]. In this study, in addition to above
mentioned features, we also used jitter and shimmer perturbation measures, harmonics to noise ratio, recurrence
period density entropy and detrended fluctuation analysis features. One of our aims in this study was to find the
effectiveness of each of the features in the classification stage. We also tested the importance of newly
introduced feature, standard deviation of fundamental frequency, for the diagnosis of PD. Standard deviation of
fundamental frequency has been shown in other studies as efficient indicator for the diagnosis of voice disorders
[21], [24]. Finally, we made comparison between three classifiers: RF, ANN and SVM.
2 Materials and Methods
Voice Records
In our study, we used voice records acquired by Little et al. [2].They were obtained from 31 males and
females, 23 with Parkinsons disease. The ages of the subjects ranged from 46 to 85 years. The records sampled
at 44.1 kHz, with 16 bit resolution. Although dataset is available at UCI Machine Learning Repository, we have
used the actual raw voice recordings of this dataset in our lab. We have added a new feature that is the standard
deviation of fundamental frequency (stevF0) to increase the accuracy of whole classification system.
In feature extraction stage, we used Matlab and Praat to analyze voice records. Praat is a software program for
speech analysis and synthesis. The traditional measures are done by using Praat [20]. These measures are
fundamental frequency (F0 or pitch period), jitter and perturbation measures, shimmer and amplitude
perturbation measures and noise-to-harmonics ratio (and harmonics-to-ratio). Seventeen features are derived by
using Praat for all 195 signals (Table 1). On the other hand, nonlinear measurements are calculated by using
RPDE , FastDFA, performing recurrence period density and detrended fluctuation analysis tools [12], [22].
Feature
Description
Fo(Hz)
Fhi(Hz
Flo(Hz)
stevFo (Hz)
Jitter(%)
Jitter(Abs)
RAP
PPQ
DDP
Shimmer
Local shimmer
Shimmer(dB)
Shimmer:APQ3
Shimmer:APQ5
APQ
Shimmer:DDA
NHR
HNR
92
RF
Random Forest [26] is one of the powerful machine learning algorithms used for classification problems. A
random forest is in fact an ensemble of unpruned classification trees, which builds a number of decision trees
and gives a prediction based on majority voting. It is useful for practical problems, because it is not sensitive to
noise in the data set. Furthermore, decision trees avoid overfitting through pruning and it reduces variance.
It works fast, and generally exhibits a substantial performance improvement over many tree-based algorithms
without need to fine-tune parameters.
Classification occurs when each tree in the forest casts a unit vote for the most popular class. The Random
Forest then chooses the classification having the most votes over all the trees in the forest. Pruning is not needed
as each classification is produced by a final forest that consists of independently generated trees created through
a random subset of the data, avoiding over fitting. The generalization error rates depend on the strength of the
individual trees in the forest and the correlation between them. This error rate converges to a limit as the number
of trees in the forest becomes large.
Random forestalso gives us an opportunity to evaluate the effectiveness of each variable. It returns measures
of what variables are important in the classification.
ANN
Articial Neural Networks (ANNs) are very powerful tools which can be utilized for pattern
recognition.Artificial neural networks represent a type of computing environment similar to the way that the
human brain performs computations. They are good at fitting non-linear functions and recognizing patterns. The
most common neural network model is the multilayer perceptron (MLP).
In this study, we used MLP. This type of neural network is known as a supervised network because it requires
a desired output in order to learn pattern in input dataset. The goal of this type of network is to create a model
that correctly maps the input space to the output space by using input data. The model can then be used to obtain
the output when the desired output is unknown. A graphical representation of our MLP structure is shown below
(Figure 1).
93
important that the training set include examples that represent the full range of possible inputs so that the
network will correctly classify any input set. Note that the new input sets do not have to be identical to any of the
training patterns for the network to correctly classify the new input pattern.
SVM
Support Vector Machines (SVMs) are also a set of related supervised learning methods that can perform
pattern recognition. SVM proposed by Vapnik [22], [26] was originally designed for classification. SVMs
translate the input data into a higher dimensional feature space by using a kernel function. In this high
dimensional feature space a linear classifier is constructed.Typical kernel functions that are used by most of the
researchers are linear, polynomial, radial basis function (RBF) and sigmoid. Basically, the idea behind the SVM
lies on the construction of optimal hyperplane, which can maximize the margin between hyperplane and the
nearest point in the input data. Actually, SVMs are looking for solution in high dimensional space (Figure 2).
y k F ( xk )
|| w ||
k=1, 2, , n
(1)
where is a margin and y is the unknown pattern, and F(x) is dened as:
F ( x) wT x b
(2)
In order to use SVM on a particular dataset, we have to specify kernel function. Our choice of kernel type for
the SVM is radial basis function (RBF) which is recommended first for modeling a nonlinear transformation.
3 Experimental Results
The comparative results are grouped into two sections. The first deals with effectiveness of features and
the second is comparison between classification methods. In order to evaluate effectiveness of the features and
methods, 195 voice records were used.
94
Feature
RAP
F0
Jitter (Abs)
Jitter DDP
Shimmer
DFA
PPQ
HNR
RPDE
Stdev F0
APQ
Fhi
Raw Score
Z-score
6,432
4,221
4,942
5,363
2,878
1,806
2,898
2,266
1,654
1,528
1,088
1,028
26,936
26,695
22,611
22,184
19,439
19,36
18,154
17,966
16,647
14,918
12,12
11,41
Table 2 shows us the best 12 features obtained by RF with their raw scores. The evaluation results shows that
seven features, which are fundamental frequency and its low value, jitter percentage, shimmer (db), shimmer
APQ3 and APQ5, shimmer DDA and noise to harmonics ratio, actually do not affect classification accuracy. We
used the most important 12 features in our study.
95
Results for RF
Our last classification method was Random forest that also gave better accuracy for 12 features.
Table 3 depicts a comparison between best and average results achieved by each method. It appears that on the
average much better results in classification were obtained with RF than SVM and ANN.
Correct Overall
ANN
SVM
RF
Average
Best
Average
Best
Average
Best
All (19)
87,57
94,6
88,17
91,75
89,7
90,76
Best (12)
87,27
91,3
91,34
94,85
91,31
92,3
Feature Set
96
Acknowledgement
We would like to thank to Max Little of the University of Oxford in collaboration with the National Centre for
Voice and Speech, Denver, Colorado for providing the dataset to us.
References
1.
2.
3.
4.
5.
6.
7.
8.
9.
10.
11.
12.
13.
14.
15.
16.
17.
18.
de Rijk, M.C. et al. (2000). Prevalence of Parkinsons disease in Europe: a collaborative study of population-based
cohorts. Neurology. 54, 2123.
Little, M. A., McSharry, P. E., Hunter, E. J., Ramig, L. O., (2009).Suitability of dysphonia measurements for
telemonitoring of Parkinsons disease. IEEE Trans. Biomed. Eng. doi:10.1109/TBME.2008.2005954.
Ishihara, L., and Brayne, C., (2006) A systematic review of depression and mental illness preceding Parkinsons
disease. Acta Neurol. Scand. 113 (4)211220. doi:10.1111/j.1600-0404.2006.00579.x.
Jankovic, J., (2008) Parkinsons disease: clinical features and diagnosis. J. Neurol. Neurosurg. Psychiatry. 79:368
376. doi:10.1136/jnnp.2007.131045.
Huse, D. M., Schulman, K., Orsini, L., Castelli-Haley, J., Kennedy, S., and Lenhart, G., (2005). Burden of illness in
Parkinsons disease. Mov. Disord. 20:14491454, doi:10.1002/mds.20609.
Elbaz, A. et al. (2002). Risk tables for parkinsonism and Parkinsons disease. Journal of Clinical Epidemiology.
55, 25-31,
N. Singh, V. Pillay, and Y. E. Choonara, (2007) Advances in the treatment of Parkinson's disease, Progr
Neurobiol, vol. 81, pp. 29-44.
Hanson, D., Gerratt, B. and Ward, P. (1984) Cinegraphic observations of laryngeal function in Parkinsons disease.
Laryngoscope 94, 348-353
Ho, A., Iansek, R., Marigliani, C., Bradshaw, J., Gates, S. (1998) Speech impairment in a large sample of patients
with Parkinsons disease. Behavioral Neurology 11, 131-37.
Little M. A., McSharry P.E., Roberts S.J., Costello D.A.E., Moroz I.M. (2007), Exploiting Nonlinear Recurrence
and Fractal Scaling Properties for Voice Disorder Detection, BioMedical Engineering OnLine, 6:23 (26 June
2007).
J. Alonso, J. de Leon, I. Alonso, and M. Ferrer, (2001) Automatic detection of pathologies in the voice by HOS
based parameters, EURASIP J Appl Sig Proc, vol. 4, pp. 275-284.
M. Little, P. McSharry, I. Moroz, and S. Roberts, , (2006) Nonlinear, biophysically-informed speech pathology
detection, in Proc ICASSP 2006. New York: IEEE Publishers.
J. I. Godino-Llorente and P. Gomez-Vilda, , (2004). Automatic detection of voice impairments by means of shortterm cepstral parameters and neural network based detectors, IEEE Trans Biomed Eng, vol. 51, pp. 380-384.
S. Hadjitodorov, B. Boyanov, and B. Teston. (2000). Laryngeal pathology detection by means of class-specific
neural maps, IEEE Trans Inf Technol Biomed, vol. 4, pp. 68-73.
B. Boyanov and S. Hadjitodorov, (1997). Acoustic analysis of pathological voices, IEEE Eng Med Biol Mag, vol.
16, pp. 74-82.
J. H. L. Hansen, L. Gavidia-Ceballos, and J. F. Kaiser, (1998). A nonlinear operator-based speech feature analysis
method with application to vocal fold pathology assessment, IEEE Trans Biomed Eng, vol. 45, pp. 300-313.
C. O. Sakar, O. Kursun, (2009) Telediagnosis of Parkinsons Disease Using Measurements of Dysphonia, Journal
of Med Syst, vol. 34, pp. 591-599.
D. Gil, M. Johnson, (2009) Diagnosing Parkinson by using Artificial Neural Networks and Support Vector
Machines, Global Journal of Computer Science and Technology, vol. 9, no 4.
97
Biographies
Yasemin ahin received the B.S. in computer engineering from Fatih University, Istanbul, Turkey in 2009, the M.S. degree
in computer engineering from Fatih University, Istanbul, Turkey in 2011-continue, working at the department of Computer
Engineering of the Fatih University as a research assistant with MS degree, the major interests are computer vision, artificial
intelligence and speech analysis.
Nahit Emanet Nahit Emanet received his B.Sc degree in 1994 at Computer EngineeringDepartment, Boazii University,
Istanbul, Turkey. He completed his M.Sc.study in 1997 at Computer Engineering Department, Boazii University,Istanbul,
Turkey. In 2004, he received his Ph.D. from the same university.His main interests are microprocessor architectures,
embedded systems, VLSIdesign, pattern recognition, computer graphics and vision. He is currently anAssistant Professor at
Computer Engineering Department, Fatih University,Istanbul, Turkey.
98
Department of Computer Engineering, Sience and Research Branch, Islamic Azad university, Tehran, Iran.
2,3
Computer Engineering Department, Iran University of Science and Technology, Tehran 16846-13114, Iran.
4
Abstract. Spam is unsolicited message with advertisement, phishing, fraud or annoyance purposes. It is sent
to many users in network. Spam is in form of email or voice. By appearing VOIP and Internet Telephony,
spam problem is continued.Spam in Internet Telephony (SPIT) is unwanted call that is sent to VOIP users.
SPIT threat is more than spam because it is a real time and has voice nature. Therefore anti-SPIT mechanism
must be attended greatly.In this paper we propose a mechanism called MSG based on PMG mechanism and
IP-Domain correlation that improve and hit PMG in identity change situations. When malicious user alters his
identity, PMG cannot detect it but MSG mechanism can detect occurrence of SPIT. MSG computes gray
levelsbased on call rate and based on identity and decides based on these parameters. Simulated scenarios
present performance of MSG.
Introduction
Spam in Internet network is unsolicited electronic mail that is sent to users inbox. Spam reading and deleting
wastes time and mood of users. For these reasons, spam is unsightly for Internet users. Spam problem grows
rapidly in Internet network. According JISC technology report [1], only 7% of all received email in 2001 was
spam but in March 2004 this increased dramatically to 63% and is rising ever further and also according
Symantec report [2] in 2008, spam average in email was 80% of emails. Although anti-Spam techniques were
better but Spam number is increased.Heavy load, throughput decreasing in servers and annoyance level
increasing are disadvantages of large amount. Spam in Internet Telephony or SPIT is unwanted call for VoIP
users. SPIT is much bigger threat for users than email spam since it will interrupt the users immediately [3]. In
the near future, SPIT problem is expected to become strong threat in VoIP networks. Users may be exposed to
unsolicited religious, political, or pornographic messages and this can be a serious problem for younger users of
these services [4].
1.2
SPIT problem is a more important threat than spam, three reasons for this assertion is discussed below. SPIT
can be even more annoying than email spam, when your phone rings you should attend and answer it but you do
not attend immediately about emails and can ignore them. Current spam filters are improved and detect spam
with heuristic mechanism but in SPIT, voice is data carrier and detection with heuristic mechanism is difficult if
99
it is not impossible.Addition to these factors, Internet calls in comparison with emails need high bandwidth.
SPIT and spam are compared in Table 1.
Table 1. Spam and SPIT comparison [5]
SPIT
Real time
Frivol time and mood for answering
Immediately answering
Voice nature
High cost for initiation sessions and
transfer media
No ability for analyzing before rings
Sending data after session initiation
Using Internet backbone
Using signaling protocol
1.3
Spam
Unreal time
Frivol time and mood for reading and
deleting
No need immediately answering
Text or picture nature
Low cost with per message
Ability for analyzing before delivery
message
Sending data with message
Using Internet backbone
No Using signaling protocol
Anti-SPIT Mechanisms
In this section we express a number of anti-SPIT techniques. Anti-SPIT techniques are divided into three
groups. Some techniques are preventive, some of them detect attacksand another may be able to handle a
detected SPIT call. An anti-SPIT mechanism uses one or more of the following techniques and frameworks in
their architecture.
content filtering
These filters check content of messages. Content filtering isnt suitable for SPIT because real time filtering
phenomenon is very hard and needs long time. Period filtering needs context analysis and voice detecting.
reputation system
In this kind of systems, reputation of user is evaluated before calling. If the reputation score is greater than a
predefined threshold then sender is connected to receiver. Reputation systems are complex and sender can
connect to a few users.
payment
In this solution, every user pays a little money per message, this amount will be very little for legal user and
high value for SPITter. SPIT sender usually send a lot of call hence his cost will be high. Then malicious user is
forced to send small counts.
100
2.3MSG
Our approach proposes a way for improving PMG mechanism in situations that SPITters change their
identities. Our algorithm is composed of two modules, gray level based on call rate and gray level based on
identity. Block diagram of proposed mechanism is shown in Fig 1.The modules adapt together for detecting
SPITters. Computation in first module is the same as PMG mechanism. In the second module, we used IP
addresses with identifiers and also domain names for detecting bad calls. We assigneda weight to SPIT source
based on suspicious situations.Authors in [7] suppose three suspicious situations:
Different usernames are associated with the same domain and IP address: This event is suspicious, but it
is also typical of enterprise traffic where proxies are used. The score assigned to this event (A) should be
low.
Different IP addresses are associated to the same SIP identity: This event is suspicious, but it is also
typical for mobile users that change IP address while moving. The medium score (B) should be assigned
to this event.
Different domains are associated to the same IP address: This is the most unlikely event for the good
traffic, so the assigned base score (C) should be high.
Base score relation is computed based on the following formula. ni is number of event repetitions in past calls.
(10)
101
OR
Data base
Decision
Norm
SPIT
MSG mechanism
Normal
Poisson
Malicious
field
Type
Range
Time of arrival
Double
0-1000000
User identifier
Int
1-9600(good),9601-19200(bad)
User domain
Int
1-204(good),205-408(bad)
IP address
Int
1-9669(good),9670-19338(bad)
Flag
String
good(normal), bad(bad)
In this simulation, we used 200,600 and 3000 calls from malicious users and compare PMG and MSG.
3.1 Scenarios
In generated traffic, we assumed that the SPITter has some valid accounts for spoofing. In these scenarios, the
number of IDs which SPITter uses called spoofed account and detection rate and false detection rate curves for
each scenario areshown.
3.2.2
First Scenario
In scenario 1, spoofed count is variable and call rate is constant and time duration of simulation is 200
seconds. Call rate setsto 1 cps. We analyzed the behavior of our algorithm in three suspicious
102
situations(position1,2and 3). In position 1 we assumed that SPITter from one system sends SPIT to network. He
uses multiple usernames for identity hiding and also normal users can lie in this situation. In position 2, we
assumed that SPITter with one account from multiple systems sends SPIT. Normal users can lie in this situation.
In position 3, user sends SPIT with one identifier from different domains. Calls from these kinds of users have
high SPIT Probabilities. In position 1 and 2, MSG acts as PMG but our proposed algorithm is successful in
position 3 while PMG does not detect SPIT effectively. Detection rate and false alarm rate of MSG is shown in
Fig 2. This curve is composed from detection rate of PMG module and detection rate of identity module.
1,000
DR,FAR
,800
,600
,400
DR
,200
FAR
,000
2,0 32,0 62,0 92,0 122,0 152,0 182,0
Spoofed Count
Fig. 2. Detectionand false detection rate vs. the number of IDs which SPITter uses (spoofed count)in Rate=1 and position=3
3.2.3
Second Scenario
In scenario 2, time duration of simulation is 200 second and rate increases to10 cps. With call rate increasing,
SPIT detection in PMG mechanism is attenuated at higher spoofed count but SPIT detection in MSG rises after
falling. DR and FAR curves is shown in Fig 3.
1,000
DR,FAR
,800
,600
DR
,400
FAR
,200
,000
2,0
Fig. 3. Detection and false detection rate vs. the number of IDs which SPITter use (spoofed count)in Rate=10 and position=3
3.2.4
Third Scenario
In scenario 3 we increases call rate from user and study algorithm in larger range of time. In this scenario, time
of simulation and number of bad calls are 600 seconds. In this duration, traffic generation program also generates
normal calls. DR and FAR curves is shown in Fig 4.
1,200
DR,FAR
1,000
,800
,600
DR
,400
FAR
,200
,000
2,0 32,0 62,0 92,0 122,0 152,0 182,0
Spoofed Count
Fig. 4. Detection and false detection rate vs. the number of IDs which SPITter use (spoofed count)in Rate=1 and position=3
3.2.5
Fourth Scenario
In scenario 4, we increase call rate to 10 calls per second with latter range of simulation (600 seconds). As it
can be seen in the Fig. 5, our proposed algorithm works well. Detection rate is improved and its gap is smaller.
103
DR,FAR
1,200
1,000
,800
,600
,400
,200
,000
DR
FAR
2,0 32,0 62,0 92,0 122,0152,0182,0
spoofed count
Fig. 5. Detection and false detection rate vs. the number of IDs which SPITter use (spoofed count)in Rate=10 and position=3
DR,FAR
3.2.6
Fifth Scenario
In scenario 5 we increases calls from user and study algorithm in larger range of time. In this scenario, time of
simulation is 3000 second and call rate is 1 cps. DR and FAR curves is shown in Fig 6.In comparison fourth
scenario, Detection rate is improved in higher rate.
1,200
1,000
,800
,600
,400
,200
,000
DR
FAR
2,0 32,0 62,0 92,0 122,0152,0182,0
spoofed count
Fig. 6. Detection and false detection rate vs. the number of IDs which SPITter use (spoofed count)in Rate=1 and position=3
3.2.7
Sixth Scenario
In this scenario, time of simulation increased to 3000 seconds. In scenario 6 we increased call rate to 10 cps.
According fig.7, MSG can detect SPIT effectively.
3500,0
TP
2800,0
2100,0
MSG
1400,0
PMG
700,0
BS
,0
2,0 32,0 62,0 92,0 122,0152,0182,0
spoofed count
Fig. 7.Comparison between PMG, Base Score (identity module) and our Proposed Algorithm (MSG)
Conclusion
MSG in different range of simulation has same behavior and detects and blocks SPIT nicety but with rate
increasing, MSG works better. Good behavior of MSG at higher spoofed counts is owing to identity gray level
module and SPIT detection at lower spoofed counts is owing to PMG mechanism.SPIT Blocking is collection of
blocking in call rate module and blocking in identity module. False detection in scenarios is very low and can be
ignored.
104
References
1.
2.
3.
4.
Salehin, S. M. A. and Ventura,N. (2007). Blocking Unsolicited Voice Calls Using Decoys for the IMS,In IEEE
Communications Society Conference, 2007, pp.1961-1966
Bower, D., Harnett, D., Long, j., Edwards, C. (2008). The State of Spam, Symentec Monthly Technical Report,
March 2008, online document: http://eval.symantec.com/mktginfo/enterprise/other_resources/SpamReport_March08.pdf
Quittek, J.Niccolini,S. Tartarelli, S. and Schlegel, R. (2006).Prevention of Spam over IP Telephony (SPIT),NEC
Technical Journal, vol. 1, pp. 114-119.
Waiting,D. Ventura,N. (2007). The Threat of Unsolicited Sessions in the 3GPP IP Multimedia Subsystem.In IEEE
Communications Magazine, 2007.
5.
Safaei Kochaksaraei, F. and Akbari,A. (2010). SPIT consequences and necessity of using anti-SPIT mechanisms in
VoIP.In Infortmation Technology Conferece. IRAN, Accepted to be published.
6.
Dongwook Shin; Ahn, J.; Choon Shim; (2006), "Progressive multi gray-leveling: a voice spam protection algorithm," Network,
IEEE , vol.20, no.5, pp.18-24, Sept.-Oct. 2006
Menna, F. (2007). SPAM Over Internet Telephony Prevention Systems: Design and Enhancements with a Simulative Approach,
PhD Thesis, degli Studi University, Italy, 2007
7.
Biographies
Fateme Safaei Kochaksaraei Fateme Safaei Kochaksaraei received the B.Sc. degree from Iran University of Science and
Technology (IUST), Tehran, Iran, in 2004, and the M.Sc. degree from Science and Research Branch of Islamic Azad
University, Tehran, Iran in 2011, all in computer engineering.Currently, she is responsible of network security of
Mazandaran Telecommunication Company, Babol, Iran. Her research interests are network security and data communication.
Hassan Asgharian Hassan Asgharian received his M.Sc. in Computer Engineering from Amirkabir University of
Technology (AUT) in 2009 and He also has got his B.Sc. from Computer Engineering Department of Iran University of
Science and Technology (IUST) in 2006. Currently he is a PhD candidate in computer engineering in IUST and working as a
research assistant in the Network Research Group (NRG) of the Research Center of Information Technology (RCIT) of
Computer Engineering Department in Iran University of Science and Technology.
Dr. Ahmad Akbari Ahmad Akbari received his Ph.D. in Electrical Engineering from theUniversity of Rennes 1, Rennes,
France, in 1995. Dr. Akbari is currently an associate professor at Iran University of Science and Technology, Iran. Dr. Akbari
works on Speech Processing related research topics (especially speech enhancement) for more than 20 years. His current
research interests include Intrusion Detection and Response Systems, VoIP Communications, Next Generation Networks,
VoIP and SIP Security and also Data Communications Networks. Dr. Akbaris work has appeared in several peer-reviewed
journals and conference proceedings.
Razmara Zakerifar Razmara Zakerifar received the B.Sc. and M.Sc. degrees from Iran University of Science and
Technology (IUST), Tehran, Iran, in 2003, and 2008, respectively, allin electrical engineering.Currently, he is working in
Mazandaran regional electric company.His research interests include power system planning, control and optimization.
105
Abstract.A warp-weft density prediction system was designed by using image processing from a given
textile fabric image. In the experimental process, 10 different woven fabric image were scanned at a high
resolution then these images were transferred to the MATLAB program. By using the vertical and horizontal
frequencies of the textile image, the FFR analyze were carried out. Consequently with 94% accuracy, the
densities were predicted only by using the images instead of counting them by hand.
1 Introduction
Warp and weft densities in woven fabrics are very important parameters on behalf of making textile
production measurements. Determination of weft and warp densities in textile woven fabric depends on the
principle of analyzing the vertical and horizontal frequencies from a textile image. Counting these parameters is
generally made by hand with the help of a loupe. So the exact density can change person to person. This paper
presents a warp-weft density prediction system from a given textile image. In order to prevent the individual
errors we used image processing technique to determine the exact weft-warp densities. For the system design 10
different textile fabrics were scanned in 1200 dpi resolution and then transformed to the MATLAB programme.
The FFR analyze was performed from the vertical and horizontal frequencies of textile image. The results show
that weft-warp densities can be determined just by using its image with the 94% accuracy.
2 Literature Review
Image processing techniques are widely used in textile sector specificly to detect structural defects. For
instance a detection algorithm which employs simple statistical features (mean, variance, median) was developed
in order to detect textile fabric defects [1]. A system that detects a linear pattern in preprocessed images via
model-based clustering was composed [2]. The digital image correlation technique was used to assess macroscale textile deformations during biaxial tensile and shear tests in order to verify the loading homogeneity and
the correspondence between the textile and rig deformation [3]. Tow geometry parameters are measured with
the help of microscopy and image processing techniques in order to determine tow deformations in a textile
preform subjected to forming forces and the influence of these tow deformations on composite laminate
properties [4]. In another research an automated searching system based on color image processing system was
improved for the forensic fiber investigations, in order to determine the connection between the crime scene and
the textile material [5]. An automated visual inspection (AVI) system was design for real time foreign fibers
detection in lint by using image processing techniques [6].
106
3 Image Processing
Images can be treated as two-dimensional data, and many of signal processing approaches can be directly
applied to image data. An image is always represents in one, or more, two dimensional arrays I(m,n). Each
element of the variableI represents a single picture element, or pixel. The most convenient indexing protocol
follows the traditional matrix notation, with the horizontal pixel locations indexed left to right by second integer
nand the vertical locations indexed top to bottom by the first integerm (Fig. 1a). Another indexing protocol
accepts non-integer indexes. In this protocol, termed spatial coordinates, the pixel is considered to be a square
patch, the center of which has an integer value. In the default coordinate system, the center of the upper lefthand pixel still has a reference of (1,1), but the upper left-hand corner of this pixel has coordinates of (0.5,0.5)
(Fig. 1b) [7].
(1)
The variables
and
frequency variables, they define spatial frequency and their units are in radians per
and
The discrete form of Esq. (1) is similar to their time domain analogs. For an image size of M by N, the discrete
Fourier transform (DFT) becomes:
(2)
107
A simple square object (30 pixels by 30 pixels) is constructed. The resultant spatial frequency function is
plotted both as a three-dimensional function and as intensity images (Figure 2-4).
100
Magnetute
80
60
40
20
0
100
50
Vertical frequency
20
60
40
80
100
120
Horizontal frequency
Vertical frequency
Horizontal Frequency
4 Methods
This paper presents a new approach for processing images of woven fabrics to determine the warp-weft
densities. This approach includes three main steps, namely, image transformation, image enhancement, and
108
analyzing signals of the image. In the first step, 10 different woven fabrics were scanned in 1200 dpi resolution
and then transformed into gray-scale images in MATLAB programme In the second step, the contrast was
enhanced and then applied to the highpass filter. Thus the undesirable lower frequencies components were
eliminated. As filter the FIR filter from twentieth degree was used. Thirdly, the FFT analyze was carried out by
getting horizontal signals from 400 different rows of the images. The average value of these signals give us the
horizontal frequency components of the image. The same process was applied on the vertical direction of the
image to determine the vertical frequency components. The images of the cotton print fabric without any process
and after preprocessing can be seen in figure 5.
Fig. 5 (a) The raw fabric (b) The fabric after preprocessing
x 10
3.5
3
2.5
2
1.5
1
0.5
0
0
10
20
30
40
50
60
Frequency (Hz)
70
80
90
100
Fig. 6 The average horizontal and vertical frequency spectrum of the image cotton print fabric
109
Table 9. Real and Predicted Values of Warp-Weft Densities in Different Fabric Types
Actual Value
Warp
Fabric Types
Printed Cotton
Batiste
Denim
Canvas
Cashmere
Madras
Damask
Georgette
Cloth
Velvet
38
44
38
32
47
50
34
28
26
26
Predicted Value
Weft
30
29
30
21
35
36
26
26
22
23
Warp
39.5
45.11
37.3
30.8
0
47.61
31.8
30
25.5
0
Weft
30.9
26.74
27.9
21.1
0
32.59
24.57
23.7
20
0
Some real and predicted values of weft-warp densities belong to 10 different fabric type can be seen from
table 1. It can be concluded that the designed image processing system can predict the densities with 94%
accuracy. As can be seen from table 1 the cashmere and velvet fabrics predicted values is very different from
the real densities. This result indicates that the fabrics having feathery surfaces cannot be applied in this image
processing system, because the feathers prevent the accurate measurement of the frequencies.
References
1. Abouelela, A.; Abbas H.M.; Eldeeb, H.; Wahdan, A.A.; and Nassar, S.M. (2005). Automated vision system for
localizing structural defects in textile fabrics. Pattern Recognition Letters 26(10):1435-1443
2. Jampbell, J.G.; Fraley, C.; Murtagh, F.; and Raftery, A.E. (1997). Linear flow detection in woven fabrics using
model based-clustering. Pattern Recognition Letters 18: 1539-1548
3. Willems, A.; Lomov, S.V.; Verpoest, I.; and Vandepitte, D. (2009). Drape-ability characterization of textile
composite reinforcements using digital image correlation. Optics and Laser in Engineering 47(3-4): 343-351
4. Potluri, P.; Parlak, I.; Ramgulam, R.; and Sagar, T.V. (2006) Analysis of tow deformations in textile performs
subjected to forming forces. Composites Science and Technology 66(2): 297-305.
5. Paulsson, P. and Stocklassa, B. (1999). A real-time color image processing system for forensic fiber investigations.
Forensic Science International 103(1): 37-59
6. Yang, W.; Li, L.; Zhu, L.; Kang, Y.; and Li, F. (2009). A new approach for image processing in foreign fiber
detection. Computers and Electronics in Agriculture 68(1): 68-77
7. Semmlow, J.L. (2004). Biosignal and Biomedical Image Processing. New York: Marcel Decker Inc.
8. Gonzales, R. and Woods, R. (1992). Digital Imae Processing. Reading, Mass,USA: Addison-Wesley.
Biographies
Kazm YILDIZ Kazm YILDIZ was born in stanbul, in 1985. He was graduated from Marmara University, Technical
Education Faculty, Computer and Control Teaching in stanbul, Turkey, in 2007. He has been working at the Electronic
Computer department as a research assistant since March 2009. He received M.Sc. degrees from Marmara University
Institutefor Graduate Studies in Pure and Applied Sciences, in 2011. He is now in PhD. His field of interests are high
dimensional data mining, dimension reduction, neural network.
Volkan Yusuf ENYREK Volkan Yusuf enyrek was born in 1979 in stanbul, Turkey. He received the B.Sc. degree
from Marmara University, Technical EducationFaculty Electronics and Computer Education Department. He received M.Sc.
degrees from Marmara University Institutefor Graduate Studies in Pure and Applied Sciences, in 2007. He has been an
research assistant with the Electronics and Computer Education Department since2004. His research interests focus on fiber
sensors and signal processing.
Zehra YILDIZ Zehra YILDIZ was born in Ankara, in 1986. She was graduated from Marmara University, Technical
Education Faculty, Textile Department in Istanbul, Turkey, in 2008. She has been working at the same faculty and
department as a research assistant since March 2009. She is now in MSc thesis work. Her master thesis is about
electromagnetic shielding effectiveness of textile composites. Her major field of studies are conductive polymers, nano
textile particles, nano composites.
110
Abstract.Complexity and extent of computer software is constantly growing. Each software product has
specific audiences.Thus, when an enterprise is written or bought a software product, logically they should
ensure that software product is acceptable or not for users, audience, customers and other stakeholders.
Software testing process efforts for such evaluation and process an application test for detection of errors and
ensure to meet customer's demand and hardware compatible with software or No. Automated software testing
can increase developers time to focus on anther aspects and improve software errors effectively. In addition,
automated software testing can be run and repeat any times and the previous tests can be re-used again. Thus,
this increase software reliability and decrease the testing time. In this paper attempts to define some
parameters of the software test and propose a method which combines accelerated automated tests for the
study of software regression and memory overflow.
Keywords: Software automated testing, Memory overflow, regression test, Mean time to overflow
1 Introduction
Software testing process that identifies the quality of computer software. Software testing process, including
implementing a program to evaluate the software bugs, but not limited to it. With this thought, testing can never
completely correct computer software to prove. One important point is that software testing, should be different
opinions to be quality assurance software, with all areas of business process is ac-companied not only the test
areas.
A software failure occurs through the following processes. When the programmer to do an error that led to
failure in the software source code and compile and run the error, in certain situations the system will produce
incorrect results that can lead to failure. Not all errors will not lead to failure. For example, an error code will not
run ever. Will never lead to failure. But if the error code to change the software environ-ment or changes in
source data or interacting with different software. Will lead to failure.
111
The principle of "Pareto" in software testing is true simply. The principle of "Pareto" states that 80 percent of
all errors discovered during testing, probably "20 percent of all program components are detectable. Issue,
isolated components testing of suspects and complete them.
Testing should be started on a small scale and expand to larger dimensions. The first tests on each of the
components are doing. Achievement test, a series of errors and then integrated component of the whole system
can be found.
Complete testing is not possible. Number of possible paths for the average program is more. Therefore, the
implementation of any combination of routes is not possible. But it is possible that the level of adequacy of
programs cover.
In order to test the efficiency is highest, must be by an impartial third party to be done. To the highest
efficiency is more likely that the errors found. Because software engineer who has developed the system, not the
best person should do all the tests.
Reliability testing software, helps software developers using the operational reliability with the aim profile
(profile of the system), to implement the tests [3].
The most important innovation in providing this technique is increasing rapidly. in order to reduce development
time and costs, various failures in the systems, In this article, regression test, and memory overflow method
is presented with implementation and evaluation aimed to simultaneously reduce costs and increase reliability
of testing software[5].
2 Literature Review
Software testing is an activity will diagnose the potential non-compliance and failure in different phases of the
construction of the software product, especially in this implementation is correct and reliable software (v & v)7,
and plays a major role.
To access the software validity, some of the codes and standards of software as a key component are used to
justify the behavior and performance of distributed systems. For example, International Standard ISO 9126, is a
model to increase the quality of software products, so that in order to optimize the product defines customer
satisfaction as well as IEEE1012 standard for software verification and validation, are created the
common framework for all activities And tasks during the software life cycle processes [7], [8].
Activity verification and validation software allows reducing the production costs and simultaneously to
increase the reliability of the software. Testing software design and other engineering products, as the
initial product design is Difficult. Any engineering product can be tested in two ways:
1) Knowing the specific function that the software is designed to do, we can design tests
that show every performance is completely true or not.
2) Knowing how to work to with internal product and process programming, we can design tests that show that
the process of implementing software to perform or not.
The first method to test the black box and the latter is called white box testing. In recent
years, an intermediate approach has also been
noted that when
the greater
access to
some internal components of the software is
used there. This method is
called gray
box testing. In this
paper, based black box testing method is presented. We describe details of these testing methods in this section:
1. Black box testing: This type of software testing as a black box approach is that no understanding of the
internal behavior does not exist. Its purpose software performance testing in accordance with requirements to see
to what extent the above mentioned requirements are met. Test designers select valid and invalid test inputs and
choose right outputs and no knowledge of the internal structure of the object is not available. This method of test
112
design in all phases of software testing can be used. Black Box Testing cannot use instead of White box testing
.It rather a complementary approach that a different class of errors than the white box methods covers [1],[2].
2. White box testing: The user may test the internal data structures and algorithms, and code is available. For
example, designers can test the design of tests that causes the program to all statements are executed at least
once.[1]
White box testing techniques allows software team to evaluate system parts that use less and ensure that very
important performance parts are tested. Another advantage is that white box testing to optimize the code, and
also helps to erase the extra lines of code. This may lead to additional lines of hidden errors. White box testing
have disadvantages, including items: this method need the knowledge and skills because of the internal structure
of code requires a skilled tester is needed to do this type of test that will increase costs and evaluate every piece
of code search and hidden errors Getting around is almost impossible and may lead to problems that are causing
application failure.
3. Gray box testing: In recent years the gray box testing was added to routine use. Which means having access to
internal data structures and algorithms for designing the test cases that we can and we are on the side or at the
level of user to use black box testing.
3 Methods
After changing Software whether improving operation or debugging an error, a regression test can be
performed all of tests that software passed successfully before again. The regression test will be sure that last
improving software not to create any new fault. At the same time, during the proposed method test, a regression
test can be performed.
Several well-known methodologies for testing general purpose software exist in literature [12].
Such techniques follow a structural or a functional approach, also identified as the white-box and the blackbox approach respectively. The proposed methodology can be classified as a black-box approach, where the
software under test has to be verified with a suitable studied set of inputs whose expected outputs are known
only on the basis of the functional specifications. In addition to the black-box approach, we propose a test
method with test sets well representative of the field behavior of the system, according to the block diagram
shown in Fig. 1. One can observe that test parameters and field data statistical distribution are inputs;
Test results, expected results, MTOF8 and final results are outputs. The software under test is the
object of the analysis and test cases generation, test cases execution, input data elaboration"," memory
occupied monitoring and comparison are activities.
The software testing approach is based on a dynamic pseudorandom generation which mixed the preliminary
test parameters, which ones came from test results, and the statistical distribution of field data [10].
8Mean
Time to Overflow
113
FIELD DATA
STATISTICAL
DISTRIBUTION
TEST
PARAMETERS
COMPARISON
TEST CASES
GENERATION
INPUT DATA
ELABORATION
FINAL
RESULTS
TEST CASES
EXECUTION
EXPECTED
RESULTS
TEST
RESULTS
MEMORY
OCCUPIED
MONITORING
SOFTWARE
UNDER TEST
SEQUENCES ITERATION
MTOF
Test case generator creates quite unlimited sequences that can be iterated on the software under test. Both the
inputs and the outputs can be txt file with dimension of about 100 KB. The test results, represented by the output
of the software under test, are compared with the expected results obtained from the test cases, at the end of the
testing phase; final results allow information to be deduced about the quality in use of the software under test and
new data for dynamic test case generation. Detailed steps concerning the comparison between obtained and
expected results are shown in Fig. 2. If no differences between obtained and expected values are present, the test
is stopped; vice versa, a fault has to be tracked and the software has to come back to the development team in
order to be analyzed and correct the bug; successively the test has to be carried out again. There is also the
possibility of locating unclear values, such as ambiguities that have to be solved manually by repeating
ambiguous sequences.
114
Input Data
Elaboration
Automated Tests
Suite
Expected Results
Test Results
Results Comparison
End Test
Faults
Tracking
Manually report of
ambiguous
sequences
CORRECT
SW Development
NOT
CORRECR
The key-step introduced in the proposed method (Fig.1), that allows important information to be obtained, is
the simultaneous evaluation of the occupied memory of the system during the execution of the automated tests.In
Fig. 1 this step is denoted as memory occupied monitor.
Through the memory occupied monitor step lists of processes, the potential memory overflow can be
diagnosed through the following parameters: [6]
Private bytes: this parameter displays the number of bytes reserved exclusively for a specific process.
Working set: represents the current size of the memory area used by the process for tails, threads, and date.
The dimension of the working set grows and decreases as the VMM9 can permit. Since it shows the memory that
the process has allocated and reallocated during its life, when the working set is too large and doesn't decrease
correctly it usually indicate a memory leak [4].
It is useful, therefore, to introduce the parameter MTOF defined as:
(11)
Where Memsize denotes the available memory size and m represents the mean increase of memory occupation
in an established time interval. According to the value of m, evaluated in KB per day, hour or sequence, we can
explain the MTOF in number of days, hours or sequence to the breakdown of the memory resources and,
9Virtual
Memory Manager
115
therefore, to the collapse of the system (system fail stop). For a complex system where independent memories
are involved for the operating functions, Eq. (1)can be modified and the System Mean Time to Overflow is
defined as:
(2)
MTOF being an index able to evaluate the software crash, it can be considered a parameter to estimate the
software reliability and, as consequence, to plan the correct system update and the quality in use according to
ISO/IEC standard 9126[11].
As an additional advantage, the proposed approach allows the change of the stress level so testing the software
to a high stress level in the smallest test time. This can lead to faults not detected with low stress levels
increasing the test plan coverage. For example, in 15 working days we can realize 12.600 test cycles (35 test
sequences/h) if we implement accelerated automated tests, instead of 5.040 cycles (14 tests sequences/h as
suggested from field data) with no accelerated tests and 600 cycles (~5 test sequences/h) with manual tests. Such
considerations are summarized in Table 1, where automated tests are compared to the traditional one (manual
testing) [9].
40
Automated
tests
336
Accelerated
automated tests
840
168
280
2.352
5.880
360
600
5.040
12.600
Days
Man-hours
Machine- hours
Manual test
24
56
15
120
4 Conclusion
The proposed approach of dynamic software automated testing showed the importance of accelerated
automated tests for the software debug and validation in a short time interval, before product distribution, with
the aim of increasing quality in use and reliability.
The application presented in this paper is able to stimulate the software under test with an established stress
level, comparable in the sequence operations to the real use but accelerated four times compared to the manual
tests. Information concerning the memory leaks and fault regression of the new software versions with respect to
the old one can be deduced, as the experimental results proved.
The proposed method represents a fundamental parameter that allows the system crash to be estimated due to
an overflow. At the same time, it will allow the estimation of the software availability in order to plan an
effective maintenance operation plan, voted to induce not only an increase of software quality in use and
customer satisfaction but also a decrease of maintenance costs.
Furthermore, the benefits of such approach based on accelerated automatic testing compared with the
traditional one (manual testing) can be reached both at a lower cost and a decrease of the testing time of the
software verification and validation. In addition, the possibility of replicating old test sequences on new future
versions (with no testing cost) can also be considered an important benefit from the industrial point of view.
116
References
1.
2.
3.
4.
5.
[Ham04] Paul Hamill, Unit Test Frameworks, O'Reilly, 2004.Baker, P.R. (1978). Biogas for Cooking Stoves.
London: Chapman and Hall.
[Chu05] HueyDer Chu, John E Dobson and IChiang Liu, FAST: A Framework for
Automating Statistics
based Testing, 2005.
ReliabilityAnalysisCenter, Introduction to Software Reliability: A State of the Art
Review, Reliability Analysis Center (RAC), 1996.
J.D.Musa.(1997).Introduction to software reliability engineering and testing. In Proceedings ofThe 8th
International Symposium on Software Reliability Engineering, 1997.
A. Birolini, Reliability Engineering-Theory and Practice, Springer-Verlag3-54040287-X 2004.
6.
7.
8.
ANSI / IEEE Std. 1012, IEEE Standard for Software Verification and Validation
Plans, 1986.
9.
10. M. Rinard, C. Cadar, D. Dumitran, D.M. Roy, T. Leu.(2004).A dynamic technique forEliminating buffer overflow
vulnerabilities (and other memory errors), In ProceedingsOf the 20th Annual Computer Security Applications
Conference, Tucson,
Arizona, USA, December 6-10 2004.
11. H. Gne Kayack, A. Nur Zincir-Heywood, M. Heywood.(2005).Evolving successful stack Overflow attacks for
vulnerability testing. In Proceedings of the 21st Annual Computer Security Applications Conference
12. H. Freeman. (2002) Software testing, IEEE Instrumentation & Measurement Magazine 5 (3): 4850.
Biographies
Maryam Haghshenas is born in Bijar on 1985 and has earnedB.Sc.degree in Computer Hardware engineering from Islamic
AzadUniversity,Tehran south branch in Tehran, Iran on 2007 and is M.Sc. degree student in IT management at Islamic Azad
University, Science &Research branch in Tehran, Iran.
She has published Information technology security & service management system in Information Society of Iran journal
on July 2010.and Comparison of protocols in Underwater Wireless Sensor Networks 4 th, Iran Society of commands control
communication computers intelligence, at Sharif University, Nov 2010.She is interested in computer Science & E-learning&
Indexing on components and Software Testing and IT System management.
Ms. Haghshenas is a member of Young Researchers Club at Islamic Azad University, Tehran south branch and also a
member of the Information Technology Management Association at Islamic Azad University, Science &Research Branch.
Khadijeh Yousefpour Jeddiis born in Tabriz on 1985 and has earned B.Sc. degree in Computer Software engineering from
Islamic Azad University, Shabestar branch on 2007 and is M.Sc. degree student in IT management at Islamic Azad
University, Science & Research branchin Tehran, Iran.
She is interested in Software Testing and Software Engineering
Ms. Yousefpour Jeddi is a member of the Information Technology Management Association at Islamic Azad University,
Science &Research Branch.
117
Abstract:Classification of satellite imagery is very useful to obtain results about land-use spatial data
collection. While unsupervised classification is easy to apply on satellite images, supervised classification
methods give better results for accuracy. In this study, a traditional supervised classification method called
Maximum Likelihood Classifier(MLC) and a newer one called Support Vector Machines(SVM) approach are
compared It is clear that SVM is superior with better accuracy compared to MLC which is a common method
based on probabilities. Results showthat SVM is a good candidate for high resolution satellite image
classification works although its application on remote sensing is a pretty new topic.
1 Introduction
In recent years, by the help of high resolution satellite images, more details about the spatial features of land
covers could be reached[1]. In literature, many techniques for the classification of high resolution remotely
sensed images are reported[2][3][4][5]. However, classification results of existing methods in the literature
indicate that there is no best classifier yet. While unsupervised classification means grouping the pixels with
similar spectral features into clusters according to some criteria without a training data, supervised classification
means classifying pixels into classes by help of training and test data. It is already known that supervised
classification gives much better results on satellite imagery; thus, in this study, supervised methods are preferred
for classification.The accuracy of remotely-sensed image classification is strongly connected to distinguisability
of spectral features and capability of the classifier to differentiate the features. While some features are easy to
separate, some are hard to discriminate because of similarity. Therefore, the success of a method which is
accuracy in this work will depend on each individual class separability as well as type of the applied
classification [6].
118
this study, Maximum Likelihood and Support Vector Machines algorithms are used to classify the high
resolution satellite images.
1.1.1 Maximum Likelihood: Maximum Likelihood classifier is a common method used in the applications of
remote sensing as a parametric statistical method where supervision occurs by selecting sample areas. These
areas are then stored numerically for the algorithm to divide the image into the respective most probable spectral
classes. It is assumed that the distribution training data is Gaussian (normally distributed). The probability
density functions of each class are used to assign an undefined pixel into the class of highest probability [8].
1.1.2 Support Vector Machines (SVMs): Support Vector Machines classification method is another new
common supervised algorithm to classify satellite images. SVM is very good at generalization and so overcome
the problem of overfitting.SVM could provide better results than other widely used pattern recognition methods,
such as the maximum likelihood and neural network classifiers for accuracy considerations. Thus, SVM is very
attractive for the classification of remotely sensed data. The aim of the SVM algorithm is to find the optimal
separating hyperplane between classes by the help of the training samples. These training samples are called
support vectors and others are discarded. In this way, an optimal hyperplane can be fitted even with fewer
training samples with high classification accuracy [9].
2 Literature Review:
Maximum Likelihood classification is a very popular method and it is often applied on satellite imagery.
Maximum Likelihood method is based on statistical calculations and probabilities, and fuzzy logic is sometimes
applied together with the method. In a paper, MLC is used for discrimination among four spectrally similar
classes. Then, each class is hierarchically subdivided into subclasses with a fuzzy classifier[10].
Application of SVM on satellite image classification is a new trend, and its results are promising. SVM is
reported to be a good candidate for hyperspectral remote sensing classification because SVM method does not
suffer from high dimensionality problem[11]. Comparison of classifiers is very common in literature. Methods
are basically compared for accuracy and performance. Artificial Neural Network (ANN) method is yet another
powerful method for classification purposes and has comparable and sometimes better results than SVM method
for accuracy consideration [12]. ANN trained by back-propagation algorithm is reported to have good results as
well [13]. Decision Tree algorithm which is not used in this study, is a simple and easy to implement method for
classification. In a study, MLC is reported as the most stable method as compared to other methods like SVM
and Decision Tree approach; MLC is least affected by training phase and training set size. SVM has good
results, but supervision part of the classification dramatically affects overall accuracy [14]. If the size of dataset
is huge, it is not a big problem for SVM. Training and testing time might be long, but it is tolerable. However,
as dataset size is too small, unexpected results might be observed.
3 Methods:
The data is a subset of an image which is acquired on 2009-08-30 by Quickbird satellite.
Mainly satellite images are classified by two supervised methods respectively Maximum Likelihood and
Support Vector Machines.
Classification accuracy analysis is achieved with 100 sample points at least for each class by a stratified
random sampling method [15]. Same training and test areas are used for comparison purposes.
Separability measure is important in decision process for formation of classes. If the classes are too similar
spectrally, then the classes could be combined.
Confusion matrix provides valuable information about accuracy of each classification method. The success of
classifier on each individual class could be observed by using confusion matrix.
119
Kappa statistics is another valuable information to decide about success of each method. The Kappa
coefficient is basically defined as the proportionate reduction in error generated by a classification process
compared with the error of a completely random classification [16]. It is reported that kappa values greater than
0.75 are considered to have a high degree of agreement beyond chance. Values below 0.40 have a low degree of
agreement and values between 0.40 and 0.75 represent a fair to good level of agreement beyond chance alone.
4 Results:
The image is divided into five classes respectively shadow, vegetation, road, building and bare land. Same
training and test areas are used for both of the supervised classification methods for comparing success of each
method on individual classes.
Color coding:
Five classes total.
red:building;
green: vegetation;
cyan:bare land;
white: shadow.
120
Maximum Likelihood Classification results are shown on the table above. As results indicate, some areas
which are mixed in each other are overwhelmed with the other one. MLC is good at finding vegetation and bare
land classes. MLC is better at finding vegetation than SVM method. However, results belonging to classes like
shadow, road and building are not very promising. Additionally, 71 Kappa statistics value (%) indicates that
there is a fair to good level of agreement beyond chance alone. Overall 77 % may be an acceptable result for a
supervised classification algorithm. However, some improvements are needed to provide better classification
accuracy.
Table 1: Confusion Matrix for Maximum Likelihood Classification
Bare
Producer
User
MLC (%)
Shadow
Vegetation
Road
Building
Land
Acc.
Acc.
Shadow
42.86
0.00
0.00
0.00
0.00
42.86
100.00
Vegetation
40.18
97.79
0.49
14.46
3.00
97.79
61.01
Road
6.25
0.00
81.28
9.64
6.00
81.28
81.68
Building
10.71
2.21
4.93
72.29
1.00
72.29
87.38
Bare Land
0.00
0.00
13.30
3.61
90.00
90.00
71.43
121
SVMs are already known for their good generalization capacity. The biggest problem of generalization which
is called over-fitting is not observed in SVM results as the table above suggests. SVM classifier is better at
finding shadow, road and building classes than MLC. As the table suggests accuracy is improved for shadow
class, from 42,86% to 76,79%; for road class, from 81,29% to 84,73%; for building class, from 72,29% to 75,5%
and for bare land, from 90% to 98%. 84.25 % overall classification accuracy is a good result. Kappa value(%)
of 80 means that there is a high degree of agreement beyond chance. As compared to maximum likelihood
method, SVM is more preferable. As classification results indicate, SVM performed better overall classification
accuracy than maximum likelihood.
Producer
User
SVM (%)
Shadow
Vegetation
Road
Building
Land
Acc.
Acc.
Shadow
76.79
0.00
0.00
0.00
0.00
76.79
100.00
Vegetation
8.93
95.59
0.49
9.64
0.00
95.59
78.79
Road
10.71
2.21
84.73
14.46
2.00
84.73
76.44
Building
3.57
2.21
6.40
75.50
0.00
75.50
90.38
Bare Land
0.00
0.00
8.37
0.40
98.00
98.00
84.48
122
5. Conclusion
High resolution images provide detailed knowledge of spatial features. High classification accuracy could be
achieved with supervised classification methods. Accuracy values of Maximum Likelihood algorithm were
compared with those obtained using the statistical Support Vector Machines Algorithm. SVM is a powerful
technique for satellite image classification as results clearly indicate. Although SVM method required too much
time for training, it is worth trying because classification accuracy is high and statistics results are good. The
situation proves that SVM is preferable over maximum likelihood method.
References:
[1] J-T Hwang, H-C Chiang. (2009), The study of high resolution satellite image classification based on Support Vector
Machine In Proceedings of the 17th International Geoinformatics Conference. Fairfax, Virginia, United States, 12-14
August. Institute of Electrical and Electronics Engineers (IEEE )
[2] D. Tuia , F. Pacifici , M. Kanevski and W. J. Emery
mathematical morphology and support vector machines, IEEE Trans. Geosci. Remote Sens., vol. 47, p.3866 , 2009.
[3] N. Thomas , C. Hendrix and R. G. Congalton A comparison of urban mapping methods using high-resolution digital
imagery, Photogramm. Eng. Remote Sens., vol. 69, p.963 , 2003.
[4] Mayunga, S. D. , Coleman, D. J. and Zhang, Y. (2007) A semi-automated approach for extraction buildings from
QuickBird imagery applied to informal settlement mapping. International Journal of Remote Sensing 28 , pp. 2343-2357
[5] Van Coillie, F. M. B. , Verbeke, L. P. C. and De Wulf, R. R. (2007) Feature selection by genetic algorithms in objectbased classification of IKONOS imagery for forest mapping in Flanders, Belgium. Remote Sensing of Environment 110 , pp.
476-487.
[6] J-F Mas. (2003), An Artificial Neural Networks Approach to Map Land Use/Cover using Landsat Imagery and Ancillary
Data, In Proceedings of the 23rd IGARSS Proceedings, Toulouse, France, 21-25 July. Institute of Electrical and Electronics
Engineers (IEEE )
[7] ERDAS Field Guide, Fifth Edition.
[8] M.R. Mustapha, H.S. Lim and M.Z. Mat Jafri Comparison of Neural Network and Maximum Likelihood Approaches in
Image Classification, Journal of Applied Sci., 10: 2847-2854.
[9] D. H. Nghi, L. C. Mai.(2008), An Object-Oriented Classification Techniques For High Resolution Satellite Imagery, In
Proceedings of the 4th GeoInformatics for Spatial-Infrastructure Development in Earth and Allied Sciences (GIS-IDEAS)
Conference. Hanoi, Vietnam, 4-6 December.
[10] A. K. Shackelford and C. H. Davis, A hierarchical fuzzy classification
approach for high-resolution multispectral data over urban areas, IEEE
Trans. Geosci. Remote Sens., vol. 4, no. 9, pp. 19201932, September. 2003.
[11] J.A. Gualtieri, R.F. Cromp. (1999) Support Vector Machines for Hyperspectral Remote Sensing Classification, , In
Proceedings of the 27th AIPR Workshop: Advances in Computer-Assisted Recognition. Washington, DC, USA,14 October.
SPIE.
[12] Y. Xiong, Z. Zhang, F. Chen. (2010), Comparison of Artificial Neural Network and Support Vector Machine Methods
for Urban Land Use/Cover Classifications from Remote Sensing Images, In Proceedings of the First International Conference
on Computer Application and System Modeling (ICCASM). Taiyuan, China, 22-24 October. Institute of Electrical and
Electronics Engineers (IEEE ).
123
Acknowledgement:
We would like to acknowledge Tubitak for providing the data belonging to the project named as A Novel
Approach of Multi-Sensor Image Fusion Based on IHS Transformation. (project code: 110Y198)
Biographies:
Ayse Ozturk: Ayse Ozturk began her undergraduate study in Fatih University Computer Engineering Department in 2001
with a full scholarship. She graduated as a computer engineer in 2005. Her research areas are Machine Learning Algorithms
and Data Mining Applications.Ms.Ozturk is a member of local GAT (Young Researchers Group).
Mufit Cetin: In 1996, Mufit Cetin graduated from Yildiz Technical University, Faculty of Civil Engineering, Geodesy and
Photogrammetry Engineering Department. In 2001, he received his MS degree from Gebze Institute of Technology and in
2007, he got his PhD degree from Istanbul Technical University. His research areas are remote sensing and image
processing.Dr. Cetin has national and international publications.
124
Abstract. The importance of the shape of granular materials is well recognized due to their mechanical
behavior. Durability, workability, shear resistance, tensile strength, stiffness, and fatigue response of concrete
and asphalt concrete is heavily depend on the shape of aggregate particles. In recent years, image analysis is
widely used to analyze the particle shape characteristics of aggregate. In this research, angle of repose and
shape properties of granular materials were compared using image analysis by determining the shape characteristics such as roundness, sphericity, angularity, convexity and fractal dimension. Angle of repose values
were determined using tilting box method (TBM). The test results indicated that there is a good correlation
between some shape properties of granular materials and angle of repose.
Keywords: angle of repose, image analysis, granular materials, roundness, fractal dimension
1 Introduction
The angle of repose is defined as the slope angle for a material pile in which the material can rest without collapse. The angle of repose is one of the most important macroscopic parameters in characterizing the behavior of
granular materials. It has been found that the angle of repose strongly depends on material properties such as
sliding and rolling frictions [1-2] and density of particles [3], and particle characteristics such as size [4] and
shape [5]. It is generally reported that the angle of repose increases with increasing sliding and rolling friction
coefficients and deviation from spheres, and decreases with increasing particle size and container thickness.
However, quantitative description of the dependence that can be used generally in engineering practice is not
available [6].
In recent years, image analysis has been used in widespread applications in many disciplines, such as medicine,
biology, geography, meteorology, manufacturing, and material science. But, there have been relatively fewer
applications of image analysis used in civil engineering. Imaging technology has been used recently to quantify
aggregate shape characteristics and several researchers had investigated the role of aggregate shape in concrete
and asphalt mixture [7-12]. However, there is no comprehensive study that investigated the effect of shape properties to the angle of repose of granular materials. Consequently, the present study was undertaken to investigate
the relationship between the angle of repose and shape properties.
125
Roundness
Value
0.715-0.740
0.830-0.845
0.910
Roundness Class
Angular-Sub Angular
Well Rounded
Well Rounded
Roundness (R): This is a shape factor that has a maximum value of 1 for a circle and smaller values for
shapes having a lower ratio of area to perimeter, longer or thinner shapes, or objects having rough edges.
4. A
P2
[14]........(2)
126
Angularity (K): This is a shape factor that has a minimum value of 1 for circle and higher values for shapes
having angular edges.
[13, 18]............(3)
Convexity (C): Convexity is a measure of the surface roughness of a particle. Convexity is sensitive to changes
insurface roughness but not overall form.
[13, 18]............(4)
where: A = projected area of particle (mm2);
L = projected longest dimension of particle (mm);
P= projected perimeter of particle (mm);
Pellipse= perimeter of equivalent ellipse (mm),
I= projected intermediate dimension of particle
Arectangle= area of bounding rectangle (mm2).
2.3 Image Analysis System and Image Processing
The image analysis system used by the authors consists of a Nikon D80 Camera and Micro 60 mm objective
manufactured by Nikon. In this study, the equipment for taking pictures of particles is set up by mounting a camera on a photographic stand as shown in Figure 2, adjusting the height of the camera (30 cm) to obtain a sufficiently large measurement area on the sample tray, and adjusting the light sources so that there is no shading of
any object placed on the sample tray. Whenever taking a picture, a black cotton cloth was placed on the background to obtain better contrast between the particle and the background pixels. The tests were performed in a
dark room. Four fluorescent light sources were positioned on the base plane to make the borders of the particles
more visible for digital measurements. The flashlight of the camera was not utilized during image acquisition [19,
20].
127
128
Figure 5. Relationship between shape properties (roundness, sphericity, convexity) and angle of repose
Figure 6. Relationship between shape properties (angularity, fractal dimension) and angle of repose
In addition, some linear relationships are found between the shape properties and angle of repose (Table 2).
The correlation coefficient shape properties and angle of repose were found to be equal to 0.68- 0.78, which
indicated a relatively high correlation between the variables.
Table 2. Relationship between fractal dimensions and shape properties
y
x
S
R
C
A
DR
Relationship
y = -46,576x + 75,4
y = -65,382x + 95,561
y = -146,7x + 148,74
y = 132,76x - 105,34
y = 46,061x - 10,298
R2
0.77
0.70
0.68
0.74
0,78
129
REFERENCES
1 Lee, J., Herrmann, H.J., 1993. Angle of repose and angle of marginal stability: molecular dynamics of
granular particles, Journal of Physics A, 26, 373.
2 Hill, K.M. Kakalios, J., 1995. Reversible axial segregation of rotating granular media, Physical Review
A: Atomic, Molecular, and Optical Physics 52, 4393.
3 Burkalow, A., 1945. Angle of repose and angle of sliding friction: an experimental study, Bulletin of
The Geological Society of America 56, 669.
4 Carstensen, J. Chan, P., 1976. Relation between particle size and repose angles of powder, Powder
Technology 15, 129.
5 Carrigy, M., 1970. Experiments on the angles of repose of granular materials, Sedimentology 14, 147.
6 Zhou YC, Xu BH, Yu AB, Zulli P. An experimental and Numerical study of the angle of repose of
coarse spheres. Powder Technol 2002;125:4554.
7 Erdogan, S.T., 2005. Determination of aggregate shape properties using X-ray tomographic methods and
the effect of shape on concrete rheology, Ph.D. Dissertation, University of Texas at Austin.
8 Masad E, Olcott D, White T, Tashman L. 2001. Correlation of fine aggregate imaging shape indices with
asphalt mixture performance. Transportation Research Record, 1757:148156.
9 Al-Rousan T, Masad E, Tutumluer E, Pan T. 2007. Evaluation of image analysis techniques for quantifying aggregate shape characteristics. Const. Build Mater, 21:978990.
10 Masad E, Button J., 2000. Unified imaging approach for measuring aggregate angularity and texture.
Journal of Computer-Aided Civil and Infrastructure Engineering, 15(4):273-280.
11 Masad E, Button J, Papagiannakis T., 2000. Fine aggregate angularity: automated image analysis approach. Transportation Research Record 1721: 6672.
12 Masad E, Saadeh S, Rousan TA, Garboczi E, Little D. 2005. Computations of particle surface characteristics using optical and X-ray CT images. Computational Materials Science, 34:406-424.
13 Arasan S., (2011). Determination of some geotechnical properties of granular soils by image analysis.
PhD Thesis, Ataturk University, Erzurum, Turkey.
14 Cox, E.A. (1927) A method for assigning numerical and percentage values to the degree of roundness of
sand grains. J. Paleontol., 1: 179183.
15 Powers, M.C. (1953). A new roundness scale for sedimentary particles. J. Sed. Petrol., 23, 117119.
16 Mandelbrot, B.B., (1983). The fractal geometry of nature, W.H. Freeman, San Francisco, CA.
17 Hyslip, J.P., Vallejo, L.E., (1997). Fractal analysis of roughness and size distribution of granular materials. Engineering Geology, 48: 231-244.
18 Arasan, S., Akbulut, S., Hasiloglu, A.S., 2010. The Relationship between the Fractal Dimension and
Shape Properties of Particles. KSCE Journal of Engineering, (accepted).
19 Arasan S., Yener E., Hattatoglu F., Hnsloglu S. Akbulut S. The Correlation between Shape of Aggregate and Mechanical Properties of Asphalt Concrete: Digital Image Processing Approach, Road Materials and Pavement Design (accepted).
20 Arasan, S., Akbulut, S., Hasiloglu, A.S., 2010. Effect of particle size and shape on the grain-size distribution using image analysis. International Journal of Civil and Structural Engineering (accepted).
21 Russell, R.D. and Taylor, R.E. (1937) Roundness and shape of Mississippi River sands. J. Geol., 45,
225267.
22 Krumbein, W.C. (1941) Measurement and geological significance of shape and roundness of sedimentary particles. J. Sed. Petrol., 11: 6472.
23 Rittenhouse, G. (1943) A visual method of estimating twodimensional sphericity. J. Sed. Petrol., 13, 79
81.
24 Pettijohn, F.J. (1949) Sedimentary Rocks. Harper and Brothers, New York, 526 pp.
25 Alshibli, K.A., Alsaleh, M.I., (2004). Characterizing surface roughness and shape of sands using digital
microscopy. Journal of Computing in Civil Engineering, 18 (1): 36-45.
130
Abstract. Nowadays, Natural Language Processing (NLP) is one of most valued topics in the field of information technologies. Many researchers have been working on NLP methods including a wide range of
studies from spell checking to natural language generation. Part of Speech (POS) Tagging is one of the
main topics in NLP researches. The main purpose of POS tagging process is to solve disambiguates occurred during text analysis. In this study, an approach to POS tagging for Turkish was presented. In this
approach, there are two phases: First phase includes the rule definitions about the relationship between
word types and the second phase includes development of the software which is used to analyze the given
sentence by applying the given rules. This software also uses pre-developed tools, Sentence Boundary
Detection and Morphological Analyzer for the text analysis. The studies were carried out both with and
without n-grams. After the analysis carried out, it was seen that more reliable results were obtained when
n-grams applied onto the text to be analyzed.
131
Ali
Alinin arabas
gzel kz
gzel
ok gzel
ok, yava, ileri, az
ok yava, daha hzl
araba ile (arabayla), okula doru,
(Ali) dt.
(Ali) okula geldi.
(Aye) Aliye kazak rd.
E= A + B + E
E= A + + E
As you see in the table, a sentence is either a noun clause or a verb clause. The reason of this is, word of
determines type of sentence is at the end of the sentence. (It is a rule of regular sentence). This word can have
noun or verb root. After finding this item, we must analyze every word itself and according to word after it.
Every word can be a clause by itself. Because of this is when multiple words can be together they behave
like one word. So we can think whole word clause like one type. In this condition one word is one clause.
In word clauses there is an iterative structure. So if two words with the same type come together, they can
create first words clause type again.
Exp: A + A A + A A
When we investigate these line-up rules, the word at the end determine type of clause. So reverse order
view brings us most accurate result, when we analyze sentences or word groups.
What Is N-Gram?
One of the required data in many NLP Works like speech determination and machine translation is basic
statistical information about the language. Simply, these systems need what is the probability of coincidence information of a word or words in that language. This information is obtained from a mechanism called
Language Model. In this topic, because of its simplicity, a popular approach is called N-Gram Language Model.
Basically, this model brings you which often one, two or more word is found side by side in a text.
Exp :
"kedi"
"tekir kedi"
"sevimli tekir kedi"
N means number of word in N-Gram word. In the examples, "tekir kedi" is 2-grams or bigram, "sevimli tekir kedi" is 3-grams or "trigram".
132
That kind of statistics or information is generally obtained with counting number of side by side words in
large-scale texts. Naturally, probability of tekir kedi is greater than tekir kpek probability. Because tekir
adjective is usually used for kedi.
N-Gram method is also used for determining a text is in which language by computer. Computer can do
this with checking a words side by side letters pattern.
Exp:
Bilgisayar 2 Gram b bi il lg gi is sa ay ya ar r
3 Gram b bil ilg lgi gis isa say aya yar r
Every languages words N-Gram patterns like 2-Gram, 3-Gram are different. So with this approach a
texts language can discover.
Importance Of N-Gram About Clauses
If we use N-Gram method with clauses instead of whole text, we can discover more meaningful and
more logical results according to general outcomes.
After reaching significant word clauses according to side-by-side rules, to analyze these clauses about NGram method will be very important to examine sentence or text, summarize, or understand the topic.
Exp:
If we do N-Gram analyze to the whole sentence with whole words, we will get some meaningless results
like dmenin terbiye or dalgalar terbiye. But when we do same operation after finding word clauses, these
insignificant conclusions cant be in your hands. This type of analyze leads us to the more correct conclusions.
Application Areas Of Clauses
Sentences Meaning Analysis
If we find clauses of a sentence, meaning analyze will be easier. Because of words have different means,
finding the sentences mean is difficult. We will have many options so to realize which one is the most correct
mean in the sentence is going to spend our too much time. However if we know clauses of sentence, reaching the
main point of sentence will be simple. Therefore it leads us to comment about sentence semantically.
Word Accents
Every sentence can have a main point. This point is told or talked about it semantically. It is called accents and it can anywhere in sentence. Knowing how to consist word clauses in the sentence will help us to find
which word or word clause has accent. In addition to sentences general accent, words can have own accent, too.
In this situation having clauses helps us again. Because if there is a subordinate in a clause or if a word specifies,
clauses show us which word or what is in the foreground.
133
according to words meaning in the text, mission in the text or even kind of words. We can use these grouping to
get statistics or reach meaning of sentence. In this point, knowing clauses helps us to understand words meaning
and kind of the word. Instead of analyzing words alone, examining clauses can help us to put the words to the
right group according to our determined grouping method.
Access Right Meaning and Right Document (Information Retrieval)
Like we previously explained at the N-Gram analyze, to use clauses is very helpful to access words of
sentence. Instead of whole words analyze, investigation of clauses gains us time for reaching to the right place in
the sentence or text. Because of clauses help us for finding meaning of word in the sentence, it will be very important to do search right thing and not to take wrong answers.
Finding Clauses
Main common point of word clauses is entitled according to its words types and that word clause is determined by its words. In this case, first of all we must define words types of sentence to find clauses in a sentence and classify these
After identifying word kinds, we will decide clauses with side-by-side rules according to these kinds.
Rule Creation
Side-by-side rules are not enough for judging, when we separate clauses according their types. The
reason of this is, if words are side-by-side just because of their types, meaningless word clauses can consist. So
we must generate usual rules for more special and detail results. May be side-by-side words suffixes can create a
general rule. Many rules should be developed like this. Rate of true return results can be increased with more
rule-based system in new versions of project.
Effect Of Punctuation Marks
In a sentence, comma between words acts for separating them. But it is not useful always for maximum
solution.
Exp: Kk Okulu sevdi.
Kk, okulu sevdi.
In above sentences, comma causes to create two different clause structures. In the first sentence if there
is no comma, kk okul is a noun clause(A). But the second one both of kk and okul are noun
clauses (A). On the other hand, in Aye gzel, alml ve bakml bir kzdr sentence, comma dont change the
structure, only separating adjectives. Because of these situations, when we examine a sentence, we should also
evaluate punctuation marks.
References
1.
2.
3.
4.
5.
6.
7.
8.
9.
Kristina Toutanova and Christopher D. Manning. 2000. Enriching the Knowledge Sources Used in a Maximum
Entropy Part-of-Speech Tagger. In Proceedings of the Joint SIGDAT Conference on Empirical Methods in Natural
Language Processing and Very Large Corpora (EMNLP/VLC-2000), pp. 63-70.
R. Garside, G. Leech and G. Sampson (eds), 1987, The CLAWS Word-tagging System. The Computational
Analysis of English: A Corpus-based Approach. London: Longman.
Levent Altunyurt & Zihni Orhan,June 2006, Submitted to the Department of Computer Engineering in partial
fulfilment of the requirements for the degree of Bachelor of Science in Computer Engineering, Boazii University
http://cogcomp.cs.illinois.edu/page/software_view/3, Illinois University
Helmut Schmid at the Institute for Computational Linguistics of the University of Stuttgart,
http://www.ims.uni-stuttgart.de/projekte/corplex/TreeTagger/
Antworth, E. L. 1990. PC-KIMMO: A Two-level Processor for Morphological Analysis. Summer Institute of
Linguistics, Dallas, TX.
Brill, E. 1992. A simple rule-based part-of-speech tagger. In Proceedings of the Third Conference on Applied
Computational Linguistics, Trento, Italy.
Brill, E. 1995. Transformation-based error-driven learning and natural language processing: A case study in partof-speech tagging. Computational Linguistics, 21(4), 543-566.
Church, K. W. 1988. A stochastic parts program and noun phrase parser for unrestricted text. In Second Conference
134
10.
11.
12.
13.
14.
15.
16.
17.
Biographies
S. Emir TURNA He was born in Sakarya in 18th February, 1987. He graduated from university in 2009 and he is in
master degree in Dokuz Eyll University, Computer Engineering Department.
His main research area is natural language processing and word clauses. This is the first publication of Mr. TURNA. Before this, there are some researches about research algorithms in Turkish sentences and some big corpus.
Mr. TURNA is a member of NLP group of Dokuz Eyll University and association of electrical engineers.
ada Can BRANT He was born in zmir in 4th August, 1982. He graduated from university in 2005.He had received B.S. degree in Dep. of Comp. Eng. from Izmir Institute of Technology and then M.S. degree in Dep.of Comp.Eng
from Dokuz Eyll University(DEU).He also has been Ph.D. students in same department .
His main research area is natural language processing and software engineering. He works in projects, some of them are
accepted by TDK. Before this, there are some researches about rule based morphological analysis.
Mr. BRANT is a member of NLP group of Dokuz Eyll University.
Yaln EB He was born in Ordu in 1st January, 1964. He received B.Sc., M.Sc. and Ph.D. degrees in Mining Engineering from Dokuz Eylul University. He received Associate Professor and Professor degrees in 2000 and 2009, respectively. He is still working at Computer Enginering Department of Dokuz Eylul University.
His main research areas are Natural Language Processing, Wireless Sensor Networks and Production Scheduling in Open
Pit Mines.
Mr.EB is a member of Chamber of Mining Engineers, Turkish Linguistics Association (TDK) Turkish World Computer Aided Linguistics Working Group, Dokuz Eylul University Natural Language Processing group.
135
Abstract. Human face is one of the major biometric characteristic utilised in order to differentiate individuals from each other. Automatic recognition systems have been developed in order to identify and differentiate individuals for various purposes. Fingerprint is the most utilised biometric characteristic in these systems.
Face characteristic is the secondary biometric characteristic. These biometrics are generally used to meet the
security needs in many systems. Fingerprint is more common than face characteristic in terms of obtainability,
widespread use and reliability. However, it is not possible to track an individual who is involved in an event
at the street automatically with cameras via fingerprint method. On the other hand, it is possible to recognise
the face of an individual from a distance via cameras. Moreover, face recognition can also be used with the
other biometrics in order to meet the tracking or security need. Therefore, face analysis is still an important
subject. Automatic face recognition software was developed in this study. Face analyses included in system
were realised utilising eigenfaces method, error curves were provided according to this method and time
analyses of the algorithms used were made in order to find out the algorithms that affect the system much
more in terms of time.
1 Introduction
The security needs of individuals and institutions should be meet in parallel with the developments in technology.
Alternative methods are also needed beside basic precautions in order to increase security. Using biometric characteristics is one of these methods. Biometric characteristics display diversity among individuals. Characteristics
such as eye retina, fingerprint, signature, face, voice, DNA are accepted as biometric data. Biometric characteristics are regarded as safer and preferred since it is impossible to be lost, forgotten and stolen. Fingerprint is a biometric characteristic which has been used for a long time and more widespread.
The comparison of biometric systems is given in Table 1 [1]. Face image is in second order with 15.4% usage
ratio in terms of widespread use. When the face recognition is compared to the other biometrics, it is seen that its
usage is low in terms of distinctiveness and performance, but on the other hand, high in terms of universality,
collectability, acceptability and circumvention [2].
136
Face
Fingerprint
Hand Geometry
Iris
Keystroke
Signature
Voice
H
M
M
H
L
L
M
L
H
M
H
L
L
L
M
H
M
H
L
L
L
H
M
H
M
M
H
M
L
H
M
H
L
L
L
Circumvention
Acceptability
Performance
Collectability
Permanence
Biometric identifier
Distinctiveness
Universality
H
M
M
L
M
H
H
H
M
M
L
M
H
H
2 PCA Algorithm
Principal Component Analysis (PCA) is used in the reduction of high dimensional data into smaller dimensions
by taking distinctive characteristics into consideration [3].
This procedure consists of two stages: 1) training stage and 2) recognition stage [4]. In training stage, eigenmatrix which is required for recognition is formed from sample data for recognition.
1 dimensional one column matrix is formed from gray coloured 2 dimensional images. If we denote the images
included in training set as 1, 2, , M the mean related to these images is computed as follows:
1
M
n 1
(1)
i i
(2)
The dimension of this high dimensional matrix requires to be reduced by finding orthonormal uk vector in M
number which denotes this high dimensional matrix optimally. Vectors are selected to maximise k.
1
M
(u
n 1
T
k
n )
(3)
These k values and uk vectors are the eigenvalues and eigenvectors of covariance matrix.
1
M
(
n 1
Tn )
(4)
C AA T
June 1 -4, Kusadasi, Aydin, Turkey http://iscse2011.gediz.edu.tr
137
C matrix is in N2 x N2 dimension for the samples which are in N x N dimensions. Since it will be difficult to
calculate the eigenvalues and eigenvectors of this matrix which is in these dimensions, the eigenvalues and eigenvectors of the matrix given below are calculated.
L AT A
(5)
It is assumed that the number of the pictures will be smaller than the image dimension. Besides, eigenvalues of
matrix which is in M x M dimension are adequate for the problem of face recognition. A new matrix consists of
V v , v , v ,...., v
A , , ,....,
1
2
3
n C covariance matrix and
1
2
3
n which symboliseigenvectors of
es matrix consists of vectors displaying the differences of each face images from mean related to face and also
U V . AT
(6)
wk u k
(7)
The image which will be used in the recognition stage should be in grayscale. Initially, the difference of T vector from mean related to face data is calculated as in Equation 8.
(8)
T T
In this stage, wT feature vector of the face which will be recognised is calculated by multiplying T difference
vector with
(9)
W w , w , w ,...., w
1
2
3
N matrix disDistance of calculated wT feature vector to wi feature vectors forming
play the similarity of the image which will be processed to the face image recorded and which was being com-
pared to feature vector. The response which will be generated by the system as recognised face is the imagei
which generates the lowest value out of di values calculated in Equation 10.
M
d i wT ( j ) wi ( j ) , i 1,2,...N
(10)
j 1
Software
C# programming language in Visual Studio .NET environment was used for face recognition system developed within this study. The screenshot of face recognition software developed is given in Figure 1. There are 3
menus in the programme: Database, Analysis and Recognition. Database menu is used in programme to select
database which will be used in training stage. User can either use ORL Database[11] or PICS Database[12]. Furthermore, he/she can utilize user-defined databases including face images. Time analyses can be displayed with
FMR and FNMR graphics utilising Analysis menu. Recognition menu is used for new face recognition procedure.
A new image can be selected with New menu and recognition procedures can be initiated with Recognition menu.
At the end of the recognition procedure, the image of the person recognised is displayed in the screen. If the procedure fails, the warning of Could Not Be Recognised is displayed.
138
x min
max min
(11)
Images pre-processed are reduced into smaller sizes with Principal Component Analysis method. Best eigenfaces in M numbers which will represent the faces in the database are searched using PCA algorithm stages.
Eigenvalue and eigenvectors related to the matrix where the values of images are stored should be obtained in the
calculation of eigenfaces. In this study, QL Method was used in the calculations of eigenfaces and eigenvectors[13]. Eigenvectors are arranged from the highest to the lowest according to corresponding eigenvalues.
Weights related to these eigenfaces are used in the recognition of the new face.
Recognition of the new face which is the second stage in face recognition is related to the issue of deciding
whether an individual is the person who is already known. The face image which will be recognised is converted
into grayscale for recognition procedure. Min-max normalisation is carried out to reduce lighting effects. Weights
are obtained by utilising PCA algorithm stages. Distance between faces is calculated with Euclidian distance
method where represent weights (=[w1,w2,...wM]).
d k k
(12)
Threshold value gains importance in the comparison procedure of the weights. Distance is divided into the
number of feature vector since feature vector used affects distance, namely, difference. Therewith, distance obtained does not display diversity according to the number of feature vector. Therefore, common threshold value
can be determined for different databases. Distance occurred exists approximately between [0-2] interval. These
values became determinative in the selection of threshold value. False Match Rate (FMR) and False Non-Match
Rate (FNMR) ratios are used in the selection of threshold value. Equal Error Rate (EER) denotes error rate ob-
139
tained on the condition of FNMR(T)=FMR(T). In other words, EER is the level where the numbers of false rejection rate and false match rate are equal.
4 Conclusions
As a result, a face recognition system based on eigenfaces method was developed in this study. Eigenfaces
method is preferred as a rapid solution approach due to the fact that it can easily be applied. It was observed that
satisfactory results were obtained under different lightings utilising the face recognition system suggested. Besides, changes in the poses which do not affect the visibility of face are also tolerated by the system. Preprocesses carried out for images, procedures carried out for calculation of eigenfaces and face image which will
be recognised were analysed and evaluated in terms of time. Results of time analyses are considered as satisfactory where there are face images in small quantities, on the other hand, the calculations of eigenvectors and eigenfaces of the system where there are face images in large quantities, for instance, 10000 images, become slowly. It
is thought that the face recognition system suggested in this study can be developed in terms of speed considering
the necessity of recalculating eigenvectors and eigenfaces for each new person who will be added to the system.
Error curves of eigenfaces method according to ORL database including face images are displayed in Figure 2(a).
Moreover, error curves according to PICS database including face images are displayed in Figure 2(b).
(a)
Fig. 2. Error curves a) ORL database
(b)
b) PICS database (66/7)
140
Moreover, time analyses related to this method are displayed in Table 2. Processing times were obtained via
desktop PC containing Intel(R) Core(TM) 2 Quad Q8300 2.50 GHz microprocessor and 4 GB RAM with Windows 7 Ultimate operation system. It is seen that the calculation of covariance matrix requires rather long processing time in Table 2. It is aimed to increase the system performance with parallel programming in our subsequent study.
References
1.
Uluda, U.; Pankanti, S.; Prabhakar, S. and Jain, A. K. (2004). Biometric Cryptosystems: Issues and Challenges. In
Proceedings of the IEEE , Special Issue on Enabling Security Technology for Digital Rights Management, Vol. 92,
No. 6, pp. 948-960.
2. Sarma, G. and Singh, P. K. (2010). Internet Banking: Risk Analysis and Applicability of Biometric Technology for
Authentication. Int. J. Pure Appl. Sci. Technol. 1(2): 67-78.
3. Kim, K. (unpublished paper) Face Recognition Using Principle Component Analysis. Department of Computer Science, University of Maryland, College Park, USA
4. Turk, M. and Pentland, A. (1991). Eigenfaces for Recognition. J. Cognitive Neuroscience 3(1): 71-86.
5. Neerja, and Walia, E. (2008) Face Recognition Using Improved Fast PCA Algorithm. Congress on Image and Signal Processing 1: 554 558.
6. Wang, L.; Wang,, X.; Zhang, X. and Feng, J. (2005). The equivalence of two-dimensional PCA to line-based PCA.
Pattern Recognition Letter 26(1):.57 60.
7. Yang, J.; Zhang, D.; Frangi, A.F. and Yang, J. (2004).Two-Dimensional PCA: A New Approach to AppearanceBased Face Representation and Recognition. IEEE Trans. Pattern Analysis and Machine Intelligence 26(1):131137.
8. Yu, H. and Yang, J.(2001). A Direct LDA Algorithm for High Dimensional Data-with Application to Face Recognition. Pattern Recognition 34:2067-2070.
9. Yambor, W.; Draper, B. and Beveridge, R. 2002. Analyzing PCA-based Face Recognition Algorithms: Eigenvector
Selection and Distance Measures. Empirical Evaluation Methods in Computer Vision. Singapore: World Scientific
Press.
10. Sahoolizadeh, A.H.; Heidari, B.Z. and Dehghani, C.H. (2008). A new face recognition method using PCA, LDA,
and neural network. International Journal of Computer Science and Engineering 2(4): 218-223.
141
Biographies
Salih GRGNOLU- Salih GRGNOLU was born in orum in 1970. He graduated from Gazi University in 1995.
He received his M.Sc. and Ph.D. degrees from the same university in 2000 and 2007 respectively. He worked as instructor at
Gazi University between the years 1999-2007. His areas of interest include biometric systems, image processing, microprocessors and neural network applications. Asst. Prof. Dr. Grgnolu still works for Karabk University, Department of Computer Engineering.
Kadriye Z Kadriye Z was born in Karabk in 1982. She graduated from Ege University in 2005. She is interested in
image processing and parallel programming. Research Assistant z still works for Karabuk University, Department of Electronics and Computer Education.
afak BAYIR afak BAYIR was born in Berlin in 1973. He graduated from Uluda University in 1998. He received M.A.
degree from Department of English Language Teaching, Social Sciences Institute, Uluda University, Bursa, Turkey in 2000.
He received Ph.D. degree from Department of Computer and Instructional Technologies Education, Educational Sciences
Institute, Ankara University, Ankara, Turkey in 2009. His major field of study is educational technology.
142
Abstract. In this study on a picture library, learning was performed using COIL-20 based on an attribute SVM (Support Vector Machine) classification method was used on the data set which will be used in
the study. In this study the attributions of the images (toys and not toys) were taught to the system with
help of SVM. And example code of this method was conducted in Matlab package program. On SVM
decision mechanism, the classification was performed using the values obtained by subtracting the average from each data set as an attribute.
Keywords: Support Vector Machine, COIL-20, Computer learning, Data mining, Optimization, Image
classification
1. Introduction
SVM is an method of classification based on optimization and with this method, we can make a classification
including linear and non linear. Support vector is a classification group including machine learning methods
[1],[2]. According to the problem first encountered, linear classification is tried first if linear classification cant
be carried, non linear classification is preferred then in this study, two class separation was performed using a
linear classification. While doing this classification, choosing the most appropriate most plane was considered
too. The most classification points from each class on the weight vector elements were selected, choosing the
most appropriate plane [3]. Taking into account these points classifying toys and non toys from the data set in the
COIL -20 Image Library [4] which includes 20 grey scale image (Fig. 1 shows an example), and was preferred
mostly, was prevailed 100 % rate linearly.
143
2. Classification
Classification in data mining is commonly used to reveal hidden patterns on database [5]. Classification is to
use the differences of an object by using these differences to be able separate different object to different classification (Fig. 2 shows an example). Basic purpose, is to separate our example space plane into two separate optimum classes {-1,1} using separator function. While creating two classes the maximum length of the two points
from the distinguishing plane was ascertained [6],[7],[8].
: Class 1
: Class 2
.x b 0
(1)
In this expression; w: the weight vector, b: constant, x represent the attribute values. This expression also can be
as follows [11],[12].
(2)
wi .xi b 0
Points remaining at the top of bounder plane meet the following inequality.
w.x b 0 y 1 (class 1)
[8]
(3)
Points remaining at the bottom of bounder plane meet the following inequality.
144
w.x b 0 y 1 (class 2)
f(x) yi i xxi b
[8]
[5]
(4)
(5)
The sign of the outcome f(x), indicates class membership of x. xi vectors show support vectors of the plane.
Signal functions are expressed as follows;
f(x) sign(w.x b)
[8]
(6)
(7)
y matrix is shown in Table 1 but in fact, y [M, 1] (20 lines and 1 column) and at first as the size of the values
assigned to -1, then images of toy is divided into classes by giving a value of 1.
145
values
1
1
1
1
1
1
H i j yi y j .(xi .x j )
(8)
[11],[13]
(9)
[14]
(10)
(11)
f sign(x*w' b)
(12)
The values in the command window show result values and these values consist of 3 columns. According to yi
classification level and this level gathered depending on f(x) function are shown and in the part of learning result,
a performance measurement was done comparing the values found with the help of f(x) function and the classification values found first hand. For the sample data set there is 100% success. The practice, of this sample data
set is given in Table 2.
Table 2. This table is showing practices results : 100% performance. yi :level of classification. f(x) : is function of classification.
yi
f(x)
1
-1
1
1
-1
1
-1
-1
-1
-1
-1
-1
1
-1
-1
-1
-1
-1
1
-1
1
-1
1
1
-1
1
-1
-1
-1
-1
-1
-1
1
-1
-1
-1
-1
-1
1
-1
Learning
results
1
1
1
1
1
1
1
1
1
1
1
1
1
1
1
1
1
1
1
1
146
4. Findings - Conclusion
According to the results obtained in this study, the data set of COIL-20 Image Library classification was carried out with toys image data and non-toys image data. Support vector machine (SVM) is a classification method
based on optimization. SVM method in the study, a classification of images in the library of COIL-20 has been
successfully performed. Using class function was performed in different categories of data in this data set. On
this functions distinctive basic points are weight vector (w), constant value (b) and attribute value (x). The other
is Hessien matrix, with the help of Lagrange multiplier (). This multiplier is one important step in obtaining
values of the function. According to the results of the experiment, SVM classification method can be used to
obtain reliable results.
147
References
1.
2.
http://www.support-vector-machines.org/
O. G. Sezer, A. Ercil and M. Keskinoz (2005) Subspace Based Object Recognition Using Support Vector Machine,
EUSIPCO.
3. Ch.Srinivasa Rao , S.Srinivas Kumar, B.Chandra Mohan (2010) Content Based mage Retrieval Using Exact Legendre Moments And Support Vektr Machine, JIMA, Vol 2, No 2.
4. S. A. Nene, S. K. Nayar and H. Murase, (996) "Columbia Object Image Library (COIL-20),"Technical Report
CUCS-005-96.
5. Dr. Yaln zkan, (2008) Veri Madencilii Yntemleri Papatya Yay.
6. Chih-Wei Hsu, Chih-Chung Chang, and Chih-Jen Lin (2010), A Practical Guide to Support Vector Classification,
, Department of Computer Science, National Taiwan University, Taipei 106, Taiwan.
7. Florian Markowetz, Max-Planck (2003), Classication by Support Vector Machines -Institute for Molecular Genetics Computational Molecular Biology, Berlin, Practical DNA Microarray Analysis.
8. Kristin P. Bennett and Colin Campbell, /2000) "Support Vector Machines: Hype or Hallelujah?", SIGKDD Explorations, 2,2, 113.
9. http://www.cs.columbia.edu/CAVE/software/softlib/coil-20.php
10. V.N. Pawar, Sanjay N. Talbar,(2009), An Investigation of Significant Object Recognition Techniques, IJCSNS International Journal of Computer Science and Network Security, VOL.9 No.5.
11. http://en.wikipedia.org/wiki/
12. http://nlp.stanford.edu/IR-book/
13. Christopher J. C. Burges, (1998), A Tutorial on Support Vector Machines for Pattern Recognition by. Data Mining
and Knowledge Discovery 2:121167.
14. March (2009) Support Vector Machines Explained, Tristan Fletcher.
Biographies
First authors name is Baha en. He received his B.Sc. Computer Science Department
from the Gazi University, Ankara/Turkey in 1996. He received his M. Sc. Degree from
Institute of Science And Technology, Gazi University in 1999, and his Ph.D. degree from
same department.
His research interests include graphics, vision, genetic algorithms, expert system, computer parallel systems, geographical information systems, 3d modeling and simulation
systems.
Second authors name is Ahmet elik. He received his B.Sc. Computer Science Department from the Seluk University, Konya/Turkey in 2003. He received his M. Sc. Degree from Institute of Science And Technology, Dumlupnar University in 2010, and he is
a Ph.D. student in Institute of Science And Technology, Karabuk University now.
His research interests include graphics, vision, computer systems, network systems,
expert system, data mining systems, simulation systems, and computer parallel systems.
148
Abstract. Nowadays, the number of students that studying at undergraduate level increases gradually.
This situation makes internet-base education a better alternative for students. Internet-base education itself
can be used as a education method as well as a method for supporting formal education. This project, with a
variety of modules developed, offers a solution to all steps in the life of a students education such as enrollment to institution, learning-teaching and assessment. In addition to this, user identification module with
RFID is developed as an alternative to username-password system login application. Thus, system login becomes much more practical with the RFID cards that are provided to students.
Keywords: learning management system education e-learning radio frequency identification RFID internet security
1 Introduction
Today, students who studying at undergraduate level, need education that independent of time and space because of their life together with training and business. This situation brought about the need for students to use
technology. So, supporting formal education or as another option to formal education, the importance of webbased education has increased steadily.
In this study is intended to bring the solution with a software development requirement, the software provides
training on the internet. This software is a solution of all educational stages as Students' system registration, education process and assessment of student achievement with a variety of test modules. So that allows students to
receive education as synchronous or asynchronous and aims to develop a solution to the negativity which brought
formal training environment.
At the same time, educational institutions, students will be served the identification cards, which has RFID specialty. Students can access the software with that cards, in-house or outside the institution by a computer that has
RFID reader, without the need for security measures provided by the computer user name and password.
149
The elearning management system(e-LMS) which uses to achieve e- learning, provides management of learning activities. The system provide functions such as providing learning materials, editing, sharing the learning
material and discussions, making assignment, take exams, providing feedback on these assignments and examinations, students, teachers, and keeping system records, receiving reports.[1]
A learning content management system (LCMS) is a related technology to the learning management system in
that it is focused on the development, management and publishing of the content that will typically be delivered
via an LMS. An LCMS is a multi-user environment where developers may create, store, reuse, manage, and deliver digital learning content from a central object repository. The LMS cannot create and manipulate courses; it
cannot reuse the content of one course to build another. The LCMS, however, can create, manage and deliver not
only training modules but also manage and edit all the individual pieces that make up a catalog of training. LCMS
applications allow users to create, import, manage, search for and reuse small units or "chunks" of digital learning
content and assets, commonly referred to as learning objects. These assets may include media files developed in
other authoring tools, assessment items, simulations, text, graphics or any other object that makes up the content
within the course being created. An LCMS manages the process of creating, editing, storing and delivering elearning content, ILT materials and other training support deliverables such as job aids.[2]
150
Module
Membership
Module
Messaging
Module
Categories
Module
Document
Module
Video
Module
Homework
Module
Quiz
Module
Examination
Module
Assessment
Module
Bookplate
Module
Voting
Module
Comment
Module
Explanation
The instructor and students to become a member of the system and to limit the authority provides access to the system.
The instructors for students in public or private, to send messages and make announcements, send a message to the students to each other.
Lesson plans and the contents of nested grouped into categories, the content of this training the students will be served.
Course content categories on the basis of the installation of
the system, allows students to download and use and reporting of students use of the document.
Installation of the system study the contents in video format,
downloaded and used by students and provides students reported the use of the document.
The system allows uploading papers of the students by students and the trainers can download and examine the papers.
During the e-learning, the system provides quizzes which as a
test form.
Students are examined on the web site. The system allows
entry of exam questions according to their degree of difficulty
by trainer with an administration panel
Provides entries of Homework, Quiz and Exam module the
evaluation by trainers by a management panel.
The module provides training contents outside of the categories and one or more categories associated with the subject
and filtered.
The contents can be voted by students, this contents how
much useful for students can be detect and reporting by the
results.
The module provides to making comments and share information to students and educators about the training content.
151
Application
Explanation
Video
Conference
Synchronicity
Education
As a result of improvements in information and communication technologies, more accurate, more detailed,
faster and more secure flow of the data have become a necessity nowadays. Because of that reason, RFID system
is one of the most promising technologies in control and monitoring of business systems in different fields. The
design of the RFID system aims to compose, to collect and to manage dynamic information, without human interaction. Its usage increases exponentially every year.
RFID is a technology that allows to tracking an object, which is microprocessor equipped and carrying a label, by the information carried by this label and works with radio frequency. With RFID, objects can be identified
and tracked throughout their entire life cycle, the chain of operations from production to distribution. With this
new technological substructure, data collection, service delivery and system management is carried out without
human intervention and the quality and speed of service is increased by reducing error rate.[3]
RFID, Radio Frequency Identification (RFID) is a method of remotely storing and retrieving data using devices called RFID tags or transponders. An RFID tag is a small object, such as an adhesive sticker, that can be attached to or incorporated into a product. RFID tags are composed of an antenna connected to an electronic chip
and are thus basically contact-less smartcards. These chips transform the energy of radio-frequency queries from
an RFID reader or transceiver to respond by sending back information they enclose.4 Finally, a computer hosting
a specific RFID application pilots the reader and processes the data it sends.[5]
Basically, all smart-cards and RFID-readers are hardware products. To acquire digital data from these cardreaders, by accessing hardware, processing the acquired results, which are turned into appropriate format, might
be needed. In that case, it is needed to develop software products which are suitable for all of these card-readers.
Hardware access is provided by unmanaged codes. The unmanaged codes which is provided for access hardware is picked as a group in a library. By creating these libraries, provides access to hardware which has software
is developed at high level languages.
152
In the Windows Operating Systems, the access to the RFID readers is provided by unmanaged dll (dynamic
link library) which is named winscard.dll. If software will be developed for Windows Operating Systems, by
using this Dll can be access to binary data which is created RFID hardware. These results are converted into digital data formats and can be shown to users who are the application's target.
If the software being developed for other platforms, libpcsclite.so.1 library can be used for LINUX, UNIX systems. Libdl.so.2 need to use the library for the Solaris operating system.
It is not necessary to make a major change at the use of unmanaged library for the platforms. The developers
of the operating systems develop this libraries for this kind of universal hardwares. Therefore, by developing a
single application, the workload will be reduced, if the platforms are taken as parameters in the application. By
installing developers libraries for the platforms, system calls, that are suitable for the operating system, in which
the application runs, can be implemented.
3.1 Flow Chart of Identification of the Reader And The Process of Getting Card Information
Project Architecture
The project developed with Microsoft .Net Framework 4.0, Entity Framework, Flash Action Script 3.0 and Microsoft SQL Server 2008 technologies. If necessary, data transfer can be made to other platforms envisaged with
the "XML" and "XML Web Services technology.
153
References
1.
2.
3.
4.
5.
6.
7.
8.
9.
10.
11.
12.
13.
14.
15.
16.
17.
18.
19.
20.
21.
22.
23.
24.
25.
DURAN N., ONAL A., (2008) renme Ynetim Sistemleri iin SCORM Uyumlu Bavuru Modeli Gelitirilmesi,
Ege Universty, Izmir
Learning and Content Management System, from http://en.wikipedia.org
YUKSEL, M.E., ZAIM, A. H., (2008). Structures Used In RFID systems And A Sample RFID System Design, Istanbul Universty
G. Avoine. Cryptography in Radio Frequency Identification and Fair Exchange Protocols. PhD thesis, Ecole Polytechnique Federale de Lausanne (EPFL), Lausanne, Switzerland, December 2005. [Retrieved December 23, 2005,
from http://lasecwww.epfl.ch/gavoine/download/papers/Avoine-2005-thesis.pdf]
FUHRER p, GUINARD d, LIECHTI O, (2006)RFID:From Concepts to Concrete Implemetation, University of Fribourg
What is the RFID? (2010), from http://www.kirbas.com/
Association for Automatic Identification and Mobility (2001), from http://www.aimglobal.org
Makal B., Bayrak S., atalkaya ., (2007) Otomatik Tanmlama Sistemlerinin Teorik Olarak ncelenmesi ve
RFIDli Sistem Tasarm, Istanbul
RFID (2010) from http://tr.wikipedia.org/wiki/RFID
RFID Etiketler (2008) from http://rfid.mantis.com.tr/2008/05/etiketler/
SOLAK, ., (2008) UHF RFID Etiketleri in SRAM Tabanl Donanmsal Rastgele Say reteci Tasarm,, Karadeniz Teknik niversitesi, Trabzon
RFID Readers, (2008) http://www.ruzname.net/rfid-reader-devre-blok-diyagram
Balkovich, E., Bikson, T. K., Bitko, G., (2005) 9 to 5: Do You Know If Your Boss Knows Where You Are? : Case
Studies of Radio Frequency Identification Usage in the Workplace, Santa Monica, CA, USA: Rand Corporation
Karygiannis T., Eydt B., Barber G., Bunn L., Phillips T., Guidelines for Securing Radio Frequency Identification
(RFID) Systems. Recommendations of the National. Institute of Standards and Technology, NIST Special Publication 800-98, National Institute of Standars and Tecnology, Gaithersburg.
Roberts, C.M., (2006) , Radio frequency identification (RFID), computers&security2
Thornton, F., Haines, B., Das A. M., Bhargava, H., Campbell, A., Kleinschmid, J., RFID Security, Syngress Publishing, Rockland, 2006
BHUPTANI M., MORADPOUR S, (2005) RFID Field Guide:Deploying Radio Frequency Identification Systems
Finkenzeller K, (2008) RFID Handbook:Fundamentals and Applications in Contactles Smart Card and Identification 3. Edition
Brown D, (2006), RFID Implementation
Shepard, S. (2004), RFID:Radio Frequency Identification
White Paper, RFID Teknolojisi, 2009
http://www.rfidturkey.com/web/index.php
http://pinvoke.net/ , access date, 11.03.2011
http://msdn.microsoft.com/en-us/default
http://www.hidglobal.com
Biographies
Ali Buldu Ali Buldu was born in 1971 in Kayseri, Turkey. He received the B.Sc. degree from Marmara University, Technical Education Faculty Electronics and Computer Education Department. He received M.Sc. and Ph.D. degrees from Marmara University Institute for Graduate Studies in Pure and Applied Sciences, in 1996 and 2003, respectively. He has been an
assistant professor with the Electronics and Computer Education Department since June 2006.
His research interests focus on Computer Hardware, Computer Aided Education, Information Technologies, and Programming Languages.
Nagihan BEKDA She was born in stanbul, Turkey in 1989. She currently is continuing education at the department of
Computer and Control Technology.
She is working on software development and analysis, control systems, Computer Aided Education and Information
Technologies. She studies to research about web based learning in education. She has a paper about learning management
systems at International Educational Technology Conference 2011.
Serdar SERT He was born in Bursa, Turkey in 1988. He currently is continuing education at the department of Computer
and Control Technology.
He is working on software development and analysis, control systems, mobile applications and Information Technologies.
He studies to research about web based learning in education. He has a paper about learning management systems at International Educational Technology Conference 2011.
154
Abstract. This paper introduces the particle swarm optimization (PSO) based trajectory tracking PID
(TTPID) controller to control the Triga Mark-II research reactor established at Istanbul Technical University.
The parameters of PID controller are determined using PSO algorithm. Object of PSO algorithm is to find optimum parameters of PID which provide maximum performance criterion value. In applications, the criterion
value is optimized successfully by PSO. To demonstrate the effectiveness of this controller, series types of
simulations are carried out. The results of the simulations demonstrate that TTPIDC efficiently provides a
simple and easy way to control the reactor. As a result, PSO could optimize the parameters of PID for this
controller and the PSO-TTPIDC could control the system successfully trajectory under various working conditions within the acceptable error tolerance.
Keywords: PSO, PID control, Trajectory Tracking, Nuclear Reactor Control.
Introduction
Proportional Integral Derivative (PID) controller has been widely used in industry due to its advantages such
as simple in structure and easy to implement. Although conventional PID is popular, it is difficult to obtain satisfying controller parameters especially when the process has nonlinear dynamics [1, 2]. Actually, much of the
effort of researchers has been concentrated on the development of new tuning rules for PID controller. Although
this is obviously a crucial issue, it is well-known that a key role in the achievement of high performance in practical conditions is also played by those functionalities that have to (or can) be added to the basic PID control law.
There are different techniques that can be used in tuning PID controller parameters, which can be classified into
the following;
Analytical methods; where PID parameters are calculated from analytical or algebraic relations between a
plant model and an objective [3, 4],Frequency response methods; where frequency characteristics of the controlled process are used to tune the PID controller [5], Adaptive tuning methods; which are for automated online
tuning [5], Optimization methods; where these can be regarded as special types of optimal control, in which PID
parameters are obtained ad hoc using offline numerical optimization methods [6] and Heuristic methods; which
are evolved from practical experience in manual tuning and from artificial intelligence [7, 8, 9, 10].
There are several reasons to develop better methods to design PID controllers. One is the significant impact
it may give because of the widespread use of controllers. The other is the enhancing performance of controllers
which can be derived from improved design methods. It is possible to obtain controller parameters with analytical
methods such as root-locus, bode analysis and pole replacement when system has linear behavior or its possible
155
for linearization of the system [3]. Artificial neural network (ANN), Fuzzy Logic controller (FLC), Genetic algorithm (GA), and PSO techniques could be used to model or control the system If the process models have nonlinear dynamics [7, 11] . PSO and GA are general-purpose optimization techniques that have been developed and
recognized as effective for combinatorial optimization problems such as the traveling salesman problem, quadratic assignment problem, and controller tuning problems with successful results [8, 9, 12-20].
Particle Swarm Optimization (PSO) is a recently invented high performance optimizer that possesses several
highly desirable attributes, especially the ease with which the basic algorithm can be understood and implemented. PSO is a technique modeling swarm intelligence that was developed by Kennedy and Eberhart in 1995 [21],
who were inspired by the natural flocking and swarming behavior of birds and insects. This technique has been
used in addressing several optimizations and engineering problems [22].
Numerous researches have been proposed a controller such as PID, FLC, adaptive control and optimal control for the power control of ITU Triga Mark-II [23-30]. The work presented here is a continuation of earlier
research on control of this nuclear reactor. In this study, TTPIDC is designed to control the nuclear research reactor Triga Mark-II at Istanbul Technical University (ITU) and constrained Partial Swarm Optimization (PSO) is
used to optimize its parameters. . Additionally, constrained cost function is used to obtain non-chattering control
surface. To demonstrate the effectiveness of proposed approach, a series of simulations are carried out on ITU
Triga Mark-II reactor. The results demonstrated that reactor power perfectly tracks the predefined trajectory under various working conditions.
This paper is organized as follows. In Section 2 reactor model and defined trajectory are given. Designed
TTPIDC is explained in Section 3. PSO algorithm is briefly explained in Section 4. Finally, experimental results
and conclusion are given in Section 5.
ITU TRIGA MARK-II research reactor is use solid fuel which obtained by homogeneously mixed enriched
uranium fuel and zirconium hydrate moderator. The reactor has steady state and pulse working mode. It is produces 250 kW powers at the steady state working mode and 1200 MW power at the pulse working mode for a
instantaneous time. The reactor is an open tank type and cooled by light water. Because of reactor has negative
heat coefficient, unexpected power increase is caused to decrease the fuel moderator heat and the reactivity. As a
result, reactor power is automatically decreased. Only TRIGA type reactor has this special characteristic. Reactor
has shock rod, safety rods and control rods which safety and control rods are controlled by electro mechanic
mechanism. In this work, to simulate the neutronic-thermal-hydraulic behaviors of reactor, YavCan code is used
[23] where reactor model could be given as follows;
dX (t )
AX(t) B
dt
(1)
Where, X is (10x1) dimension vector and represents the neutron density, precursor concentration, fuel temperature, coolant temperature, and coolant velocity. A is a coefficient matrices and has (10x10) dimension. B is
(10x1) dimension input vector. Detailed reactor model can be found in Can et al [23, 30].
In order to define control signal, it is necessary to determine the trajectory function that has to be tracked by
the reactor power. The trajectory has three regions which given as follows;
(2)
Where, PT is power trajectory, t is time, 3 , 2 , 1 , 0 are trajectory coefficients and can be calculated
from the initial, required power and trajectory slope () [14]. There are 14 unknown coefficients in the above
equations.
The tuning of digital PID controllers for industrial plants could frequently be difficult and time-consuming,
even if expert systems are used to automate the tuning process. Indeed, the use of such expert systems for the
156
tuning of digital PID controllers often creates more problems than it solves because of the difficulties inherent in
the process of knowledge elicitation [8]. Therefore, constrained PSO based TTPIDC was proposed as an alternative means of tuning digital PID controllers. In fact, these algorithms provide a much simpler approach to the
tuning of such controllers than the classical approach. Since its invention, the PID control scheme has found wide
applications in industry. Owing to its simplicity and unexpectedly good performance provided by these three
actions defined as a positional form as follows;
s
(3)
r 0
Where e(k) is error signal, Kp is proportional gain, Ki is integral gain, Kd is derivative gain of controller.
Object of constrained PSO algorithm is to find optimum parameters of PID controller which provide maximum performance criterion value while smooth control signal achieved. Realized TTPIDC block diagram is
shown in Fig.1.
Trajectory
Power
+ PT
e
_ PR
Reactor Model
PID Controller
Kp
e(t)
s
e(t)
Ki
t=0
de
dt
Kd
1-
Point Kinetics
Model
PR
Thermo -Hydroulic
Model
Z-1
PSO Algorithm
The Best
PID
Parameters
Optimization
Of
Parameters
Cost
Function
(J)
ITAE
ts
t | PT (t ) PR(t ) |
(4)
t t0
Where, t is the time vector, PR(t) is reactor power and PT(t) is trajectory power.
t1
t2
ts
t t 0
t t1
t t 2
(5)
Where,
J J ;
(u( t n ) u( t n 1) ) u
J J
(u( t n ) u( t n 1) ) u
is penalty value.
(6)
PSO is one of the evolutionary computation which used group intelligence generated by cooperation and
competition between group particles to guide the optimization search. Its unique memory function enables it to
157
adjust its searching strategy according to most current search situation. Thus, PSO can be more easily realized
and is a more effective searching algorithm. At present, PSO have already made breakthrough progress in some
practical application fields [31].
Like other evolutionary computation techniques, a population of potential solutions to the optimal problem
called particles is used to probe the search-space. In PSO, every particle has a fitness value determined by object
function, and is assigned a randomized velocity which determines their movement. The concrete steps of PSO are
as below:
1. Random initiate particle population. Randomly generated particle population, each particle represents a candidate solution of a satisfaction utility preference value set.
2. Evaluate all particles according to fitness degree value and update individual extreme value according fitness
degree value function.
3. Update global extreme value according to fitness degree function.
4. Iterate speed and position according to update set of speed-displacement.
5. Repeat step 2 to 5, until iterative condition is satisfied.
Update set of speed-displacement is:
Vi (k 1) WVi (k ) c1r1 ( Pi (k ) X i (k )) c2 r2 ( Pg (k ) X i (k ))
X i (k 1) X i (k ) Vi (k 1)
W Wmax
W Wmax T max
Tmax
(7)
(8)
(9)
In these sets, position X stands for satisfaction utility preference value set, initial position X i is the initially
generated satisfaction utility preference value set; V stands for speed, P stands for individual extreme value of
each particle, g stands for global extreme value, c1 and c2 are study factors, W is inertia weight, r1 and r2 are random numbers in (0,1). In our study, this parameters are selected as; c1=c2=2, W linear decreasing 0.9 to 0.4,
swarm size=35, number of iterations=35.
5
This paper introduced the PSO based TTPID controller to control the Triga Mark-II research reactor established at Istanbul Technical University. The parameters of PID controller were determined using constrained PSO
algorithm. The best fitness value developed by PSO is 6045. For this system, this value is very successful. Then
to test the designed controller, various initial and desired power levels are applied to MARK II reactor. In the first
experiments, reactor has started up for PTd=50 kW, 150kW and 250kW when initial power PTi=500 W and
trajectory slopes 1=50, 2=20, 3=50. At these experiments reactor is reached the set point without overshoot
and 40W steady-state error is obtained.
Figure 2-a. Power Output and Error when (1=50, 2=20, 3=50, PTi=500W,
PTd=50 kW,150 kW, 250kW) Selected.
158
Figure 2-b. Control Signal and Rod Position when ((1=50, 2=20, 3=50,
PTi=500W , PTd=50 kW,150 kW, 250kW) Selected.
Reactor outputs and errors between trajectories and outputs are given in fig. 2-a. It is clearly observed from
these figures that reactor outputs are perfectly tracked the defined trajectories. Control signals and rod positions
are also given in fig. 2-b.Control signal and rod position gradually changed so there is no destructive signal to
send to the electro mechanic apparatus of the rod.
In the second experiments, reactor outputs at the three initial power level PTi=100 W, 1kW and 5kW are
observed. Reactor outputs, errors, control signals and rod positions are given in fig. 7-a. and fig. 7-b. respectively.
Reactor is reached the set point without overshoot and a smooth control signal is also achieved. The maximum
tracking error is 175W when PTi=100W selected. Deviation of initial power has no major effect on control signal
or reactor output but, on the other hand setting time is gradually changes.
Figure 3-a. Power Output and Error when (1=50, 2=20, 3=50, PTd=250 kW,
PTi=100 W,1000 W, 5000W) Selected.
Figure 3-b. Control Signal and Rod Position when (1=50, 2=20, 3=50,
PTd=250 kW, PTi=100 W,1000 W, 5000W) Selected.
As a result, the all simulation results show that the reactor power successfully tracks the given trajectory and
TTPIDC could control the system successfully under all conditions within the acceptable error tolerance.
159
References
1.
2.
3.
4.
5.
6.
7.
8.
9.
10.
11.
12.
13.
14.
15.
16.
17.
18.
19.
20.
21.
22.
23.
24.
25.
26.
27.
28.
29.
30.
31.
G. Chengyi, S. Qing, and C. Wenjian, "Real-Time Control Of AHU Based On A Neural Network Assisted Cascade Control System," in Conference on Robotics, Automation and Mechatronics, Singapore, 2004.
J. E. Seem, A New Pattern Recognition Adaptive Controller with Application to HVAC Systems, 1998.
K. J. strm and T. Hgglund, PID Controllers Theory, Design and Tuning. USA: Instrument Society of America, 1995.
Y. Lee, S. Park, M. Lee, and C. Brosilow, "PID controller tuning for desired closed-loop responses for SI/SO systems,"
AIChE, vol. 44, pp. 106-115, 1998.
P. Comino and N. Munro, "PID controllers: recent tuning methods and design to specification," in Control Theory and
Applications. vol. 149, 2002, pp. 46-53.
W. Feng and Y. Li, "Performance indices in evolutionary CACSD automation with application to batch PID generation,"
in Int. Symp. on Computer Aided Control System, Hawai, USA, 1999.
M. nal, "Optimization Of PID Controller Using Ant Colony / Genetic Algorithms and Control Of The Gunt RT 532
Pressure Process," in Institue for Graduate Studies in Pure and Applied Sciences. vol. Master stanbul: Marmara University, 2008.
H. Jones and P. B. DeMouraOliveira, "Genetic Auto-Tuning of PID Controllers," in Genetic Algorithms in Engineering
Systems: Innovations and Applications, 1995.
H. A. Varol and Z. Bingul, "A New PID Tuning Technique Using Ant Algorithm," in American Control Conference, Boston, Massachusetts, 2004.
Y.-T. Hsiao, C.-L.Chuang, and C.-C.Chien, "Ant Colony Optimization For Designing of PID Controllers" in lnt. Symp. on
Computer Aided Control Systems Design, Taiwan,2004.
V. Topuz, "Bulank Genetik Proses Kontrol," in Institue for Graduate Studies in Pure and Applied Sciences Istanbul,
Turkey: Marmara University, 2002.
R. K. Ahuja, J. B. Orlin, and A. Tiwari, "A greedy genetic algorithm for the quadratic assignment problem," Computers &
Operations Research, vol. 27, pp. 917-934, 2000.
E. Bonabeau, M. Dorigo, and G. Theraulaz, "Inspiration For Optimization From Social Insect Behaviour," Nature, vol.
406, 2000.
Colorni, M. Dorigo, and V. Maniezzo, "Distributed Optimization by Ant Colonies," in the First European Conference on
Artificial Life, Paris, France, 1992.
M. Dorigo, "Ant Algorithms for Discrete Optimization," Artificial Life, vol. 5, p. 137, 1999.
D. E. Goldberg and K. Sastry, "A Practical Schema Theorem For Genetic Algorithms Design and Tuning," University of
Illinois, USa 2001017, 2001.
E. A. Gonzalez and F. S. Caluyo, "Normal Ziegler-Nichols-Based PID Retuning Using Sequential Ant Colony Optimization (SeqACO)," in 27th Annual Philippine-American Academy of Science and Engineering Meeting & Symposium
(APAMS 2002) Manila, Philippines, 2007
M. Jeffers, "A Genetic Algorithm Based Fuzzy Logic Controller," B.Eng. in Electronic Engineering, Dublin 2001.
M. Salami and G. Cain, "An Adaptive PID Controller Based on Genetic Algorithm Processor," in Genetic Algorithms in
Engineering Systems: Innovations and Applications, 1995.
G. Zhang, B. E. Patuwo, and M. Y. Hu, "Forecasting with artificial neural networks: The state of the art," International
Journal of Forecasting, vol. 14, pp. 35-62, 1998
Kennedy, J., Eberhart, R.C. Particle swarm optimization Proc. IEEE int. Conf. on neural Networks,Vol. IV, pp. 1942
1948. Piscataway, NJ, 1995.
H. Yoshida, K. Kawata, Y. Fukuyama, Y. Nakanishi, A particle swarm optimization for reactive power and voltage control considering voltage security assessment, IEEE Transactions on Power Systems 15 (2000) 12321239.
Can, B., Yavuz, H., Akbay, E.: The Investigation of Nonlinear Dynamics Behavior of ITU Triga Mark-II Reactor, Eleventh European Triga Users Conference, The Germen Cancer Research Institute Heildelberg, Germany, (1990) 11-16
Can, B.: The Optimal Control of ITU Triga Mark II Reactor, The Twelfth European Triga Users Conference, NRI Bucuresti-Pitesti, Romania, 22-25, (1992)
Erkan, K., Can, B.: Self-Tuning Control of ITU Triga Mark-II Reactor, First Trabzon International Energy and Environment Symposium, KTU, Turkey (1996) 29-31
Baba, A.F., Can, B., Gok, I., Akbay, E.: Design of a Fuzzy Logic Controller for the ITU Triga Mark-II Reactor, Proceedings of the Nuclear Plant Instrumentation Control And Human-Machine Interface Technologies, The Pennsylvania State
University, USA, (1996)
Erkan, K., Butun, E., Can, B.: Triga Mark-II Reactor Controller Design Using Genetic Algorithm, Fuzzy Logic and Intelligent Technologies for Nuclear Science and Industry, World Scientific, Singapure, (1998)
Can, B., Baba, A.F., Erdal, H.: Optimal Controller Design by Using Three Zones Trajectory for ITU Triga Mark-II Reactor, Journal of Marmara University for Pure and Applied Sciences, Vol.15, Istanbul, Turkey, (1999) 187-196
Akbay, E., Can, B, Baba, A.F.: Fuzzy Logic Control for ITU TRIGA Mark-II Reactor, 3rd International Symposium on Intelligent Manufacturing Systems, Sakarya, Turkey, (2001)
Coban, R., Can, B.: An Expert Trajectory Design for Control of Nuclear Research Reactors, Expert Systems with Applications 36 (2009) 1150211508
J. Kennedy, R.C. Eberhart, Y. Shi, Swarm Intelligence, Morgan Kaufman Publishers, San Francisco, 2001.
160
Abstract. Face and facial expressions express the feelings of a human. Human-computer interaction systems highly depend on face and facial expression recognition. We are proposing a new algorithm for person
dependent facial expression recognition from single frontal face image. Our proposed algorithm uses 3D wire
frame face models to model six basic facial expressions of the face: Anger, disgust, fear, happiness, sadness,
surprise. We have tested our algorithm on BU-3DFE database. Proposed algorithm provides acceptable recognition performance with only one known face image.
Keywords: face modeling, facial expression analysis, facial expression recognition, facial features.
1 Introduction
Face and facial expressions play a crucial role in human computer communication. Recognition of face
and facial expressions is essential in order to provide natural dialogue with a computer. During last ten years, the
developments in the field of multimedia signal processing and 3-Dimensional (3D) imaging methods have attracted the utmost attraction of the researchers to focus on a robust solution of facial expression recognition problem.
Recognition of facial expressions is identified by the classification problem between the expression faces and
researchers develop feature based approaches to the classification of facial expressions. By the development of
facial expression recognition systems, a number of applications in the field of human computer interaction can be
developed.
Recognition of facial expressions can be studied in two different categories: depending on a person or it
can be person independent. Person dependent algorithms focus on the problem of facial expression classification
of a known face whereas person independent algorithms work without knowing the person.
The problem of facial expression recognition is studied as a classification problem. Therefore, extraction
of the discriminative features affects the overall classification process and improves the recognition performance.
Several facial features can be extracted from face images depending on the classification problem. In our study,
we consider 2-Dimensional (2D) geometrical facial feature points.
Recent research studies in the field improved automatic facial expression systems at reasonable recognition performances. Unfortunately, most of the systems developed have several limitations that make them unsuitable for real-time applications. Therefore, automatic recognition of facial expressions is still a challenge for the
digital computers and many researchers are working on this concept. The limitations of facial expression recognition performance are the result of high variability in face images including facial expressions. These limitations
can be listed as pose, lighting, resolution and orientation of the face images. Thus, a robust facial expression
recognition system should take into account these limitations, resulting in an extremely large variation in the face
images.
Early researches in facial expressions field lead to the classification of human basic facial expressions in
seven categories: Anger, Fear, Disgust, Happiness, Neutral, Sadness and Surprise [1]. Later on, facial movements
are defined on the face in order to code facial expressions [2].
161
In our study, we focus on the problem of facial expression recognition for a known person. The aim of
the study is to recognize the facial expressions of a known person at acceptable recognition rates with only single
frontal face image in the known set. Our approach is based on 3D face model. We develop 3D facial expression
models for each person to recognize unknown facial expressions. The 2D geometrical facial feature points are
used in the modelling process, that they are less sensitive to lighting and pose variations.
The paper introduces a novel algorithm for recognition of facial expressions. First, a literature review is
given in section 2, then, face and facial expression modelling is summarized based on our previous studies in
section 3 [6, 7, 9], and model based expression recognition algorithm is explained in section 4. Finally, we conclude with our experimental results and conclusion.
2 Literature Review
The early studies of Ekman in 1970s have been accepted as the reference for the classification of basic
facial expressions [1]. The basic facial expressions according to Ekmans studies are anger, fear, disgust, happiness, neutral, sadness, and surprise. Later on, Facial Action Coding System (FACS) was developed by Ekman and
Friesen to code facial expressions in which the movements on the face are described by action units [2]. Many
researchers have worked on these concepts then. Most of the researchers analyze facial expressions in 2D, and
were tracking the facial movements with the help of facial feature point locations in images and video. In 1999,
MPEG-4 standard defined a model for neutral face with 84 feature points. MPEG-4 standard also defined 68
Facial Animation Parameters (FAPs) for the animation of the face by the movements of the feature points.
MPEG-4 FAPs are popular in most of the research labs to synthesize or recognize facial movements and expressions [3].
In the field of face modeling, one of the most important works done is the face model created by Stromberg, named as Candide, which is popular in many research studies [4]. Some of the methods are using two orthogonal face images to generate the facial texture or on single frontal face image. In our previous studies, we
have developed a new face model for three different resolutions, high, low and medium, and applied two orthogonal face images to obtain the face texture, with our rotation adaptive texture mapping method [5]. Later on,
we developed a face mask in order to synthesize a face from a single frontal face image [6, 7]. Face modelling is
also used in face recognition systems [10]. In this study, we are using the artificial face and facial expression
models generated using our previously proposed model based facial expression synthesis system [9] in order to
recognize the basic six expressions of a face, for a known individual.
model, where
Vm, i
x,
Vm, i
V
y , and m, i z are x, y, and z coordinate values in 3D space [6, 7, 9].
162
,V
T V , V
m m,1 m,2 m,.3
m 1
V
V
Vm,i
Vm,i
m, i m,i x
y
z
(a)
(b)
(1)
(d)
(c)
Figure 1. Wire frame model used: side views (b),(c), front view (a). 20 facial feature points used for modeling (d).
Neutral generic face model is then adapted to the facial characteristics of the individual. To achieve
adapted face models, 3D face model undergoes a model matching procedure. Model adaptation algorithm considers the subdivision of the local regions of the face: silhouette, nose, mouth, right eye and left eye. The detailed
explanation about the synthesis of 3D face models can be found in [6, 7, 9].
3.2 Facial Expression Modeling
Facial expression synthesis is achieved by the deformation of 3D wire frame model with the use of
MPEG-4 Facial Animation Parameters (FAPs). MPEG-4 feature points and FAPs are used for facial animation.
FAPs represent a set of basic facial movements including head motion, tongue, eyes and mouth control. FAPs
obtain displacement functions for each vertex. After computing these displacement functions, the new vertex
coordinates are calculated and applied to deform and model the facial movements as shown in Eq. (2) [3].
The weight Wi is representing the distance of the vertex from the feature point affected from the FAP. Wi
decreases from the centre of the influence area to the less affected area. Wi = 0 for the vertices which are not
affected by the FAPs. Facial expression models generated for an individual is shown in Figure 2 for the six basic
facial expressions: Anger, disgust, fear, happiness, sadness and surprise. The detailed explanations and outputs
can be found in [9].
i x Wi x
'
i y Wi y * W j * FAPx , y , z
i Wi
z z
(a)
(c)
(b)
(
2)
(d)
163
(e)
(g)
(f)
(h)
Figure 2. (a) Original frontal face image, (b) 3D wire frame model of the individual, artificial facial expression models generated: (c) anger (d) disgust (e) fear (f) happiness (g) sadness
(h) surprise.
Face
Model Synthesis
20 Facial Features
for Unknown Expression
Expression Recognition
Expression
Model Synthesis
Unlike the most of the algorithms, our proposed algorithm is based on only single frontal face image
with 20 facial feature points located, making it suitable for real-time applications. Currently, the algorithm is
tested without any training stage, only one frontal face image is known for each person. This gives some limitations in our classification performance, but the algorithm shows promising results with only single face image.
In this study, classification of facial expressions is done in 4 main classes: Anger or Disgust for the first
class, Fear or Surprise in the second class, Sadness in the third class and Happiness in the fourth class. Due to
overlaps in 3D wire frame models for Anger and Disgust, and for Fear and Surprise, we classify them in one class
currently. The next step of the study will be the within class separation of the first two classes reported in this
paper. Also, addition of a training phase will further improve the classification performance as well as within
class separation of class 1 and class 2.
164
5 Experimental Results
Our facial expression recognition algorithm is tested on BU-3DFE database. Database includes face
images and 83 facial feature point locations in 3D space for 56 female and 44 male samples. We have used 10
female and 10 male samples in order to model and generate artificial expressions. Then, we have tested our algorithm for six basic expressions of each face: Anger, Disgust, Fear, Happiness, Sadness and Surprise. The inputs to
the system are the 20 facial feature point locations in 2D space provided in the database. Totally, we have generated 20 neutral face models, and 120 expression models for 20 input facial feature point sets. Then the recognition
performance of the system for the 120 input feature point sets are measured and classified in 4 different classes:
Class 1 as Anger or Disgust, class 2 as Fear and Surprise, class 3 as Sadness, and class 4 as Happiness. Table 1
summarizes the preliminary test results.
As it is observed from the Table 1, surprise and anger expressions are classified with high recognition rates, whereas , the other expressions have relatively average classification accuracy. In fact, the performance results indicate that there are too much similarity between artificial happiness, anger, disgust and sadness models.
This is the disadvantage of using Euclidean distance metric as the classifier, furthermore, measuring the distances
between the whole wire frame models. Thus, one of our future research directions will be the subdivision of the
face models to the region of interests in the recognition phase. This will reduce the similarities between the face
models of different expressions and improve the recognition performance of the system.
Table 2. Recognition performance of our algorithm.
Expression
Correct
Classification
Incorrect Classification
Classification
Rate (%)
Anger
16
80
Disgust
15
75
Fear
16
80
Happiness
13
65
Sadness
15
75
Surprise
20
100
Overall
95
25
79.16
6 Conclusion
In this study, a model based person dependent facial expression recognition system has been proposed.
To test our system performance, we used BU-3DFE facial expression database [8]. Our algorithm produces acceptable recognition rates in BU-3DFE database. The overall recognition rate is around %79. The proposed
model based algorithm produces promising results by using only single face image for an individual whereas,
most of the other recognition systems are using number of face images in the training phase. By the improvements
in the training phase, higher recognition rates can be achieved with the proposed model based system.
Future directions of the study will be the within class seperation of highly correlated face models for the
expressions anger and disgust in class 1, and fear and surprise in class 2. Also, improvements in the classifier will
improve the system performance. Currently we are using Euclidean distance metric between the wire frame models to measure the similarity. Further improvements of the classifier with a training phase and optimization of
wire frame vertices for best seperability among expression classes will highly improve our recognition rates.7
165
References
1.
Ekman, P. & Friesen, W. (1976). Pictures of Facial Affect. Palo Alto, CA: Consulting, Psychologist.
2. Ekman, P. & Friesen, W. (1978). The Facial Action Coding System: A Technique for the Measurement of Facial Movement, Consulting Psychologists Press, San Francisco.
3. G.Abrantes, F.Pereira, MPEG-4 facial animation technology: survey, implementation and results, IEEE Transactions
on Circuits and Systems for Video Technology, vol.9, no. 2, pp. 290 305, March 1999.
4. Ahlberg, J.: Candide-3: an updated parameterised face. Report No. LiTH-ISY-R-2326, 2001 (Linkoping University,
Sweden).
5. Kamil Yurtkan, Hamit Soyel, Hasan Demirel, Hseyin zkaramanl, Mustafa Uygurolu, and Ekrem Varolu, Face
Modeling and Adaptive Texture Mapping for Model Based Video Coding, A. Gagalowicz and W. Philips (Eds.): CAIP
2005, LNCS 3691, pp. 498 505, 2005.
6. Kamil Yurtkan, Turgay elik and Hasan Demirel, Automatic Facial Synthesis from Single Frontal Face Image, The
13th IASTED International Conference on Computer Graphics and Imaging, CGIM 2010, February 15-17 2010, Innsbruck, Austria.
7. Kamil Yurtkan ve Hasan Demirel , Dsk Bit Hizli Video Kodlama Uygulamalari iin MPEG-4 Uyumlu 3B Yz Modellemesi ve Sentezi ,3.Haberlesme Teknolojileri ve Uygulamalari Sempozyumu,Habtekus 2009,9-11 Aralik, Yildiz
Teknik niversitesi.
8. Yin, L. Wei, X. Sun, Y. Wang, J. & Rosato, M.(2006). A 3d facial expression database for facial behavior research. In
Proceedings of International Conferance on FGR, pp. 211-216, UK.
9. Kamil Yurtkan and Hasan Demirel, "Facial Expression Synthesis from Single Frontal Face Image", 6th International
Symposium on Electrical and Computer Systems, 25-26 November 2010, Eueropean University of Lefke,
Gemikona,TRNC.
10. Kamil Yurtkan and Hasan Demirel , Model Based Face Recognition Under Varying Facial Expressions, International
Science and Technology Conference, ISTEC-2010, 27-29 October 2010, Gazimausa, TRNC.
166
1 Introduction
During the last two decades the Internet and mobile communications have evolved to become the main networking platforms that people and organizations use in their online activities.Hundreds of millions of people all
over the world connect to the Internet to perform such diverse tasks as Web surfing, talking to others, video sharing,game playing, gambling,e-commerce , online banking, conducting business etc. All these and similar activites
are essentially made possible by the developments in hardware, software and internetworking technologies. In
recent years, online social networking have become very popular which constitutes a new type of social life for
most people. The services offered by the sites like Facebook, Orkut, MySpace,LinkedIn etc. allow people to host
and expand their online social networks.The users create their profiles in these platforms and exchange information with their friends and strangers.Within the last few years,the number of users of online social networks
has grown at an unprecedented rate. For example, Facebook now attracts more than five hundred million users
daily, an average user has 130 friends and over 30 billion pieces of content are shared monthly[1].
Internetworking exploison have also created serious security and privacy risks.The risks are especially considerable for search engines and mobile and cloud computing platforms.It is a well known fact that, active and
passive profiles of most network users contain private information which is aggregated over time. The utilization
purposes and actual uses of accumulated information can be harmful for the individuals or organizations concerned. New technologies including Deep Packet Inspection have further expanded potential sources for obtaining sensitive data and almost automated the entire process[2].
This paper,investigates the privacy risks involved in currently popular internetworking environments and offers
an evaluation of various types of related practices.First,the concept of privacy is explored within the context of
internetworking applications and then the negative impact of social networking exploison on privacy is discussed.Privacy related problems caused by cloud computing , mobile computing and deep packet inspection are
167
also considered and an up to date evaluation of the inherent privacy risks brought by these technologies are presented.Finally,current privacy protection measures and practices involving internetworking platforms are discussed.
2 Privacy Concerns in a World of Networks
Classical definitions of privacy involve the ability of an individual controlling the exposure of sensitive information about himself or herself.Sensitive personal information that is obtained in networking environments is
covered to some extent in this definition. Therefore,lesser privacy means a reduction in a persons freedom to
control private information.In a more IT-related definition, privacy is considered as The right to be free of
unreasonable personal intrusions and the right to be left alone[3].Accordingly, privacy is a condition in which
people have limited access to the personal affairs and sensitive information of others.However this classical privacy framework is beginning to receive opposition stimulated by the latest developments in technology and expansion of social networking[4]. In fact, increased numbers of channels and incentives to share personal information with others make it more difficult to maintain the division between public and private lives of people.
Furthermore,what is sensitive for someone may not be so for others and what could be considered sensitive in
one context may not be in another.The prevalence of mobile technology and social networking have changed
traditional notions of public and private in such a way that now private lives becoming more and more public[5]. Mobile devices and increasing availability of location-based applications further bring public eye into
the private realm. So,the expectations of privacy in a networked world should be different than those of preInternet era.Social networking supporters point to this difference to justify further openness and sharing of personal data for various motives .
The Internet was originally designed to be a borderless dumb network of networks that operates without
any central governing authority.However, applications and developments expanded the structure and role of the
Internet to involve various aspects of daily life.Cultural and legislative differences of countries make various
Internet related activities illegal in one country and legal in another.Current extent of the internetworking causes
ethical and legal privacy problems in this new and borderless cyber world.Not only information privacy, but
security also is a growing concern for individuals,organizations and other public authorities.The need for protecting public safety is considered to have far higher priority than respecting the privacy rights of network users
in most situations.Indeed,governments and public security authorities use terrorism and crime fighting as excuses for limiting or even violating the privacy rights of citizens in many countries[6].
Governments ,corporations and other responsible bodies compile and accumulate huge volumes of personal data about individuals from online sources.Some data is gathered during e-government dealings,such as
filling out an application form for a payement or for demanding a service.Some data is collected when using a
credit card at a supermarket , at a travel agency or for payements during Internet dealings etc.Online data is also
picked up by omnipresent surveillance cameras, radio-frequency identification chips in products and ID cards and
cell phone tracking devices.However, according to the law and the rules of ethics, sensitive personal information
should only be collected,used and disclosed with the knowledge, and consent of the individual and only for
legitimate purposes.Normally, security and confidenciality of the personal information should be guarantied by
the data collecting entity.A discouraging fact in this respect is the speed and ease of copying data from one server
and moving it to another,irrespective of location ,border or political system.Outsorcing practices also have potential to increase cross-border privacy related problems.
There are a number of technology related factors that increase the scope and complexity of privacy and security problems.Some important factors :Ever expanding volumes of electronic data that are being stored , processed and communicated over internetworks,ever increasing complexity of computer hardware and software
components,fast spread of portable devices with high processing capabilities,diminished longtime storage,processing and communication costs and applications of data mining and other business intelligence technologies.The combination of all these factors render protecting privacy and maintaining security in internetworks
difficult tasks.Concerning these environments,the definition and enforcement of common privacy rules and
detection of their violations are becoming even more difficult.So,a paradigm shift about the definition and protection of privacy in modern internetworks is needed.This could be the first step towards resolving the dilemma
of choosing between personal privacy and public safety.
168
169
170
communications at the request of law enforcement ,enforcing copyright laws, giving transmission priority to
some packets over others and identifying computer viruses and spam traffic.
Obviously,the privacy of information is always at stake in a network environment where DPI is being performed[11].Since most Internet traffic is unencrypted, DPI enables network providers to interfere with their customers' Internet activities,including email, social network communications,web surfing and downloads. Entering
this way into packets reduces the effectiveness of privacy preserving measures as extracting sensitive information
is easy.DPI is popular for governments that wish to control the network traffic of the country.If an oppressive
regime wants to control the flow of information across its borders, blocking Web sites hosted elsewhere is no
longer sufficient.Because, online information sharing through social networking sites are now easy and widespread.Since DPI techniques are very effective for scanning message contents, they can be used for network
surveillance and blocking communications. The DPI process can extract a real time copy of any type of information; including email messages,forum postings,Internet phone calls,URL and IP addresses[12].Because of its
great power,the technique is also used for network wiretapping by public authorities in most countries for law
enforcement purposes.
DPI technologies permit the collection and use of personal information about internet subscribers,based on
their online traffic.The information accumulated this way is utilized to form detailed profiles of people to infer
their desires and interests about various commercial service and product categories. The profile is then used for
creating targeted advertisements. For example, Googles Gmail displays advertisements related to the text of email traffic of users.It is usually very difficult for a subscriber to opt-out or even to have knowledge about this
kind of information collection and targeted advertising.Obviously,most users whose Internet traffic is being
exploited,do not consent for the inspection and use of their mostly private data in such a pocess.
Although DPI is controversial and has been facing criticisizm from privacy and network neutrality circles,content inspection and filtering is common among corporations, schools and other institutions.The technology is mainly used to block viruses ,spam ,pornography,gambling and to enforce IT usage policies to students
and employees.However,the real threat to privacy is posed by Internet service providers with their dominating
technological and financial powers.They do not usually reveal their motives for DPI and do not ask the subscribers prior consent for content inspection.In fact, average users cannot be expected to have enough power
to protect their privacy or to easily switch provider.The DPI software is generally proprietary and privacy protecting setups are too complex for most users to properly understand and configure the services or to opt-out
parts of it.The result is the capability of singling out a subject,tracking every move on the cyberspace,recording
the details of every e-mail, following every Web search and even physical trajectory of someone.All these facts
are alarming and have increased privacy related objections from the public and calls for legal protection in countries like the US,Canada and GB.Certainly new forms of protective legistlation and self-regulation from network
operators are needed to protect individuals from privacy risks of DPI
6 Conclusions
The sizes and scopes of internetworking platforms have expanded to cover considerable percentages of populations and organizations in most countries.Especially,social networks,CC and newer DPI technologies create
increasingly higher privacy and security concerns which are difficult to overcome using older preventive
measures.Social networks have great potential for the leakage of sensitive private information as they aim to
creating and expanding social relationships similar to real life among their members.Active and passive user profiles that are formed in these networks are not only utilized for targeted advertising but also open to abuses and
misuses for malicious purposes.The privacy settings offered by social networks are usually very complex to be
fully grasped or mastered by average users.Therefore,simple sets of rules and sandards are needed for preventing
abuses and protecting privacy rights in social networks.
CC on the other hand creates different types of privacy problems as huge volumes of Internet traffic and personal data are managed by the cloud operator on behalf of customers.Cloud operators,public authorities or hackers can interfere with the data to utilize it for various intentions.The owner has very little control over the data
once it is trusted to the cloud operator.Stringent rules procedures and eventually standards are needed for protecting cloud customers from misuses of private data.
DPI capabilities of latest network control and monitoring technologies are useful for improving performance,but also bring direct privacy risks for network users.Because,proprietary packet inspection devices and
software can examine,change and copy the contents of tansmitted packets.This is a great power which can be used
for law enforcement,crime fighting ,network surveillance and profit making.A network user does not have too
many options to prevent deep DPI process as the operational procedures and hardware/sofware components are
beyond the control of network users. In reality, DPI introduces additional complexities to already complicated
problems of protecting privacy and security in large internetworking platforms.Standards and rules are needed
171
that would force network operators to get permission from customers before inspecting and tracking their online
activities.
A major difficulty arises from the fact that the internetworking technologies are borderless and their application areas are expanding very rapidly.These conditions make it difficult to develop appropriate policies,standards and legislation which will prevent inherent privacy and security risks. Because the lifetimes of
related technologies and applications are so short,the usage types and procedures never reach to stability.Intelligent software aided privacy protection may offer solution to most problems, but research in that direction has not been productive enough so far.
References
1. Facebook Statistics(2011). http://www.facebook.com/press/info.php?statistics
2. Facebook Privacy Policy(2011). http://www.facebook.com/policy.php
3. Brey, P. (ed.) (2006). Privacy and Surveillance. Special issue of the journal Ethics
and Information Technology, 7(3) .
4. Jones, H. and Soltren, J.H..(2005). Facebook :Threats to privacy.Technical report,
Massachusetts Institute of Technology.
5. Felt,A. and Evans, D.(2009). Privacy Protection for Social Networking Platforms, in
Proc. of Web 2.0 Security and Privacy.
6. Grimmelmann, J. T. (2009). Facebook and the Social Dynamics of Privacy. Iowa
Law Review,95(4).
7. He, J ,Chu,W and Liu,V.(2006).Inferring privacy information from social networks.
In Proceedings of Intelligence and Security Informatics, volume LNCS 3975.
8. Knight,J. (2010). Exposing the inherent privacy design flaws in Facebook
Online Conference on Networks and Communities.
9. European Network and Information Security Agency (2009).Cloud Computing:
Benefits, Risks and Recommendations for Information Security. Report.
10. Allan A. F. and Darrell M. W.(2010). Privacy and Security in Cloud Computing.
http://www.brookings.edu/techinnovation
11. Clarke, R. (2009). Deep packet inspection: Its nature and implications. Office of
thePrivacy Commissioner of Canada Deep Packet Inspection Essays Website.
12. Bendrath, R. and Mueller, Mi. (2010). The end of the Net as we know it? Deep
packet inspection and Internet governance. Working paper series, SSRN.
13. Daly, A(2010).The legality of deep packet inspection.First Interdisciplinary 07-19,
Workshop
on
Communications
Policy
andRegulation
Amherst
No. 2007.
172
Technical
Report
Abstract. Information system applications are basically designed to administer business processes
faster and more comprehensively. Today, institutions are tendency in the information systems projects to
use predominantly web-based architectures. These projects must comprise certain qualified features including authentication, authorization, logging, error handling/reporting, presenting a user-friendly interface and
a healthy account management administration. In practice, using all of these features separately in each developed application makes the application development process longer, disrupts the integrity of business
systems and forces to user to login separate addresses to access applications. This study presents a proposal
on managing different servers and server services under a single framework. This proposal promotes security level by allowing login control operations under one-single point and thus prevents recurrence of same
operations. In this study; besides regarding web-based projects into a service unity, it is also aimed for carrying out instant used services needs such as authentication, authorization, logging, error handling/reporting and a healthy account management administration through a user-friendly interface.
Keywords: Campus Automation, Service Unity, Web Based Projects, One-Single Point
Control
This work was supported by Sakarya University Scientific Research Foundation (Project number:
2007.01.10.001)
1 Introduction
Information system applications are basically designed to administer business processes faster and more comprehensively. Today, institutions are tendency in the information systems projects to use predominantly webbased architectures. These projects must comprise certain qualified features including authentication, authorization, logging, error handling/reporting, presenting a user-friendly interface and a healthy account management
administration. In practice, using all of these features separately in each developed application makes the application development process longer, disrupts the integrity of business systems and forces to user to login separate
addresses to access applications. In the light of these findings, we should highlight the importance of web portals.
The World Wide Web continues to be the superior application on the internet because it has regularly reinvented
itself. In fact, for most people, the World Wide Web has become synonymous with the Internet. With the introduction of web portals, the web is in the process of reinventing itself once again. This change may prove to be
more far-reaching than any other change to hit the Web, and it will change the way that university and corporate
Web pages are built, the organizational structures used to build them, and the essential way that people use the
173
web. There have been many different definition sentences about web portal. Like so many of the terms that we
encounter in our industry, the word portal has come to mean nearly anything; in fact, it is probably the most overused term on the web today. But what does it mean? What do you do when your boss comes to you and says he
wants you to build a portal? Most people will tell you that a portal is some kind of entry page that represents a
larger section of content [1]. Wikipedia, for example, defines it as follows:
A web portal is a website that provides a starting point or gateway to other resources on the internet or intranet
[2]. So, given a definition as loose as that, what isnt a portal? Each of the following examples would fit the definition of a portal quite nicely:
http://Google.comas a search engine, it could easily be considered as the starting point for the entire Internet.
The front page of your companys intranetit has links that lead to all of your companys content.
Any page on the Internet that contains hyperlinksthese would pretty much satisfy those definitions.
Given such a simple definition, each of the three listed items must be a portal because each provides a place from
which to start an excursion on the Internet. However, when the term portal is used to describe websites, it implicitly refers to other, more specific characteristics beyond simply being a page with hyperlinks on it. So well need
to come up with a definition of our own which better describes the type of websites that we think of when we
refer to portals.
Specific application portals are typically built on common sets of core services, so reusability of these services
is a key problem in portal development [3]. In this study we address the reusability problem by presenting a comprehensive view of an interoperable portal architecture, beginning with a set of core services built using the web
services model and application metadata services that can be used to build campus automation front ends out of
these core services, which in turn may be aggregated. This study also presents a proposal on managing different
servers and server services under a single framework. Proposed single framework might be regarded as a portal
architecture. This proposal promotes security level by allowing login control operations under one-single point
and thus prevents recurrence of same operations. In this study; besides regarding web-based projects into a service unity, it is also aimed for carrying out instant used services needs such as authentication, authorization, logging, error handling/reporting and an healthy account management administration through a user-friendly interface.
2 Literature Review
With the introduction of Web portals, the Web is in the process of reinventing itself once again. This change
may prove to be more far-reaching than any other change to hit the Web, and it will change the way that university and corporate web pages are built, the organizational structures used to build them, and the fundamental way
that people use the web. Portals are not a fad or a new name for something that weve been doing all along. They
will turn the Web from an institution-centric repository of information and applications to a dynamic user-centric
collection of everything useful to a particular person in a particular role. Instead of a single home page that proclaims identically to all who visit how grand the institution is, portals will give nearly every user a customized,
personal-specific, unique web page.
With the appearance of the internet and the development of advanced internet portal technology [4], the design
of internet-based information systems which could serve common sector/chain information interests seems to be
within reach. There are many confusing and often contradictory definitions about portals. It is useful to divide
portals into two groups: horizontal portals, or HEPs (Horizontal Enterprise Portals, also called mega-portals), and
vertical portals, or VEPs (Vertical Enterprise Portals). A horizontal portal is a public Web site that attempts to
provide its users with all the services they might need [3].
The HEPs combine a wide range of services, information, etc. and their building is generally simple, because
the main principle valid in this process is to make the site present anything for many types of users [5],[6]. A VEP
is a portal that delivers organization-specific information in a user-centric way. A university VEP should also
deliver all the information a HEP delivers. Whereas a HEP looks the same to all who first enter it, a VEP looks
quite different. Unlike a HEP, a VEP requires authentication for access. When a user logs on to a VEP, it produces a customized portal page, tailored to the user who logged on. It knows a great deal about the user who logged
on because the user is a member of the organization that produced the VEP[7],[8]. It knows, for example, what
174
cohort a user belongs to (for example, student, faculty, staff), what role a user plays (for example, help desk manager, department chair, lacrosse team member), what projects a user is involved with, how many vacation days a
user has taken this year, and much more.
The present situation in European countries as well as in the US is characterized by the emergence of horizontal portals, some of them highly publicized and with strong backing by interested groups [9] However, the majority of them has a narrow focus on one or a few sectors (products). Vertical portals are still rare and on an experimental stage. The experimental vertical portals are usually managed by a single member (enterprise) of a chain
and might only be accessible by members of the chain.
Examples of horizontal portals at farm level in different countries include machinery trade, plant production,
commodity trade, meat production, etc. [10]. Examples of vertical portals include a portal in meat production
where farms are linked with meat processors, restaurants and
consumers and a chain-internal portal in
grain/flour production where farms are linked with mills and bakeries [11]. In this study, proposed automation
architecture falls into vertical portal category due to vertical portal that delivers organization-specific information
in a user-centric way.
Many universities are considering what they call student portals or course portals or financial information portals. Although starting with a portal that has a limited constituency may make sense, the goal of a university
should be to move as quickly as possible to a single portal that serves everyone: students, faculty members, staff
members, alumni, and parents of students, prospective students, trustees, donors, and anyone else who would
access a university home page. This work comprises beyond these necessities. Beyond expected necessities find
its meaning in proposed comprehensive framework that also combine all other campus automations under a single
framework named as CAWIS.
3 Methods
3.1 Design Preferences
During all development processes of this automation system, ultimate web technologies have been traced and
investigated closely. At early times in process of development, necessities of university have been revealed openly. There have been occurred comparative performance tests among alternative technologies and results have
been reported. Among different programming languages, database systems and operating systems, following
preferences have been selected.
3.1.1 Operating System
Due to satisfied results from our tests especially in performance and security category, Linux was selected as
main operating system.
3.1.2 Web Server
Apache web server was chosen as web server because of unique harmony with Linux operating system[12].
3.1.3 Programming Language
Today, PHP [13], ASP [14], CFM [15], JSP [16] etc. are used for server-side scripting languages. By the light
of our test, it has been shown that PHP is speedier, more flexible and has more performance than other languages.
PHP was chosen as programming language.
175
3.1.5 Standardization
For the aim of getting rid of programming complexities and defining some orders about quality manners, it
was chosen a Web Software Standards cover at early times in the process of development.
3.1.6 Database Design
Currently used databases in other departments are in major purposes of this portal. That means created service
unity architecture comprises a flexible database structure over all projects used in university automation. This
database model satisfies one-point control on all projects. Especially, this model increases the efficiency of transaction durations and makes one-point security phenomenon easy to implement.
3.1.7 Security
128 bit SSL crypto was designed to use for protecting each created web page [18]. User specific information
was also protected with MD5 crypto-algorithm [19]. Every input was saved with an IP number and detailed transaction types. Authorization mechanism comprises due assignment for each users with different access level on
each services.
3.1.8 Data Integrity with Other Automation Systems in University
Data for providing available services is obtained from the other automation systems in university. Other automations include Medico Automation, Goods-Stock Automation, Circulating Capital Automation, Salary Automation, Student Affairs Automation, Santral Automation, Staff Automation, and General Documents Automation.
Figure-1 sketches relational flow-chart between CAWIS and all other automation systems in university. Web
Page box in Figure-1 resembles CAWIS interface.
Fig. 1. Relational Flow-Chart between CAWIS and All Other Automation Systems in University
CAWIS holds 9 different services namely WebGate, WebMail, WebObis, WebAbis, WebPbis, WebMenu,
WebRehber, WebForm and WebAnket. All these services are integrated with other automation systems.
176
WebGate (User Information Service) carries out access control, security logs, password change and personal
preferences settings.
WebMail (E-mail Service) carries out rich text support mail receive/send service, file and addresses manipulations.
WebObis (Student Information Service) carries out demonstrating course schedules, selected courses, grade results, transcript view, and executing course enrollments and some edit manipulations.
WebAbis (Academic Information Service) carries out demonstrating course information, activitiy workload definitions, scores and attendance lists, and executing relative evaluations and score entries of courses for academic
staff.
WebPbis (Staff Information Service) carries out demonstrating salary details, additional payments, charge lists,
telephone bills, personal specific preferences and pdf print choices for all staff.
WebMenu (Monthly Meal Information Service) carries out demonstrating normal and diet meal lists, meal calorie
values and ingredients of meals.
WebRehber (Telephone and E-mail Search Service) carries out searching with name, surname, telephone,
username, and department in electronic staff guidebook.
WebForm (Form Transfer Service) carries out holding satisfactions, critiques and suggestions about whatever
university manners.
WebAnket (Survey Service) carries out survey creation, management and reportings.
177
References
1. Zabir, O.A., Building a Web 2.0 Portal with ASP.Net 3.5, O'Reilly Media; 1 edition, ISBN-13: 9780596510503, January 11, 2008.
2. http://en.wikipedia.org/wiki/Web_portal
3. C., Young, Web Services Based Architecture in Computational Web Portals, Doctorate Dissertation,
Syracuse University December-2003.
4. B., Detlor, The Corporate Portal as Information Infrastructure: Towards a Framework for Portal Design.
International Journal of Information Management, Year 2000, Vol. 20 (2): 91-101.
5. D., Barnick, D., Smith, G., Phifer, Q&A Trends in Internet and Enterprise Portals. Gartnet Group.
Bartels, A., 1999. What's a portal and Why It Matters Giga Information Group.
6. Milutinovic, V., Infrastructure for Electronic Business on the Internet. 2001.
178
7. Kashyap, V., A. Sheth, Information Brokering Across Heterogeneous Digital Data, 2000.
8. http://www.dkms.com/ html.
9. Fritz, M., Kreuder, A.C., Schiefer, G. (eds.) 2001. Information Portals and Information Agents for Sector and Chain Information Services. Report A-01/4. University of Bonn-ILB.
10. Schiefer, G., Helbig, R., Rickert, U. (eds.) 1999. Perspectives of Modern Information and Communication Systems in Agriculture, Food Production and Environmental Control. Proceedings of the 2nd European Conference of the European Federation for Information Technology in Agriculture, Food and the
Environment. University of Bonn-ILB, Bonn, ISBN 3-932887-07-7.
11. Boeve, A.D. 1999. Integrated Veal Information. In: Schiefer et al. (1999): 835-844.
12. http://www.apache.org/
13. http://php.net/
14. http://www.asp.net/
15. http://www.adobe.com/products/coldfusion
16. http://java.sun.com/products/jsp/
17. http://www.mysql.com/
18. http://www.openssl.org/
19. http://en.wikipedia.org/wiki/MD5
Biographies
zkan CANAY He was born 1977 in Sakarya-TURKEY. He has graduated from Sakarya University Industrial Engineering department in 2002. He has received master science degree-2005 from Sakarya University Industrial Engineering department. He has already been as PhD student at Computer and Information Systems Engineering department in Sakarya University since 2005. His interests are basically in internet and web programming, database management systems design, data
mining, artificial intelligence, and linux based application development.
He is currently working as a manager at Computer Sciences Research and Application Center in Sakarya University.
Msc. Eng. Canay, currently is an executive board member of Computer Sciences Research and Application Center, a
member of University Information Technologies Commission, an advisory board member of University, an executive board
member of Information Technology Association of Sakarya.
Selim MER He was born 1978 in stanbul-TURKEY. He has graduated from Sakarya University Electrical-Electronics
Engineering department in 2003. He has already been as master science student at Electrical-Electronics Engineering department in Sakarya University since 2009. His interests are basically in pic and pcd programming, electronics circuits, internet
and web programming, database management systems design.
He is currently working as an electronics engineer at a research and development company.
Bs. Eng. Meri, currently is a member of Electrical Engineers Chamber in Turkey.
Hayrettin EVRGEN He was born 1956 in Tekirda-TURKEY. He has graduated from Seluk University Physics Engineering department in 1986. After getting master science degree from Sakarya University-TURKEY in 1997. He has studied
as researcher at Manchester University in United Kingdom-1999. He gained his PhD degree at Computer Engineering in
Sakarya University in 2002. His interests are basically in computer and educational technologies, programming, database
management systems design, project management.
He is currently working as a manager at distance learning based Vocational School of Adapazari in Sakarya University.
Assist.Prof. Evirgen, currently is an executive board member of Computer Sciences Research and Application Center, a
member of University Information Technologies Commission, an advisory board member of University, and honorary president of Information Technology Association of Sakarya.
Metin VARAN He was born 1981 in Bingl-TURKEY. He has graduated from Sakarya University ElectricalElectronics Engineering department in 2006. He has received master science degree-2008 from Sakarya University ElectricalElectronics Engineering. He has already been as PhD student at Electrical Engineering Department in Sakarya University
since 2009. His interests are basically in Electric Power System Analysis, Application of Chaos Theory in Electric Engineering, Voltage Stability Analysis in Power Systems, Embedded System Programming, API Programming, and Object Oriented
Programming.
He is currently working as a software developer specialist at Computer Sciences Research and Application Center in Sakarya University.
Msc. Eng. Varan, currently is a member of Electrical Engineers Chamber in Turkey.
179
1 Introduction
Today, computer and electronics take more space in our lives and take the mission of man power in many
tasks which are impossible, dangerous or time consuming for human being to perform. Previously, it could not
pass beyond imagination for using computerized electronic devices instead of human being to collect data physically from a dangerous environment. This is because the devices and mechanisms to perform these operations
were not previously available. By the time they were available, they had very big costs. However, by the developments in circuits, signal processing, other some sub electronics and wireless data communication techniques,
procedures appeared to be difficult and costly have become extremely easy and low cost to perform.
Wireless Sensor Networks have been taking the mission of human being in many areas such as performing
control functions in offices, fabrics; early diagnosis of faults in ships, trains, cars and airplanes; forest fires prevention with the help of early diagnosis; realization of efficient agricultural watering systems and instant notification of attacks at places which are critical in terms of security [1, 2, 3, 4].
One of the biggest advantages of Wireless Sensor Networks is to have the ability to be able to be selforganized, in other words, is that the ad-hoc structure. However, in order to have those capabilities such as without being tied to any structure and being able to be self-organized, appropriate communication protocols and
techniques need to be improved. Although it is conceivable to use the traditional techniques used for Wireless
Ad-hoc Networks also for Wireless Sensor Networks, in fact, the development of techniques specific to Wireless
180
Sensor Networks is more suitable. To better understand the reasons for this, let's look at the differences between
Wireless Sensor Networks and Wireless Ad-hoc networks [5]:
The number of nodes working in Wireless Sensor Networks is much higher than the number of nodes
working in Ad-hoc.
In Wireless Sensor Networks, communication is performed based on the broadcast logic in which within
the coverage area of a node all others can obtain the data. However, in Ad-hoc Networks data transmission mostly occurs in a point-to-point manner between two nodes.
Wireless sensor nodes are more limited in terms of energy, data processing capacity and memory compared to others.
Since the number of nodes in Wireless Sensor Networks is too large, it is not possible to assign each
node an individual ID number. The reason for that is large amount of nodes, large ID space. Large ID
space means larger ID numbers which bring additional burden in terms of energy consumption and delay
[5].
Small size of the sensor nodes has given the opportunity of cheaply utilizing them in many areas. However,
another significant problem emerged is that sensor nodes have limited resources. The biggest hardship brought by
the limited power supply is this: there are situations where it is impossible to replace the energy exhausted nodes
with the new ones. According to the application's purpose, Wireless Sensor Networks is expected to take the life
of many months and sometimes years. For these reasons, the researchers have been continuing their work on
communication techniques that lead to less energy consumption.
It is observed that, the amount of energy expended during transmission of a single bit between two wireless
sensor nodes is equal to the amount of energy spent by a sensor node performing a thousand transactions [6].
Another important leading energy consuming unit is the sensing unit that is used for gathering physical data from
environment. In many cases, depending on the sensor type used, the amount of energy spent by that unit varies,
however it is still negligible relative to the energy spent during data transmission [7].
In this paper, we propose an application-specific complete architecture by including other researcherss successful methods blended with our new techniques. Other methods such as LEACH [8, 9], the brainchild of clustering idea and its followers HEED [10] and PEGASIS [11] dealt only with intra-cluster communication. However, besides the intra-cluster communication, another and more energy consuming challenge to be dealt is the
inter-cluster communication. We propose here a TDMA-based, localized, energy aware next-hop selection idea
which provides fair load balancing and fair order of next hop selection between adjacent clusters resided in the
same layer. As it is discussed in [12, 13], employing multiple sinks in wireless sensor networks provides significant gains in terms of fair load distribution and prolonging lifetime of the nodes located in the Hot-spot region.
In addition, multi-layer approach is applied in corporation with multiple sinks idea. With the aim of alleviating
the burden of the nodes located in the so-called hot-spot region, the network structure is divided into multiple
nested layers. Each node in the number of clusters created in a layer varies according to the layer the cluster located. Since, the workload of nearby nodes is higher than the farther ones, clusters located nearby the sinks contain more nodes regarding to the ones located in the inner layers. Multi-channel communication, data aggregation, in-node processing are other energy conservative methods employed in our architecture in order to improve
energy conservation efficiency.
2 Related Work
Low-Energy Adaptive Clustering Hierarchy (LEACH) [8, 9] is an energy-efficient clustering based protocol
that has created a base for many studies. In LEACH, the topology is divided into clusters as it is depicted in Figure 1. Each sensor node belongs to a cluster with a cluster head (CH) selected randomly at the beginning of each
round.
181
182
Data
0
(1)
Next layer in the coverage area with the minimum cost value is selected as the next hop. There appears another challenge that has to be dealt while selecting the next hop. The CHs of the two clusters that are adjacent in
the same layer can select the same next CH at the same time. Therefore, although it is possible to select two distinct CHs as next hop, without applying any control mechanism, load balance and fairness may not be achieved.
In our project, a token mechanism is included in order to keep the next hop selection crash under control. For the
first round, the node with the smallest ID number is given the opportunity to select firstly the next hop CH. After
calculating and defining next hop CH, sending node writes the id of the selected CH in the token and passes to the
adjacent CH. Adjacent CH gets the token and starts to calculate and define a next hop CH with the ID that does
not appear in the list stored in the token. So, while selecting a next hop, the criteria is finding with the minimum
cost value that is not selected by the other CHs in the same layer until the token arrives. However, there emerges
another issue to be considered. That is, if every round the same CH is given the opportunity to define next hop,
how can fairness is achieved? The solution we suggest here is giving the next hop selection first opportunity to
the CHs in a round robin manner. That means, when a node is the first for the present round, it is the last for the
next one. Token structure is depicted in Figure 4.
ClusterID1 NextClusterID1
ClusterID2 NextClusterID2
ClusterIDn NextClusterIDn
Fig. 1. Token Structure.
183
ResidualEnergyTable is stored in each node and contains the energy level of the nodes recently. Hence, when
a node starts to calculate the next hop, firstly checks the ResidualEnergyTable and calculates the CostFactor parameter for each record. While selecting the one with the minimum cost value, the aim is selecting the one thats
ID is not present in the token for the present round.
For intra-cell communication, every cluster uses the frequency assigned at the setup stage and it is fixed. Like
cellular technology, frequency reusing approach is applied. CH in a layer transmits its intra-cluster data to the
next layer over the frequency used for intra-cluster communication. Thus, next hop CH assigns its receiving radio
to the frequency of the sending CHs frequency. So, every node in the topology has two radios. One is used for
intra-cluster and inter-cluster sending operations. Other one is used for receiving packets incoming from inner
layers.
Inter-cluster packets differ from each other according to the emerging layer number. That is, a packet originated in a cluster located at layer 1 do not have the same structure and size with the packet originated in a cluster
located at second layer. Inter-cluster packets have the structure as follows:
3.ClsID
SrcClsID
Data1.2
3.ClsCHID
SrcClsCHID
Data1.1
2.ClsID
SrcID1
Data2.1
2.ClsCHID
SrcID2
Data2.2
3.ClsCHID
SrcID2
Data2.1
SrcClsID
Data1.2
Data2.2
SrcClsCHID
Data1.1
SrcClsCHID
Data1.1
SrcID1
Data2.1
SrcID2
Data2.2
184
4 Performance Evaluation
Simulations performed for 100 rounds each round taking 18500 s. For the first step, our next hop selection
criteria, cost factors performance regarding to two other next hop selection criteria is discussed. CHs forward
their data towards the closest sink in the network in order to spend minimum energy. Hence, next hops have to be
selected between the nodes on the way to the closest sink. One of the next hop selection criteria we compared
with is the selection of the next cluster head that is closest to the sink. Second method is the selection of the next
CH which is closest to the sender CH. As it is clear in Figure 8 and 9, other two methods approximately show the
same performance. However, our cost factor criteria prolongs the lifetime of the network as it is depicted in Fig.
9. After 60 rounds, there is not any alive node in the network when NextHop-Cls-To-Sink and NextHop-Cls-ToSender routing criteria are applied. With the NextHop-WithSmallestCostFactor method we propose, there are still
alive nodes even after 70 rounds.
160
NextHop-Cls-To-Sink
140
NextHop-Cls-To-SENDER
120
NextHopWithSmallestCostFactor
100
80
60
40
20
0
10
20
30
40
50
60
70
80
90
100
Round
Fig. 5. Comparison of three different routing schemes according to the number of alive nodes remain in the network after
various number of rounds.
20
NextHop-Cls-To-Sink
NextHop-Cls-To-SENDER
Lifetime
15
NextHopWithSmallestCostFactor
10
0
Routing1 Methods
Lifetime of the network ends with the energy depletion of the first node in the network. Since nodes consume less
energy with the routing scheme SmallestCostFactor, lifetime of the network is longer than the networks using
other routing schemes as it is presented in Fig. 9.
The amount of energy dissipated by the nodes during data transmission is proportional with the distance between the sender and the receiver. Hence, in our architecture, nodes intend to direct their data to the closest sink
between the sinks surrounding the topology. Furthermore, instead of a CH sending directly to the sink, relaying to
a CH located on the direction to that sink is less energy consuming. If CHs send the data aggregated in their clusters directly to the sink, after 30 rounds all of the nodes in the network deplete of energy as shown in Fig. 10.
However, with multi-hop transmission, CHs relay their aggregated data to a next hop CH located at the outer
layer. Thus, they dissipate less energy and lifetime of the network prolongs inevitably that can be seen clearly in
Fig. 10 and 11.
185
140
100
80
60
40
20
0
10
20
30
40
50
60
Round
70
80
90
100
MultiHop Transmission
4000
DirectTransmissionFromCHsToSinks
3500
3000
2500
2000
1500
1000
500
0
10 20 30
40 50 60 70
80 90 100
Round
20
Lifetime
DirectTransmissionFromCHsToSinks
120
15
10
0
Transm itting
1 Techniques
186
5 Conclusion
In this paper, we presented an energy-efficient cluster based architecture containing various energy-efficient
techniques in order to decrease energy dissipation of the nodes thereby prolonging the network lifetime. Clusterbased structures are before proved to be more energy efficient with respect to others. In our simulations we observed this reality with brief measurements. As it is discussed in previous sections, hot-spot problem emerges in
wireless sensor networks because of the reason that sensor nodes located near the sink have to relay other sensor
nodes data that are far from the sink and do not have a possibility to transmit data directly to the sink. Thus,
multi-layer structure is employed with clustering mechanism. In this approach, clusters located near the sinks are
sized larger and contain more nodes than the ones far from the sink. That is because, nodes in the nearby clusters
transmit more data and being cluster-head turn must come less frequently in order to prevent energy depletion.
Another method, improving the energy efficiency is the multi-sinks idea. By setting a network to a single sink
located at one side of the topology, all data traffic from the nodes flow over the nodes located at the same side
with the sink. Hence, those nodes quickly deplete of energy and network lifetime reduces. However, by employing multiple sinks with surrounding the all topology, instead of relaying the data to the sink located at the other
side of the topology, CHs can forward their data to the sink as close as possible.
Furthermore, another challenge discussed in this study is finding a cost factor to be applied during the next
hop selection process. Cost Factor proposed in our architecture helps nodes to select the optimum next hop nodes
by means of the residual energy of the next node and the distance between the sender and the receiver. By employing cost-factor-considering-routing method, network lifetime prolongs with a factor of 50%.
By using multi-channel communication collisions are prevented thereby, undesirable energy consumptions
caused by retransmissions of packets are precluded. Besides, after physical data collection form environment and
transmission of that data to the CHs, it is inessential for the plain nodes to stay awake until the next round start
up. Thus, a periodic sleep-wake-up schedule is employed with the aim of redundant energy consumption.
In conclusion, none of the methods described above is alone enough for minimizing the energy consumption
in the network. Therefore, the smartest thing is to use the above-mentioned methods blended together as we did in
our study.
References
1.
2.
3.
4.
5.
6.
7.
8.
9.
10.
11.
12.
13.
14.
15.
Sohraby, K.; Minoli, D.; Znati, T. (2007). Wireless Sensor Networks, Wiley.
Seminar, University of Massachusetts. USA. (2003).
Whitehouse, K.; Sharp, C.; Brewer, E.; Culler, D. (2004). Hood: A Neighborhood Abstraction for Sensor Networks.
In Proceedings of the ACM International Conference on Mobile Systems, Applications, and Services (MobiSys04).
Whitehouse, K.; Karlof, C.; Culler, D. (2004). Getting Ad-Hoc Signal Strength Localization to Work. Technical
Report, University of California-Berkeley.
Akyildiz, F.; Su, W.; Sankarasubramaniam, Y.; Cayirci, E. (2002). A survey on Sensor Networks, IEEE Communications Magazine.
Pottie, G.; Kaiser, W. (2000). Wireless integrated network sensors, Communication of ACM.
Anastasi, G.; Conti, M.; Francesco, M. D.; Passarella, A. (2009). Energy Conservation in Wireless Sensor Networks: A survey. Elsevier Ad Hoc Networks.
Heinzelman, W.; Chandrakasan, A.; and Balakrishnan, H. (2000). Energy-Efficient Communication Protocol for
Wireless Microsensor Networks. In Proceedings of the Hawaii International Conference on System Sciences. Island of Maui, Hawaii.
Heinzelman, W.; Chandrakasan, A.; and Balakrishnan, H. (2002). An Application-Specific Protocol Architecture
for Wireless Microsensor Networks. IEEE Transactions on Wireless Communications 1(4): 660-670.
Younis, O.; Fahmy, S. (2004). HEED: a hybrid, energy-efficient, distributed clustering approach for ad hoc sensor
networks. IEEE Transactions on Mobile Computing 3(4): 366-379.
Lindsey, S.; Raghavendra, C. S. (2002). PEGASIS: Power-Efficient Gathering in Sensor Information Systems. In
Proceedings of the IEEE Aerospace Conference.
Cevik, T.; Zaim, A. H.; Yiltas, D. (To Appear). Localized Power Aware Routing with an Energy Efficient Pipelined
Wakeup Schedule for Wireless Sensor Networks, Journal of Electrical Engineering & Computer Sciences.
Yu, W.; Wei, S. S.; and Shie, B. J. (2009). Activating the Hot Spot Problem by Switching Multiple Sinks in Wireless Sensor Networks, In Proceedings of the Fifth International Conference on Mobile Ad-hoc and Sensor Networks.
Hanzalek, Z.; Jurcik, P. (2010). Energy Efficient Scheduling for Cluster-Tree Wireless Sensor Networks With
Time-Bounded Data Flows: Application to IEEE 802.15.4/ZigBee. IEEE Transactions on Industrial Informatics.
IEEE P802.15 Wireless Personal Area Networks: Proposal for Factory Automation, Working Draft Proposed
Standard. (2009).
187
Abstract. Recently, Mobile Ad-hoc Networks (MANETs) have drawn of many researchers.
One of the
main subjects accepted by researchers studying on Mobile Ad-hoc Networks is developing routing protocols
for wireless systems. Routing protocol development is related with dealing such as complexity, scalability,
adaptability, productivity and battery life in wireless systems. Routing protocols based on wireless systems are
developed in order to cope with these problems. In this paper, AODV, DSDV and Beeadhoc algorithms are
empirically compared to investigate large-scale behavior. The results are presented as graphs and evaluated.
Keywords: Mobile Ad-Hoc Networks; Routing Protocols; Wireless; Swarm Intelligence; Simulation
1 Introduction
In recent years, the development of communication technology has made wireless device smaller, less expensive and more powerful. Such rapid technology advance has contributed great growth to mobile devices connected to the Internet [1]. Nodes communicate each other without any center in Mobile Ad Hoc Networks (MANETs) and have a constantly changing network topology. Routing protocols are used to find the paths between
the nodes. Due to the constantly changing network topology, Mobile Ad Hoc Networks use different routing
protocols but wireless networks use. Constantly changing network topology leads to updating of routing tables
and thus increases the number of control packets. In Mobile Ad-Hoc Networks it is the most important point is to find
routes between source and destination nodes. In this process traffic of network should not be increased and should lose lesser
packet and delay. Wireless Mobile Ad Hoc Networks used to provide communication existing infrastructure networks does
not work because of war and natural disasters such as earthquakes and flood.
Fig.1 [1]
Fig.2 [1]
In infrastructure wireless network, there is a base station or an access point to be the portal of wireless devices. Ad-hoc network [2, 3, 4] is a self-organized, dynamically changing multi-hop network. All mobile nodes in
188
an ad-hoc network are able to communicating with each other without any established centralized controller. The
mobility of wireless nodes will cause the change of network topology.
In this paper, AODV, DSDV and Beeadhoc algorithms are empirically compared to investigate large-scale
behavior. The results are presented as graphs and evaluated.
2 MANETs Routing Protokocols
Many routing protocols have been proposed for ad-hoc Networks [1, 5, 6, 7, 8, 9, 11, 12]. These routing
protocols can be divided basically into two types, table-driven and on-demand routing protocol. Table driven
routing protocols each node has one or more tables to include routing information for all nodes in the network. A
change in the network updates routing tables of all t nodes. On-Demand the routing protocols, Unlike the TableDriven driven routing protocols there is no requirement to keep continuously the routing tables of the nodes.
Instead of this, by running variety of mechanism routs are created when it is necessary. The generated paths are
kept until packet reaches the destination node or certain period of time.
2.1 On-Demand Distance Vector Routing Protocol
Recently, the on-demand routing protocols for ad-hoc network are demanded because of their low routing
overheads and effectiveness in the case of the frequency of route re-establishment and the demand of route queries are not high. In the on-demand routing protocols the route is created only when is desired by the source node.
Many on-demand routing protocols have been suggested in [8, 11, 7, 10]. High routing overheads usually have
important impacts on performance in low-bandwidth wireless links. Therefore, the reactive on-demand routing
algorithms where routing paths are established only when desired are the recent liability in ad-hoc networks, such
as the Ad-hoc On-Demand Distance-Vector (AODV) routing protocol. In the AODV protocol, there is only a
single path established during the transmission of the packet. Therefore, when the transmission route fails, data
packets are simply dropped by nodes along the broken route.
2.2. Destination Sequenced Distance Vector Routing Protocol
Destination-Sequenced Distance-Vector routing (DSDV) [8], tries to keep a global picture of network topology and responds to topological changes by dispersing update messages along the wireless network. One or
more tables are required to maintain continuous, up-to-date routing information for each node in the wireless
network. In a highly mobility network environment, to maintain the routing information causes heavy overheads.
2.3. Beeadhoc Routing Protocol
Beeadhoc[14] algorithm has been taken into consideration in the comparison as an energy efficient routing
algorithm which is inspired from foraging principles of honey bees. The bee behavior was instrumental in designing efficient mobile agents, scouts and foragers, for routing in Mobile Ad-Hoc Networks. Scouts discover new
routes from their launching node to their destination node. A scout is transmitted using the broadcasting principle
to all the neighbors of a node with an expanding time to live (TTL) timer. Foragers are the main workers in our
Beeadhoc algorithm. They receive the data packets from transport layer and then deliver them to the destination.
Foragers are of two types: delay or lifetime. The delay foragers collect the delay information from the network
while the lifetime foragers collect the remaining battery capacity of the nodes that they visit. Bees usually are
forced to go long distances to find food. The areas of foraging bee finds a source of food for notice to the other
members of the colony and return to the hive after a while they start to fly around the other bees. Honey bees are
deaf, and therefore will not be able voice communications with each other. Establish communications with each
other with different shapes. These shapes are called dance. This dance has source distance, direction, information
about the quality and quantity of food available [13].
189
Topography
Value
Area (m2)
Number of
nodes
10
50
100
200
Mobility (m/s)
300 x 300
500 x 500
700 x 700
1500 x 1500
1 ~ 20
Topology genera-
512
SETDEST
Traffic generator
Traffic
Simulation time (s)
Computer
Operating system
CBRGEN
FTP/TCP
50
Core 2 Duo, 2 GB RAM
Fedora
tor
190
(a)
(b)
(c)
(d)
Fig.3 Effect of varying speed and number of the nodes.
Fig 3 shows that network throughput values are different in four different sizes and at different speeds.
While AODV and DSDV almost the same efficiency in all velocity values, productivity of Beeadhoc protocol
remains low. In addition, at the speed of 10 m / s, the DSDV protocol is running more efficiently. Fig. 3 b, c and
d show that the scale is running better performance scalability of AODV. Especially 200 nodes and 10 m / s
speeds, protocols, virtually no throughput other than AODV does. This result shows that, the scalability of
AODV is more efficient than DSDV and Beeadhoc routing protocols are.
191
4. Conclusion
The purpose of this study is to investigate the effects of the efficiency of protocols by changing the speed
and the number of nodes. The network topology changes more quickly in high speeds. Thus, it is difficult rapid
adaptation of the protocols to new situation. The network protocols did not work in the largest and the most high
speed scenario. Increasing the number of nodes has been decreased the throughput of the network .
References
1.
2.
3.
4.
5.
6.
7.
8.
9.
10.
11.
12.
13.
Wei K.Lai, S.Hsiao, Y.Lin, Adaptive backup routing for ad-hoc networks, 2006.
Internet Engineering Task Force (IETF) Mobile Ad-Hoc Networks (MANET) Working Group Charter, Chaired by
Joseph Macker and M. Scott Corson, http://www.ietf.org/html.charters/manet-charter.html.
J. Jubin, J.D. Tornow, The DARPA packet radio network protocols, Proceedings of the IEEE 75 (1) (1987) 21
32.
C.E. Perkins, Ad Hoc Networking, Addison Wesley (2001).and RSSI information, IEICE E88-B (9) (2005) 3588
3597.
S. Corson, J. Macker, Mobile Ad-Hoc Networking (MANET): routing protocol performance issues and evaluation
considerations, RFC. 2501 (1999).
D.B. Johnson, D.A. Maltz, Dynamic source routing in ad hoc wireless networks, in: Mobile Computing, Kluwer
Academic Publishers, 1996, pp. 153181.
C.E. Perkins, E. Royer, Ad hoc on-demand distance vector routing, Proceedings of IEEE WMCSA (1999).
C.E. Perkins, P. Bhagwat, Highly dynamic destination-sequenced distance-vector routing (DSDV) for mobile computers, Proceedings of the ACM SIGCOMM. (1994) 234244.
E.M. Royer, C.-K. Toh, A review of current routing protocols for ad hoc mobile networks, IEEE Personal Communications 6 (2) (1999) 4655.
R. Dube, C.D. Rais, K. Wang, S.K. Tripathi, Signal stability based adaptive routing (SSR alt SSA) for ad hoc mobile networks, IEEE Personal Communication (1997).
C.E. Perkins, Royer, and S. Das, Ad Hoc On-Demand Distance Vector (AODV) Routing, Internet Draft, draftietf-manet-aodv- 13.txt, February 2003.
H. F. Wedde and M. Farooq, The wisdom of the hive applied to Mobile Ad Hoc Networks, University of
Dortmund, 2005.
K. von Frisch. The Dance Language and Orientation of Bees. Harvard University Press, Cambridge, 1967.
Biographies
Zafer Albayrak is received MSc degree from the Department of Computer Science at University of North London, London, UK. He is a PhD student in the Electronics and Computer Science Department, University of Sakarya, Sakarya, Turkey.
He is also a Lecturer at Gaziosmanpasa University, Tokat, Turkey. His main interest in wireless networks.
Ahmet Zengin is PhD at Sakarya University, Turkey. His experience with modeling and simulation includes a one-yearstay in ACIMS Lab at the Arizona State University. His research topics include DEVS theory, multi-formalism modeling,
parallel and distributed simulation, modeling and simulation of large scale networks, distributed systems management,
biologically-inspired optimization schemes. His main research interest lies in parallel and distributed simulation and the High
Level Architecture.
Mehmet Recep Bozkurt is an Associate Professor in the Electronics and Computer Science Department, University of
Sakarya, Turkey. His research topics include Electronic Circuit Design, Electronic Devices, Pre-processing and classification
of Biomedical signals, Artificial Intelligence and MatLab.
192
Abstract. Fault tolerance is a necessary part of software with high reliability. As software systems
evolve, fault tolerance is more substantial. Nowadays, many software systems are implemented with
high level programming languages. One of these programming languages is Java. Java technology enables researchers to develop for many huge software systems. However, this may require interoperability
of Java and native language such as C/C++. In Java, we need Java Native Interface (JNI) calls to operate hardware and running mathematical intensive operation, which requires performance needs, written
in native code. These parts of software need utmost care about preventing faults and are the main source
of the critical errors in projects. In traditional exception handling in Java we cannot catch those major
errors caused by memory leaks and low-level hardware faults. Those faults in native code can crash
whole Java Virtual Machine (JVM). In order to prevent these failures we need a fault-tolerant system to
catch those native code errors. Separation of operating system processes brings us the fault tolerance we
need by using the memory address space protection. In this paper we analyzed the details of this process
separation approach and proposed some architecture that isolates the faults using separate processes.
We also conduct experiments on a ballistic imaging system with one of the appropriate architectures.
Keywords: reliability, fault tolerance, Java, JNI
1 Introduction
Software reliability is the probability that software will not cause the failure of a product for a specified time
under specified conditions [1]. It is also an important factor affecting system reliability. Software fault tolerance
is a necessary part of a system with high reliability. It is a way of handling unknown and unpredictable software
system failures [2]. With the development of the complex software systems, fault tolerance is unavoidably significant. Therefore, researchers have intensively studied to perform fault tolerant software and systems. Golden G.
Richard III and Tu [3] presented their experiences with Java as a portable solution for developing completely
fault tolerant programs. They also argued Java based fault tolerant programming patterns that allow development
of fault tolerant Java solutions for a large class of sequential and distributed applications. Xin et al. [4] designed
an object oriented developing framework of fault tolerance system using Java to solve some problems. They used
JRMI technique as the communication mechanism and also used Object Oriented Design method to improve
software usability. Pleisch and Schiper [5] presented Java-based fault tolerant mobile agent system called
FATOMAS. Liang et al. [6] provided a fault tolerance mechanism based on Java exception handling, Java thread
status capturing technology and mobile agent to handle Grid fault in Linux platform. Kumar [7] proposed a fault
tolerant system design and implementation for CORBA that employs object replication. Kumar benefited from
Java serialization to capture objects state. Main objective for this study to advance availability of Java objects in
CORBA 2. Li and friends [8] proposed a fault tolerance sandbox to support the reconfigurable fault tolerant
mechanism on application servers to improve availability and reliability of reconfigurable applications. Lastly
193
Yang et al. [9] implemented a tool named McC++/Java for enabling monitoring and fault tolerance in C++/Java
programs on multi core applications.
Recent software projects often use different components that have been implemented in various programming
languages, so language interoperability gains great importance. In order to interoperate different software components in different languages, there must be a technology to make those languages to communicate with each other.
One technique to solve this issue is to use a foreign function interface [10]. Java also uses this technique and it is
called Java Native Interface (JNI). JNI is a powerful feature of Java platform that allows programmers to interoperate Java code and native code written in programming languages such as C, C++ and assembly 10. Although
JNI allows programmers to take advantages of Java platform, care should be taken in native coding. Java is a type
safe language and has advanced memory management features. On the other hand, memory management in
C/C++ is very hard [12]. Writing multithread programs in C/C++ is difficult and these can cause dead-locks that
crash JVM [12]. Also, hanging in hardware method calls results in unresponsive applications. Therefore, memory
leaks in native code and low-level hardware faults may cause system failures and crash of whole system. As a
result, all these faults must be handled by the software and do not let them to put the system in an unresponsive or
unstable state.
When Firearms are used, they mark on cartridge and bullet. Similar marks are created by the firearm on different cartridges and bullets that are thrown by same firearm [13, 14]. By examining these marks, firearms may be
identified. While examining of marks was performed using microscope early, this process has been computerized
with the development of ballistic imaging system [14], [19], [20]. Thus, fastest and more reliable data entry and
high performance matching process can be achieved. Ballistic imaging systems have software and hardware components. Hence, the system has potential to create faults. Therefore these systems must be fault tolerant to be able
to run for a long time.
Balistika 2010 is a ballistic imaging system that has been developed for automated firearm identification [18].
This system controls hardware to take images of cartridges and bullets. Then it produces data using these images.
This phase needs mathematical intensive operation that runs fast when written in native code.
In this paper, we have explained methodologies of fault tolerant software systems that interoperate Java and
C/C++ in a single computer. After that we have described our case and implemented selected method among
these methodologies for Balistika 2010. Eventually, experimental results are given.
2 Methodologies
There are some methods that interoperate Java and C/C++ programming languages in a single computer. If native methods are directly called from a main process with JNI as shown in figure 1, JVM of Main Process can be
crashed immediately because of the drawbacks of the native code.
Main Process
(Java)
JNI
Native Code
(C,C++)
In order to avoid this challenge and carry out a fault tolerant software system, address space protection approach can be applied [11]. Native code that can create JVM crash and Main Process could be separated as
shown in figure 2. Main Process forks a Child Process which handles all communication with the native code.
When native code crashes, main process can live.
Main Process
(Java)
Run
Child Process
(Java)
JNI
Native Code
(C,C++)
Fig. 2. Separating processes and calling native method from child process
June 1 -4, Kusadasi, Aydin, Turkey http://iscse2011.gediz.edu.tr
194
Although above approach relatively allows fault tolerant structure, some drawbacks can be created concerning
communication between two process. For example, data that are created in Child Process cannot be transferred to
Main Process easily. There are some solutions to cope with this trouble. One of the solutions is shown in figure
3. In this case child process and main process can communicate through hard disk drive (HDD). Communicating
through HDD is an easy and reliable way but reading and writing performance of HDD is clearly poor.
Main Process
(Java)
Run
ReadWrite
Child Process
(Java)
JNI
Native Code
(C,C++)
ReadWrite
HDD
Fig. 3. Calling native methods from child process and communication through HDD
Using Ram Disk as shown figure 4 can be proposed for data transfer between main and child process as an alternative to the HDD. This method that uses the block of RAM like HDD has better performance in reading and
writing data compared to the HDD. Ram disk can be created using methods of operating system or third party
programs. There are also points to take into consideration about using Ram Disk. Total memory of system decreases as amount of size that Ram disk uses.
Main Process
(Java)
Run
ReadWrite
Child Process
(Java)
JNI
Native Code
(C,C++)
ReadWrite
Ram Disk
Fig. 4. Calling native methods from child process and communication through ramdisk
Another solution for transferring data between two processes is to use inter process communication (IPC).
There are some IPC methods. Socket is one of the IPC methods and frequently used in Java [15]. Socket technology that is easy and reliable way for network communication has been used for many years. As shown in figure 5,
main process and child process communicate by the help of Socket connection. Socket communication has relatively better performance than HDD.
Main Process
(Java)
Run
Child Process
(Java)
JNI
Native Code
(C,C++)
Data
Socket
Socket
Fig. 5. Calling native methods from child process and communication through Java socket
The last solution for data transfer between main and child Process is to use other IPC methods [16] [17] as
shown in figure 6. These IPC mechanisms can be performed with collaboration of Java and C, C++ languages by
means of JNI. In this method, developers must be more careful because main process can call the native method
via JNI.
195
Main Process
(Java)
JNI
Run
Child Process
(Java)
IPC methods
-Shared Memory
-Pipe
-Semaphore
-Message Queue
JNI
Native Code
(C,C++)
JNI
Fig. 6. Calling native methods from child process and communication through other IPC methods
3 Our Case
Balistika 2010 ballistic imaging system is composed of hardware and software components as shown in figure
7. User controls CCD camera that captures 4MB images with 20 frames per second and communicates over
Ethernet. This brings 80MB data transfer every second between native code and Java. This system will be used
24 hours a day. Moreover, user controls table motors and LEDs that is linked with a control card and serial cable
to the computer by means of graphical user interface, while imaging cartridge and bullet. User also performs
mathematical intensive operation on images of cartridge and bullet.
CCD
Camera
Ethernet
cable
LED
Serial
cable
User
Control
Card
X,Y,Z
Tables
Our Ballistic imaging system has generally three parts that are controlled with calling native codes: Camera,
custom hardware and mathematical intensive operation. Camera is the main component of our ballistic imaging
system. We purchased camera with API and firmware. We do not know update frequency of the API or the firmware of the camera. In our case, last two firmware update resolves two important bugs which can cause segmentation faults and dead locks. This shows we must expect fault in our system and nothing guarantees that camera will
not bring any problem.
Another component of our ballistic imaging system is the custom hardware. Custom hardware has motor and
LED control drivers. APIs of these drivers were implemented with C++. It is very hard to make sure that those
components will not fail.
Last part of the imaging system that is controlled with calling native code is mathematical intensive operations.
Images that are captured by the camera are processed with these operations. These mathematical operations are
performed using pointers. Usually these are the main sources of failure.
As number of cartridges and bullets increase, Ballistic imaging system needs to be reliable, robust and fault
tolerance. That is this system must be long running system so user should be affected as little as possible because
of system faults mentioned above. For this reason, main process and child process that controls camera and custom hardware and performs mathematical intensive computations have been separated in our system as shown in
196
figure 8. As seen figure 8, child process was implemented in Java. Thus we minimized C, C++ codes in our system. Transferring of created data by the mathematical computation has been carried out using Ram disk. There
are several reasons to select this method. Firstly RAM is faster than HDD so we benefited from writing and reading speed. Second, both the child and the main processes can read the data at the same time with this approach. In
order to achieve this with socket we need to save the data in both processes memories which needs to allocate
twice memory space. We can achieve this synchronous reading if we use IPC mechanism of OS. But in this case
we have to use native code which can crash main process.
Ballistic Imaging System User Software
User
Recording
Software
(Main Process)
(Java)
Run
Native Code
(C,C++)
Camera
JNI
Custom
Hardware
Mathematical
Computation
ReadWrite
ReadWrite
Ram Disk
4 Experimental Results
Our ballistic imaging system was tried by two ballistic experts for 1504 cartridges. We logged about size of
6.5 MB log file. Number of calling of native method by Java with JNI on child process is shown in Table 1. As
shown in Table 1, during recording 1504 cartridges, 18 faults that caused crash of the child process occurred.
Also 23 faults that caused hanging of the camera occurred and we resolved the hanging problem by exiting from
child process. Figure 9 and 10 show scatter of JVM crash and camera hangs according to date respectively. Main
process was not affected negatively and maintains its execution. Thus ballistic experts could be continued recording cartridge without making any additional operation.
Table 3. Number of Calling of Native Method by Java Language with JNI on Child Process
Number
of
Calling
Native
Method
(Per
Cartridge)
Camera
Custom Hardware
Mathematical Computation
Total Number of
Calling Native Method (1504 Cartridges)
Number
Crash
of
40
60160
11
10528
1504
197
5 Conclusion
In this paper, we proposed a fault tolerance design using Java accessing native code in a single computer. Also, we have explained methodologies of fault tolerant software systems that interoperate Java and C/C++. Main
concept of our design is separating processes to provide address space protection. In our case, we used a child
Java process to interoperate with the native libraries and a software ram drive to overcome the communication
problem. We implemented the proposed design on our ballistic imaging system and conducted experiments. We
succeeded to encapsulate the crashes and hangs in the child process. Crashes and hangs caused by the native
codes do not lead to failure of whole system. Thus, our system is reliable and it continues to operate without any
data loss.
References
1.
2.
3.
4.
5.
6.
Krajcuskova, Z. (2007). Software Reliability Models. Radioelektronika International Conference. Brno, 24-25 April.
Pan,
J.
(1999).
Software
Reliability.
Carnegie
Mellon.
From
the
World
Wide
Web:
http://www.ece.cmu.edu/~koopman/des_s99/sw_reliability.
Richard III, G. G. and Tu, S. (1998). On Patterns for Practical Fault Tolerant Software in Java. Proceedings of IEEE
Reliable Distributed Systems Symposium.West Lafayette, IN 20-23 Oct.
Xin, Z., Daoxu, C. and Li, X. (1998). An object-oriented developing framework of fault-tolerant system. In Proc. of the
31st International Conference on Technologies of Object-Oriented Language and Systems.
Pleisch, S. and Schiper, A. (2001). FATOMAS: A Fault-Tolerant Mobile Agent System Based on the Agent-Dependent
Approach. Proc. Intl Conf. Dependable Systems and Networks.
Liang, J., WeiQin, T., JianQuan, T. and Bo, W. (2003). A Fault Tolerance Mechanism in Grid. Proceedings of Industrial
Informatics IEEE International Conference.
198
Kumar, A. (2008). A fault-tolerant system for Java/CORBA objects. Parallel and Distributed Processing IEEE International Symposium.
Li, J., Huang, G.,Chen, X., Chauvel, F., Mei, H.(2009). Supporting Reconfigurable Fault Tolerance on Application
Servers. Parallel and Distributed Processing with Applications, IEEE International Symposium.
Yang, L., Yu, L.,Tang, J., Wang, L., Zhao, J. and Li, X.(2010). McC++/Java: Enabling Multi-core Based Monitoring
and Fault Tolerance in C++/Java. Engineering of Complex Computer Systems (ICECCS) EEE International Conference.
Tan, G., Appel A. W., Chakradhar, S., Raghunathan, A., Ravi, S. and Wang, D. (2006). Safe Java native interface. In
Proc. 2006 IEEE International Symposium on Secure Software Engineering.
Liang, S. The Java Native Interface: Programmer's Guide and Specication. Addison-Wesley, 1999.
Hoff, A. V. (1997). The case for JAVA as a programming language
From the World Wide Web:
http://computer.org/internet/
Sakarya, U., Lelolu, U.M., Tunali, E. (2008). Three-dimensional surface reconstruction for cartridge cases using photometric stereo. Forensic Science International, 175 (2-3), pp. 209-217.
Huang, Z. and Leng, J. (2010). An online ballistics imaging system for firearm identification. Signal Processing Systems
(ICSPS), 2010 2nd International Conference
Java socket (2011). From the World Wide Web: http:// download.oracle.com/ javase/tutorial/ networking/sockets/
Immich, P.K., Bhagavatula, R.S. and Pendse, R. (2003). Performance analysis of five interprocess communication
mechanisms across UNIX operating systems. In Proceedings of Journal of Systems and Software. 27-43.
Inter process Communications (Windows) (2011). From the World Wide Web: http://msdn.microsoft.com/enus/library/aa365574(VS.85).aspx
Sakarya, U., Es, S., Leloglu, U., Tunali, E. and Birgul, O. (2010). A Case Study of an Automated Firearms Identification System: BALISTIKA 2010. 1st Annual World Congress of Forensics 2010.
Smith, C. L. (1997). Fireball: a forensic ballistics imaging system. 31st Annual 1997 International Carnahan Conference,
Canberra.
IBIS (Integrated Ballistics Identification System). From the World Wide Web: http://www.ftiibis.com/.
Biographies
Ebubekir Temizkan - He was born in 1982 in Kayseri. He earned his B.S. degree in 2007 from Maltepe University, Department of Computer Engineering.
He is interested in graphical user interface technologies on Java, usability and 3D face recognition and 3D reconstruction
from 2D images technologies.
Mr. Temizkan is currently working for The Scientific and Technological Research Council of Turkey Space Technologies
Research Institute (TBTAK - UZAY).
Kerem nal - He was born in 1978 in Ankara. He earned his B.S. degree in 2000 and MSc. in 2004 from METU, Department of Electrical and Electronics Engineering.
He is interested in HCI; especially on usability, usability evaluation, eye tracking, input devices, human centered design
and visualization.
Mr. nal is currently working for The Scientific and Technological Research Council of Turkey Space Technologies Research Institute (TBTAK - UZAY).
199
ilker.korkmaz@ieu.edu.tr, 2senem.kumova@ieu.edu.tr,
orhan.dagdeviren@izmir.edu.tr 4fatihtekbacak@iyte.edu.tr
Abstract. The cache configuration is an important concern affecting the performance of sorting algorithms. In this paper, we give a performance evaluation of cache tuned time efficient sorting algorithms. We
focus on the Level 1 and Level 2 cache performance of memory tuned versions of both merge sort and quicksort running on data sets having various sizes and probability distributions. The algorithms are simulated in
different cache configurations, including a Pentium 4 processor configuration, by using Valgrind simulator.
Besides, we propose a new merge sort algorithm design which aims to decrease run time by reducing Level 1
cache misses.
Keywords: merge sort, quicksort, cache effected sorting, simulation
1 Introduction
Sorting operation [1-13] is commonly used in mathematics and computer sciences as well. Sorting functions are
generally implemented as top-level definitions to be called whenever required in any software program or mathematical application. There are two main performance issues for a sorting algorithm: time and space complexity.
Although there is a theoretical trade-off between those two issues, practical sorting implementations try to balance the effects and provide efficient running time results. There may be some different hardware configurations
for the sorting codes to be executed in more performance-effective manner. In this concept, cache configuration is
an important concern affecting the performance of sorting algorithms.
In this paper, we introduce a comprehensive performance evaluation of cache tuned time efficient sorting algorithms. We cover the Level 1 and Level 2 cache performance of memory tuned versions of both merge sort [2]
and quicksort [8] running on data sets having various sizes and probability distributions. The algorithms are simulated in different cache configurations, including a Pentium 4 processor configuration [14]. The data and instruction miss counts on Level 1 and Level 2 caches are measured with the time consumption for each algorithm by
using Valgrind simulator [15]. Besides, we give the design of a new proposed merge sort algorithm which aims to
decrease run time by reducing Level 1 misses.
2 Related Work
Sorting algorithms effect run time and efficiency of operations in which they are used. If the algorithms are
qualified, it may result with a decreased running times. The traditional method which is accepted to be the most
200
effective way of qualifying algorithms to reduce run time is decreasing the number of instructions in the algorithm. Although reducing the instruction number of any algorithm has a positive effect on its running time, it is
not possible to state that it supplies enough improvement for sorting algorithms. The approach in which the reality
of cache usage in sorting algorithms is considered to reduce running time of sorting operations was stated firstly
by LaMarca and Ladner [9-11], restudied by Xiao et al. [12].
Basic merge sort algorithm which is designed as divide and conquer paradigm is a comparative sort algorithm
with O(nlogn) complexity. Multi merge sort and tiled merge sort algorithms take account of cache properties and
reduce running time by reducing the memory access rates or cache miss rates. Multi merge sort algorithm merges
all sub arrays in one operation. In tiled merge sort algorithm, the data is divided in two sub arrays and the arrays
are sorted between each other. Conceptually based on those algorithms, Xiao et al. [12] proposed their padding
versions. Tiled merge sort with padding organizes data locations to decrease the collision miss rate. Multi merge
sort with TLB padding aims to reduce TLB misses created by multi merge sort algorithm. As the initials of cache
effected quicksort algorithms, LaMarca and Ladner [9-11] proposed memory tuned quicksort and multi quicksort.
To further improve the quicksort performance, Xiao et al. [12] examined flash quicksort and inplaced flash quicksort, which are basically some kinds of integration of flash sort and quicksort algorithms. For the procedural details of these cache effected merge sort and quicksort algorithms, related studies [9-12] can be dissected.
In a general cache configuration design, main properties to decide on are basically as follows: capacity, line
size, associativity, multi-level design. Considering those configuration issues, cache effected algorithms can be
simulated by Valgrind [15] and Cachegrind which is a Valgrind tool that detects miss rates by simulating Level 1
and Level 2 caches.
3 Experiments
3.1 Approach
To investigate the cache performances and effect of cache usage on merge sort and quicksort derivations, the
following algorithms are examined in in Valgrind simulator: base merge sort, tiled merge sort, multi merge sort,
and tiled merge sort with padding; base flash sort, flash quicksort, inplaced flash quicksort, and memory tuned
quicksort. We simulated the algorithms in 2-level cache with different configurations.
As a top-view prescription of our approach, we have 3 phases: cache configuration, simulation of the corresponding sorting algorithms within the selected configuration with different data set distributions, simulation of
the algorithms with Pentium 4 cache configuration.
Each different experiment is repeated 5 times and the average values are used to evaluate the corresponding algorithm. The data miss rates of Level 1 and Level 2 caches of each algorithm are used to observe the cache performances. The running times of the implementations are also measured. It is expected that although all merge
sort implementations have the same time complexity of O(nlogn), the algorithms, which have more misses in the
last level of the cache and have more accesses to the main memory, result with higher values of running times. It
is also expected that flash sort supported quicksort algorithms run faster than merge sort variations.
201
Base merge sort, tiled merge sort and multi merge sort algorithms are utilized on the selected cache configuration. Table 1 gives L1 and L2 miss rates and sorting run times (in microseconds) per element for random data sets
with N=1024, 10240, 102400, 1024000. L1 misses in Table 1 include both instruction and data misses.
In Table 1, it is shown that miss rates are lower in tiled merge sort and multi merge sort compared to base
merge sort as expected. Multi merge sort algorithm gives the lowest miss rate in randomly distributed data set. In
order to examine the effect of distribution of data sets according to possible miss rates, the two best of three algorithms in Table 1, tiled merge sort and multi merge sort algorithms, are applied to randomly, equilikely, poisson
and zero distributed data sets. Zero distribution refers the distribution in which all data values are equal to 0 (zero) and equilikely distribution is the uniform distribution. Since there is not enough space in this paper to present
the detailed measurement results, only a concise summary of the results is given as follows: Multi merge sort
gives better run time results only in poisson distribution and for all other distributions tiled merge sort generates
better run times.
multi
tiled
base
merge sort merge sort merge sort
Table 1. The results of several merge sort algorithms on selected cache configuration
Average Miss Rate
and Run Time
0.08
0.27
L2 miss rate
L1 miss rate
1024
0.1
0.27
10240
0
0.2
102400
0.1
0.3
1024000
0.1
0.3
Run Time
L2 miss rate
L1 miss rate
31.22
0.1
0.27
27.84
0
0.1
34.05
0
0.1
38.96
0
0.1
33.02
0.03
0.14
Run Time
L2 miss rate
L1 miss rate
32,74
0.1
0.27
23,74
0
0.1
33.99
0
0
41.96
0
0.1
33.11
0.03
0.12
Run Time
32.28
28.95
46.32
65.32
43.22
Besides, we simulated the quicksort and flash sort variations to compare the cache effected quicksort algorithms. It is expected that flash sort supported quicksort algorithms run faster than merge sort variations. The
comparisons are evaluated regarding to cache miss rates and run times of the algorithms.
Flash sort, flash quicksort, inplaced flash quicksort and memory tuned quicksort algorithms are simulated on
randomly distributed data sets. Fig. 1 and Fig. 2 include the sorting time and miss rate results of the simulations
for several different sorting algorithms on randomly distributed different sized data sets, respectively.
Fig. 1. Sorting run times (in microseconds) per input element of various sorting algorithms
202
In Fig. 1, the vertical axis is the sorting time in microseconds per one input element; whereas in Fig. 2, the values of the vertical axis refer to miss rates for corresponding size of data sets. The horizontal axis is the corresponding total element size of input data set. As mentioned in the study of Xiao et al. [12], although flash supported quicksort algorithms produce high miss rates (shown in Fig. 2), their sorting running times are lower compared to other sort alternatives in large data sets (shown in Fig. 1). Flash sort, flash quicksort, and inplaced flash
quicksort algorithms have very close results. It is also shown in Fig. 2 that memory tuned quicksort algorithm
gives the lowest miss rates as expected.
A more realistic case. In the last experimentation step, a more realistic case, 2-level cache structure of Northwood core Pentium 4 processor architecture [14] is simulated. Level 1 cache is configured as 4-way set associative with a total capacity of 8 K, which has a line size of 64 Bytes; Level 2 cache is used as 8-way set associative
with a capacity of 256 K, and with the line size of 128 Bytes. Base merge sort, tiled merge sort, multi merge sort
and tiled merge sort with padding algorithms are simulated with several data sets of 1 K to 4 M. The detailed
simulation results can be found in related previous study of authors [13]. Among those merge sort alternatives,
multi merge sort algorithm offers better miss rate performances for data sizes less than 1 M; however, tiled merge
sort is more scalable for larger data sets; on the other side, the best run time performance is obtained by tiled
merge sort with padding algorithm.
203
j1=k;
jj2=j2+w;
j2=j*w;
jj3=j2+w;
while ((j1<jj2)&&(j2<jj3)){
if (source[j1]<=source[j2])
j1++;
else{
temp=source[j1];
source[j1]=source[j2];
source[j2]=temp;
j2++;
}
}
if (j1==jj2){
while (j2<jj3)
source[j1++]=source[j2++];
}
if (j2==jj3) j1=jj2;
}
}
if (target==a){
for (i=0; i<n; i++)
a[i]=b[i];
}
end
5 Conclusions
In this paper, we first provided a survey of cache effected sorting algorithms. Secondly, we provided simulation
results including basic merge sort algorithm, tiled merge sort algorithm, multi merge sort algorithm and tiled
merge sort with padding algorithm using Valgrind simulator. We also simulated those various merge sort algorithms on different input data set distributions as random, equilikely, poisson, and zero distributions. As a general
conclusion of those simulation experiments among merge sort variations, tiled merge sort performance seems
appropriate based on the average of sorting time and cache miss rate results.
Its known that basic quicksort wins against basic merge sort in general runnig time. To observe the performance differences between alternative quicksort and flash sort variations, we simulated the miss rates and running
times of flash sort, flash quicksort, inplaced flash quicksort, and memory tuned quicksort. Among them flash sort
has greater miss rates, however it leads better running times than quicksort,
We also provided the simulation results of various merge sort alternatives based on Pentium 4 configuration.
The instruction miss rate on L1 cache performance is nearly same for merge sort algorithms. Tiled merged algorithm performs best among other algorithms considering L1 data cache miss rate. Although multi merge sort algorithms performs best considering L2 data cache miss count, it has the worst performance when L1 data cache
count is considered thus its overall run time performance is worst among other merge sort algorithms. Overall,
tiled merge sort with padding algorithm performs best in run time. Overall findings in our simulations are consistent with the findings and evaluations pointed in the studies by LaMarca and Ladner [9-11], and Xiao et al.
[12].
We also proposed a new algorithm to further reduce the wall clock time and showed its design in this paper.
Our work is ongoing and we are planning to implement the proposed algorithm and compare with the other merge
sort algorithms.
Acknowledgment. The authors thank Prof. Dr. Mehmet Emin Dalkilic for his advises to focus them on cache
effected sorting area.
204
References
1.
2.
3.
4.
5.
6.
7.
8.
9.
10.
11.
12.
13.
14.
15.
Biographies
Ilker Korkmaz Ilker Korkmaz received the BSc. degree in Electrical and Electronics Engineering and MSc. degree in
Computer Science from Ege University. He is a PhD candidate at International Computer Institute at Ege University. He is
also an instructor at the Department of Computer Engineering at Izmir University of Economics. His research interests include computer networks, network security, and wireless sensor networks.
Senem Kumova Metin Senem Kumova Metin received her BSc. degree in Electrical and Electronics Engineering from Ege
University and MSc. and PhD degrees from International Computer Institute at Ege University. She is currently working as a
lecturer in Izmir University of Economics. She is mainly interested in natural language processing applications and computational linguistics.
Orhan Dagdeviren Orhan Dagdeviren received the BSc. and MSc. degrees in Computer Eng. from Izmir Institute of Technology. He received PhD degree from Ege University International Computing Institute. He is an assistant professor in Izmir
University, Department of Computer Engineering. His interests lie in the computer networking, distributed systems and wireless sensor networks areas.
Fatih Tekbacak Fatih Tekbacak received the BSc. and MSc. degrees in Computer Engineering from Izmir Institute of
Technology. He is a PhD candidate in Computer Engineering at Ege University. He is also a research assistant in Izmir Institute of Technology. His interests lie in the software agents, access control of role based agents on multi agent systems and
semantic web.
205
Ataturk University, Engineering Faculty, Computer Engineering Department, 25240, Erzurum, TURKEY
1
2
ddal@atauni.edu.tr, atolga@atauni.edu.tr
Abstract. When computing, there are three basic approaches for improving the performance: using a better algorithm, using a faster computer or dividing the calculation among multiple processors/computers. A
computer cluster might be employed as a means to achieve the desired performance improvement for the third
approach. A cluster is a collection of interconnected stand-alone computers cooperatively working together as
a single, integrated computing resource. In the context of computer clusters, every stand-alone computer forming the cluster is also called as a compute node. When all of the compute nodes of a cluster have exactly the
same hardware and software components, this cluster is referred to as a homogeneous cluster. If the cluster is
formed by using computers with different hardware and software specifications, this cluster, on the other
hand, is referred to as a heterogeneous cluster. This paper compares the performance of those two cluster options in terms of execution times. The conducted experiments over two different sets of benchmarks show an
average improvement of 31% in favor of the heterogeneous Windows cluster.
Keywords: Parallel Computing, Computer Cluster, Dedicated Cluster, Linux Cluster, Windows Cluster, Homogeneous Cluster, Heterogeneous Cluster
1 Introduction
There are several scientific applications needing to be solved in a reasonable amount of time and requiring huge
amount of computing power such as weather forecasting. To overcome such a problem, in general, three basic
approaches are employed for improving the performance: using a better algorithm, using a faster computer or
dividing the calculation among multiple processors/computers [1]. A computer cluster might be used as a means
to achieve the desired performance improvement for the third approach. A cluster is a type of parallel or distributed processing system, which consists of a collection of interconnected stand-alone computers cooperatively
working together as a single, integrated computing resource [2]. Clusters serve a number of purposes. A single
system with several computers can be used to obtain the high performance need of a scientific application (High
Performance Clusters). Another cluster may be recruited to create a high-availability in case of a failover for a
critical web serving environment (High Availability Clusters). A cluster might also be used to evenly distribute
the processed data among several computers to provide better performance for a web search engine (Load Balancing Clusters).
Clusters are generally formed by using off-the-shelf, commodity computers, hardware and open source software.
By using commodity computers and connecting them over a switching device enables us to obtain a processing
power once only supercomputers had the opportunity to employ. The flexibility of including a new processing
unit when one of the computing nodes from the cluster fails makes clusters a highly reliable and scalable solution for system administrators. Due to their above-mentioned advantageous, there are many computer clusters that
are on the worlds top 500 supercomputers list along with SMP style supercomputers.
206
As far as the computer clusters are concerned, every stand-alone computer forming the cluster is also called as a
compute node. When all of the compute nodes of a cluster have exactly the same hardware and software components, this cluster is referred to as a homogeneous cluster. If the cluster is formed by using computers with different hardware and software specifications, this cluster, on the other hand, is referred to as a heterogeneous cluster.
One of the important design steps for building up a computer cluster is the selection of the operating system and
the other software that will be used. Linux would be the top operating system choice that runs on each computing
node of a cluster since it is open source software and does not require a licensing cost. In the context of this paper, a homogeneous Linux cluster is referred to as the cluster where each node of the cluster has Linux installed
on it (same version of Linux distribution, kernel and software packages) and has the same CPU and memory configuration. On the other hand, a heterogeneous Windows cluster is the one where every machine from the cluster
has any Windows operating system installed on and has different CPU and memory configurations. This paper
compares the performance of those two cluster options in terms of execution times. The conducted experiments
over two different sets of benchmarks show an average improvement of 31% in favor of the heterogeneous Windows cluster.
The rest of the paper is organized as follows: Section 2 explains the motivation of this paper. Section 3 is a survey
of related work. Section 4 introduces the process of building up our homogeneous and dedicated Linux Cluster
and our heterogeneous Windows cluster in two separate subsections. Section 5 discusses the experimental results.
Finally, Section 6 concludes the paper with remarks on future research
directions.
2 Motivation
This paper is inspired by a discussion held on a graduate level parallel programming course. In this course, students form a heterogeneous Windows cluster by connecting their personal laptops (all have windows OS installed
on) together over the WLAN of engineering faculty to run parallel programs written in C++ and Windows
MPICH2 [3] implementation of MPI (Message Passing Interface) during the academic semester. The questions
listed below were arisen in one of the lectures while the concept of dedicated Linux clusters were being studied:
In the scope of this paper, the fourth question will be under scrutiny.
3 Related Work
Some real life problems scientists face today need to be solved as fast as possible. Weather forecasting is one of
such problems and by using parallel computing, people can predict the weather and the climate changes two or
three weeks in advance. Another important problem requiring huge amount of processing power is the human
genome project [4]. Parallel computing is simply either using more than one processor sharing the same memory
bank (supercomputing) or using more than one computer with distributed memory (cluster) to achieve higher
processing power. Clusters offer high performance, high reliability and flexibility with very little cost over their
supercomputer counterparts. When talking about computer clusters, for many people, the most likely thing coming into the mind is a Beowulf Cluster [5]. A Beowulf cluster is a popular variant of a compute cluster that is
based on off-the-shelf, commodity PC hardware and free open source software [6]. There are also commercial
cluster solutions available [7], [8]. Clusters can be either homogeneous or heterogeneous. A heterogeneous parallel and distributed computing environment using PVM is introduced in [9]. Load balancing for heterogeneous
clusters of PCs is examined in [10]. To the best of our knowledge, this work is the first one in the literature that
compares a homogeneous dedicated Linux cluster with a heterogeneous Windows cluster in terms of performance.
207
4 Experimental Setup
4.1 Building up a Homogeneous and Dedicated Linux Cluster
The schematic view of our homogeneous and dedicated Linux Cluster is shown in Figure 1. This cluster is composed of four nodes: one node for administrative purposes and three dedicated nodes for computing. It also has a
10/100 Mbps ethernet switch (ASUS GigaX1024) for intra-node communications. The software and hardware
specifications of the cluster computers are given in Table 1 and Table 2-3, respectively. Since it is a dedicated
cluster, there is no need for a monitor, a keyboard and a mouse for each individual compute node. (One monitorkeyboard-mouse set is plugged into the head node all the time and also used for troubleshooting when there is a
networking problem and hence, accessing to the compute nodes are impossible.)
The head node of the cluster serves as the administrative node and has two network interface cards (eth0 and
eth1). This helps us to isolate the compute nodes from the outside world. The cluster can be accessed over the
internet through eth1 and the head node can communicate with the compute nodes through eth0. All cluster
software are installed on the /opt directory of the head node. The head node also runs as an NFS server to share
its /opt and /home directories over the internal network and those two directories are mounted by all the
compute nodes. Linux MPICH2 implementation of MPI (version 1.3.2.) for message passing and g++ (mpicxx
208
wrapper) for compiling C++ programs are installed on the head node. Programs are compiled with O3 flag to
increase the speed. All user accounts are set in such a way that they allow passwordless ssh over the cluster machines. Since there is no need for graphical environments for the compute nodes, they run at init level 3 to reduce
the load of the machines while the head node runs at init level 5 to support X window.
4.2 Building up a Heterogeneous Windows Cluster
Three personal laptops with different versions of Windows OS installed on are employed to build up a heterogeneous Windows cluster for the performance comparison. Those computers are selected in such a way that the
average CPU speeds and the average memory sizes of those computers will match its homogeneous counterpart.
The schematic view for this version of the cluster is very similar to the previous one except the fact that there is
no head node for this scheme. Here, compute nodes are replaced with the laptops and the same ethernet switch is
used for message passing. The software and hardware specifications of the cluster computers are given in Table 4
and Table 5-6, respectively.
Table 4. The Software Specifications of the Heterogeneous Windows Cluster Nodes
Brand Name
Operating System (OS)
Dell Inspiron 6000
Windows XP-SP3
Laptop 1
Dell Latitude D600
Windows Vista-SP1
Laptop 2
Sony Vaio VGN-FS500P
Windows 7 Enterprise
Laptop 3
Table 5. The Hardware Specifications of the Heterogeneous Windows Cluster Nodes (1)
CPU Model Name
CPU Speed (GHz)
FSB Speed (MHz)
Intel(R) Pentium(R) M 740
1.73
533
Laptop 1
Intel(R) Pentium(R) M 725
1.6
533
Laptop 2
Intel(R) Pentium(R) M 760
2.0
533
Laptop 3
AVERAGE
1.78
Table 6. The Hardware Specifications of the Heterogeneous Windows Cluster Nodes (2)
Cache Size
Memory Size
Hard Drive Capacity
(MB)
(MB)
(GB)
2
512
60
Laptop 1
2
512
40
Laptop 2
2
512
100
Laptop 3
AVERAGE
512
All three Windows laptops are equipped with Microsoft Visual Studio 2010 Ultimate for compiling C++ programs and Windows MPICH2 implementation of MPI (version 1.3.2.) for message passing. /O2 flag is used
during compilation to increase the speed of the programs. A new Windows user account with the same username
and the password is created on each laptop and this account is registered with the MPICH2 software. Finally, the
virus protection software (if any exists) and the Windows firewalls of all laptops are disabled.
5 Experimental Results
To compare the performance of two above-mentioned cluster schemes, two different sets of benchmarks were
used. The first set is composed of three embarrassingly parallel benchmarks (the first column of Table 7): Cardano Triplets problem [11], a naive Prime Numbers problem [12] and a Maximum Neighboring Value problem
[13]. Neighboring value of a number (all digits must be unique) is obtained by multiplying each successive digits
of that number and the distances between them together and by taking the sum of all those partial products. For
example, the neighboring value of 7952468 is calculated with the following formula:
4*5*1+5*6*2+6*7*4+7*8*5+8*9*4=816. The third embarrassingly parallel benchmark tries to find the number
that has the maximum neighboring value. The second set of benchmarks (the first column of Table 8) multiplies
two square matrices with different sizes. The second and the third columns of Table 7 and Table 8 depict the
execution times of the first and the second set of benchmarks, respectively. The fourth columns of both tables list
the performance improvements of heterogeneous Windows cluster over homogeneous Linux cluster.
209
As shown in Table 7, a 35% improvement on average was obtained by using the heterogeneous Windows cluster
in comparison with the homogeneous Linux cluster over the first set of benchmarks. When it comes to the second
set of benchmarks, a 26% improvement on average was also achieved in favor of the heterogeneous Windows
cluster. (The average improvement over two sets of benchmarks yields a value of 31%.)
Architectural differences of the computers (processor families, FSB speeds and cache sizes), the compilers and
their optimization capabilities as well as the operating systems used in those two clusters might have played a role
in the improvements observed. The investigation of all those possibilities could be an avenue for a future research.
6 Conclusions and Future Work
Connecting commodity computers over a switching device and making them act as a single processing unit offer a
cost effective and highly reliable solution for parallel computing needs. This is called clustering and is widely
recruited to find solutions for various kinds of problems within a reasonable amount of time. Clusters can be
either homogeneous or heterogeneous in terms of hardware and software. In this paper, a homogeneous and dedicated Linux cluster was compared with a heterogeneous Windows cluster in terms of performance. The conducted
experiments over two different sets of benchmarks showed an average improvement of 31% in favor of the heterogeneous Windows cluster. As mentioned in the last paragraph of Section 5, several possibilities exist for yielding such a result. The investigation of all those possibilities could be an avenue for a future research.
References
1.
2.
3.
4.
5.
6.
7.
8.
9.
Bohn C. A. and Lamont G. B. (2002). Load Balancing for Heterogeneous Clusters of PCs. Future Generation
Computer Systems, Volume 18, Issue 3.
Burkardt J., http://people.sc.fsu.edu/~jburkardt/cpp_src/prime_mpi/prime_mpi.C
Halici E. (2011), Puzzle Question. The Science and Technology Journal of The Scientific and Technological Research Council of Turkey, Volume 2.
http://h20338.www2.hp.com/hpc/cache/276360-0-0-0-121.html
http://www.mcs.anl.gov/research/projects/mpich2/
http://www.sun.com/blueprints/1201/beowulf-clstr.pdf
http://www-03.ibm.com/systems/clusters/
Jin, H., Buyya, R. and Baker, M. (2000). Cluster Computing Tools, Applications, and Australian Initiatives for
Low Cost Supercomputing. Monitor, The Institution of Engineers Australia (IEAust), Volume 25, No 4.
Project Euler, http://projecteuler.net/index.php?id=251§ion=problems
210
Biographies
Deniz Dal He received the B.S. degree in Electrical Engineering from Istanbul Technical University, Istanbul, Turkey, in
1996 and the M.S. and Ph.D. degrees in Computer Engineering from Syracuse University, Syracuse, NY, USA in 2001 and
2006, respectively. His research interests include computer-aided design digital (CAD), high-level synthesis of digital circuits, power optimization and high performance computing systems.
Since 2007, he has been with the Computer Engineering Department, Engineering Faculty, Ataturk University, Erzurum,
Turkey, where he is currently working as an Assistant Professor.
Dr. Dal is a member of IEEE, ACM and SIGDA.
Tolga Aydn He received the B.S. (May 1998), M.S. (October 2000) and Ph.D. (February 2009) degrees in the field of
Computer Engineering from Bilkent University, Ankara, Turkey. His major fields of study are data mining and machine
learning.
He has several conference and journal publications (Knowledge-Based Systems, Lecture Notes in Computer Science) on
data mining and machine learning fields. His current research interests concentrate on spatio-temporal data mining and parallel data mining.
Since September 2009, he has been with the Computer Engineering Department, Engineering Faculty, Ataturk University,
Erzurum, Turkey, where he is currently working as an Assistant Professor.
Dr. Aydn is a member of MOVE (COST Action IC0903) action.
211
Abstract. Author identification is used in identifying the corresponding author of an anonymous text, in
areas such as criminal justice, plagiarism detection, computer security etc. In our previous study 12 style
markers based on word n-gram analysis had been collected. In this study, we have collected 12 new style
markers. Column articles of 16 authors have been tested using the old and new markers. While some of these
markers produced better identification results, others did not improve. A method was searched to increase the
success rate of identification by keeping and complementing the successful markers, and eliminating the unsuccessful. The genetic algorithm is one such method which uses feature selection for obtaining better results.
Therefore, genetic algorithm's feature selection principles have been applied to select the successful markers
in our study, which produced much better results.
Keywords: Author Identification, Feature Selection, Genetic Algorithms
1 Introduction
Author identification deals with the task of identifying the corresponding author of an anonymous text, among
a given set of predefined author candidates. The aim is to automatically determine the author of a given text with
approximately zero failure rates. The main idea under computer-based author identification is defining characteristics of documents to determine the writing style of authors. Multivariate analysis techniques which combine
lexical and syntactic analysis techniques, can achieve desirable success ratios.
Related with innovations in the identification technologies of computer science, author identification has been
used in areas such as criminal justice, corporate litigation, plagiarism detection, computer security and verifying
the authorship of e-mails and newsgroup messages.
Recent studies in the author identification area can be exemplified by Mortons study focused on sentence
lengths [1], and Brinegars study focused on word lengths [2]. Other important studies belong to Stamatatos et al.,
who applied Multiple Regression and Discriminant Analysis using 22 style markers [3]; and Corney, who gathered and used a large set of stylometric characteristics of text including features derived from word and character
distributions, and frequencies of function words [4].
Some of the leading Turkish researchers are Tan, who developed an algorithm by using 2-grams [5], and Diri
& Amasyal who formed 22 style markers to determine author and type of a document [6]. In another study, the
same authors used 2 and 3-grams to determine the author, author's gender and type of a document [7]. In our
previous study; linguistic statistics were used based on frequently used word n-grams (1 n 6) collected for all
authors [8].
In our present study 24 style markers are formed to characterize authors. However, not all of the 24 attributes
affect author attribution, positively. Genetic algorithm can be used to discard the unsuccessful attributes, thus
increasing success ratios.
212
Genetic algorithm is a programming technique that mimics biological evolution as a problem-solving strategy.
Given a specific problem to solve, the input to the genetic algorithm is a set of potential solutions, encoded in
some fashion, and a metric called a fitness function that allows each candidate to be quantitatively evaluated.
Genetic algorithm has been applied to a broad range of subjects like pipeline flow control, pattern recognition,
classification, structural optimization, and has been used to solve a broad variety of problems in extremely diverse area of fields. Some examples of these fields are electrical engineering, chemistry, mathematics, military
and law enforcement, astronomy and astrophysics, geophysics, molecular biology, data mining and pattern recognition, routing and scheduling, and author identification.
One of the genetic algorithm studies for author identification was carried out by Schlapbach et al., who extracted 100 features from a handwriting sample to isolate the author out of a set of writers [9]. By applying genetic algorithm as a feature selection and extraction method on set of features, subsets of lower dimensionality are
obtained. The results show that, significantly better writer identification rates can be achieved, if smaller feature
subsets are used. In 2006, Gazzah and Amara, used genetic algorithm for feature subset selection in order to eliminate the redundant and irrelevant features in study of identifying the writer, using off-line Arabic handwriting
[10]. In 2007, Lange and Mancoridis, developed a technique to characterize software developers' styles using a
set of source code metrics to identify the likely author of a piece of code from a pool of candidates [11]. They
used the genetic algorithm to find good metric combinations for effectively distinguishing developer styles. In
2009, Shevertalov et al., presented a genetic algorithm to discretize metrics of improving source code author
classification. The algorithm evaluated the approach with a case study, involving 20 open source developers and
over 750,000 lines of Java source code [12].
In recent studies, feature selection by genetic algorithm was applied on handwriting identification and source
code identification. In our study, we deal with identifying the author of a newspaper article among 16 columnists,
using 24 style markers characterizing the 16 authors. A total of 33,666 training-set and 7,137 test-set articles are
collected for 16 authors. All training-set articles are used to obtain authors statistics. Randomly selected 100
training-set articles and 100 test-set articles for each 16 authors are used for experimental studies.
In section 2, preliminary works will be explained. Used methods will be detailed in section 3. Experimental
studies will be given in section 4 and finally, section 5 deals with obtained results.
2 Preliminary Work
2.1 General Terms
Measurements of vocabulary diversity play an important role in language research and linguistic fields. The
common measures used are based on the ratio of different words (Types) to the total number of words (Tokens).
This is known as the Type-Token Ratio (TTR) and can be calculated with the Formula 1.
(1)
The Hapax Legomena Ratio (HR) is the ratio in percent between once-occurring types (hapax legomena) and
the vocabulary size. This ratio is calculated by using Formula 2 given below.
(2)
The entropy is a statistical parameter which measures how much information is produced on the average for
each letter of a text, in the language. If the language is translated into binary digits (0 or 1) in the most efficient
way, entropy H is the average number of binary digits required per letter of the original language. Entropy values
of n-gram series are calculated by using the Formula 3. x is the each n-gram observed in the corpus, p is the
probability of the n-gram.
(3)
213
Training
Test
#13
23.4
19.9
#14
11.8
12.6
#15
13.2
11.4
#16
15.9
15.1
#17
14.3
14.1
#18
18.1
15.3
#19
9.4
8.8
#20
8.8
10.7
#21
6.7
7.4
#11
37.8
24.4
Entropy
#10
60.7
49.2
#22
9.1
8.5
#23
23.2
18.4
6-Gram
#12
24.4
11.2
Relative
Entropy
5-Gram
4-Gram
3-Gram
#9
74.9
66.6
: Freq.
#8
84.9
77.4
? Freq.
#7
80.1
76.0
; Freq.
2-Gram
1-Gram
6-Gram
#6
22.9
8.6
! Freq.
#5
37.9
23.4
Method 2 (SVM)
, Freq.
5-Gram
4-Gram
#4
61.0
50.3
AWL
#3
77.6
69.4
ASL
3-Gram
2-Gram
#2
88.1
83.3
HR
#1
90.1
87.1
WPA
Training
Test
TTR
1-Gram
#24
11.9
12.7
3 The Method
A genetic algorithm (GA) is a search heuristic that mimics the process of natural evolution as a problemsolving strategy and routinely used to generate useful solutions to optimization and search problems. In GA, an
individual represented by a chromosome is any possible solution, population is group of all individuals and
search space is all possible solutions to the problem. Given a specific problem to solve, the input to the GA is a
set of potential solutions to corresponding problem. These potential solutions are generated randomly to form
initial population of GA.
A metric which is called as fitness function allows each candidate to be quantitatively evaluated. Promising
candidates are kept and allowed to reproduce a new population of candidate solutions. The expectation is that the
average fitness of the population will increase each round, and so by repeating this process for hundreds or thousands of rounds, appropriate solutions to the problem can be discovered.
In this study, a potential solution is a 24 bits chromosome that encodes the appropriate attributes to author
identification process. Search space has 224 possible solutions. This space is so large to search sequentially.
Therefore an evolutionary algorithm GA is needed for optimization of feature set. At the beginning, an initial
214
population including parameterized number of chromosomes is generated randomly like on Table 2. Population
sizes (P) of 10, 20, 30, 50, 100 and 200 are tested in this study.
Each bit of a chromosome represents corresponding attribute to be used for identification process. If value of
a bit is 1, the attribute in the corresponding index will be used, otherwise not.
Table 2. Sample population includes randomly generated potential solutions
#
1
2
...
P
Chromosomes
011011011001110111011101
101100100010010111101011
110001101010011101110110
Fitness value to quantitatively evaluate a chromosome is author identification success ratio that is obtained
when attribute selection is made according to chromosomes encoding. Two different types of fitness function,
Weighted Fitness and Weightless Fitness are used in this study. In Weighted Fitness function predicted top
three authors are weighted according to their ranking. Top three authors are selected because top five and top two
are also tested, and best results are obtained when 3 authors are used. For example, as given on Table 3, while
fitness value of a chromosome is being calculated, the top authors predicted by features corresponds to a gene 1
have coefficient 3, the second ranked authors have coefficient 2 and third ranked authors have 1. In Weightless
Fitness function, all authors predicted by features corresponds to a gene 1, have coefficient value 1, otherwise
0. Finally, in both cases the author having maximum total coefficient assumed as predicted author. This operation
is repeated for randomly chosen 100 training-set articles and 100 test-set articles of each 16 authors and average
authorship success ratios are calculated for each chromosome.
Table 3. Top three author id predicted for each 24 features and sample chromosome encodes used features
2 11 12 11 11 2
2 10 8
7 10 9
3 13 15 10 2
0 1 1
Rel. Ent.
9 10 16 13 4
1 1
5
1 0
: Ferq.
5 7 11 11 2
Entropy
1 12 12 3 16 5
; Ferq.
? Ferq.
6-Gram
, Ferq.
5-Gram
! Ferq.
4-Gram
AWL
3-Gram
HR
2-Gram
ASL
1-Gram
TTR
6-Gram
WPA
5-Gram
1 1 1
Chr. Predicted
Author ID
4-Gram
Method 2
1-Gram
g
2-Gram
3-Gram
Method 1
4 15 6
6 16 9
6 11 7
Linear Ranking Selection is used for selection of chromosomes which are used for reproduction operation. In
linear ranking selection, population is sorted from best to worst according to fitness. Elite individuals of parameterized count are copied to new generation. Reproduction operation includes crossover and mutation operations.
In crossover operation, two new chromosomes (offspring) are created from each of two parents in mating pool
that consists of best chromosomes. In mutation operation, a randomly selected chromosome is mutated by changing a randomly selected bit 1 to 0 or 0 to 1. Mutation probability is also a parameter of the GA. Termination condition of the evaluation process is a pre-determined number of generations (in our example it is 1000) or
when a satisfactory solution has been achieved.
4 Experimental Studies
The algorithm was tested on both training-set articles and test-set articles with the two different fitness functions mentioned above and six different population sizes (10, 20, 30, 50, 100, and 200). Each run continued over
a thousand iterations. Best individual feature sets and success ratios for all these parameters are given on Table 4.
All population sizes produced same individuals as fittest solution for each four cases. Thus, it can be said that
growth of population does not improve obtained results.
Table 4. Best individuals (feature sets) and best fitness values (author identification success ratios) obtained by using GA for
all population sizes (10, 20, 30, 50, 100, 200)
215
Best Individual
Best Fitness
Best Individual
Best Fitness
Weighted Fitness
Training-set Articles
Test-set Articles
111100110000001000000000
110100111000110000000000
94.6
90.5
Weightless Fitness
Training-set Articles
Test-set Articles
111100111000000000000000
111000111100100000000000
89.3
85.2
Genetic algorithm reaches to the same fittest individuals for all population sizes. But individuals (feature sets)
vary for each four cases given on Table 4. 7 of 24 features are selected for training-set articles, while 8 of 24
features are selected for test-set articles. Four features are common for all cases, these are 1-gram and 2-gram
versions of two n-gram based methods mentioned on Section 2. Other important findings are those Hapax Legomena Ratio is a useful feature for training-set articles when weighted fitness function used, Word per Article is
useful for both training-set and test-set articles when weightless fitness function is used and finally Type/Token
Ratio is a useful feature for test-set articles when weighted fitness function used according to the chromosomes
given on Table 4.
Recently, average success ratio 90.1% was obtained for most successful feature #1 and 6.7% was obtained for
worst feature #21 on training-set articles as can be seen on Table 1. Average was 37.8% for all 24 features on
training-set. By combining features selected with weighted fitness function, average success ratio increases to
94.6% for training-set articles.
In the same way, when test-set articles were handled, average success ratio was equal to 87.1% for feature #1
and 7.4% for feature #21. Average success ratio was 32.5% for all 24 features. By combining features selected
with weighted fitness function, average success ratio is obtained as 90.5% for test-set articles.
As seen from Table 4, although weightless fitness results are worse than our previous studys best results, it
exceeds average success ratios 37.8% and 32.5% by 89.3% and 85.2% for training-set and test-set articles respectively.
4 Conclusion
In this study, 24 style markers are formed to characterize authors. But not all of the 24 attributes are successful for author identification. The aim of the study is to obtain more successful results by eliminating weak features
and combining the useful ones. Genetic algorithm is used to increase success ratios, because the cost of sequential
searching of all possible solutions is too expensive.
By using genetic algorithm, an initial population consisting of possible solutions is created. Candidate solutions are kept and allowed to reproduce a new population of candidate solutions. As a result, the average fitness
of the population is increased in each round. By repeating this process for a thousand of rounds, correct author for
the given anonymous text is found, within acceptable success ratios.
For training-set articles, the average success ratio is increased to 94.6% from 90.1%. For test-set articles the
average success ratio is increased to 90.5% from 87.1%, by using weighted fitness function. After a 90.1% high
success ratio in our previous study, the improvement of success ratio gets harder. Therefore, 4.5% improvement
is a satisfactory result. This study shows that, genetic algorithm is an effective and useful attribute selection method to obtain optimal feature subsets, in author identification. The success ratio is increased by using the optimal
feature subset.
In future work, new features for authors will be collected and used to increase the identification success ratio.
Also, parameters of the genetic algorithm and fitness function will be enhanced. Other genetic algorithms like
hybrid genetic algorithm can be applied to investigate the success ratios.
References
[1] Morton A.Q. (1965). The Authorship of Greek Prose, Journal of the Royal Statistical Society, Series A, 128: pp. 169-233
[2] Brinegar C. (1963). Mark Twain and the Quintus Curtius Snodgrass Letters: A Statistical Test of Authorship. Journal of
the American Statistical Association, Vol.58, pp. 8596.
216
Biographies
Feritah rc was born in Aydn in 1984. She had received the B.S. degree in 2006 and the M.S. degrees in 2009 in
Computer Engineering from Dokuz Eylul University, Turkey. She is still a Ph.D. student in Computer Engineering from
Dokuz Eylul University, Turkey. Her fields of studies are NLP, data mining and business intelligence.
Gkhan Dalkl was born in zmir in 1975. He had received the B.S. degree in Computer Eng. from Ege University, Turkey, the M.S. degrees in Computer Science from USC, Los Angeles, and from Ege University ICI, Ph.D. degree in Computer
Eng. from DEU, Turkey. He had been a visiting lecturer in UCF, Orlando. He has been an Assistant Professor of the Department of Computer Engineering of DEU. His fields of studies are cryptography, and NLP.
217
Abstract. The popularity of real-time applications such as video streaming and Voice over IP (VoIP) have been
increasing in recent years with the development of wireless technology. It is difficult to support Quality of Service
(QoS) for real-time applications in wireless networks. WiMAX is the broadband wireless technology and it supports
heteregenous traffic in wireless medium and QoS for real-time applications. In WiMAX, packet scheduling implementation is left open to vendor and it directly affects the performance of the WiMAX networks. In this study, a TTLbased packet downlink scheduler is proposed. The developed scheduler is tested in the simulation enviroment. The
results show that developed TTL-based scheduler algorithm outperforms good performance with the Earliest Deadline
First (EDF) algorithm.
Keywords: Time to Live (TTL) , EDF, WiMAX, Packet Scheduling, QoS
1.
Introduction
WiMAX, abbreviated from Worldwide Interoperability for Microwave Access, is the commercial name of
products IEEE 802.16 wireless metropolitan area network (MAN) standards [9, 10]. It provides broadband wireless services in MANs and it is an alternative last-mile internet access technology in urban or suburban areas.
WiMAX has economically low cost deployment and it facilitates installation of internet access where cable-based
infrastructures are very expensive or infeasible.
WiMAX is an emerging broadband wireless network and it supports Quality of Service (QoS) in MAC level
[6]. It supports class-based QoS for applications and it has several components for supporting QoS such as admission control, scheduling, classification, etc. The scheduling algorithm, one of the most important components
in WiMAX networks, is not standardized by IEEE 802.16. In the literature, many scheduling algorithms and
schemes are proposed, but most of these algorithms are developed for uplink schedulers. However, there is no so
much effort for downlink schedulers in the literature.
In this study, a novel TTL (time-to-live) based downlink scheduler for WiMAX is proposed. The main objective of the TTL based downlink scheduler is decreasing the delay of packets. The TTL label in Internet Protocol
(IP) lets us know the number of hops (hop count) between the source and the destination. It is a well-known fact
that there is a correlation between the hop count and the delay of the packet in todays networks. When the hop
count increases, the delay of packet also increases. Backlogged packets (packet available to transmission) are
differentiated in Base Station (BS) according to their hop count, and urgent packets are transmitted to their destination before their deadlines are not missed. To the best of authors of this article knowledge, almost nothing has
been published on the subject of this article.
The paper is organized as follows: Section 2 summarizes WiMAX broadband technology and its standard in
details. In Section 3, literature review and the proposed TTL-based packet scheduler are proposed. The simulation environment and the experimental results are presented in Section 4. Finally, Section 5 reports the final evaluation and the future work
218
UGS
T1/E1,
VoIP WoSS
Maximum
Sustained
Traffic Rate
Minimum
Reserved Traffic
Rate
Maximum
Latency Tolerance
Tolerated Jitter
Traffic Priority
rtPS
Video Streaming
nrtPS
FTP, Telnet
BE
HTTP, SMTP
IEEE 802.16 defines many components in the standard. For example, radio resource management, packet
scheduling for uplink/downlink data, and call admission control are some of them. However, implementation
details are left open for vendors. The packet scheduling algorithm is the most important component in the standard, because it directly affects the performance of the WiMAX networks.
219
2.
Real-time applications, such as Voice over IP (VoIP) and video streaming, generate symmetric traffic in the
network. Therefore, when a new scheme is proposed for these applications both uplink and downlink scheduler
must be taken into account in WiMAX. Also, a real-time application can compensate certain amount of packet
loss but they are sensitive to jitter and delay of packets. In order to improve the delay performance, appropriate
packet scheduler algorithms should be used. Earliest Deadline First (EDF) [1] algorithm is one of them, and it
shows good performance in real-time applications [2, 3]. The EDF scheduling algorithm marks each incoming
packet with deadline and calculation of this deadline is a critical issue. In the Wimax networks, SS informs the
BS about QoS parameters of flow during the admission control phase. Latency and jitter are some of these parameters. The classical deadline calculation of a packet in EDF is as follows: Latency of the flow is added to the
arrival time of packet in the flow. Although EDF shows good performance in the real-time applications, it can be
improved for downlink scheduler in WiMAX networks with using delay-awareness.
Classical EDF does not differentiate packets that are far or near from/to destination (more hop). It takes only
arrival time and latency into account. However, the packets that far to the source generally face more delay than
near one. Therefore, when a packet coming from far enters to the BS, it should be send to the SS as soon as possible even if its deadline is greater than near destination packet. The main contribution of this study is giving
more precedence to the packet which far to the source. The proposed architecture is shown in Figure 1,. In backlogged real-time applications, such as ertPS and rtPS, the traffic is differentiated according to their hop count
values. Hop count value is a derived attribute that can be calculated from IP labels TTL value. The differentiated packets are assigned to dedicated queues according to their hop counts. There are three queues for each realtime service classes (rtPS, ertPS), HC < 10, 10 < HC < 20, and HC >30, where HC means Hop Count. Therefore,
the precedence of higher HC is greater than the lower one.
The classification component is defined in the standard [9, 10], but the implementation details also left to the
vendors. There are efforts about classification in the literature, such as RFC [12, 13], but these studies can be
improved with adding new types of classifiers. Furthermore, one of the important favours of this study is the conservation of the WiMAX protocol stack. There is no need to decompose of TCP/IP protocol stack. Only the new
classification module will be adapted in WiMAX BS, and this module will be classified incoming real-time packet according to their hop counts.
220
hop count. It consists of a Time to Live (TTL) 8-bit field. Each router (hop) where the packet that comes through
decrements this value by one. At the destination, the hop count of the packet that reaches from source to destination can be calculated in the final TTL. But different operating systems (OS) assign different default values for
TTL. Therefore, there cannot be a single initial TTL value for each sender IP. Todays modern Operating Systems set the initial values as 30, 32, 60, 64 and 512 as initial values. In the literature, it is observed that these TTL
values are between 30 and 32, 60 and 64, and 32-64 [4]. This assumption is used in this study, for example if the
final TTL value is 110; it is obvious that initial value TTL will be 128.
Packets are differentiated by BS according to the hop count (final TTL values) and packets that are come far to
destination have more precedence than the near ones. In this study, the packet delay in WiMAX network is balanced.
3.
The real-time TTL based downlink scheduler algorithm is implemented in house simulator for WiMAX. The
developed simulator is a discrete-event simulator. For the simplicity, only ertPS service is implemented and tested. The real-time application traffics are generated from video traces that are available [14, 15] in the simulation.
Video traces are encoded in different formats such as H.263 and MPEG. The qualities and packet sizes of the
traces vary between 15 and 256 Kbps.
The simulation topology developed is shown on Fig.2 in details. There are six video traffic generator components in the Internet which their hop counts vary. There are two video streaming traffic sessions where hop counts
are greater than 20, one session between 10 and 20, and three sessions less than 10 hop counts. Also, WiMAX
Base Station schedules the backlogged real-time packets to the Subscriber Stations (SSs) (there are six SSs).
There are two scenarios in the simulation: in the first scenario, the WiMAX Base Station default scheduler algorithm uses the Earliest Deadline First (EDF) algorithm. EDF selects the packet which has the smallest deadline.
The deadline is equal to the arrival time plus latency parameter which is defined in the mandatory QoS parameter.
In this scenario, EDF does not take Hop Count of the destination of packets into account. All the packets have
same priority in the downlink scheduler. In the second scenario, the scheduler is again EDF, but it takes backlogged packets Hop Count into account. The packets are differentiated into different EDF queues and the scheduler selects greater HC packets first, even if their deadlines are greater than the others. It means that the greater
HC packets have higher precedence.
The simulation results are shown in Figure 3. In this simulation, the Earliest Deadline First (EDF) downlink
packet scheduler algorithm is used. And also, to be aware of the HC effect, the Classical-EDF Scheduler and HCAware-EDF Scheduler are implemented and tested. Since we differentiate packets according to their HCs and put
them in appropriate queues, we have three cluster queues where HC<10, HC>10 and HC <= 20, and HC > 20. In
the simulation result, only two clusters, HC<10 and HC>30, are shown. It is obvious that HC-aware scheduler
balances the delay of the backlogged packets. The cumulative delay of the first cluster (HC<10) increased more
221
than classical EDF but cumulative delay of third cluster (HC>30) decreased. It can be said that this is a compensation mechanism that compensates the delay of near (HC<10) backlogged packets and far (HC>30) backlogged
packets.
4.
Conclusions
WiMAX is the broadband wireless technology and it supplies ubiquitous high speed connectivity in the last
mile. The packet scheduler algorithm allocates resources to SSs according to their QoS requirements. Scheduler
selection is also critical for the performance of the network. In this study, a new scheme whose name is TTLbased (Hop Count) downlink scheduler is developed. It is implemented in the house simulator. The proposed
scheme outperforms the classical EDF scheme in the downlink scheduler. In addition to TTL-awareness, channel
states of subscriber stations will be considered in future work of this study.
References
1.
2.
3.
4.
5.
6.
7.
8.
9.
10.
11.
12.
13.
C. L. Liu and J.W. Layland. Scheduling algorithms for multiprogramming in a hard-real-time environment,
J. ACM,20(1):4661, 1973.
Melek Oktay, Hac Ali Mantar, "A Real-Time Scheduling Architecture for IEEE 802.16 WiMAX Systems", 9th IEEE International Symposium on Applied Machine Intelligence and Informatics, Slovakia/Slomonice,SAMI 2011,
N. A. Ali, P. Dhrona, and H. Hassanein. A performance study of uplink scheduling algorithms in point-tomultipoint wimax networks.Comput. Commun., 32(3):511521, 2009.
H. Wang, C. Jin, and K. G. Shin. Defense against spoofed IP traffic using hop-count filtering. IEEE/ACM
Trans. Netw. 15, 1 (February 2007), 40-53.
A. Fei , G. Pei, R. Liu and L. Zhang Measurements on delay and hop-count of the internet, Proceedings of
IEEE GLOBECOM'98, , January. 1998
M. S. Kuran and T. Tugcu, "A Survey on Emerging Broadband Wireless Access Technologies," Computer
Networks, Vol. 51, No:11, pp 3013-3046, August 2007.
http://www.isi.edu/nsnam/ns/
http://code.google.com/p/ns2-wimax-awg/
IEEE 802.16-2004, IEEE Standard for Local and Metropolitan Area Networks Part 16: Air Interface for
Fixed Broadband Wireless Access Systems, October 2004.
IEEE 802.16e-2005, IEEE Standard for Local and Metropolitan Area Networks Part 16: Air Interface for
Fixed and Mobile Broadband Wireless Access Systems Amendment 2: Physical and Medium Access Control
Layers for Combined Fixed and Mobile Operation in Licensed Bands, February 2006.
Speeding Up WiMAX http://www.itbusinessedge.com/cm/community/features/interviews/blog/speeding-upwimax/?cs=40726
http://www.faqs.org/rfcs/rfc5154.html
http://www.rfc-editor.org/rfc/rfc5121.txt
222
14. Fitzek, F.H.P.; Reisslein, M.; ,MPEG-4 and H.263 video traces for network performance evaluation, IEEE
Communications Society , Network, IEEE, 2001
15. http://www.tkn.tu-berlin.de/research/trace/trace.html
Biographies
M. Oktay received the BS and MS degrees in Computer Engineering from Fatih University, Turkey, , in 2004, and 2007
respectively. He is a PHD candidate in Computer Engineering, Gebze Institute of Technology, Kocaeli, Turkey.
He is currently working at Fatih University as an Instructor. His recent work focuses on Wireless and Mobile Networks, IEEE
802.16 (WiMAX) Metropolitan Area Networks: PMP, Mesh and Relay modes, Load Balancing, Vertical & Horizontal
Handover, Quality of Service in Computer and Wireless Networks
H. A. Mantar received the BS degree in Electronics and Telecommunication engineering from Istanbul Technical University,
Turkey, the MS and PHD degrees
In Electrical Engineering and Computer Science from Syracuse University, Syracuse, NY, U.S.A., in 1993, 1998 and 2003
respectively.
From 1997 to 2000 he worked as an instructor at Syracuse University. Between 2000 and 2003, he was supported by Graduate Students Research fellowship from National Science Foundation. He is currently working at Gebze Institute of Technology (GIT), Turkey, as an Associate Professor. His research interests include wide area network management, QoS, routing, IP
telephony and wireless networks.
223
Abstract. This paper presents a sequential machine learning approach for disambiguation of morphological analysis of Turkish words in a given sentence. Given that a Turkish word, a morphological analyzer produces faulty results
as well as the correct ones. In this paper we focus on filtering the wrong results which a morphological parser generates by a sequential machine learning model namely Conditional Random Fields. Through experiments performed on
METU-Sabanci Treebank, we show that a 92% accuracy in disambiguation is achieved with using only morphosyntactic features obtained from morphological analyzing task.
Keywords: conditional random fields, morphological disambiguation, natural language processing
1 Introduction
Morphological Analyzing is a major field in Natural Language Processing for a long time. It consists of
analyzing the structure of morphemes, word forms and affixes. In agglutinative languages such as Turkish, Czech,
Finnish, and Hungarian, morphemes are attached together in complex order to form words. However analyzing of
such ordering is cumbersome and most of the time produces more than one result. For instance, in Turkish, the
word "masali", has three different morphological analyses: "masal+Noun+A3SG+Pnon+Acc", "masal+Noun+A3sg+P3sg+Nom", and "masa+Noun+A3sg+Pnon+Nom^DB+Adj+With". In this example there are
two different word roots and three different morphological analyses. In a given context (sentence) dissolving such
ambiguities is called morphological disambiguation.
Morphological disambiguation is an important task in syntactic parsing, word sense disambiguation,
spelling correction and semantic parsing for all languages. For Turkish, rule-based and statistical approaches have
been proposed to this effect. In rule based approaches, a large number of hand-crafted rules are organized to
obtain the correct morphological analysis [10]. Statistical work done on Turkish involves a purely statistical
disambiguation model [3] and a hybrid model of rule-based and classification approaches [12]. For other agglutinative languages, several methods have been proposed: unsupervised learning for Hebrew [6], a maximum entropy model for Czech [4] and a combination of statistical and rule-based disambiguation methods for Basque [2].
The aim of this work is to obtain the correct form of morphological analysis from several morphological
analysis offered by the rule-based parser. In this scope, what we propose differs from alternative statistical approaches in at least two respects. Firstly, we use n-grams to recognize repeating sequences of morphemes instead
of manually organized rule sets. Secondly, rather than confining ourselves to a single word, we also take into
consideration the neighboring words in the learning process. In this process, we use n-gram language models and
Mallet framework to build CRF models. Combining these two methods, we achieve satisfactory accuracy results
on dataset obtained from METU-SABANCI Treebank [1][8].
224
In this example, we recognize the most common morphemes that occur together in a morphological analysis of
a given word. From this example, we can infer that the probability of co-occurring morphemes above a threshold
is treaded as patterns. Some of these patterns supply a good heuristics to determine the part of speech of the word.
For example features Pos-Imp and Imp-A2sg often appears in a verb, while A3sg-Pnon appears in a noun.
Please also note that A3sg alone appears in a verb while A3sg-pnon is essential for a noun.
In order to calculate the score of each pattern, we used simple consistency tables and the mutual information
score. An example for a bi-gram model is given below.
Count(A,B)
Count(~A,B)
Count(A,~B)
Count(~A,~B)
From the consistency table we can calculate the probabilities of P(A,B), P(A), and P(B). P(A) and P(B) represents the probability of having a morpheme A at the first place, and a morpheme B at the second place respectively. The mutual probability gives us the final score which is described in the following equation.
(1)
If the mutual probability score is higher than for a given threshold, we can conclude that the two morphemes
are related and should form a single morpheme by merging.
225
In the model given in Figure 2, each label state is conditionally dependent on its features and the previous state.
According to linear chain CRF model,
can be written in the following equation [14, 15].
(2)
where
is a parameter vector,
function. In our experiments to calculate the model parameters ( ) we used MALLET1 framework to train and
test CRF models with Limited Memory BFGS optimizer [17].
In Figure 2, a more complex version which is a second order CRF is demonstrated. In this model, the number
of conditional dependencies is increased via increasing the connections between sequential states and features.
The problems of higher models are sparseness along with over-fitting. Here we refer over-fitting as losing the
ability to generalize the learning process by intrinsically learning noise.
4 Method & Experimental Results
Our method comprises two steps: First, we compile a dataset containing alternative analysis results for given
words. Second, we use a CRF model to train and test the learning on the samples of the prepared dataset. In order
to build our dataset, we analyze the words in METU-SABANCI Treebank using a morphological analyzer [19].
Analysis results are checked against the METU-SABANCI Treebank to detect the correct ones. Once the dataset
is formed, we apply the n-grams language model on it to obtain groups consisting of frequently co-occurring
morphological tags. These groups of tags are then used as features in the CRF model where the correct result is
signaled by an appropriate label. In this approach, given the morphological analysis results, we have two independence assumptions:
1
226
1) We assume that different morphological results are independent from each other.
2) We assume that features are independent from each other.
In our experiments, with 95% confidence rate in 10-Fold Cross Validation tests, we observe 92% accuracy for
morphological part-of-speech (POS) tagging and a 88% accuracy in average for morphological disambiguation of
Turkish words. In these experiments we confined ourselves only to 500 words with a 986 different morphological
analysis results (with a ratio of 1.972 morphological results per word). Accuracy results for morphological disambiguation of Turkish in several combinations of CRF models and corresponding n-gram lengths are given in
the following table:
1st Order CRF
Unigram
78.86%
81.88%
Bigram
89.90%
92.52%
Trigram
92.01%
92.85%
In this table we observe that the accuracy is increased with the increase in the order of models up to 92.85%. In
2nd Order CRF model, we consider two previous words and their morphological features to determine the current
result. In the following table, accuracy of POS Tagging results is given.
1st Order CRF
Unigram
91.86%
89.13%
Bigram
92.90%
92.52%
Trigram
92.01%
92.45%
In POS Tagging of words, we achieve 92.45% success in terms of accuracy with 2 nd Order CRF model
and trigram model.
5 Conclusion & Future Work
As it turns out, incorporating neighboring morph-syntactic tags as well as neighboring words into disambiguation
process increases the accuracy of finding the correct morphological analyses form. Moreover, if we consider the
patterns of co-occurring morph-syntactic tags in disambiguation process, we achieve a significant increase of 11%
from 81% to 92%. This result suggests that instead of using single morph-syntactic tags as features, a valid compounding is a better feature in POS Tagging and morphological disambiguation process.
For the future work, we will investigate the effects of certain morph-syntactic tags to the learning process and
eliminate the noisy features in the training set. Moreover we will integrate the probability of sequence of morphsyntactic tags as a feature to our CRF models in order to build a hierarchical CRF model.
References
1.
2.
3.
4.
Atalay N. B., Oflazer K. and Say B., (2003). The Annotation Process in the Turkish Treebank, in Proceedings of the
EACL Workshop on Linguistically Interpreted Corpora - LINC, April 13-14, Budapest, Hungary.Baker, P.R. (1978).
Biogas for Cooking Stoves. London: Chapman and Hall.
Ezeiza, N. et. al. (1998). Combining stochastic and rule-based methods for disambiguation in agglutinative languages. In
Proceedings of the 36th Annual Meeting of the Association for Computational Linguistics (COLLING/ACL98), pages
379-384.
Hakkani-Tur, D. Z., Oflazer, K. and Tur, G. (2002). Statistical Morphological Disambiguation for agglutinative languages. Computers and the Humanities, 36:381-410.
Hajic, J. and Hladka, B. (1998). Tagging inflective languages: Prediction of morphological categories for a rich, structured tagset. In Proceedings of 36th Annual Meeting of the Association for Computational Linguistics
(COLING/ACL98), pages 483-490, Montreal, Canada.
227
13.
14.
Lafferty J., McCallum Andrew, and Pereira Fernando. (2001). Conditional Random Fields: Probabilistic Methods for
Segmenting and Labeling Sequence Data. In Proceedings of the Eighteenth International Conference on Machine Learning.
Levinger, M., Ornan, U. and Itai, A. (1995). Learning morpho-lexical probabilities from an untagged corpus with an
application to Hebrew. Computational Linguistics, 21(3):383-404.
McCallum, and Kachites A. (2002). MALLET: A Machine Learning for Language Toolkit. http://mallet.cs.umass.edu.
Oflazer K., Say B., Hakkani-Tr D. Z. and Tr G. (2003). Building a Turkish Treebank, Invited chapter in Building and
Exploiting Syntactically-annotated Corpora, Anne Abeille Editor, Kluwer Academic Publishers.
Oflazer, K. (1994). Two-level description of Turkish morphology. Literary and Linguistic Computing, 9(2):137-148.
Oflazer, K. and Tur, G. (1997). Morphological disambiguation by voting constraints. In Proceedings of the 35th Annual
Meeting of the Association for Computational Linguistics (ACL97, EACL97), Madrid, Spain.
Thomas G. Dietterich. (2002). Machine Learning for Sequential Data: A Review. In Structural, Syntactic, and Statistical
Pattern Recognition; Lecture Notes in Computer Science, Vol. 2396, T. Caelli (Ed.), pp. 15-30, Springer-Verlag.
Deniz Yuret and Ferhan Ture. (2006.) Learning morphological disambiguation rules for Turkish. In Proceedings of the
main conference on Human Language Technology Conference of the North American Chapter of the Association of
Computational Linguistics (HLT-NAACL '06). Association for Computational Linguistics, Stroudsburg, PA, USA, 328334.
Dietterich, T. G. 2002. Machine Learning for Sequential Data: A Review. In Proceedings of the Joint IAPR International Workshop on Structural, Syntactic, and Statistical Pattern Recognition (4)
Lafferty, J., McCallum, A., and Pereira, F. 2001. Conditional Random Fields: Probabilistic models for Segmenting and
Labeling Sequence Data. In Proceedings of ICML-01, 282-289.
15. Wallach, H. M. 2004. Conditional Random Fields: An Introduction. Technical Report MS-CIS-04-21. Department of
Computer and Information Science, University of Pennsylvania.
16. Sha, F., and Pereira, F. 2003. Shallow Parsing with Conditional Random Fields. In Proceedings of HLT-NAACL, 213220. (20)
17. Sutton, C. 2006. "GRMM: Graphical Models in Mallet." http://mallet.cs.umass.edu/grmm/.
18. Manning C. D. and Schtze H. 1999. Foundations of Statistical Natural Language Processing, MIT Press, 191-196.
19. Eryiit G. and Adal E. 2004. An affix stripping morphological analyzer for Turkish. In Proceedings of the International
Conference on Artificial Intelligence and Applications, pages 299-304, Innsbruck.
Biographies
Hayri Volkan Agun
Birthdate: 1982, Birthplace: Erzurum.
MSc in Computer Engineering, Trakya University 2008.
BSc in Computer Engineering, Anadolu University 2005.
He has been contributing to the research projects at Cognitive Science Society of Trakya since 2006. Research interests include Natural Language Processing, Text Mining and Information Retrieval. Mr. Agun is a professional member of
ACM since 2010, and IEEE since 2006.
Ylmaz Klaslan
Birthdate: 1969, Birthplace: Simav.
PhD in Cognitive Science, Edinburgh University 1998.
MSc in Cognitive Science, Edinburgh University 1994.
BSc in Computer Engineering, Middle East Technical University 1992.
He is a lecturer in Computer Engineering and Science, at Trakya University. Dr. Klaslan is teaching a course on
semantics at Boazii University. Research interests include Computational Linguistics, Natural Language Processing and
Machine Learning.
228
Abstract. Alcoholism is an addictive disorder with effects on individuals social, physical and psychiatric
lives. Additionally, it causes hard neurological damages on brain of the victims. In this paper, Global Field
Synchronization (GFS) of multi channel Event Related Potential (ERP) signals in Delta, Theta, Alpha, Beta
and Gamma frequency bands were used as discriminating feature vectors in the classification of alcoholic and
non-alcoholic control subjects. GFS is a feature extraction method which is used to acquire the functional
connectivity of the brain as a response to a given stimuli type. A channel optimization algorithm was applied
during GFS feature extraction stage to determine the most significant channel sources in order to improve
classification accuracy of the system. Probabilistic Neural Network (PNN), Support Vector Machine (SVM)
and Random Forest (RF) were used as classifiers. SVM is the best classifier with the classification accuracy
of %89.
Keywords: Alcoholism, Event Related Potentials, Probailsitic Neural Network, Support Vector Machine,
Random Forest
1 Introduction
Alcoholism is an addictive disorder which is defined as excessive and unconscious alcohol usage. Genetic and environmental factors trigger the development of alcoholism. Although some researches show that genes
inherited from the family provide a common risk factor for alcoholism; the environmental risk factors and risk
factors related with lifestyle play the supporting risk factor for its progression [1]. It has social, physical, psychiatric and neurological harms on individuals. Diagnosing an alcoholic subject is very important for his rehabilitation. In order to diagnose a subject, it is useful to know the strength of damage that the addiction causes in patients brain. For this reason the diagnosis is done by using electrical activity of the brain.
Electrical activity of the brain is recorded by using multiple electrodes placed at the surface of the scalp
is called electroencephalogram (EEG). The amplitude of EEG signal can vary between -100 V and +100 V and
its frequency ranges up to 40 Hz or more. Event related potentials (ERPs) that are specific parts of the ongoing
EEG records are the negative and positive voltage changes related to the brains response against to auditory or
visual stimulus [2].
ERP has been widely used tool in almost all areas of psychological research to make correlations between brain activity and cognitive deficits caused by various neurological disorders. Even though ERP consists of
lots of components which carry different information about the brain activity, P300 or P3 is the most commonly
used task-relevant component in the literature. The significant maximal positive peek between 300 and 900 ms is
characterized as P300 which is generated over parietal/central area of the brain. The P300 component of ERPs
provides features to investigate the neurological disorders including alcoholism [3]. Functional connectivity of
the brain regions as a response against to a given stimuli was investigated before by using PNN as a classifier [4].
229
Unfolding the relation between different brain regions by some distinct synchronization measurements helps us to
investigate brain functioning in alcoholism and other disorders.
In this paper, SVM and RF classifiers are added into proposed system that uses PNN as a classifier in
order to distinguish alcoholic subjects from the non-alcoholics by using multichannel ERPs by identifying the
regions of the brain that gives the most significant information for classification and efficiency of the three classifiers are compared.
2 Material and Methods
In the second experiment, the subjects were asked to press a mouse key in matching conditions. In the
third experiment, subjects were asked to press in no matching conditions.
Finally, ERPs were averaged as artifact-free trials with correct responses for three cases: S1, S2 match
and S2 nomatch.
70 subjects were randomly selected from EEG data set. Both alcoholic and non-alcoholic groups consist
of 35 subjects for each three experiment. Before applying feature extraction and classification, the ERP data were
arranged in order to be ready for processing.
230
GFS ( f )
E ( f )1 E ( f ) 2
E ( f )1 E ( f ) 2
0 GFS ( f ) 1
(1)
(2)
GFS value varies between 0 and 1. If the GFS value approaches to one, this emphasizes coherence among the
channels of ERP signals in terms of functional connectivity. Otherwise, there is not such coherence [9].
GFS values were calculated for each five frequency bands for each channel using ERP signals as shown in Table 1. Then, GFS values that come from each channel were averaged for each subject. Finally, five GFS values
from each frequency band for each subject were extracted as feature vector.
Table 6. Frequency bands that are used in GFS calculation
Frequency Band
Delta
Theta
Alpha
Beta
Gamma
Frequency Range
1-4 Hz
4.5-7.5 Hz
8-12 Hz
12.5-30 Hz
35-45 Hz
r radbas distance
(3)
Result is passed to the summation layer of the PNN. The summation layer sums the outputs of the pattern
layer and calculates the probabilities of input for each class as a vector. The outputs of the summation layer for a
testing data are Bayesian posterior probabilities that show the probabilities of each output belonging to each class.
The classification for PNN is done in decision layer as given in equation 4.
hi ci f i ( x) h j c j f j ( x)
231
(4)
where hi and hj are the probabilities of the input for class i and j, ci and cj are the misclassification cost
for two class and fi(x) and fj(x) are probability density functions of the classes [12]. The input is classified based
on the maximum probability value, and the decision layer produces a logical one to the corresponding class and
zero for all other classes.
c ai k ( s i , x ) b
(5)
A kernel function k(si,x) is selected based on the application. In default, dot product is used for linear
classification. After constructing the classification function, the new examples are classified by applying the function. If c 0, the instance x is classified as a member of one group, otherwise it is classified into other group.
k (si , x) exp( si x )
2
(6)
SVM is used for non-linear classification by selecting different kernel functions. Gaussian Radial Basis
Function is commonly used non-linear classification kernel function for SVM.
232
3. At each node, n out of the N variables are selected randomly instead of choosing the best split
4. Gini node impurity criterion is used to determine the best split on these n variables.
Reducing n causes a reduction on the strength of the individual trees in the forest and the correlation between them. On the other hand increasing causes increase on both. An optimum value of n can be found by using
the OOB error rate. The value n is used the only adjustable parameter which specify the sensitivity of RF
The computational complexity of each tree in RF is O((Slog(S)N1/2), where S is the number of the training cases. Therefore, it can handle very large number of variables with moderate number of observations.
Random forest is also used to measure the effectiveness of each variable in the data set. This means that
significance of each feature is the given data set is measure and compared by RF [15].
Our channel optimization algorithm is a simple heuristic algorithm. Two channel sets are defined as L
and G. L represents local best channel set and G represents global best channel set. Initially, L and G contain all
61 channels. The algorithm first tries to classify subjects by using G channel set and find classification accuracy
as global best classification accuracy. Then a channel is omitted such that without it the global accuracy is maximized with the remaining channels. This process continues until omitting a channel does not improve the global
classification accuracy [4].
In order to formulate the algorithm, we can define accuracy improvement of L with respect to G for
omitting a single channel x G.
(7)
233
3 Results
The proposed optimization algorithm is applied with PNN, SVM and RF separately to 70 subjects for
each three experiment cases; single stimulus (S1 Subject), S2 matching (S2 match) and S2 no matching (S2 nomatch) conditions. The test performance of the system and classifiers are determined by the computation of sensitivity, specificity and accuracy measurements based on classification results as shown in Table 2. In order to
compute these statistical parameters, true positive, false positive, true negative and false negative results are calculated.
Table 7. Experimental Results by Applying Channel Optimization Algorithm
Probabilistic
Neural
Network
Support
Vector
Machine
Random
Forest
Classifier
# of channels
Specificity
Sensitivity
Accuracy
# of channels
Specificity
Sensitivity
Accuracy
# of channels
Specificity
Sensitivity
Accuracy
S1 Subject
18
0,74
0,89
0,81
20
0.72
0.86
0,79
26
0.79
0.76
0.77
S2 Match
8
0,83
0,80
0,81
18
0.92
0.86
0,89
18
0.85
0.81
0.83
S2 Nomatch
13
0,69
0,80
0,74
19
0.74
0.66
0,70
11
0.76
0.75
0.76
Classification accuracies are improved from %50 to %90s by applying channel optimization algorithm.
While the classification accuracy is being improved, the number of channels that are used to record ERPs is reduced. The used channel number is reduced from 61 to nearly 15-20. So the main reason of this inversely correlated pattern between the classification accuracy and number of channels can be unimpaired brain regions. The
deficits of the alcoholism on brain can affect on some specific regions of the brain. Finding these regions and
recording ERPs by using these channels can improve the classification accuracy of the system as proved in this
paper. Moreover, the brain regions that have the most significant features for classification are found by the proposed system.
The resulting channel electrode places are shown by using Standard Electrode Positions (Fig. 1 shows an
example) for each classifier results and each data set. If the regions of the brain are examined in terms of the
results of PNN classifier; right frontal, right central, middle central and left parietal sites have valuable information for classification of S1 subjects; middle central and left central regions for S2 match subjects and finally
right frontal, middle central and middle parietal regions for classification of S2 nomatch (Fig. 3 shows an example).
Fig. 3. Channel electrode places for PNN classifier results respectively belongs to S1 subjects, S2 match and S2 nomatch data
sets
In the results of SVM classifier; right and middle frontal, middle central, right and left parietal sites have
valuable information for classification of S1 subjects; right central and left and right parietal regions for S2 match
234
subjects and finally right frontal, middle central right parietal regions for classification of S2 nomatch (Fig. 4
shows an example).
Fig. 4. Channel electrode places for SVM classifier results respectively belongs to S1 subjects, S2 match and S2 nomatch
data sets
In the results of RF classifier; right frontal, right, left and middle central, left parietal and occipital sites
have valuable information for classification of S1 subjects; middle central and left and right parietal regions for
S2 match subjects and finally right central, left and right parietal regions for classification of S2 nomatch (Fig. 5
shows an example).
Fig. 5. Channel electrode places for RF classifier results respectively belongs to S1 subjects, S2 match and S2 nomatch
data sets
In order to generalize the regions in terms of data sets: right frontal, right and middle central, right and
middle occipital regions have more significant information for classification of S1 subjects; left and middle central, left parietal for S2 match subjects; right and middle central, right parietal regions have more significant information for classification for S2 nomatch subjects.
Although the classification accuracies of the three classifiers are nearly similar, SVM is seem to be best
classifier with the classification accuracy of %89. After that, RF is the second classifier with the classification
accuracy of %83. PNN has %81 classification accuracy.
Additionally, it can be interpreted that S2 match data set and test environment is seem to be more efficient to obtain ERPs from the subjects in order to classify them into alcoholic.
References
1.
2.
3.
Micheal, D.K. (2008). Approach to the genetics of alcoholism: A review based on pathophysiology. Biochemical pharmacology 75: 160-177.
Micheal, G. H. C. and Micheal, D. R. (1995). Electrophysiology of Mind: Event-Related Brain Potentials
and Cognition. Oxford University Press, 1-23.
Porjesz, B and Begleiter, H. (1990) Event-related potentials in individuals at risk for alcoholism. Alcohol
7(5): 465-469.
235
4.
5.
6.
7.
8.
9.
10.
11.
12.
13.
14.
15.
Cokyilmaz, M. and Emanet, N. (2010). Classification of Alcoholic Subjects using Multi Channel ERPs based
on Channel Optimization and Probabilistic Neural Network. In Proceedings of Ninth Applied Machine Intelligence and Informatics, Smolenice, Slovakia, 27-29 January.
L.Ingber http://kdd.ics.uci.edu/database/eeg/eeg-full.tar
Zhang, X. L.; Begkiter, H.; Porjesz, B.; Wang, W.; and Litke, A. (1995). Event Related Potentials during
Object Recognition Tasks. Brain Research Bulletin 38(6): 531-538
Koenig, T.; Lehmann, D.; Saito, N.; Kuginuki, T.; Kinoshita T.; and Koukkou, T. (2001). Decreased functional connectivity of EEG theta-frequency activity in rst episode, neuroleptic-naive patients with schizophrenia: preliminary results. Schizophrenia Research 50(1): 5560.
Koenig, T.; Prichep, T.; Dierks, T.; Hubl, D.; Wahlund, L. O.; John, E. R.; and Jelic, V. (2005). Decreased
EEG synchronization in Alzheimer's disease and mild cognitive impairment. Neurobiology Aging 26(2):
165171.
Kikuchi, M.; Koenig, T.; Wada, Y.; Higashima, M.; Koshino, Y.; Strik W.; and Dierks, T. (2007). Native
EEG and treatment effects in neuroleptic-nave schizophrenic patients: Time and frequency domain approach. Schizophrenia Research 97(1): 163-172.
Specht, D. F. (1990). Probabilistic neural network. Neural Networks 3(1): 109-118.
Wahab, N. I. A.; Mohamed, A.; and Hussain, A. (2008). Comparing least squares support vector machine
and probabilistic neural network in transient stability assessment of a large-scale power system In Power and
Energy Conference. Johor Bahru, 1-3 December.
Hammond, M. H.; Riedel, C. J.; Rose-Pehrsson, S. L.; and Williams, F. W. (2004). Training set optimization
methods for a probabilistic neural network Chemometrics and Intelligent Laboratory Systems 78(1): 73-78.
Cortes, C. and Vapnik, V. (1995). Support-Vector Networks, Machine Learning 20(3): 273-297.
Breiman, L. (2001). Random Forests, Machine Learning 45(1): 5-32.
Emanet, N. (2009). ECG Beat Classification by Using Discrete Wavelet Transform and Random Forest Algorithm. In International Conference on Soft Computing, Computing with Words and Perceptions in System
Analysis, Cyprus, 2-4 September
Biographies
Mehmet okylmaz Mehmet okylmaz received his B.Sc degree in 2008 at Computer Engineering Department, Fatih
University, Istanbul, Turkey. He is now studying his M.Sc at Computer Engineering Department, Fatih University, Istanbul,
Turkey. His main interests are pattern recognition, heuristic optimization. He is currently a Research Assistant at Computer
Engineering Department, Fatih University, Istanbul, Turkey.
Nahit Emanet Nahit Emanet received his B.Sc degree in 1994 at Computer Engineering Department, Boazii University, Istanbul, Turkey. He completed his M.Sc. study in 1997 at Computer Engineering Department, Boazii University,
Istanbul, Turkey. In 2004, he received his Ph.D. from the same university. His main interests are microprocessor architectures, embedded systems, VLSI design, pattern recognition, computer graphics and vision. He is currently an Assistant Professor at Computer Engineering Department, Fatih University, Istanbul, Turkey.
236
Abstract.This paper is the final consequence of studies and investigations on producing an appropriate
authoring tool in order to construct a standard interactive multimedia content in the process of Elearning.Although many tools have been developed to produce content only some of them are compatible to
standards, such as SCORM and IMS. Because of this fact we intend to introduce a content constructor engine
designed in our workgroup and scrutinize its capabilities. In addition to producing standard content, this
engine employs knowledge management in content as well.
Keywords: standard learning authoring tools, metadata ,asset ,SCO,IMS Global consortium, knowledge
management in content, reusable learning object
1 Introduction
E-learning content authoring tool is software designed to create E-learning curriculum. These authoring tools
develop pages comprised of text, audio, image, video and animation. By organizing the pages, produced content
will ease learners to track on the process of learning and appraise their progression themselves and this will
apparently lead to self study. With a proper scenario, mentors and educators could observe the progression of
their learners as well. So if we recognize the standards, we will be able to choose an appropriate authoring tool
and provide congenial content to employ in E-learning process [1], [4].
237
packaging, content and learning material consortium, appraising educate and interactive content, included in
IMS 1and SCORM2 standards, will be mentioned [2], [3].
In SCORM standard, E-learning content is created on the basis of reusable learning objects. The model which
presents content in this standard is composed of three main parts:
1- Assets including media (text, audio, image, video, animation and web pages) which are introduced as
valuable objects.
2- SCO3 that contains some assets and are introduced as the lowest level of component in content, in
order to provide reusability.
3- Content packaging that provides methods of packing contents in the form of connectable learning units
(course, chapter and modules).
This standard, for each of the three parts, puts metadata in XML 4 format so that they could be described. The
most important E-learning standard is the one related to descriptive metadata, so it could provide the following
abilities:
The defined data model for metadata in SCORM standard is divided into 9 main sections and internal parts of
each section is depicted in detail (description of name, data type, maximum length and number of permitted
repetitions). After the declaration about components name and structure involved in data model, the method of
components implementation accompanied by XML standard is defined. The descriptions provided for each
component in metadata include:
In order to explain the content and its constructing process some suggestions are propounded in IMS standard as
well. We are going to describe limited number of them briefly:
IMS content packaging specification to create and share the object, content and learning elements with
the capability of being reusable
IMS question test specification to share test items and other appraisable devices
IMS learning design specification to define learning scenarios and interactions for the use of contents
and curriculums
It must be mentioned that these are only some of the categories for constructing E-learning contents in IMS [4],
[6], [8], [9], [10].
1IP
Multimedia Subsystem
Content Object Reference Model
3 Sharable Content Object
4 Extensible Markup Language
2Shareable
238
239
These are only some of the points about choosing a content authoring tool. So we are ought to make a list of our
requirements in the beginning, and then we can go for selecting an appropriate authoring tool [1], [4], [5].
To simulate books
Outmost efforts were made in order to make the similarity of this work with original book to build
confidence in end user. Due to respect the principles of interaction, we had to change some exercises in
this software and in this process we talked with some reliable experts to keep the exercise originality.
To add signs
This feature allows the user to mark the desired page with a text label, so that he becomes able to
access the same page whenever is required.
To add notes
The user is able to add notes in each page and in the next reference he could find the attached note
under a notation.
Quick access to book contents
Using a hierarchical menu, different parts are classified properly.
To zoom in pages
In order to obtain a better view of pages, the user can employ this feature.
To complement the book contents
In addition to keep similarity between the software content and book content, the efforts were made to
add more educational items as complementary, so that a suitable assessment could b provided.
To create an interactive environment
Tenets in this software are based on multimedia and interactive deportment. In order to increase the
motivation and interest of the user to keep studying and doing exercises, interactive practices are
created.
Reportage of the user activities
If we use this feature it helps us to have an accurate report of the user performance on content.
Creating username
Each user must create his username to observe the notes, allegories and reports related to him.
Searching abilities
This software allows user to search a word on the text. After the process of searching a list containing
desired word will be shown. By choosing one of them, the user can jump to the page directly.
240
User Registration
User
User Identification
Simulating Books
Search
<<include>>
Thematic List
<<include>>
Navigate On Pages
List Of Results
<<extend>>
Book Presentation
<<extend>>
Managing Notes
Interactive Activities
<<extend>>
Selective T ests
<<include>>
<<extend>>
Managing Signs
Zoom In
<<extend>>
Zoom Out
Add Note
Show Reports
Delete Note
Course Test
Comprehensive Test
Delete Sign
Component based
The produced engine is distributed in a collection of standard components. So, modular components
could be easily added or deleted from this collection. By using Flash CS4 and Action Script 3.0
capabilities, we can develop our desired components and customize them congruous to our
requirements.
Usage of XML
In order to make communication between internal and external components, XML format is employed.
IMS and SCORM compliance
Produced content via this engine, supports SCORM standard completely and the text, audio, image,
video, animation and pages are known as SCO assets. Besides the format used in assessments and tests
are based on IMS standard, so that the assessments are relevant to educational content and could be
used in appraisal reports to provide reportage for each role congenial to a particular knowledge
management overview.
Knowledge management ability
In content coding scenario, the content is classified in order to help user to gain his required knowledge
depending on his role.
To present content in the form of online and offline
Content constructor engine is designed to use HTML as its main format and Cascading Style Sheet as a
pattern to present content. These patterns are examined by script languages as well and each page will
be presented after being rendered. Hence in the online presentation, when a page is opened it will be
loaded and shown at the same moment. This attribute will also make able the content to perform on a
low bandwidth.
241
8 Conclusion
Although many tools have been developed to produce content but only some of them are compatible to
standards. Because of this fact, we intend to present the criteria and properties to choose best tools. In addition,
we introduced a content constructor engine which is designed and produced in our academic group. This engine
demonstrates standard contents and besides some new features is included. The most significant characteristic
used in this software is to display the produced content in form of both offline and online. When the considered
band width is not available for the user (learner or mentor), the offline phase is recommended in the process of
learning and for reaching higher performance a web service is employed to transfer users information. Some of
other attributes in the engine are: simulating variable books, adding notes, interactive environment, searching
abilities and reportage of user actions. It must be mentioned that the engine we made has capability to use
knowledge management in content, as well.
References
1.
2.
Biographies
Hossein Keynejad Is born in Tehran on 1965 and has earned M.Sc. degree in applied mathematics from
Shahid Bahonar University, Kerman, Iran in 1992 and B.Sc. degree in applied mathematics from Shahid
Bahonar University, Kerman, Iran in 1985.
He has published Multimedia Environments book in Persian. He is interested in E-learning and multimedia
fields.
Mr. Keynejad is a faculty member in Islamic Azad University, south Tehran branch and Project manager on
producing contents in Soroush Mehr Company.
Maryam Khademi is born in Boroujerd on 1967 and she received B.Sc. in Pure Mathematics from
TarbiatMoaalem University, Tehran, Iran in 1989, M.Sc. in Pure Mathematics from Islamic Azad University
North Tehran Branch, Iran in 1991, M.Sc. in Applied Mathematics from Iran University of Science and
Technology, Tehran, Iran in 1993, and Ph.D. degree in applied mathematics from University of Tehran in 2004.
242
She has very publication in different fields and her CV is located on link:
http://old.azad.ac.ir/faculty/khademi89.htm. She is interested in finite groups Theory, Graph Theory, Computer
Science and Artificial Intelligence.
Dr. Khademi is the faculty member of Islamic Azad University, Tehran south branch.
Maryam Haghshenas is born in Bijar on 1985 and has earned B.Sc. degree in Computer Hardware
engineering from Islamic Azad University, south Tehran branch on 2007 and is M.Sc. degree student in IT
management at Islamic Azad University, Science & Research branch.
She has published Information technology security & service management system in Information Society of
Iran journal on July 2010. She is interested in E-learning & Indexing on components of multimedia & authoring
tools and IT System management.
Ms. Haghshenas is a member of Young Researchers Club at Islamic Azad University, Tehran south branch
and also a member of the Information Technology Management Association at Islamic Azad University, Science
&Research Branch.
Hoda Kabir is born in London on 1985 and has earned a bachelors degree in Computer Software Engineering
from Islamic Azad University, south Tehran branch on 2010.
She is interested in producing E-learning content, artificial intelligent & Multimedia fields.
Ms. Kabir is a member of Young Researchers Club at Islamic Azad University, South Tehran branch.
243
mkose@ciu.edu.tr, 2yercan@ciu.edu.tr
Abstract. In this study, the possible solutions of the Tabu Search algorithm is investigated for the solution
of the University Course Timetabling Problem (UCTP). The diffuculty of the UCTP is the coordination of
lectures, classrooms and teachers according to some constraints. The replacement of courses and lectures to
periods and classrooms is the main task that have to be performed for each academic semester. This
replacement requires new solution for each semester because of part-time lecturers, changes on workloads of
lecturers and the changes inthe number of students. The main aim of this study is to solve the UCTP with the
use of optimum number of classroom assignment by reducing the unused periods in classrooms timetable.
Tabu Search incremental strategy is studied within the Faculty of Engineering at Cyprus International
University (CIU).
Key words: timetabling problem, tabu search algorithm, metaheuristics
1 Introduction
The University Course Timetabling Problem (UCTP) has been studied by many researchers since 1965. It is a
NP-hard problem and it is a real-world application of combinatorial optimization problems. Timetabling
problems are mainly worked in high schools and universities. Usually, these problems are solved manually.
Depending on the search space of the problem these manual solutions often require a few weeks or more to
solve, because during the construction of timetabling always some changes are done. After some number of
modifications, the result is mostly not optimal. Over the past few years, researchers have been investigated new
algorithms or techniques to find general solution of timetabling problem. However, the UCTP solutions may
vary from university to university because of their own curriculum and thereby they require their indivual
solutions.
Timetabling problems are studied with several techniques for more than four decade. Some of the recent studies
include Constraint-Based Reasoning [1], Integer Programming [2], Tabu Search Algorithm [3], Memetic
Algorithms [4], Genetic Algorithms [5]-[7-9]. Other popular techniques including Graph Coloring Algorithms
[6] and Linear Algorithms [10]. Acceptable solutions were reported in use of all of these techniques. However,
most of the time these solutions require man touch.
A mathematical model of university timetables is introduced as a linear program [10]. The timetabling problem
is a type of multi-objective satisfaction problem. Mainly the constraints are classified into two as hard and soft
constraints. The violation of these constraints denotes the quality of the developed solution. Therefore, the
solution space of the problem is huge and finding the optimal solution is a complex task. Studies show that
244
Genetic Algorithms find optimal solutions [5],[7]-[9] but they are complicated with respect to time and memory.
Morever, recently a computer tool is developed to solve the UCTP with tabu search algorithm and efficient
solutions were reported [3]. Another computer tool is developed for the generalized solution of UCTP using
Genetic Algorithms at Yeditepe University called TEDI [7].
2 Tabu Search Algorithm for UCTP
Tabu search is one of the commonly used metaheuristic algorithm for solving combinatorial optimization
problems, such as traveling salesman problem (TSP), university course timetabling problem (UCTP), etc. It
provides optimal or near optimal solutions for the practical problems. Tabu search algorithm searches the
solution domain by exploiting the neighbors of the current solution using a tabu list. Some of the basic
ingredients of the tabu search are a tabu list, neigborhood and aspiraiton criteria.
A tabu list keeps the list of not allowed and visited neighbors during the search process. The main aim of tabu
list is to avoid reversal of moves and cycling. Moves are performed between adjacent solutions that can be
reached from the current solution. If a tabu move results in better solution than any previously performed move
so far, than this tabu move may be ignored. A condition that allows such a move is called an aspiration criteria.
The timetabling problem constraints are mainly considered as hard constraints in which every timetable must
satisfy all these constraints, and soft constraints which are not essential for a timetable solution but it makes the
solution better when they considered.
Almost in all of the past studies similar hard and soft consraints are listed [1]-[3]. Depending on the universitys
limitations these constraints may vary. In this study, the number of classroom assignment is aimed to minimize
to reduce the university resource usage cost.
In this study, a novel tabu search algorithm is proposed to solve the timetabling problem with incremental
strategy. The proposed algorithm starts with a randomly generated solution to the UCTP and this initial solution
is improved by using moves that are searched in timeslots of classrooms incrementally where a best movement is
choosen in a classroom time table by satisfying the hard and soft constraints.Possible moves of tabu search
algorithm is performed for all courses of the active department incrementally. If there is any hard and soft
constraints violation, this solution gets worst fitness values. All periods in a classroom timetable is searched first
to fullfill the classroom periods. If there is no suitable period available, then a new classroom period is started to
search.
2.1 Parameters and Constraints of UCTP
The course timetabling problem deals with finding the exact time slots of courses within a limited periods,
classrooms and teachers. The problem parameters are given in Table 1. Courses in each department usually
organized in number of semesters. Constraints regarding to the semesters of each courses and about teachers are
organized as hard constraints and soft constraints. Hard constraints are the ones that must be obeyed. Soft
constraints are those constraints that make the solution better but they are not fully essential. The list of used
hard constraints in this study are:
No classroom can be assigned to more than one course at the same timeslot.
No teacher can be assigned to more than one course at the same timeslot.
Same semester courses cannot be assigned to the same timeslot.
245
Parameters
Number of courses
Number of rooms
Number of teachers
Number of days
Number of periods
Value
108
16
40
5
9
The proposed tabu search algorithm is illustrated in Fig. 1. Algorithm starts with a randomly generated solution
and searches the solution space for better solution based on hard and soft constraints within a tabu list. Moves
are incrementally choosen from the assigned timeslots. Two types of moves are used in the proposed tabu search
algorithm; simple move and simple swap which are illustrated in Fig. 2 and Fig. 3, respectively.
Start
Does thestoping
condition meet?
Y
Stop
Fig. 1.Tabu search algorithm: Incremental strategy
The simple move shifts the choosen course timeslot into another empty slot if there exits any in current room, if
not searches in other used rooms.
Periods
09:00
10:00
11:00
12:00
13:00
14:00
15:00
16:00
17:00
Mon
Tue
Wed
Thu
MAT101
MAT101
MAT101
MAT101
246
Fri
However, the simple swap swaps two courses blok of timeslots where they are assigned. Moves of the proposed
tabu search algorithm are marked as tabu once they have been performed. This leads to protect correct timeslot
assignments and to avoid recycling. Courses are incrementally choosen from the list of courses those are
violating the hard constraints for moves both simple move and simple swap.
Periods
09:00
10:00
11:00
12:00
13:00
14:00
15:00
16:00
17:00
Mon
Tue
ENG101
ENG101
Wed
Thu
Fri
CPE112
CPE112
Objective
Violation of hard constraints
Violation of soft constraints
Number of used rooms
Overall timeslots utilization
Man-made
solution
0
6
16
70,63%
Tabu Search
produced solution
0
8
12
80%
In order to compare solutions of UCTP similar parameters must be used. The number of courses that are tought
by instructors and the nunber of part-time instructors effect the hard constraints of the UCTP. However, when
the timeslots utilizations are considered with studies [1] and [3], the performance of these algorithms are given
as 80% and 95% respectively. Constraint-based reasoning strategy is investigated with 16 rooms for 136 courses
at the University of Yemen. There is no given information about the number of teachers in [1], therefore to
compare the performances of the proposed tabu search algorithm with [1] is difficult. In another tabu search
algorithm for the solution of UCTP is given in [3] having 57 courses, 6 rooms and 27 teachers. The workload of
the teachers in Statistics Department of Hacettepe University seems that 2 on the average. However, the
proposed tabu search algorithm applied with the workload of 3 different courses on the average for each teacher.
This may cause that the proposed tabu search algorithm has lower timeslot utilization than [3].
The developed software GUI screen shot is shown in Fig. 4. The implementation of tabu search incremental
strategy has been performed in C# programming language by using a laptop which has Intel Core 2 Duo 2.2 GHz
processor with 4 GB RAM.
247
5 Conclusion
The proposed tabu search algorithm is investigated for the solution of timetabling problem in Faculty of
Engineering at Cyprus International University (CIU). Timatabling problem is solved department by department
and all forms the timetabling solution to a faculty. Thereby, the whole problem search space is reduced to
consecutive departments curriculum courses.
One of the aim of this study was to develop a computer automated software which solves UCTP for the CIU
with optimum classroom usage. A user friendly computer graphical user interface developed to solve other
faculties timetabling problems in CIU.
The result of the proposed algorithm is compared with the man made solution of CIU Faculty of Engineering
timetable solution and also with other studies in other universities. The solution of the proposed tabu search
algorithm for the UCTP shows encouraging results withrespect to optimum resourse allocation and timeslots
utilization according to teachers workload as 3 courses on the average for each semester.
References
1. Irene, H.S.F., Safaai-Deris, Hashim, S.Z. (2009). Investigating Constraint-Based Reasoning for University
Timetabling Problem. In Proceedings of the International MultiConference of Engineers and Computer
Scientists. Hong Kong, 18 20 March. IMECS 2009.
2. Bakr, M.A., and Aksop, C. (2008). A 0-1 Integer Programming Approach to a University Timetabling
Problem. Hacettepe Journal of Mathematics and Statistics. 37(1): 41-55.
3. Alada, .A., and Hocaolu, G. (2007). A Tabu Search Algorithm to Solve a Course Timetabling Problem.
Hacettepe Journal of Mathematics and Statistics. 36 (1):53-64.
4. zcan, E. and Alkan, A. (2007). A Memetic Algorithm for Solving a Timetabling Problem: An Incremental
Strategy. In Proceedings of the 3rd Multidisciplinary Int. Conf. On Scheduling: Theory and Application.
Paris, France. 28-31 August. P. Baptiste, G. Kendall, A. M. Kordon, F. Sourd (ed.)
5. Nuntasen, N. and Innet, S. (2007). A Novel Approach of Genetic Algorithm for Solving University
Timetabling Problems: a case study of Thai Universities. In Proceedings of the 7th WSEAS International
Conference on Applied Computer Science. Venice, Italy, 21-23 November.
6. lker, ., zcan, E., Korkmaz, E.E. (2006). Linear Linkage Encoding in Grouping Problems: Applications
on Graph Coloring and Timetabling. In Proceedings of the 6th International Conference on the Practice and
Theory of Automated Timetabling. Brno, Czech Republic, 30 August- 1 September. LNCS 3867, Springer.
7. zcan, E. And Alkan, A. (2002). Timetabling using a Steady State Genetic Algorithm. In Proceedings of
the 4th International Conference on the Practice and Theory of Automated Timetabling. Gent, Belgium, 2123 August. LNCS 2740, Springer.
248
8. Burke, E., Elliman, D., Weare, R. (1994). A Genetic Based University Timetabling System. In Proceedings
of the 2nd East-West International Conference on Computer Technologies in Education.Crime, Ukraine,
September. EW-ED94.
9. Colorni, A., Dorigo, M., Maniezzo, V. (1990). Genetic Algorithms and Highly Constrained Problems: The
Time-Table. In Proceedings of the 1st International Workshop on Parallel Problem Solving from Nature.
Dortmund, Germany. LNCS 496, Springer.
10. Akkoyunlu, E.A. (1994). A Linear Algorithm for Computing the Optimum University Timetable.
Computational Journal. 16 (4):347-350.
Biographies
Mehtap KSE ULUKK- Born in Silivri in 1975. Graduated from Computer Engineering Department of
Eastern Mediterrenean University, Famagusta, T.R.N.C at 1998. Master and PhD programs were completed in
Computer Engineering Department from the same university in September 2000 and January 2009, respectively.
Artificial intelligence is the main research topic.
She has several publications about metaheuristic searches, expert systems and robot control.
Yasemin ERCAN- Born in Gaziantep in 1978. Graduated from Computer Studies in Information Technology in
Eastern Mediterrenean University, Famagusta, T.R.N.C at 2001. Master program was completed from
Information Systems Department of Faculty of Fine Arts in the same university in June 2003. She is currently
studiying in Management Information System Department of Cyprus International University for her PhD
degree. Internet programming and database systems are the main research topics.
She has some publications about expert systems, web programming with database systems and tabu search
algorithm.
249
Abstract. The Self Organizing Map (SOM) is a particular type of artificial neural network that is generally
used for unsupervised clustering of the input data. SOM algorithm is generally implemented as sequential in
which one thread carries out one instruction at once and this structure works fine in domains where the
dataset is small. In large-scale domains, it is necessary to deal with tens of thousands records and at this point,
the application of sequential SOM can become unfeasible in terms of computational time. This paper
proposes a novel SOM model, called PSOMDM that is parallel implementation using a special division
method. The PSOMDM is different from existing architectures in that it processes the entire data in parallel
by continuously dividing spaces into quarters. The comparison results show that novel approach improves the
accuracy and speed.
Keywords: self organizing map, neural networks, machine learning
1 Introduction
The Self Organizing Map (SOM) [1] is a neural network model that is capable of clustering high-dimensional
input data. It produces a two-dimensional feature map that can be useful in detecting and analyzing features in
the input space. SOM model has been successfully applied in a number of disciplines including pattern
recognition, image classification, and document clustering.
SOM algorithm requires the high execution times to train the map and this situation put a limit to its
application in many high-performance data analysis application domains. For this reason, it is necessary to
develop a parallel implementation of SOM by either partitioning the network among the processors (network
partitioning) or by partitioning the input data across threads or multi-core processors (data partitioning).
This paper proposes a novel approach for faster clustering by using of SOM, called PSOMDM (Parallel SOM
by using Division Method). PSOMDM is different from existing architectures in that it divides the map area
constantly and lower number of data is clustered on different neurons by parallel method. PSOMDM has many
advantages over conventional SOM based methods. The most remarkable advantage of PSOMDM is in saving
training time for clustering large and complicated data sets by using special division method. Furthermore, the
proposed division method provides higher accuracy by decreasing the number of unstable data points and
internal errors.
Experimental results with various test data setsand comparison results show that PSOMDM can be efficiently
used for clustering large datasets with SOM, somehow making technique move toward the more powerful
250
unsupervised technique. In the experimental studies, PSOMDM is implemented by dividing data processing into
a number of parallel threads and multi-core processors. The threads or multi-cores of parallel SOM model
process over the same neurons by partitioning data.
2 Related Works
The SOM performs clustering by means of unsupervised competitive learning. The neurons in the SOM are
usually arranged in a two dimensional lattice and each neuron gets information from the input layer and from the
other neurons in the map. The SOM has the complexity O(NC), where N is the input vector size and C is the
number of dataset presentation cycles. N contains n2w as the multiply of the map size n2 and the number of
weights w. C contains n2a as the multiply of the map size n2 and the number of attributes a. Usage of parallel
SOM supplies lower complexity than O(N2).
Several studies have been done related to parallel SOM by using different techniques and different
architectures. Garcia et al. proposed a speculative approach which fits better than traditional Map-Partitioning
strategies due to a better exploitation of the memory hierarchy [2]. Yu et al. proposed a multistage modular selforganizing map (SOM) model which can be used for parallel web text clustering [3]. Gorgonio and Costa
presents an approach for efficient cluster analysis in distributed databases using SOM and K-Means [4][5]. They
also proposed a SOM-based strategy for cluster analysis in distributed databases [6]. Another study of parallel
SOM is about mining massive datasets by an unsupervised parallel clustering on a GRID [7]. Liping et al.
analyzed a variety of software and hardware environment of designing artificial neutral networks on clusters [8].
Dagli et al. proposed a SOM model, called ParaSOM, which utilizes a feature called a cover region, in which
individual neurons cover whole regions of the input space, and not just a single vector [9].
Parallel SOM and its derivations were used in many different areas. For example, the study about dynamics of
soil organic matter and mineral nitrogen in soil was given as an application which used parallel SOM for
geological researches [10]. The study of M. Takatsuka and M. Buis presents the parallel implementation of
SOMs, particularly the batch map variant using Graphics Processing Units (GPUs) through the use of Open
Computing Language (OpenCL) [11]. Parallel SOM was also used for the computer security, the healthcare,
ecological modeling, the financial sector and other areas which need clustering.
Differently from the previous studies, this study proposes a new SOM model which is implemented in a
parallel way and using a different division method to gain greater computational efficiency. To the best of our
knowledge, this paper is the first on proposing a division the area into small areas for parallel processing and
different SOM architecture as defined in the next section.
3 PSOMDM Approach
SOM consists of two layers of artificial neurons: an input layer and an output layer. The input layer is fed into
feature vectors, so it is the same as the number of dimensions of the input feature vector. Output layer, also
called the output map, is usually arranged in a regular two dimensional structure such that there are
neighborhood relations among the neurons. Every neuron in input layer is fully connected to every output
neuron, and each connection has a weighting value attached to it.
The major goal of SOM is to determine the suitable weight values for the neurons according to dataset. The
count of these weight values is equal to the number of attributes in dataset and each weight value corresponds to
an attribute. At the beginning of the SOM algorithm, these weight values in all neurons are initialized randomly.
Secondly, the best matching unit as the winner neuron is found by calculating Euclidean distance from each
weight to the chosen sample vector which consists of the weights set in one of neurons. After finding the best
matching unit, all vectors of the SOM are updated by using Gaussian function. These processes are repeated until
a certain number of iterations. The loop accrues over the all neurons.
Usage of Parallel SOM supplies lower complexity than O(N2). According to the number of parallel threads or
multi-core processors, parallel SOM accelerates the process. However, the threads or cores of parallel SOM
process over the same neurons; therefore, the solutions which come from parallel SOM are at risk about lower
accuracy. This new approach of parallel SOM by using division method has higher accuracy and speed. Because
the random choosing the data in the training phase of SOM is abandoned and the data are chosen in the same
251
order for each time. The accuracy problem is passed over at connecting the subparts of maps.
This new approach firstly divides the area by four and this standard SOM processes for all little areas by
parallel processing. Thus, datasets are divided for all processes and the complexity becomes lower.
PSOMDB consists of standard SOM (SSOM). This algorithm is used for 2x2 neurons stably in each phase.
PSOMDB starts training with SSOM with 2x2 neurons and usage of all dataset. After that, the recursive structure
of PSOMDB is activated and recursively, SSOM for 2x2 neurons is processed with usage of divided datasets.
The following Fig. 1 shows the process-flow of PSOMDB for 4x4 neurons. There are four parallel maps and
they are trained for 2x2 maps. At the end of the PSOMDB, a 4x4 map is obtained after combining operation in
the same order before splitting.
The complexity of SSOM is O(N2). However, this formula is obtained by the assumption of the multiply of
the map size and weight numbers equal to the multiply of the tuple number and attribute number in the dataset.
Therefore, if N2 is splitted to NxC as N is the total size of map and C is the total size of dataset, the changes of
SOM speed according to the different datasets and its effects appears in more detail. When this formula is
splitted sub-components, these following formulas in Eq.1 and Eq. 2 are obtained;
T x A = C.
(1)
M x W = N.
(2)
whereT is the tuple number in the dataset, A is the attribute number for a tuple in the dataset, M is the total
neuron number in the map and W is the total number of weight variables for a neuron in the map.
T x A x M x W = .
(3)
(4)
whereX and Y are tuples in the dataset or neurons in the map and d is the attribute number if X and Y are tuples
or d is the total number of weight variables if X and Y are neurons.
Four pieces of dataset are used for training by SSOM algorithm in parallel. SSOM is processed for 2x2 neurons
(M=4) for each piece again. After these pieces are trained, these pieces are joined and the map with 4x4 neurons
is obtained. The process time calculation is done the following formulas:
(5)
252
(6)
(7)
If it is assumed that SSOM is processed for all dataset and 4x4 neuron, process time of SSOM is found by Eq. 3.
For 4x4 neurons, PSOMDB is processed for 2x2 neurons firstly and the process time of this phase is calculated
by Eq. 5. The result shows that this phase is equals to quarter of SSOM. However, PSOMDB continues to be
trained for 4x4 neurons and in parallel and for four pieces of datasets, PSOMDB is processed for 2x2 neurons
again. The process time of this phase is calculated by Eq. 6. Totally, the process time is calculated by Eq. 7 and it
shows that PSOMDB takes less time than SSOM as 5/16 times or nearly 1/3 times. This difference is kept even
if the size of the dataset gets larger. Also, if the map size gets larger, new components add to Eq. 7 and because
the dataset is divided again and again, these process times are approximate to zero a lot like. Thus, this
difference is kept. Theoretically, this is possible; however, the machine, where PSOMDB is processed, must
have enough numbers of cores to supply the parallel processing.
In section 4, experimental studies, these formulas are compared to the results when the real world datasets are
used for PSOMDB.
4 Experimental Studies
Our proposed approach was proved by three different datasetsfrom UCI machine learning [12] and successful
results are observed. The datasets are Iris, CPU and Segment- challenge. These datasets are for
classification operation; however, for accuracy tests, clustering algorithms are tested by usage of classification
datasets. Because training phase takes the dataset without the target attribute, then test phase compares the
derivation of pattern to the correct values in target attribute.
Iris, CPU and Segment-challenge have different features: Iris is a small-sized dataset. It has 150 tuples
and 5 attributes. The last attribute is the target attribute; therefore, this column is deleted in the training phase.
Thus, PSOMDB is trained with 4 attributes. CPU is a middle-sized dataset. It has 209 tuples and 7 attributes.
Like the content in the Iris dataset, the last attribute is the target attribute; therefore, training operation of
PSOMDB is implemented without this column. The third dataset, Segment-challenge is a larger dataset. It has
1500 tuples and 21 attributes. Like the content in the other two datasets, the last attribute is the target attribute
and PSOMDB is trained with 20 attributes.
PSOMDB and SSOM algorithms are implemented in Visual Studio 2010 platform by usage of C#
programming language. The implementation gives the results about speeds as Milliseconds with the result map.
The comparison of result maps shows that there are the same result maps with the same weight values.
Therefore, the accuracy does not change with PSOMDB. However, it is observed that the performance of
PSOMDB is better than the performance of SSOM.
Table 1 shows that there are three different datasets as Iris, CPU and Segment-challenge and two different
clustering algorithms as PSOMDB and SSOM. The algorithms are processed for 4x4 neurons. The performance
results of PSOMDB are given for all components. The first five rows in Table 1 are results for PSOMDB and the
last row is the results for SSOM because SSOM is processed for 4x4 neurons at one phase. However, the results
of PSOMDB are taken for 2x2 neurons firstly. Then, for each cell, {(0,0),(0,1),(1,0),(1,1)}, PSOMDB is
processed in parallel and for each cell, different results are obtained. The process time of the cell, which takes
the biggest time, determines the total process time. For example, (0,1) cell for Iris takes the biggest time as 376
ms, (1,0) cell for CPU takes the biggest time as 641 ms and (0,1) cell for Segment-challenge takes 3021 ms.
Finally, Table 1 shows that the total performances for each datasets are more successful than the performances of
SSOM.
253
Datasets
Iris
CPU
Segmentchallenge
441
747
3427
353
490
2990
376
562
2986
336
641
3021
339
454
2998
817
1388
6448
SSOM (ms)
1784
2755
11870
Fig.2 shows that PSOMDB is more successful than SSOM because the performances decrease nearly half and
half. For Iris dataset, the process time decreases from 1784 to 817, for CPU dataset, the process time decreases
from 2755 to 1388 and for Segment-challenge dataset, the process time decreases from 11845 to 6448. The most
important gain is that when there is a large dataset, PSOMDB increases the training performance and gives the
same result map in less time.
ms
th
nd
rd
In this paper, parallel implementation of SOM has been devised by using a special division method to attempt for
improving the performance of SOM in large-scale domains. It introduces a new clustering approach PSOMDM.
The significant difference between PSOMDM algorithm, the standard SOM and parallel SOM is that PSOMDM
divides the map area constantly and lower number data are clustered on different neurons by parallel method.
PSOMDM algorithm has many advantages over conventional SOM based methods. The most remarkable
254
advantage of PSOMDM is in saving training time for clustering large and complicated data sets by using
division method. Furthermore, the rate of unstable data points decreases and internal error decreases. For future
work, the proposed algorithm, PSOMDM, can be used for the computer security, the healthcare, ecological
modeling, the financial sector and another area which needs clustering its large data on a map successfully at
accuracy, consistency and speed.
References
1.
2.
Biographies
Yunus Doan received his MSc degree in Computer Science from Dokuz Eylul University in 2009. He completed MSc
thesis on data mining. He has been doing PhD degree in Computer Science in Dokuz Eylul University for 2009. Currently, he
is an assistant at the Department of Computer Engineering, Dokuz Eylul University in Turkey. His research interests include
data mining, bioinformatics, data warehousing, machine learning and knowledge-based systems.
Derya Birant received her PhD degree in Computer Science from Dokuz Eylul University in 2006. She completed PhD
thesis on data mining. Furthermore, she worked on five major projects related to data mining. Currently, she is an assistant
professor at the Department of Computer Engineering, Dokuz Eylul University in Turkey. Her research interests include data
mining, data warehousing, machine learning and knowledge-based systems.
Alp Kut is a full professor of computer engineering at Dokuz Eylul University. He is head of the Department of Computer
Engineering of Dokuz Eylul University since the fall of 2003. His works include data mining in databases, database
management systems and distributed systems. He has many publications on a variety of topics, including, data mining,
artificial intelligence, web-based systems and parallel systems.
255
Abstract.The aim of this study is to create height maps that will be used in 3-dimensional applications.
Height maps are created with 2 ways. The first method aims to create height maps with random data. Random
algorithms have various controls to create a model which is similar to the real world. In the second method,
height maps are created through real-world data. For this purpose, modeling related to real world data was
realized utilizing the most commonly used geographic map files like USGS DEM, DTED and SRTM. In
addition, users were allowed to do various real time modifications on this model. Furthermore, options such
as increasing or decreasing the height value of any point on the model and saving the changes with the same
format were provided to users.
Keywords: Geographical Information System, A-Real Time Application, Height Map Generation and
Modification, Random Fault Algorithm, Random Circle Algorithm
1 Introduction
Digital elevation maps are structures which are used to represent the topographic surface of the earth.
Creation and usage areas of these maps have increased with the development of geographical information
systems. Moreover, height maps are data structures which are also used for 3D applications such as military or
civilian simulators, games and other applications aiming to create virtual worlds [1].
Kamal and Udin have presented a technique which controls minimum level to produce height maps
model on demand. This technique produces the terrain by creating continuous random polygons [2]. Lencher et
al. have described a method for procedurally generating typical patterns of urban land use using agent-based
simulation. Users give a terrain description; their system returns a map of residential, commercial, industrial,
recreational park and road land uses including age and density of development [3]. Doran et al. explore more
controllable system that uses intelligent agents to generate terrain elevation height maps according to
designer-defined constraints. This allows the designer to create procedural terrain those specific
properties [4].
Li et al. provide a semiautomatic method of terrain generation that uses a four-process genetic
algorithm approach to produce a variety of terrain types using only intuitive user inputs [5]. They gave the
sample terrain dataset to system which will produce new height map. Brosz et al. present a technique of terrain
256
synthesis that uses an existing terrain to synthesize new terrain. They use multi-resolution analysis to extract the
high-resolution details from existing models and apply them to increase the resolution of terrain [6].
Gkta et al. have devised an algorithmic system which automatically generates virtual cities supporting
several different map formats (i.e. DTED, HTF, SRTM, and DEM). Depending on the conditions submitted, the
system generates different layouts for the cities even on the same geographical site [7], [8]. Wiggenhagen have
produced software for the inspection and quality control of digital terrain models and orthoimages. He uses data
standards like USGS and BMP-Format [9]. Errors found can be corrected through interpolation methods.
Dewen et al. discussed the implementing methods for the representation of 3D terrain realistic scenes
based on OpenGL, such as the digital terrain model mapping, the blend of water and terrain data, the blend of
thematic information and terrain [10].
Chi et al. presents an improved method called property-control IDW (PCIDW) for 3D modeling of
complex terrain. This method regards property of the sampling points and the unknown points to be estimated as
a parameter to control whether the sampling points involves in modeling [11].
Noghani et al. describe a variety of methods that can be used to create realistic, random 3D
environments for serious games requiring real-time performance. These include the generation of terrain,
vegetation and building structures [12]. An alternative method to create randomized terrain is the use of Linden
Mayer systems.
L-systems were originally created to study organic growth, such as is the case with plants,
but they can easily be adapted to cover other self-similar structures, such as mountainous terrain [13].
2 Algorithms of Creating Random Terrains
We use two algorithms for creating random terrains .First algorithm is Fault and second is Circle Algorithm.
Both of algorithm starts with nxn dimensional matrixes which have an initial zero value. Then Fault algorithm
selects a random line which divides the terrain in two parts. Points to one side of the line will have their height
displaced upwards, whereas the points on the other side will have their heights displaced downwards [14], [15].
But Circle algorithm creates different size of circles on the height maps [15]. Fig. 1 shows the result of these
algorithms.
Fig. 1 Sample terrains a is created Fault and b is created circle (1000 iterations)
Modeling water flow or mass movement (for example avalanches and landslides)
Creation of relief maps
Rendering of 3D visualizations.
3d flight planning
257
LIDAR
Real Time Kinematic GPS
Topographic maps
Doppler radar
Focus variation
The SRTM data sets result from a collaborative effort by the National Aeronautics and Space
Administration (NASA) and the National Geospatial-Intelligence Agency (NGA - previously known as the
National Imagery and Mapping Agency, or NIMA), to generate a near-global digital elevation model (DEM) of
the Earth using radar interferometry. The SRTM instrument consisted of the Space borne Imaging Radar-C (SIRC) hardware set modied with a Space Station-derived mast and additional antennae to form an interferometer
with a 60 meter long baseline [18], [19] and [20].
GTOPO30 is a global digital elevation model (DEM) with a horizontal grid spacing of 30 arc seconds
(approximately 1 kilometer). GTOPO30, completed in late 1996, was developed over a three year period through
a collaborative effort led by staff at the U.S. Geological Survey's EROS Data Center (EDC) [21], [22], [23] and
[24].
With this study we can simulate SRTM, DTED and GTOPO30 files. Fig 2 shows the step of simulation.
First step is creating the height values from the GIS based files. It is the important step because all GIS file have
different standards. After getting the height values program simulate it.
258
the triangle is found out. For this aim, 3 dimensional coordinate values of all points in the height map are
converted into OPENGL unit volume reference system. Fig 3 shows the formula.
P00
x
P
y
Vunit
P 10
P20
z
w
P30
P01
P11
P02
P12
P21
P31
P22
P32
P03
M 00
M
P13
xM 10
M 20
P23
P33
M 30
M 01
M 11
M 02
M 12
M 21
M 31
M 22
M 32
M 03
x
y
M 13
xV
z
M 23
M 33
1
Fig 3.This equitation shows the transformation of Vertex values into the OPENGL unit cube values V unit:
OPENGL unit volume reference system, P: OPENGL Projection Matrix; M: OpenGL Modelview Matrix; V: any
point.
After converting process, the point selected by users was being passed through an intersection test
considering all triangles on the model. At this stage, a new equation based on Barycentric coordinate plane was
obtained in order to see whether the point is on any triangular [25], [26]. Formula 1 shows any point on the
triangle and formula 2 shows any point on the straight. (Fig.4)
( )
(
[Fig.4.
] Barycentric
[ ]
[coordinate
]
[27].
)( )
The values of equation obtained are transformed into matrix form. Gauss Elimination method is utilized in order
to solve Equation-1. If the values of equation solved with this method match with the Barycentric coordinate
plane, the position of the point selected is found out on the relevant triangle.
After the process of finding triangle we need to create a new point on this terrain. For this aim we use a
resolution value which is need to resize terrain. With this resolution value, Algorithm divides terrain into small
parts and creates new points on so we can create a high resolution terrain from low resolution. For resizing the
terrain we use formula 3. D variable in formula 3 means dimension of terrain.
( )
After creating new points, Algorithm uses bi-linear interpolation (formula 4) to find the height values of them
[28], [29] and [30].
259
( )
(
(
)
(
)
(
(
)
)
(
)
( )
Changes on the terrain dimension are shown Fig.6. Left picture shows a part of terrain before and right is show
after modification of height map.
260
Fig.7 summarizes the system. After entering the system two options can be chosen. First option creating terrain
random and creating terrain from GIS based option. With the help of render mode desired modification can be
done.
Findings and Conclusion
With this paper, USGS DEM and SRTM files are modeled 3D. But these files have low resolution and this
causes disadvantage. But with this study, users are able to create high resolution height maps from low resolution
height maps. Unlike other studies, users are allowed to make modifications in real time on this model. Users can
increase or decrease the height of any point on this model. Besides, users can save their changes on this model
with the same file format and reuse them whenever they want. For future works, we propose to increase
performance with parallel programming methods that we have planned to use NVDIA CUDA technology.
References
[1] GNDODU, K. S. (2003) Saysal Ykseklik Modellerinin Arazi Boy kesitlerinin karlmasnda Kullanm Uluda
niversitesi, Ziraat Fakltesi Dergisi, 17(1): 149-157, Bursa.
[2] Kamal, K. R. and Udin, Y. S. (2007) Parametrically controlled terrain generation. In GRAPHITE'07: Proceedings of the
5th international conference on Computer graphics and interactive techniques in Australia and Southeast Asia, pages 17-23,
New York, NY, USA, ACM.
[3]LECHNER T., REN P., WATSON B., BROZEFSKI C.,WILENSKIU.: Procedural modeling of urban land use. In ACM
SIGGRAPH 2006 Research posters (2006), ACM, p. 135.
[4] Doran, J. and Parberryy, I; (2010) Controlled Procedural Terrain Generation Using Software Agents Dept. of
ComputerScience & EngineeringUniversity of North Texas January 19
[5] Li, Q. ; Wang, Q.; Zhou, F.; Tang X. and Yang, K. Example-based realistic terrain generation. Lecture Notes in
Computer Science, 4282:811, 2006.
[6] Brosz, J. ; Samavati, F.F. and Sousa, M.C. (2006) Terrain Synthesis By- Example, Proc. First Intl Conf. Computer
Graphics Theory and Applications (GRAPP ).
[7] GKTA, H.H.; AVUOLU, A.; EN, B.; GRGNOLU, S. ; (2006) "Simlasyon Sistemleri in 3 Boyutlu
Sanal ehirlerin GIS Haritalar zerinde Oluturulmas", Teknoloji/Z.K.. Karabk Teknik Eitim Fakltesi Dergisi, 9(1);
27-38;.
[8] GKTA, H. , AVUOLU, A., EN, B., (2009)"AUTOCITY : A System for Generating 3D Virtual Cities for
Simulation Systems on GIS Map", Auto Soft - Intelligent Automation and Soft Computing, Vol.15, No:1, pp.29-39, USA,.
[9] Wiggenhagen, M. (2000) Development Of Real-Time Visualization Tools For The Qualty Control Of Dgtal Terran
Models And Orthoimages International Archives of Photogrammetry and Remote Sensing. Vol. XXXIII, Part
B3.Amsterdam.
[10] Dewen, S; Hongxia, W; Realistic Real-time Rendering of 3D Terrain Scenes Based on OpenGL
[11] Chi, M; Liu, Z; Wei, Z.; (2010) An Improved Method for 3D Modeling of Complex Terrain 2nd Conference on
Environmental Science and Information Application Technology
[12] Noghani, J; Liarokapis, F; Anderson E. F (2010) Randomly Generated 3D Environments for Serious Games Second
International Conference on Games and Virtual Worlds for Serious Applications
[13]. Prusinkiewicz, P;Lindenmayer, A. (1990). The Algorithmic Beauty of Plants, Springer-VerlagNew York
[14] http://www.lighthouse3d.com/opengl/terrain/index.php?fault
[15] Che Defu; Wu Lixin; Xu Lei;(2006); Study on 3D Modeling Method of Faults Based on GTP Volume, Geoscience and
Remote Sensing Symposium, IGARSS 2006. IEEE International Conference on 10.1109/IGARSS.2006.411 Page(s): 1594
1597
261
Biographies
Serpil EROLU was born in Diyarbakr, 1987. She received her PhD degree in Computer Science from HarranUniversity.
She interested in geographical information systems, 3D modelling and simulation systems. She works for KarabukUniversity,
Department of Computer Engineering.
Baha EN received his B.Sc. in Computerfrom GaziUniversity, Ankara/Turkey in 1996. He received his M.Sc. degree from
Institute of Science and Technology, GaziUniversity in 1999, and his PhD degree from same department. His research
interests include graphics, vision, genetic algorithms, expert systems, geographical information systems, 3D modelling and
simulation systems. Asst. Prof. Dr. EN works for KarabukUniversity, Department of Computer Engineering.
Salih GRGNOLU- was born in orum, 1970. He graduated from GaziUniversity in 1995. He received his M.Sc. and
Ph.D. degree from the same university. In 2000 and 2007 respectively. He worked as instructor at GaziUniversity between
the years 1999-2007. His areas of interest include biometric systems, image processing, microprocessors and neural network
applications. Asst. Prof. Dr. Grgnolu still works for KarabukUniversity, Department of Computer Engineering.
262
Abstract. The aim of this project is to integrate a system to a Command and Control Center(CaCC) and
ambulances, for automatize the managerial functions of ambulance redirection, using Geographic Information
Systems (GIS) supported with Computer-Aided Design (CAD). To achieve this, first the execution phases of
the entire model is determined. Service-based (oriented) architecture (SOA) was chosen for the integration of
the systems. Thereby, a web-service is implemented to ensure communication with GPRS. When an on-site
assignment is made from CaCC, this web-service notifies the coordinates of the event taking place to the GIS
application. This application specifies the ambulance's Geographical Position with GPS information and
draws the shortest-path to the destination's location. Selected two-dimensional map is composed of nodes and
links. The main idea is to determine the shortest path or fastest path between the ambulance and the hospital.
Keywords: GIS Application, CAD, Dijkstra Algorithms, Public Health Care, Navigation, Fastest Path,
Shortest Path
1 Introduction
In 1736, graph was used by Euler to solve the Koningsberg seven bridges problem and then graph theory came
into being. But it was not until middle and later ages in 20 th century did mathematics and computer scientists
attach much importance to graph theory with the help of appearance and developing of computer [1], [2].
Nowadays, many applications have been developing while using graph theory, which is applied to many fields
such as GIS network successfully.
GIS can be used to analyzing, processing, capture, store, handle, and geographically integrate large amounts of
information from different sources, programmes, and sectors; including epidemiological surveillance, census,
environment, and others [3], [4].
With the popularity of the computer and the development of the geographic information science has been
increasingly extensive and in-depth applications for its powerful functions [5] and also usage of Health Care
Systems.
Public Healthy policy covers decisions in any sector or at any level of government and is characterized by an
explicit concern for health and accountability for health impact [6]. It is an established fact that most developing
countries are not spending more than 2% of their gross national product (GNP) on health, resulting in poor
coverage of public health services [6].
In this paper, several technological keys are established the road topology structure is discussed, and a
algorithm for finding the shortest and fastest paths for public health care system based on GIS which integrated
the communication service. Then, many tests were conducted to show the propose system performance.
263
2 Background
Nowadays, GIS aided systems are used all areas as health sector. Using computer-aided systems on the
contrary performing all the processes needed in a health center on paper provides in terms of work power gain
and to minimize the possibility of making mistakes. In addition, it is much more easy and quick to have access to
any information and data held on computers. Due to, this type of advantages the computer-aided systems have
been used in the health sector. Health care processes require a race against time and adapting computers instead
of traditional methods into our business and our lives.
Due to their structures Emergency Medical Services (EMS) systems must have some characteristics like realtime response, seamless roaming and decentralization. In order to provide real-time response of medical services,
several components are essential, such as dispatching, routing, communications between emergency facilities,
and ability to attain patients medical history quickly. With the advancement of communication technology and
GISs, all these processes can be automated.
GIS research and its applications in public health field have many examples in different countries. GIS has
been widely applied in Chinas public health, serving in disease surveillance, epidemiology, environmental
health research, research of health care utilization and decision making, respond to health emergencies and other
aspects [7]. Another research investigates the appropriate model for road traffic accident management is applied
in Chon Buri province of Thailand [8]. As a similar project report the development of a geographic information
system (GIS) for Texas-Mexico border disease surveillance and environmental health research [9]. In these type
of applications, most common approach is to select commercial 3D maps in GIS to maximize the visual
efficiency. Bu on the other side choosing the "Open API" maps such as Google Maps Open API can be an
alternative way. Implementation of a Real-time Ambulance Location Monitoring System using Open API based
on web 2.0 shows that Google maps were offered more detail geographical information than other maps [10]. It
were also easily integrated visual studio programming.
Based on this view, GIS application was implemented for ambulances which integrated with Command and
Control Center. A tablet PC with a mobile operating system is selected to host proposed application.
3 Method
EMSs are consist of four main entities. All communication takes place between a Command and Control
Center(CaCC), ambulances, hospitals and unfortunately accidents. To speed up the communication between
these assets dramatically improves the chances of the patient's chance of survival.
As illustrated in Figure 1; everything starts with the phone indicating the status of the accident to the CaCC
(112 in Turkey). CaCC automatically redirects the nearest ambulance to the crash scene. Ambulances are
equipped with tablet PCs to operate GIS applications. Assigned address shown within the GIS application and
the route is automatically determined due to traffic and distance factors. When ambulance reaches to the scene,
ambulance doctor performs the first diagnosis to the patient. Then enters the injury details to the application.GIS
applications finds the nearest most suitable hospital and draws a route. On the way of hospital, injury details and
medical status of the patient is transferred to the hospital via CaCC.
264
DO event assign;
GO end point location
WHILE <patient in ambulance>
function<DetermineHospital>(< preliminary diagnosis>)
CHOOSE Hospital;
IF <No doctor related branch>
function<DetermineHospital>(< preliminary diagnosis>)
ELSE IF <no empty beds>
function<DetermineHospital>(< preliminary diagnosis>)
ELSE IF <no empty operating room>
function<DetermineHospital>(< preliminary diagnosis>)
returnCaCC;
265
Because of routing is the most effective part to the elapsed time, selection of the route becomes the most
important phase of the proposed model. At this point, the algorithms used in GIS applications plays critical role.
In GIS, there are many algorithms to determine shortest path; one of the most classic is Dijkstra algorithm.
Proposed system has been performed by using Dijkstras algorithm that consists of a weighted graph and a
source vertex. Each edge of the graph is an ordered pair of vertices and the cost of an edge can be thought if as
the distance between the two vertexes. Based on this information for calculating distance with minimum cost,
look at steps to find out how to calculate the shortest path;
1.Set initial node as current. (Weights have to be assigned for all possible nodes).
2.All nodes are marked as unvisited.
3.For current node, consider all its unvisited neighbours and calculate their distance.
4.After finish the considering all neighbours of the current node, mark it as visited.
5.If all nodes have been visited, finish the search and choose the smallest distance considering all nodes
in graph.
References
1.
2.
3.
4.
5.
6.
7.
8.
9.
Chen W.K. (1997). Graph theory and its engineering applications. Advanced Series in Electrical and
Computer Engineering.
Deng-feng C., Deng-rong Z. (2001). Algorithm and its application of N shortest paths problem. In Infotech and Info-net, 2001. Proceedings. ICII 2001 - Beijing. 2001 International Conferences. Issue Date:
2001 Volume: 1 On page(s): 387 - 391 vol.1
Hussain M, Arsalan M.H., Mehdi M.R. (2006). Role of GIS in Public Health Management in Pakistan,
In Advances in Space Technologies, 2006 International Conference. Issue Date: 2-3 Sept. 2006 On
page(s): 102 - 107
Xuri Y. (2010). GIS-based Simulation System for Wartime Military Highway Transportation,
Computer Science and Education (ICCSE), 2010 5th International Conference, 978-1-4244-6002-1
Fuhao Z., Jiping L. (2009). An Algorithm of Shortest Path Based on Dijkstra for Huge Data. In Fuzzy
Systems and Knowledge Discovery, 2009. FSKD '09. Sixth International Conference. Issue Date: 14-16
Aug. 2009 On page(s): 244 - 247
Voncina L., Brborovic O., Pavlekovic G. (1986). Charter adopted at the First International Conference
On Health Promotion: The move towards a new public health, November 1721, 1986 Ottawa, Ontario,
Canada. 1986. Geneva. World Health Organization, 1986 (WHO/HPR/HEP/95.1).
Zhen Z., Jing-min J., FangMayer, L. (2010). The application of geographic information system (GIS)
in the field of public health. In Geoscience and Remote Sensing (IITA-GRS), 2010 Second IITA
International Conference Issue Date: 28-31 Aug. 2010 On page(s): 442 - 445
Rudtanasudjatum, K. (2008). Road traffic accidents management model: The successful application of
stakeholder involvement in Chon Buri, Thailand. In ITS Telecommunications, 2008. ITST 2008. 8th
International Conference. Issue Date: 24-24 Oct. 2008 On page(s): 195 - 198
Zhan, F.B.; Yongmei Lu; Giordano, A.; Hanford, E.J. (2005). Geographic information system (GIS)
as a tool for disease surveillance and environmental health research. In Services Systems and Services
Management, 2005. Proceedings of ICSSSM '05. 2005 International Conference. Issue Date: 13-15
June 2005 On page(s): 1465 - 1470 Vol. 2
266
10. Doyoon K, Jungchae K., DongKeun K., Minhong C., Sun K.Y. (2008) Real-time Ambulance Location
Monitoring using GPS and Maps Open API. In Engineering in Medicine and Biology Society, 2008.
EMBS 2008. 30th Annual International Conference. Issue Date: 20-25 Aug. 2008 On page(s): 1561 1563
Biographies
Fatma Patlar Fatma Patlar was born in Istanbul, Turkey, on May 14, 1981. She graduated from Istanbul Kltr University
in 2005 with a degree in Computer Engineering. After graduating with a B.S., Patlar worked for three years as a software
developer and participated in speech recognition and GIS projects. At the same time, she started a masters degree in
Education at Istanbul Kltr University in the Department of Computer Engineering. After finishing her MS degree, she was
recruited as a Technical Analyst by Aras Holding. After working almost a year at this company, she worked as a Business
Analyst at Garanti Technology. Now, She is working as a Research Assistant at Istanbul Kltr University in Istanbul. Her
research areas include software engineering, signal and speech processing and its applications. She has 5international paper
conference publications.
Akhan Akbulut BSc in Computer Engineering (Istanbul Kltr University 2001), MSc in Computer Engineering (Istanbul
Kltr University 2008), PhD candidate in Computer Engineering (Istanbul University 2009 - ~). He has worked as
senior/lead software developer and project manager in software sector and participated in B2B, B2C, e-commerce, e-state
and portal projects. Currently, he is working in Computer Engineering Department of Istanbul Kltr University as a research
assistant. Research areas includes Software Engineering, Smart Applications (agents, spiders, bots), Security, Web-Services
and Database Management Systems topics. He has 10 international and 2 national paper conference publications.
267
Abstract. Starting from a review of available literature, the effects of various factors on the duration of
software process improvement programs are investigated. The hypotheses formulated based on the
experiences reported in the literature provide a baseline for the qualitative research carried out on the CMMI
Level 3 organization. Within the framework of this study it has been observed that Staff Involvement,
Awareness, Management Commitment and Management Involvement have had a shortening effect on SPI
duration.
1 Introduction
The last decade has seen many organizations striving to achieve software development process maturity
through certification within the Capability Maturity Model (CMM) framework. The resources required for such
improvement have been studied extensively in the literature (see e.g. [1],[2],[3]) but the duration of process
improvement and the factors that affect the time span for reaching the next level still seem to be relatively less
investigated.
There are some successful cases that have managed to decrease the time to move up from one CMM level to
another drastically [4], [5], [6]. Analysis of these success stories may be helpful in identifying a relationship
between various factors and SPI duration. Such a relationship may help managers in planning their SPI effort;
strengths and weaknesses of the organization may be determined and resource allocation for the program can be
enhanced.
There have been a number of studies that try to identify factors that effect SPI success. However, these studies
do not provide answers to questions about the effect of these factors on SPI program duration. This study aims to
study the factors that effect SPI duration. The results of this study may provide guidelines to organizations that
wish to accelerate their CMMI work.
This paper is organized as follows. The relevant literature is reviewed and the hypotheses derived from the
literature are presented in Section 2. Details of the case study design like sample profile, data collection and data
analysis methods are explained in Section 3. Findings of the case study carried out to examine the hypotheses
derived from the literature are presented in Section 4. Section 5 concludes the paper.
268
2 Literature Review
Wilson et al. [7] propose a framework for the evaluation of SPI success. They adapt and apply a framework
which was previously developed for evaluation of metrics programs, to SPI. They identify the success factors as
the management commitment, respect, initial process definition and explanations.
Niazi et al. [8] aim to propose a maturity model for the implementation of SPI programs. Niazi et al. [9] in a
follow-up study, propose a framework that will provide companies with an effective SPI implementation
strategy. In the maturity model and framework proposed in these two studies, critical success factors are
identified as management commitment, training, staff involvement, awareness, experienced staff, formal
methodology, and reviews.
Petterson et al. [10] have developed a light-weight process assessment framework. While developing the
framework, within the several critical success factors mentioned in the previous studies reviewed, the ones
related to the study are given as SPI initiation threshold and commitment and involvement.
Cares et al. [11] propose an agent-oriented process model and identify the most frequently cited thirteen
critical success factors, whereas Berander and Wohlin [12] name the key factors for successful management and
evolution of the software process as baselining, synchronization, user involvement, management commitment,
change management, and documentation.
Dyba [13] has proposed a conceptual research model to investigate the relationship between SPI success and
the factors defined in the model. Based on that model, it is concluded that employee participation, business
orientation, concern for measurement, exploitation of existing knowledge, involved leadership, and exploration
of new knowledge effect the success of SPI.
In addition to the studies discussed above, experience reports about CMM/CMMI studies also provide detailed
information about the settings and conditions in which various SPI exercises have been carried out. Six success
and two failure stories reported in those studies are examined in detail [4],[5],[6],[14],[15],[16],[17],[18]. The
common points reported in these research papers were observed as quality environment, management
commitment, SPI awareness, participation, training, and the existence of experienced people involved with the
process improvement endeavors. On the other hand, the reasons that lead to failure were identified as lack of
management commitment, lack of quality environment, lack of SPI awareness, and lack of training. The aim of
the present study is to investigate the relationship between these factors and SPI duration, so the dependent
variable of our model is selected as time to complete SPI successfully. Based on the literature review, the
independent variables were selected as Management Commitment, Awareness, Staff Involvement, Training,
Experienced Staff, Metrics and Measurement, Process Documentation, and Quality Environment.
3 Research Design
3.1 Sample Profile
The case was selected from the CMMI-Level 3 awarded companies in Turkey. Organization A, which is a global
firm that provides solutions and services for information and communication technologies, was awarded as
CMMI-Level 3 in 12 months time.To safeguard privacy, the names of the organizations have been withheld in
the sequel.
3.2 Data Collection Method
Data was collected in two phases. The first phase consisted of formulating the initial hypotheses based on the
literature. In the second phase of the research, data was collected through semi-structured interview that lasted
around 45 minutes. There was at least one question about the existence of each identified factor in the
organization in the interview. These factors were Quality Environment, Experienced Staff, Management
Commitment, Awareness, Staff Involvement, Process Documentation, Training and Metrics, as selected based
on the literature review. In addition to these factors, sub-factors were also identified to analyze reasons for
acceleration of CMMI. For example, Quality Environment is associated with the sub-factors Parallelism between
269
Standards, Frequency of Assessments, Gap Analysis and Class-B Appraisal. Sub-factors were defined based on
the literature review and interviews.
3.3 Data Analysis Method
During the analysis of the interview, for each factor and sub-factor, a score taking either one of the three values:
none-low, medium, or high was assigned according to the answers to interview questions. A score of none-low
means that the factor does not exist in the organization or exists only weakly. A score of medium means that
the organization somehow practices the factor. A score of high means that the organization utilizes the
practices of the factor at a significant level. The score matrix is given in Table 1.
Organisation A
Quality Environment
high
medium
Frequency of Assessments
high
Gap Analysis
high
Class-B Appraisal
high
Experienced Staff
high
CMM/CMMI Experience
none or low
medium
Consultant
high
Reviews
high
Staff Involvement
high
Awareness
high
Resistance to Change
medium
Rewarding
medium
Training
high
Annual Training
high
Training Plan
medium
medium
Metric Analysis
medium
270
medium
Management Commitment I
high
Management Commitment II
high
Management Involvement
high
Process Documentation
high
4 Findings
Organization A had previously utilized ISO 9001:2000, AQAP 160 and internal standards. Being part of a
global company enabled them to benefit from previous experiences. Prior to the CMMI initiative, around 4050% of the process definitions were already compatible with CMMI.
Frequency of Assessments & Reviews: Prior to SCAMPI A assessment, Organization A performed Class B
and Readiness Review. In addition to these formal assessments, internal reviews were performed every two
weeks. Moreover, progress reports were prepared and staff was informed about the progress of the CMMI work
by team leader, group manager, and department manager.
Experienced Staff: SPI team members did not have previous SPI experience but they were enthusiastic about
SPI. They had been working in the company for 2-3 years and they had, on the average, about five years of
experience on software project development. Within the team members, the group manager had 10-year
experience on software development which helped the team members a lot during the CMMI program. Prior to
SPI, staff opinion on SPI was positive.
Separation of Process and Products: SPI team members were responsible from defining the rules, processes,
standards, template, etc. While selecting the team members, it was important that they were experienced, open
minded and adopted the idea of continuous SPI. SPI team was composed of a group manager, a configuration
manager, a quality manager, a project manager and software developers. Since they were experienced on the
related process areas, group diversity had positive effects on CMMI studies.
While preparing the work plan, resource allocation was done based on the experience of the employee. Each
team member was responsible from the process area in which s/he was experienced. Resource allocation to
process and product activities was done with alternating priorities such that when it was more critical to deliver
the product, the SPI team were assigned to product development, in other times they focused on process
improvement.
Staff Involvement and Awareness: Prior to the SPI program, management was aware of the benefits, outcomes
and required budget; staff was partially knowledgeable about the benefits of SPI.
During the interview it was mentioned that staff were aware of their responsibilities and workload. It was also
added that the SPI team worked overtime when necessary without questioning.
During the program, the parallelism between the improvements and corporate culture was taken into account.
Improvements which were not parallel with the corporate culture were withdrawn.
Training: Besides mandatory CMMI trainings, a systematic training schedule was not available. However,
after CMMI certification, a regular training program was planned.
Metrics and Measurement: Prior to the CMMI program, Organization A did not collect metrics. Even though
some of the project groups had collected some internal data, they did not use that data in project or SPI planning.
Therefore, it can be said that there was no practice of measurement and analysis.
Even though they did not have a previous metrics practice, during the metrics selection, they had the
advantage of being a global company. They used the metric documentation template which was prepared by the
main office. The availability of such documentation enabled them to overcome the difficulties of metric selection
and analysis.
Management Commitment and Involvement: Management provided full support and commitment for CMMI
activities. When necessary, they approved the budget revisions related to training program and resource
allocation. They also attended the meetings, which was a sign of their participation in CMMI studies.
271
5 Conclusions
In this paper, factors that affect the duration of CMMI certification programs were investigated. First of all,
factors of SPI success were identified from the literature. In the second phase, interview with a CMMI-Level 3
company was held in order to identify the factors that effect CMMI program duration.
Organization A had no previous CMM-CMMI experience. However, the interviewee from Organization A
mentioned that even though they were not experienced about CMM/CMMI, but they took the advantage of being
were enthusiastic about SPI processes and knowledgeable about software development processes.
When the factor Management Involvement is examined, in Organization A management actively participated in
CMMI meetings and provide necessary feedbacks when necessary. It can be concluded that when the
management take on responsibility, an accelerated CMMI program can be achieved. When it comes to
Management Commitment, Organization A stated that management provided a strong leadership and
commitment for SPI program and approved the changes in the budget when necessary.
Within the framework of this study it has been observed that Staff Involvement, Awareness, Management
Commitment and Management Involvement have had a shortening effect on SPI duration. A qualitative research
method has been applied throughout the study. Further validation of effect of factors on CMMI duration would
be possible through extensive and quantitative research, which would provide valuable guidelines to the
industry, but it is outside the scope of the current study.
References
1.
2.
3.
4.
5.
6.
7.
8.
9.
10.
11.
12.
13.
Diaz, M., & Sligo, J. (1997). How Software Process Improvement Helped Motorola, IEEE Software 14,
75-81.
Herbsleb, J., Carleton, A., Rozum, J., Seigel, J., & Zubrow, D. (1994). Benefits of CMM-Based
Software Process Improvement: Initial Results, (CMU/SEI-94-TR-013), Pittsburg, Pennsylvania:
Carnegie Mellon University, Software Engineering Institute
Herbsleb, J., Zubrow, D., Goldenson D., Hayes, W., & Paulk, M. (1997). Software quality and the
Capability Maturity Model, Communications of the ACM 40, 30-40.
Akmenek, F., & Tarhan, A. (2003) The Leap to Level3: Can It Be a Short Journey?, SEI- ESEPG
Conference, London, England.
Tufail, Z., Kellum, J., & Olson, T. (2006). Rapidly Defining a Lean CMMI Maturity Level 3 Process,
6th Annual CMMI Technology Conference & User Group, Denver, Colorado.
Zeid, S. S. (2004). Moving from CMM level-2 to CMM level-3, Egypt -SPIN Newsletter, (Issue 6), 3-8.
Retrieved April 17, 2008 from http://www.secc.org.eg/SPIN%20Newsletter.asp
Wilson, D., N., Hall, T. & Baddoo, N. (2001). A Framework for evaluation and prediction of software
process improvement success, The Journal of Systems and Software 59, 135-142.
Niazi, M., Wilson, D., & Zowghi, D. (2005a). A maturity model for the implementation of software
process improvement: an empirical study, The Journal of Systems and Software 74, 155-172.
Niazi, M., Wilson, D., & Zowghi, D. (2005b). A framework for assisting the design of effective
software process improvement implementation strategies, The Journal of Systems and Software 78,
204-222.
Petterson, F., Ivarsson, M., Gorschek, T., & Ohman, P. (2007). A practitioners guide to light weight
software process assessment, The Journal of Systems and Software 81, 972-995.
Cares, C., Franch, X., Mayol, E., & Alvarez. E. (2006). Goal-Driven Agent oriented Software
Processes, Proceedings of the 32nd EUROMICRO Conference on Software Engineering and Advanced
Applications, Cavtat/Dubrovnik, Croatia, 336-342.
Berander, P. & Wohlin, C. (2003). Identification of Key Factors in Software Process Management A
Case Study, Proceedings of the 2003 International Symposium on Empirical Software Engineering
(ISESE03), Rome, Italy, 316-325.
Dyba, T. (2005). An Empirical Investigation of the Key Factors for Success in Software Process
Improvement, IEEE Transactions on Software Engineering 31, 410-424..
272
14. Guerrero, F., & Eterovic, Y. (2004). Adopting the SW-CMM in a Small IT Organisation, IEEE
Software 21, 29-35.
15. Olson, T. G., & Sachlis, M. (2002). Aggressively Achieving CMM Level 3 in One Year. SEPG 2002,
Phoenix, AZ.
16. Jackelen, G. (2007). CMMI Level 2 Within Six Months? No Way! CrossTalk The Journal of Defense
Software Engineering, (February 2007), 13-16.
17. Iversen, J. and Ngwenyama, O. (2006). Problems in measuring effectiveness in software process
improvement: A longitudinal study of organizational change at Danske Data, International Journal of
Information Management 26, 3043.
18. Balla, K., Bemelmans, T., Kusters, R., & Trienekens, J. (2001). Quality Through Managed
Development and Implementatin of a Quality Management System for Software Company, Software
Quality Journal 9, 177-193.
Biographies
Yasemin Karagul has received her BSc from the Middle East Technical Universtiy Department of Chemical Engineering
(1999), and her MSc (2003) and PhD (2009) from Informatics Institute, Middle East Technical University, Ankara, Turkey.
She is an Assistant Professor at Dogus University, Istanbul, Turkey. Her areas of interest are software quality and software
process improvement
E-mail address: ykaragul@dogus.edu.tr;Alternative e-mail address for the same author ysalihoglu@gmail.comPostal
address: Dogus University, Zeamet Sokak, No:21/G904-B Acibadem/Kadikoy/Istanbul/Turkey Telephone: +90- 216- 544 55
55/1183, Fax: 0216 327 96 31
Semih Bilgen has received his BSc from the Middle East Technical University Department of Electrical and Electronics
Engineering (1973), MSc from Rensselaer Polytechnic Institute, Department of Computer Science (1976) and PhD from the
University of Manitoba, Department of Electrical Engineering (1982). He has been with the Middle East Technical
University since 1976, teaching in the Management, EEE Departments and in the Informatics Institute. He was among the
founders of that institute in 1997. His main research and consulting areas are Computer Networking, Software Engineering
and Information Systems.
273
Abstract. In this paper, a developed evolutionary algorithm, neuroevolution of augmented topologies (NEAT) for
simultaneous connections and weights of neural network training is described. A distinctive feature of the algorithm is
flexible and effective evolutionary search and a balanced resulting neural network structure due to weight and
connection mutation operator. In NEAT, neural network structure changes, caused by mutation operators, as well as
mutation rate are defined independently for each individual. A comparison of obtained results with results of other
algorithms is presented.NEAT is used to evolve a controller for the double pole balancing problem with non-Markovian
version by unidentifying the velocity information. If the neural network structure is evolved with connection weights such
that topologies are minimized and grown incrementally, signicant performance gains result A comparison of obtained
results of different levels of mutation operators is presented.
1 Introduction
Controlling unstable nonlinear systems with neural networks can be problematic. The successful
application of classic control design techniques usually requires an extensive knowledge of the system to be
controlled, including an accurate model of its dynamics. The challenge is to control a system without a priori
information about its dynamics. One way to achieve this is to evolve neural networks to control the system using
only sparse feedback from the system. is no longer difcult enough to serve for measuring the learning efciency
of these systems.
The double pole case, where two poles connected to the cart must be balanced simultaneously is much more
difcult than the classic pole balancing problem, especially when velocity information is not available.. Most
neuroevolution (NE) systems that have been tested on pole balancing evolve connection weights on networks
with a xed topology (Gomez and Miikkulainen 1999; Moriarty and Miikkulainen 1996; Saravanan and Fogel
1995; Whitley et al.1993; Wieland 1991). On the other hand, NE systems that evolve both network topologies
and connection weights simultaneously have also been proposed (Angeline et al. 1993; Gruau et al. 1996; Yao
1999). A major question in NE is whether such Topology and Weight Evolving Articial Neural Networks
274
(TWEANNs) can enhance the performance of NE. On one hand, evolving topology along with weights might
make the search more difcult. On the other, evolving topologies can save the time of having to nd the right
number of hidden neurons for a particular problem (Gruau et al. 1996). In a recent study, a topology-evolving
method called Cellular Encoding (CE; Gruau et al., 1996) was compared to a xed-network method called
Enforced Subpopulations (ESP) on the double pole balancing task without velocity inputs (Gomez and
Miikkulainen 1999). Since ESP had no a priori knowledge of the correct number of hidden nodes for solving the
task, each time it failed, it was restarted with a new random number of hidden nodes. However, even then, ESP
was ve times faster than CE. This study aims to demonstrate evolving structure along with adding and
removing of connection weights can signicantly affect the performance of NE.
2 NEAT
NEAT provides a principled methodology for implementing a complexifying search from a minimal starting
point in any such structures. NEAT is designed to take advantage of structure as a way of minimizing the
dimensionality of the search space of connection weights.
As summarized in Yao's overview of TWEANNs, algorithms for finding optimal architectures fall into two
broad categories, constructive and destructive [1]. For the purposes of this paper, we refer to these dynamics as
complexification and simplification, defined in the following ways:
Acomplexification algorithm is one that begins with a minimal architecture and includes mutation operators that
incrementally add topological features, such as hidden nodes and connections.
Asimplification algorithm generally begins with an overabundance of topological features and incrementally
removes features.
In the context of evolutionary algorithms there are inherent problems with each of these approaches taken
alone. As Angeline is quoted in Yao's overview: "Such structural hill climbing methods are susceptible to
becoming trapped at structural local optima." In addition, they "only investigate restricted topological subsets
rather than the complete class of network architectures".[12]
NEAT, a complexifying TWEANN approach in particular, retains variation (and hence simpler structures) in
the population through the use of speciation [3]. This approach allows for simpler individuals in the population
to act as starting points for reinvestigation of the solution space through complexification.
However, though niching techniques such as speciation avoid the susceptibility of being trapped at local
optima, they do not address Angeline's second objection, that the full range of architectures is not being
investigated.
A problem arises, in spite of these dynamics, in the choice of starting topology. NEAT defines a minimal
architecture as "a uniform population of networks with zero hidden nodes." This concept of starting minimally,
though, makes assumptions about ideal topologies for solutions to given domains. In purely complexifying
algorithms there is no way to remove unnecessary features. If those features exist in the starting topology they
represent an impediment to finding the optimal architecture.
With simplifying algorithms, even if great care is taken to insure variation and efficiency in search, to find an
optimal architecture for a given problem the initial topologies must be a superset of that optimal architecture. In
the vast majority of cases, the optimal architecture for a given problem's solution is not known (hence the use of
TWEANNs in the first place).
3 Experimental Setup
NEAT is distinguished by allowing crossover between networks with different topologies. Also, NEAT
uses speciation to divide the population into morphologically similar subgroups. The algorithm has demonstrated
the ability to outperform other neuroevolutionary approaches, and perform well at a variety of tasks [3], [6], [9].
The version of NEAT used here was an open source version, ANJI (Another NEAT Java Implementation)
[http://anji.sourceforge.net/], written in Java. According to the NEAT paradigm, the initial neural network
architecture consisted of only input and output nodes, fully connected with only feed-forward connections. Initial
weight values were taken from a uniform distribution between -1 and 1. Input node activation functions were
linear, output, and hidden nodes are sigmoid.
275
Each generation, upon receiving a fitness score as mentioned above, the best performing 20% of the population
was selected for survival and reproduction. For all experiments, a population size of 100 was used, so after
selection there were always 20 survivors. The population was then replenished back to 100 individuals: the 20
survivors, plus 20 mutated versions of those survivors, plus 60 "offspring" the result of both crossover and
mutation.
NEAT is a system for evolving both the connection weights and topology of ANNs simultaneously. It does so
by means of crossover and types of mutation defined below:
Value(s)
200-500-1000
100
0.25-0.50-0.75
0.1-0.3
1.0
1.0
0.04
0.2
0.01-0.05
0.01-0.05
0.01-0.05
As seen from Table 2, when the mutation rate and population size is low, to find the fitness target; the number
of connections, the number of hidden nodes, the number of increases, the number of generations needed and the
276
number of genes of best solution increases. For low mutation rates; the performance values are very much
sensitive due to the size of the population. As the mutation rate increases, the averaged values do not change so
effectively.
Examining the add neuron mutation rate and survival rate; the increase of the survival rate increases the
performance values. By comparison, at the same level of the survival rates, add neuron mutation rate does not
have much effect on the results.
Table 2. Averaged Values For Different Levels of Weight Mutation Rate and Population Size (10 Runs Each)
Avg. #
Average
of
number of
Weight
genes generation
mutatio Pop. of best
s to find
n rate
size solution solution
Average
number of
connections
Average
number of
hidden
nodes
0.25
200
28,00
18,88
16,25
3,75
0.25
500
18,31
6,50
9,19
1,13
0.25
1000
18,63
7,81
9,38
1,44
0.50
200
21,31
14,63
10,94
2,38
0.50
500
20,63
9,06
10,69
1,94
0.50
1000
17,75
7,00
8,50
1,25
0.75
200
17,56
19,94
8,56
1,00
0.75
500
18,75
11,56
9,06
1,69
0.75
1000
19,00
8,00
9,69
1,31
Table3. Averaged Values For Different Levels of Add Neuron Mutation Rate And Survival Rate (10 Runs Each)
Average
Add
Avg. # number of
Neuron Surv. of genes generations
Mutatio
of best
to find
n Rate Rate solution
solution
0.01
0.01
0.05
0.05
0.1
0.3
0.1
0.3
17,00
19,00
19,42
24,56
7,58
14,00
9,94
14,42
Average
number of
connections
Average
number of
hidden
nodes
8,22
9,69
9,72
13,36
0,86
1,31
1,69
3,19
277
Table 4.Averaged Values For Different Levels of Add And Remove Connection Mutation Rates (10 Runs Each)
Avg. #
Average
Rem.
of
number of
Add
Con. genes generations
Average
Con.Mu Mut. of best
to find
number of
t rate
Rate solution solution
connections
0.01
0.01
0.05
0.05
0.01
0.05
0.01
0.05
19,17
17,11
23,94
19,75
9,17
17,25
9,44
10,08
Average
number of
hidden
nodes
9,44
7,75
13,58
10,22
1,81
1,36
2,36
1,53
Comparing the connection mutation rates; the scenarios having the same values, there is little difference in
the averaged results. If remove connection rate is bigger than the add connection mutation rate, the number of
connections decreases and the number of generations increases logically. On the contrary, as add connection
mutation rate gets higher, the number of connections increases and the number of generations decrease.
Connection mutation rates do not have much effect on the number of hidden nodes.
Another important consideration is the time spent searching. All experiments were carried out on a
desktop PC with a 2.00 GHz AMD Phenom Quad-Core processor, 4GB of RAM, running the Ubuntu
11.04 operating system of a Linux environment. As the weight mutation and population size increases, the time
needed for the runs increases due to complexification.
Table 4.Averaged Computation Time For Double Pole Balancing Problem For Different Levels of Weight
Pop. size
0.25
200
137,5
0.25
500
72,56
0.25
1000
193,50
0.50
200
61,56
0.50
500
289,88
0.50
1000
587,44
0.75
200
140,25
0.75
500
360,81
0.75
1000
677,13
278
4 Conclusion
The experiments for the double pole balancing problem have demonstrated that as the weight mutation rate
and population size increases, more simplified neural network structures can be gained. On the other hand, there
is a tradeoff to this simplification: more time is needed to reach the target value. As the aim of ANJI is to
simplify the neural network structure, there is no need to get the connection mutation rates higher.
References
1.
Biographies
Asil Alkaya Mr.Alkaya was born in Istanbul on September 14, 1977 that had earned M.Sc. degree of industrial engineering
from Dokuz Eylul University, The Graduate School of Natural and Applied Sciences, Izmir, Turkey in 2001 and Ph.D.
degree of industrial engineering from the same institution in 2010. Major field of study is neroevolution.
He made publications on evolutionary algorithms, neural network related time series analysis, and parameter optimization.
Interest areas are data modeling, information systems, neural networks, evolutionary algorithms, classification and clustering.
279
Abstract. Let set [n] = f1; 2; :::; ng be given. Find the minimum cardinality of a collection Sof subsets of [n]
such that, for any two distinct elements x; y 2 [n], there exists disjointsubsets Ax;Ay2 S such that x 2 Ax, y 2 Ay,
and Ax \ Ay = _. If S _ 2[n], thenone can _nd such a set S in polynomial time. If S ( 2[n], then _nding such set S is
anNP-Complete problem.
Keywords: Extremal Set; Complexity
1 Introduction
280
281
282
283
284
References
[1] Atici, M., and Vince, A.: Geodetics in Graphs, an Extremal Set Problem, and Perfect
Hash Families,Graphs and Combinatorics 18, 403-413 (2002)
[2] Atici M.: On the Edge Geodetic Number of a Graph, Intern. J. Computer Math.,Vol
80, No 7, 853-861(2003)
[3] Garey M. R. and Johnson D. S.:Computers and Computers and Intractability: A Guide
to the Theory of NP-completeness. Freeman (1979)
285
Abstract. A recent discrete wavelet transform (DWT) image watermarking paper embeds a pseudo
random number (PRN) sequence as a watermark in three bands, excluding the low pass subband, using
coefficients that are higher than a given threshold. During watermark detection, all the coefficients higher
than another threshold are chosen for correlation with the original watermark. Our another paper, we
extend the idea to embed the same watermark in two bands (LL and HH). There are two parameters in all
this algorithms: Scaling Factor and Threshold. These parameters have been determined by Try and Use
method. In this work we used Simulated Annealing (SA) method to determine OPTIMUM scaling factor
and threshold values for high degree of robustness and invisibility. In a result, automated selection gives
better result than Try and Use method in semi-blind PRN embedding algorithms in gray and color image
watermarking.
1 Introduction
Digital watermarking has received increasing attention in recent years. Distribution of movies, music, and
images is now faster and easier via computer technology, especially on the Internet. Hence, the content owners
(e.g., movie studios and recording companies) are concerned about illegal copying of their content [1].
Watermarking is a pattern of bits (logo or noise) inserted into a digital image, video, text or audio which
identifies the copyright information.
The most important properties of a watermarking system are robustness, invisibility, data capacity, and security.
An embedded watermark should not introduce a significant degree of distortion in the cover multimedia element.
Robustness is the resistance of the watermark against normal A/V processes or intentional attacks. Data capacity
refers to the amount of data that can be embedded without affecting perceptual transparency. The security of a
watermark can be defined to be the ability to thwart hostile attacks such as unauthorized removal, unauthorized
embedding, and unauthorized detection. There are basically two approaches to embed a watermark: spatial
domain and transform domain (e.g., DCT, DFT, or DWT). In the spatial domain, the watermark is embedded by
modifying the pixel values in the original image. Transform domain watermarking is similar to spatial domain
watermarking; in this case, the coefficients are modified. Both methods have advantages and disadvantages: One
disadvantage of spatial domain watermarking is that the cropping attack might remove the watermark [2].
286
There are several criteria to classify watermarking techniques. Table 1 shows some fundamental categories [3].
Criteria
Types
Type of Document
Human Perception
Visible, Invisible
Working Domain
Watermark Type
Information Type
a. Original Image
b. Visible Watermarked
c. Invisible Watermarked
Image
Image
2 Literature Review
There are three major schemes in multimedia watermarking. The first is spatial domain watermarking, which
basically embeds a visible logo or a PRN sequence directly to selected pixels in host image. The second is
transform domain watermarking such as DCT, DWT or DFT. Least Significant Bit (LSB) embedding is the
earliest and also the simplest technique. Since the last binary bits are the least significant bits, their modification
will not be perceived by human eyes. However, the information carried in the least significant bit is difficult to
survive various attacks.
Schyndel [4] et al. proposed two methods by using the least significant bit embedding. In the first method, they
compress the original 8 bits to 7 bits by adaptive histogram manipulation so as to enable the LSB to carry the
watermark information. Bender et al. [5] provided two methods, patchwork and texture block coding, that
change the data directly in host image. Pitas [6] presented a technology for casting digital watermarks on images
by embedding a predetermined small luminance value to randomly selected images. The luminance values are
small enough to be undetected by the human eye. The seed of the random pixel generator is essentially the
copyright holder watermark. Wang et al [7] proposed a watermark embedding algorithm, which adds the
watermark to the host in the spatial domain instead of the transform domain, reducing the complexity
considerably.
287
Cox et al. [8] proposed secure spread spectrum watermarking algorithm. This algorithm uses the Discrete Cosine
Transformation (DCT) in gray scale image. Piva et al. [9] presented DCT based watermark recovering without
resorting to the uncorrupted original image. This algorithm provides extra robustness against intentional and
distortions. Dugad et al. [10] proposed wavelet based scheme for watermarking images. Kusyk et al. [11]
proposed a semi-blind logo watermarking for color images in the DFT domain.
Hsu and Wu [12] proposed a DCT-based watermarking algorithm for video sequences. In this method,
embedding occurs in both intra and inter frames. Chan and Lyu [13] proposed a DWT algorithm in video
watermarking with Error Correcting Code. In the experiments, the authors split the video into two components:
video sequences and audio sequences, and embedded watermark in both separately. Serdean et al [14] proposed
a DWT based blind watermarking method that is invariant to geometric attacks such as shifting, rotation, scaling
and cropping. Lin and Delp [15] proposed a method using spatial domain in uncompressed video and finite state
machine (FSM) for watermark key generator. One of the main ideas is the key generator in this work. Feature
extraction allows the key schedule to be video dependent.
Mobasseri [16] discussed the spread spectrum in the form of CDMA (Code Division Multiple Access) in
uncompressed digital video. This scheme appears to be resistant to noise as well as attacks on destroying
synchronization. Swanson [17] proposed an object based watermarking technique. Individual watermarks are
created for objects within the video. The watermark is embedded in the DCT domain. Swanson et al. [17]
propose an object based watermarking technique. Individual watermarks are created to insert object in the video.
This watermarks created by shaping and video dependent pseudo random sequence according to the perceptual
masking characteristics of the video.
Watermark detection:
1.
2.
3.
4.
Convert the NxN watermarked (and possibly attacked) RGB image to YUV.
Compute the DWT of the luminance layer.
Select all the DWT coefficients higher than T2 in LL and HH bands.
Compute the sum
, where i runs over all DWT coefficients > T 2, yi represents
either
the
real
watermark
or
a
fake
watermark,
represents the watermarked and possibly attacked DWT coefficients. Choose
a predefined threshold.
5.
In both band, if z exceeds Tz, the conclusion is that the watermark is present.
288
n n0; x F(n)
// Initial state
nbest n; xbest x
y0
xnew F(nnew)
return nbest
To calculate threshold values T1 ,T2 and scaling factor alpha values we applied SA algorithm and found out
OPTIMUMscaling factor and threshold values. This optimum values provide high robustness, invisibility and security in
wavelet watermarking scheme.
3 Experiments
In our experiments, we obtained the first level decomposition in wavelet domain using the Haar filter. The values
of and the threshold for each band are selected by SA algorithm given in Table 2.
Parameters/Bands
LL
HH
0.4
3.5
T1
15
45
T2
25
55
In figure 2-a shows that watermarked image with try and use method parameter selection and figure 2-b shows
that watermarked image with SA based parameter selection. SA based scaling factor and threshold selection
gives high image quality in watermarked Lena.
289
a. Watermarked Lena
(PSNR: 39.256)
Matlab was used for all attacks. The chosen attacks were JPEG compression, resizing, adding Gaussian noise,
low pass filtering, rotation, histogram equalization, contrast adjustment, gamma correction, and cropping. This
algorithm resists all types of common attacks.
JPEG compression
Resizing
(Q=25)
(512256512)
Rotation
(window size=3x3)
(5 )
290
Contrast adjustment
Gamma correction
(1.5)
Gaussian noise
Histogram equalization
(automatic)
3 Conclusion
In a DWT-based semi-blind image watermarking, a watermark is embedded in three bands, leaving out the low
pass subband, using coefficients that are higher than a threshold T 1. During watermark detection, all the high
pass coefficients higher than another threshold T 2 (T2 T1) are chosen for correlation with the original
watermark.
In this research:
1.
We have extended the idea by embedding the same watermark in two bands (LL and HH) using different
scaling factors and thresholds for each band in color images.
2. We used simulated annealing (SA) method to determine optimal scaling factor and threshold values for
high degree of robustness and invisibility.
Our experimental results show that in gray and YUV color images for one group of attacks (JPEG compression,
resizing, adding Gaussian noise, low pass filtering, and rotation), the correlation with the real watermark is
higher than the threshold in the LL band, and for another group of attacks (histogram equalization, contrast
adjustment, gamma correction, and cropping), the correlation with the real watermark is higher than the
threshold in the HH band. In the end, automated selection gives better result than Try and Use method in semiblind PRN embedding algorithms in gray and color image watermarking.
References
1.
2.
3.
4.
5.
6.
291
7.
8.
9.
10.
11.
12.
13.
14.
15.
16.
17.
18.
19.
20.
H. Wang, Z. Lu, J. Pan, S. Sun, Robust Blind Video Watermarking with Adaptive Embedding
Mechanism, International Journal of Innovative Computing, Information and Control Volume 1,
Number 2, June 2005.
I. J. Cox, J. Kilian, T. Leighton and T. Shamoon, Secure Spread Spectrum Watermarking for
Multimedia, IEEE Transactions on Image Processing, 6(12), December 1997, pp.1673-1687.
A. Piva, M. Barni, F. Bartolini, V. Cappellini, DCT-based Watermark Recovering without Resorting to
the Uncorrupted Original Image, Proceedings of International Conference on Image Processing,
Washington, DC, October 26 - 29, 1997, pp. 26-29.
R. Dugad, K. Ratakonda and N. Ahuja, A New Wavelet-Based Scheme for Watermarking Images,
Proceedings of 1998 International Conference on Image Processing (ICIP 1998), Vol. 2, Chicago, IL,
October 4-7, 1998, pp. 419-423.
J. Kusyk and A. M. Eskicioglu, A Semi-blind Logo Watermarking Scheme for Color Images by
Comparison and Modification by Comparison and Modification of DFT Coefficients Optics East 2005,
Multimedia Systems and Applications VIII Conference, Boston, MA, October 23-26, 2005.
C. Hsu, J. Wu, DCT-Based Watermarking for Video, IEEE Transaction on Consumer Electronics,
Vol. 44, No. 1, February 1998, pp. 206-216.
P. Chan, M. R. Lyu, A DWT-Based Digital Video Watermarking Scheme with Error Correcting Code
ICICS 2003, pp. 202-213
C. V. Serdean, M.A. Ambroze, M. Tomlinson, G. Wade, DWT Based Video Watermarking for
Copyright Protection, Invariant to Geometrical Attacks.
E. Lin and E. Delp, Temporal synchronization in video watermarking, in Proc. SPIE Security and
Watermarking of Multimedia Contents IV, vol. 4675, San Jose, CA, Jan. 2124, 2002, pp. 478490.
B. Mobasseri, Exploring CDMA for Watermarking of Digital Video, Proceedings of SPIE Vol. 3657,
pp.96-102, 1999.
M. D. Swanson, B. Zhu, A. H. Tewfik, Multiresolution Scene-Based Video Watermarking Using
Perceptual Models, IEEE Journal on Selected Areas in Communications, Vol. 16, No.4, May 1998.
E. Elbasi, A. M. Eskicioglu, A Semi-Blind Watermarking Scheme for Color Images, SKM
Workshop, Sep 2006, New York.
www.wikipedia.org
Ersin Elbasi, Multimedia Security: Digital Image and Video Watermarking, PhD Thesis, Graduate
Center, CUNY, 2007.
Biographies
Ersin ELBASI received MS degree in Computer Science at Syracuse University (2001); MPhil (2006) and PhD (2007)
degrees in Computer Science at Graduate Center, The City University of New York. His research interests include
multimedia security, event mining from video sequences and medical image processing. Dr. Elbasi currently working for
Turkish Scientific and Technological Research Council and part time instructor at Cankaya University in Ankara.
292
Turkish Scientific and Technological Research Council,Tunus Caddesi No: 80 Kavaklidere, Ankara
2
The Graduate Center, City University of New York365 Fifth Avenue, New York City, NY
3
Brooklyn College, The City University of New York2900 Bedford Avenue, Brooklyn, NY
1
Abstract. In subjective evaluation of distorted images, human observers usually consider the type of distortion, the
amount of distortion, and the distribution of error. We recently proposed an image quality measure, M-SVD, for
gray-scale images that can express the quality of distorted images numerically or graphically. As a graphical tool, it
predicts the distortion based on the three factors used by human observers. As a numerical tool, it evaluates the
overall visual quality of the distorted image by computing a single value. It performs better than two state-of-theart metrics, Q and MSSIM, especially when we compute the correlation with mean opinion score across different
types of noise. Each test image was degraded using six types of noise (JPEG, JPEG 2000, Gaussian blur, Gaussian
noise, sharpening and DC-shifting), each with five different levels of intensity [1]. In this work we tested M-SVD
quality assurance and compared with other assurances (Q, MSSIM and PSNR) in watermarked image and videos.
Results show that M-SVD is a highly trustable quality assurance for wavelet based watermarked multimedia
elements.
Keywords: multimedia, M-SVD, quality, evaluation, watermarking
1 Introduction
Measurement of image quality is a challenging problem in many image processing fields ranging from lossy
compression to printing. The quality measures in the literature can be classified into two groups: Subjective and
objective. Subjective evaluation is cumbersome as the human observers can be influenced by several critical
factors such as the environmental conditions, motivation, and mood. The most common objective evaluation
tool, the Mean Square Error (MSE), is very unreliable, resulting in poor correlation with the human visual
system (HVS). In spite of their complicated algorithms, the more recent HVS-based objective measures do not
appear to be superior to the simple pixel-based measures like the MSE, Peak Signal-to-Noise Ratio (PSNR) or
Root Mean Squared Error (RMSE). It is argued that an ideal image quality measure should be able to describe
[2]: Amount of distortion, type of distortion, and distribution of error.
There is a need for an objective measure that provides more information than a single numerical value. Only a
few multi-dimensional measures exist in the relevant literature today [2].
293
1 Recent Works
Image quality measures can be classified using a number of criteria such as the type of domain (pixel or
transform), the type of distortion predicted (noise, blur, etc.) and the type of information needed to assess the
quality (original image, distorted image, etc.). Table 1 gives a classification based on these three criteria, and
includes representative examples of recently published papers. Measures that require both the original image
and the distorted image are called full-reference or non-blind methods, measures that do not require the
original image are called no-reference or blind methods, and measures that require both the distorted image
and partial information about the original image are called reduced-reference methods.
Domain
Type
Publication\Criterion
Type of distortion
Type of
information
Pixel
Fullreference
DWT
Fullreference
DCT
JPEG compression
No-reference
DFT
JPEG compression
No-reference
Pixel
JPEG compression
No-reference
Marziliano, Dufaux,
Winkler and Ebrahimi,
2002 [8]
Pixel
No-reference
Pixel
No-reference
Pixel
JPEG compression
No-reference
Pixel
JPEG compression,
JPEG 2000
compression
Reducedreference
294
1.
2.
3.
4.
2 Methodology
Every real matrix A can be decomposed into a product of 3 matrices A = USVT, where U and V are orthogonal
matrices, UTU = I, VTV = I and S = diag(s1, s2, ...). The diagonal entries of S are called the singular values of A,
the columns of U are called the left singular vectors of A, and the columns of V are called the right singular
vectors of A. This decomposition is known as the Singular Value Decomposition of A. It is one of the most
useful tools of linear algebra with several applications to multimedia including image compression and
watermarking [15].
The graphical measure is a bivariate measure that computes the distance between the singular values of the
original image block and the singular values of the distorted image block:
Where si is the singular values of the original block, are the singular values of the distorted block, and n is the
block size. If the image size is k, we have (k/n) x (k/n) blocks. The set of distances, when displayed in a graph,
represents a distortion map.
The numerical measure is derived from the graphical measure. It computes the global error expressed as a single
numerical value depending on the distortion type:
Where Dmid represents the midpoint of the sorted Di, k is the image size, and n is the block size.
In [12], the measure was applied to 512x512 gray scale Lena, one of the most widely used test images. In our
experiments, we used six distortion types (JPEG, JPEG 2000, Gaussian blur, Gaussian noise, sharpening and
DC-shifting) at five distortion levels using 8x8 blocks. The distortion types, the distortion levels, and the
associated parameters are shown in Table 2 [15].
295
Type \ Level
Level 1
Level 2
Level 3
Level 4
Level 5
JPEG
10:1
20:1
30:1
40:1
50:1
JPEG2000
10:1
20:1
30:1
40:1
50:1
Gaussian blur
Gaussian noise
12
15
Sharpening
10
20
30
40
50
DC-shifting
12
16
20
Local error expressed as a 3-dimensional graph (provides the amount and type of error as well as its
distribution in the image).
Global error expressed as a single numerical value (provides overall error based on the distortion).
JPEG
Distortion Map
JPEG2000
Distortion Map
Gaussian Blur
Distortion Map
Gaussian Noise
Distortion Map
Sharpening
Distortion Map
DC-Shifting
Distortion Map
296
Fig. 1.presents the distorted images and distortion maps at level 5 for one test image. Each distortion map,
which provides the amount of distortion, the type of distortion, and the distribution of error, is obtained as a
gray-scale image by mapping the Divalues to the range [0,255].
3 Experimental Results
In this work we applied M-SVD for watermarked images and videos. Before quality assurance we used
following methods in watermark embedding [14].
Watermark embedding algorithm for images:
6.
7.
8.
Process:
1. Split the video sequence into frames.
2. Convert the NxM RGB frame to YUV in I frames.
3. Compute the DWT of the luminance layer (Y) for each I frame.
4. Modify the DWT coefficients Vijin the LL, LH, HL, and HH bands in all frames.
Vw,ij = Vi,jk + k. Wi,j where i = 1,.,n ; j = 1,.,m and k=1,2,3,4.
5. Apply inverse DWT to obtain the watermark cover frame I w for each I frame.
Output: Watermarked cover video sequence.
Watermarked Video
Frame
Watermark
Table 3 displays the overall performance of four measures using two criteria: Correlation and Root Mean
Squared Error (RMSE) between Mean Opinion Score (MOS) and objective prediction. It can be observed that
297
M-SVD outperforms all three measures. In particular, the correlation is improved by approximately 20%, and
the RMSE is reduced by almost 58%, relative to the state-of-the-art metrics Q and MSSIM.
Criteria\Measure
PSNR
MSSIM
M-SVD
CC
0.617
0.785
0.802
0.941
RMSE
9.22
7.53
7.84
4.82
4 Conclusions
In this work we tested M-SVD quality assurance and compared with other assurances (Q, MSSIM and
PSNR) in watermarked image and videos. Results show that M-SVD is a highly trustable quality assurance
for wavelet based watermarked multimedia elements.
References
A. M. Eskicioglu and P. S. Fisher, Image quality measures and their performance, IEEE Transactions
on Communications, Vol. 43, pp. 2959-2965, December 1995.
2. A. M. Eskicioglu, Quality measurement for monochrome compressed images in the past 25 years,
Proceedings of IEEE International Conference on Acoustics, Speech, and Signal Processing, Vol. 4, pp.
1907-1910, Istanbul, Turkey, June 5-9, 2000.
3. D. Van der Weken, M. Nachtegael and E. E. Kerre, A new similarity measure for image processing,
Journal of Computational Methods in Sciences and Engineering, Vol. 3, No. 2, pp. 209-222, 2003.
4. A. Beghdadi and B. Pesquet-Popescu, A new image distortion measure based on wavelet
decomposition, 7th International Symposium on Signal Processing and Its Applications, Paris, France,
July 1-4, 2003.
5. A. C. Bovik and S. Liu, DCT-domain blind measurement of blocking artifacts in DCT-coded images,
Proceedings of International Conference on Acoustics, Speech, and Signal Processing, Salt Lake City,
UT, May 7-11, 2001.
6. Z. Wang, A. C. Bovik and B. L. Evans, Blind measurement of blocking artifacts in images,
Proceedings of IEEE 2000 International Conferencing on Image Processing, Vancouver, BC, Canada,
September 10-13, 2000.
7. Z. Wang, H. R. Sheikh and A. C. Bovik, No-reference perceptual quality assessment of JPEG
compressed images, Proceedings of IEEE 2002 International Conferencing on Image Processing,
Rochester, NY, September 22-25, 2002.
8. P. Marziliano, F. Dufaux, S. Winkler and T. Ebrahimi, A no-reference perceptual blur metric, IEEE
2002 International Conference on Image Processing, Rochester, NY, September 22-25, 2002.
9. E.-P. Ong, W. Lin, Lu, Z. Yang, S. Yao, F. Pan, L. Jiang and F. Moschetti, A no-reference quality
metric for measuring image blur, 7th International Symposium on Signal Processing and Its
Applications, Paris, France, July 1-4, 2003.
10. L. Meesters and J.-B. Martens, A single-ended blockiness measure for JPEG-coded images, Signal
Processing, Vol. 82, pp. 369-387, 2002.
11. M. Carnec, P. Le Callet and D. Barba, An image quality assessment method based on perception of
structural information, 2003 International Conference on Image Processing, Barcelona, Spain,
September 14-17, 2003.
1.
298
12. Z. Wang and A. Bovik, A universal image quality index, IEEE Signal Processing Letters, Vol. 9, No.
3, pp. 81-84, March 2002.
13. Z. Wang, A. C. Bovik, H. R. Sheikh and E. P. Simoncelli, Image quality assessment: from error
measurement to structural similarity, IEEE Transactions on Image Processing, Vol. 13, No. 4, April
2004.
14. Ersin Elbasi, Multimedia Security: Digital Image and Video Watermarking, Ph.D Thesis, Graduate
Center, CUNY, 2007.
15. Ahmet M. Eskicioglu, Quality Measure using Singular Value Decomposition, US Army project
proposal, 2005.
Biographies
Ersin ELBASI received MS degree in Computer Science at Syracuse University (2001); MPhil (2006) and PhD (2007)
degrees in Computer Science at Graduate Center, The City University of New York. His research interests include
multimedia security, event mining from video sequences and medical image processing. Dr. Elbasi currently working for
Turkish Scientific and Technological Research Council and part time instructor at Cankaya University in Ankara.
Peining TAO was born in Shanghai, China. She received the bachelor degree in clinical medicine from the Medical Center
of Fudan University (Former Shanghai Medical University) in 1998, the master degree in computer information science from
Brooklyn College, The City University of New York in 2001. She received PhD degree in computer science at The Graduate
Center, the City University of New York in 2008. Her research interests include digital watermarking, image/video quality
measurement and image recovery.
Ahmet M. ESKICIOGLU received the B.S. degree from the Middle East Technical University (METU), Ankara, Turkey,
and the M.S. and Ph.D degrees from the University of Manchester, England. He was with the Computer Engineering
department, METU from 1983 to 1992, the department of Computer Sciences, University of North Texas from 1992 to 1995,
and Thomson Multimedia Corporate Research, Indianapolis from 1996 to 2001. He is with the department of Computer and
Information Science, Brooklyn College of the City University of New York. Prof. Eskicioglus teaching and research
interests include data security, conditional access, digital rights management, copy protection, digital watermarking, and
multimedia applications. He has been a guest lecturer at several universities and research organizations. Dr. Eskicioglu is a
senior member of the IEEE.
299
1 Introduction
Document image analysis has been the topic of research for almost three decades. A large number of the
research in this subject have already been published on Latin, Chinese and Japanese characters but there is a little
work on Persian or Arabic documents and scripts because of the less efforts in these direction of researches.
Therefore, this area of research is still an open field.
With the progress in information and communication technology and the increased call for information, the
amount of documents containing information has increased more and more. Nonetheless the use of electronic
documents has increased, the amount of printed or handwritten documents has never decreased and most of the
people prefer to read the printed documents. Increasing of the printed documents causes to a lot of problems
appeared into the storing and retrieving these documents. So it is important to find ways to segmentation and
finally recognition printed documents and words in order to change these documents into electronic document
[1], [2].
There are some approaches with different algorithms for segmentation of printed documents. We can classify
these methods into three categories such as top-down, bottom-up and spectral [3], [4], [5].
In this paper we will propose a method as a bottom-up method for printed page segmentation by using
connected component and clustering. In Section 2 related works is introduced. In section 3 connected
components analysis is introduced. In section 4 feature extraction and selection would be described. In section 5,
k-mean clustering will be presented. In section 6 we will propose our algorithm and finally the paper will be
ended by experimental result.
300
2 Related Work
Document analysis and understanding are relevant techniques for the automatic processing of paper
documents. Extraction of the structure or the layout from a document is referred to as document analysis and
mapping the layout structure into a logical structure is referred to as document understanding. These techniques
allow the recognition of document contents and the simplification of a number of complex tasks, such as reediting, storage, maintenance, retrieval and transmission. Document image analysis plays an important role in
the field of document processing and provides techniques for partitioning a document into a hierarchy of
physical components (pages, columns, paragraphs, words, tables, figures, halftones, etc.)
So far different document image segmentation methods have been proposed in the literature, both for printed
pages and for hand written pages [6]. In our work, at the first case, with connected component the page is
divided into bounding box and with clustering algorithm that are successively merge into big blocks, in an
iterative fashion, to obtain the final text, graphics or image segments. In connected component analysis, single
pixels are gathered on the basis of a low level analysis, to constitute blocks that can be merged into successively
larger blocks [7]. A number of heuristics approaches may be devised to produce region segmentation and a
classical approach is to employ generic rules, based on thresholding [8]. The adoption of fuzzy logic has been
proposed as a complementary approach to existing methods [9]. In fact, the concept of a fuzzy set allows a
gradual transition from membership to non-membership, providing the management of a greater degree of
abstraction, thus overcoming the threshold-based dimension of classical approaches. A number of works
demonstrate that the introduction of fuzzy techniques can be very successful in the area of document image
processing, [10].
3 Connected Component
Connected component labeling or connected component analysis is an algorithmic application of graph
theory, where subsets of connected components are uniquely labeled based on a given heuristic. Connected
component labeling is not to be confused with segmentation. Connected component labeling is used in computer
vision to detect connected regions in binarydigital images, although color images and data with higherdimensionality can also be processed. When integrated into an image recognition system or human-computer
interaction interface, connected component labeling can operate on a variety of information.
One of the basic step in document segmentation is finding the connected components in a document image.
The points in a connected component form a candidate region for representing an object. In this research we use
the sequential algorithm usually requires two passes over the document image. This algorithm is given as
algorithm 1.
Algorithm 1. Sequential connected components algorithm using 4 connectivity:
1.
2.
3.
4.
5.
The figure 1 is shown result of sequential connected components algorithm on a sample of Farsi document.
301
Fig1.Result of connected component analysis on the one sample of Farsi document with bounding box.
is given by
(1)
The position of a connected component in a document image plays an important role in many applications.
There are different ways to specify the position of a connected component, such as using its enclosing rectangle
or centroid. To calculate the position of the object, we use:
(2)
(3)
Where and are the coordinates of the center of the region measured with respect to the top left pixel and
m, n is size of bounding box of each connected components.
Calculating the orientation of a connected component is a little more complex than calculating its position.
We use following relation to calculate the orientation:
302
(4)
(5)
Where
(6)
(7)
(8)
Pay attention that for each point (X,Y), X and Y would be calculated by
and
These features are used in next step for grouping connected component.
5 K-Mean Clustering
Cluster analysis is the grouping of individuals in a feature vectors in order to classify structure in the image
documents. In some sense, we would like the individuals within a group to be close or similar to one another, but
dissimilar from individuals in other groups. We use clustering method with feature vectors to classification
document image into homogenous regions. The K-mean is a powerful tools to clustering of document images. KMeans is a least-squares partitioning method that divide a collection of connected components into K groups.
The K is number of homogenous regions such as texts, images and graphics regions [11]. The algorithm iterates
over two steps:
1.
2.
3.
The initial assignment of feature vectors to clusters can be done randomly. In the course of the iterations, the
algorithm tries to minimize the sum, over all groups, of the squared within group errors, which are the distances
of the feature vectors to the respective group means. Convergence is reached when the objective function (i.e.,
the residual sum-of-squares) cannot be lowered any more. The groups obtained are such that they are
geometrically as compact as possible around their respective means. Using the set of feature images, a feature
vector is constructed corresponding to each pixel, where d is the number of feature images used for the
segmentation process. The figure 2 is shown result of K-Mean clustering on a sample image with k=3.
303
6 Propose Algorithm
In this paper, we propose a methodology that aims at segmenting a Farsi/Arabic document image into
coherent and homogeneous regions containing text, graphics and backgrounds. Following a bottom-up approach
base on connected component and clustering, the overall methodology executed of consecutive steps. Initially, a
connected component analysis is applied to the document image to classify each connected components to
bounding box part of text, graphics or background. Successively, clustering procedures group the identified
bounding box into coherent regions using a set of local feature vector of connected component of each bounding
box. To refine the classification results, an additional region level analysis is performed, based on shape
regularity and a region skew angle. At the end of stage, a final accurate classification of the document regions is
obtained.
7 Experimental Result
In this paper, an effective method for Farsi/Arabic document image segmentation and classification has been
proposed. The bottom up method based on connected component and clustering are exploited to perform
classification tasks on two different levels of details. The final results, obtained by segmenting pages from
journal and newspapers articles, correspond with human perception and compare well with those reported in the
literature. Moreover, the segmentation process is completely unchanged by page skew, as the skew angle of the
document image is automatically determined during the region analysis step. As guidelines for future work, the
overall methodology could be further enhanced by employing a wavelet coefficient analysis as feature vector for
clustering. Moreover, it could be useful to introduce more classes during the connected component classification
step (titles, headings, graphs, etc.). This could avoid the misclassification of article headings and titles as
graphics, instead of text.
References
1.
2.
M. Khademi, Z.Bani, (2009), Applying Persian online handwriting recognition in e-learning, international
conference of E-learning, Elmo Sanat university of Tehran, Iran(In persian).
Sh. Pirnia, M. Khademi, A. Nikookar and Z. Bani,( 4-5 Dec. 2010), A Feature-Based Approach to Segmentation
of Persian Online Cursive Script, The 2010 International Conference on Computer and Software Modeling
(ICCSM), Manila Philippines.
304
J.shanbezadeh, A.broumandnia, ( 2007), "segmantation of printed Farsi dicuments", the fourth conference on
M.V.I.P 2007, Ferdosi Mashhad university,Iran (In persian)
4. A.Broumandnia, J.Shanbehzadeh, M. Nourani, (2007), Segmentation of Printed Farsi/Arabic Words
,IEEE2007,761-766.
5. A. Broumandnia, J. Shanbehzadeh, M. Rezakhah Varnoosfaderani, (2008), Persian/arabic handwritten word
recognition using M-band packet wavelet transform, ELSEVIER,Image and Vision Computing 26 (2008) 829
842.
6. Y. Sun, T.S. Butler, A. Shafarenko, R. Adams, M. Loomes, N. Davey,(2007),Word segmentation of handwritten
text using supervised classification techniques, Appl. Soft Comput. 7 (1) (2007) 7188.
7. A. Fletcher, R. Kasturi, (1998) , A robust algorithm for text string separation from mixed text/graphics images,
IEEE Trans. Pattern Anal. Mach. Intell. 10(1998) 294308
8. L. Cinque, L. Lombardi, G. Manzini, (1998) , A multiresolution approach for page segmentation, Pattern Recognit.
Lett. 19 (1998) 217225.
9. Z. Shi, V. Govindaraju,( 2004), Line separation for complex document images using fuzzy runlength, in:
Proceedings of the First InternationalWorkshop on Document Image Analysis for Libraries (DIAL04).
10. L. Caponetti, C. Castiello, P. Gorecki, (2008) , Document page segmentation using neuro-fuzzy approach,
ELSEVIER, Soft Computing 8 (2008) 118126 .
11. Andrew R. Webb, 2002, Statistical Pattern Recognition, Second Edition, John Wiley & Sons Ltd, The Atrium,
Southern Gate, Chichester, West Sussex PO19 8SQ, England.
Biographies
Zahra Bani was born in Tehran, Iran in 1985. She received B.Sc. in Computer Hardware Engineering from Islamic Azad
University South Tehran Branch in Tehran, Iran, in 2008, and she is studying M.Sc. in Computer Software Engineering at
Islamic Azad University South Tehran Branch in Tehran, Iran since 2009.
She is interested in pattern recognition, document image analysis, and Handwriting Recognition. Her publications are two
papers Applying Persian online handwriting recognition in e-learning and A Feature-Based Approach to Segmentation
of Persian Online Cursive Script that have been published in international conferences.
Ms. Bani is a member of Young Researchers Club at Islamic Azad University, South Tehran Branch.
Ali Broumandnia was born in Esfahan, Iran in 1967. He received B.Sc. in Computer Hardware Engineering from Esfahan
University of Technology in Esfahan, Iran in 1992, M.Sc. in Computer Hardware Engineering from Iran University of
Science and Technology in Tehran, Iran in 1995, and Ph.D. degree in Computer Engineering from Islamic Azad University,
Science and Research Branch in Tehran, Iran, in 2006.
He was a lecturer of Computer Engineering at the Islamic Azad University, South Tehran Branch, Iran in 1996. He has very
publication in different fields and his CV is located on link: http://old.azad.ac.ir/faculty/bromandniai.htm. He is interested in
pattern recognition, document image analysis, and wavelet analysis.
Dr. Broumandnia is reviewer of Machine Vision and Application journal, Pattern Recognition journal, and some
international conferences.
Maryam Khademi was born in Boroujerd, Iran in 1967. She received B.Sc. in Pure Mathematics from TarbiatMoaalem
University, Tehran, Iran in 1989, M.Sc. in Pure Mathematics from Islamic Azad University North Tehran Branch, Iran in
1991, M.Sc. in Applied Mathematics from Iran University of Science and Technology, Tehran, Iran in 1993, , and Ph.D.
degree in Applied mathematics from University of Tehran, Iran in 2004.
She has very publication in different fields and her CV is located on link: http://old.azad.ac.ir/faculty/khademi89.htm. She is
interested in finite groups Theory, Graph Theory, Computer Science and Artificial Intelligence.
Dr. Khademi is the faculty member of Islamic Azad University, South Tehran Branch.
305
dozkan@posta.oyak.com.trand 2bilgen@iku.edu.tr
Abstract. In this study, it was analyzed some parameters on the number of Goldbach partitions up to 10 9.
Goldbach equivalence classes, consisting of those even numbers with the same number of Goldbach
partitions, were determined. For each equivalence class, class size, minimum and maximum elements, sum of
elements, maximum elements to minimum elements ratio and the average element size values has been
calculated and investigated for patterns. Characteristic properties of edge members of equivalence classes
were examined by factorization. It was detected that lowest members prime factors of Goldbach equivalence
classes had numerous and small while largest members prime factors of Goldbach equivalence classes had
few and large. Besides, the proposition that G:2NN surjective is verified for 1G(2N)4,020,437 interval.
Keywords: Goldbach partitions, goldbach conjecture.
1 Introduction
Prime numbers are very important elements in the research area of the Cryptology. These numbers are used to
improve encoding techniques. Goldbach conjecture is seen as one of the biggest problem of the prime numbers
and mathematics. Although it seems to be correct, it is still an open problem which has not been proved yet .
New encryption algorithms will be developed, if the accuracy of Goldbach conjecture can be proved.
Goldbach conjecture arose in a letter which had written from Christian Goldbach to Euler at the 7th of June
1942. By this letter, mathematician Goldbach wanted Euler to prove the premise of the Goldbach conjecture if it
is completely correct or false in any case [2]. This conjecture claims that the given premise:
every even number which is greater than number two is summation of the two prime numbers.
Several researchers have been tried to solve this conjecture up until now. With the support of computers, several
experiments were done up to the number of 1,1x10 18 to determine the correctness of the conjecture [5]. But it has
not been proved yet.
It has been intended to examine new patterns that are related to Goldbach conjecture by using Goldbach
conjecture data of Hartoka [1]on numbers up to 109. The equivalence classes that constituted by set of even
numbers were determined. Then various parameters of these equivalence classes have been examined to
determine if they match with a certain pattern or not.
306
2 Golbach Conjecture
Goldbach partitions show the different ways of obtaining even number 2N from prime summations.
Goldbach partitions which is also expressed with the symbol G(2N) was formulated as below:
The graphical view of G(2N) for the even numbers between [4;100,000] are shown below. The
characteristics of the obtained graphical view looks like a stray comet which firstly realized by Fliegel and
Robertson (1989). Because of this property, it is called Goldbach Comet.
Generally it is obtained that G(2N) values increase with increasing 2N values. On the other hand, an idea[2]
about the relation between 2N~G(2N) is explained below:
it is very likely the value of G(2N) depends on prime factors of 2N. For example, when 120 = 2.2.2.3.5 then
G(120) = 12. On the other hand when its close neighbours 118 = 2.59 , 122 = 2.61 then G(118) = 6 and G(122)
= 4 respectively.
The formula of the results which expresses the relation between G(2N) and the prime factors has already
defined by Sylvester. The expression is given below [2]. In the expression, when value p is the greatest one of
the equivalence classes, the effect of (p-1)/(p-2) is small and it is big when p is the smallest one. The value of
C2 is 0.6601618 as being constant of twin prime number.
/
2
N
2
N p
p
1
G
(
2
N
)
2
C
(2)
2 2
p
2
log
2
N
2
log(
2
N
)
2
2
N
307
3 Method
(3)
The Goldbach partition values used in this study are shown in the following table. By using below table,
Goldbach equivalence classes were determined and some parameters of them were examined.
Table 1: Goldbach partitions values in study
2N
G(2N)
1
................
999,999,996
3,407,072
999,999,998
1,705,026
1,000,000,000
2,274,205
For each equivalence class, class size (|Gi|), minimum (min(Gi)) and maximum elements (max(Gi)), sum of
elements (Ti), maximum elements to minimum elements ratio (max(Gi)/min(Gi)) and the average element size
(Ti/|Gi|) values has been calculated. Then tables of Goldbach equivalence classes were generated.
Table 2 :Tables of Goldbach Equivalence Classes
Gi
|Gi
|
min(Gi)
max(Gi)
Ti
max(Gi)/min(Gi)
Ti /|Gi|
12
30
7.5
10
68
244
6,8
27.11111
11
22
128
582
5.818182
52.90909
.......................................................................
16
22
246
1,112
13,278
4.520325
603.5455
17
25
288
998
18,114
3.465278
724.56
18
35
240
1,412
30,544
5.883333
872.6857
308
29
210
1,202
21,708
5.72381
748.5517
20
26
324
1,448
25,864
4.469136
994.7692
21
25
300
1,718
22,156
5.726667
886.24
22
27
360
1,532
29,096
4.255556
1077.63
.......................................................................
From the obtained table, it is generally observed for Gi>Gi-1 that all examined parameters of the equivalence
classes of Goldbach partitions has increased. The first odd case in which all Gi-1 parameters are smaller than
those of Gi's has been observed for G18.
min(G19)<min(G18)
max(G19)<max(G18)
T19<T18
max(G19)/min(G19)<max(G18)/min(G18)
T19/|G19|<T18/|G18|
The only G(2N) value which equals to 7,531,017 among even numbers of 2N in the range of
[4;1,000,000,000] is 999,068,070. This case reminds us that there may be even numbers greater than 10 9 whose
G(2N) values equal to 7,531,017. There are also 370,442 rows in Goldbach equivalence classes table whose |Gi|
values equal to 1. By considering this information, it was concluded that taking |Gi|= 7,531,017 as max Gi value
could cause deficiency of obtaining parameters such as |Gi|, max(Gi), Ti, max(Gi)/min(Gi) and Ti/|G(i). The
max(Gi) value of the chosen Gi should ideally be very close to 109. In order to determine this value the graphical
view of |Gi| versus Gi in the range i=[1;7,531,017] is shown in Fig.2. In this graphic, there is a sharp decreasing
in |Gi| for Gi>1,7x106 which makes graphic degenerated. It has been concluded that max(Gi) for Gi<=1,7x 106
is smaller than 109. The following graph supports this opinion. As a result, it has been decided to examine Gi
and other parameters for 2N=[4;1,700,000] range.
309
310
Table 3: max and min value of studied parameters of Goldbach equivalence classes for [1;1,700,000] range.
l
a
r
g
e
s
t
l
o
w
e
s
t
Gi
|Gi
|
min(Gi)
max(Gi)
Ti
max(Gi)/min(Gi)
Ti/|Gi|
1,700,000
229
246,333,360
998,494,42
4
189,485,964,596
4.05
827449627.05
1,692,234
312
245,274,120
993,774,39
8
253,716,888,492
4.05
81395155.42
1,585,153
218
378,048,264
923,882,04
8
169,897,572,782
2.44
779346664.13
1,699,860
240
231,718,410
999,434,008
198,417,809,604
4.31
826740873.35
1,692,234
312
245,274,120
993,774,398
253,716,888,492
4.05
813195155.42
10
68
244
6.8
27.11
1,695,491
236
270,886,560
995,687,138
202,797,551,780
3.67
859311660.08
12
30
7.5
12
30
7.5
12
30
7.5
12
30
7.5
12
30
7.5
591,686
189
129,900,468
307,102,468
49,524,358,308
2.36
262033641.84
12
30
7.5
Remarkable point in table of Goldbach Equivalence Classes has been stated as; Ti/|Gi| shows continuous
growth, whereas max(Gi)/min(Gi) value has remained under value 7 and not exceeded value 6 except for G2 and
G51. Max (Gi) / min (Gi) 's average in table of Goldbach Equivalence Classes has been 3.63.
The max of the G(2N) value for even numbers up to 500,660,160 had already been found as 3,977,551 by
Jrg Richstein [2]. In this study, max of the G(2N) value for even numbers up to 109 has been found as
7,531,017.
3.2 Creating Table of Prime Factor of Edge Members
Edge members of Goldbach Equivalence Classes have been examined to determine what kind of a
connection exists between their prime factors. Prime factors of the largest elements of Goldbach equivalence
classes have been found. The table of max(Gi)'s prime factors has been created. The largest Goldbach partitions
value in table of max(Gi)'s prime factors has been taken as 1,7.10 6. Studied properties of largest members
prime factors for Goldbach equivalence classes are expressed in the following equation.
max(Gi)= p1k1p2k2p3k3 prkr (p1<p2<< pr)
|max(Gi)|= k1+k2..+kr
(4)
|max(Gi)|*= k1+k2..+kr kr 2
{n1, n2, n3 nr}={p1, p2, p3 pr}
311
Prime factors of the smallest elements of Goldbach equivalence classes have been found. The table of
min(Gi)'s prime factors have been created . The largest Goldbach partitions value in table of min(Gi)'s prime
factors has been taken as 7,531,017. Studied properties of lowest members prime factors for Goldbach
equivalence classes are expressed in the following equation.
min(Gi)= m1k1m2k2m3k3mrkr (m1<m2<<mr)
(5)
|min(Gi)|= k1+k2..+kr
|min(Gi)|*= k1+k2..+kr kr 2
{n1, n2, n3 nr}={m1, m2, m3 mr}
The largest and lowest values of some of studied parameters for max (Gi)'s prime factors for i=[1;1,700,000]
range and min (Gi)'s prime factors for i=[1;7,531,017] range have been shown the table below.
Table 4: The largest and lowest values of some of studied parameters of max (Gi)'s prime factors and min
(Gi)'s prime factors
|max(Gi)
|
|max(Gi)|
*
pr
|min(Gi)
|
|min(Gi)|
*
mr
largest
21
499,639,501
25
16
33,330,679
lowest
These results have shown that number of min(Gi)'s prime factors different from 2 have been much more
compared to max(Gi). However, this study has supported with finding alleged by Jrg Richstein that G(2N) is in
correlation with the prime factors of the 2N (Richstein,2000)[2]. It has been observed that prime factors of
min(Gi) have been numerous and small. It has been also observed that prime factors of min(Gi) average has
been 5.77. Besides, it has been observed that prime factors of max(Gi) have been few and large. It has been also
observed that prime factors of max(Gi) average has been 1.12.
Table 5: Table of max(Gi)'s prime factor
Gi
max(Gi)
max(Gi)
|max(Gi)
|
|max(Gi)|
*
n1
n2
n3
n4 n5
n6
n7
n8 n9 n10
pr
12
{2;2;3}
68
{2;2;17}
17
17
128
{2;2;2;2;2;2;2;2}
.......................................................................................................
1,699,99
8
998,559,454
{2;499,279,727}
499,279,727
499,279,727
1,699,99
9
99,432,012
{2;2; 24,858,003}
24,858,003
24,858,003
1,700,00
0
998,494,424
{2;2;2;10,163;12,281}
10,163
12,281
12,281
312
min(Gi)
min(Gi)
|min(Gi)
|
|min(Gi)|*
n1
n2
n3
n4
n5
n6
n7
n8
n9
n10
pr
{2;2}
10
{2;5}
22
{2;11}
11
11
.................................................................................................
7,401,84
3
979,668,690
{2;3;5;7;11;13;19;101}
1
1
1
3
17 19
101
101
7,401,89
3
995,337,420
{2;2;3;5;11;17;19;23;29}
10
1
1
1
7
19 23
29
29
7,531,01
7
999,068,070
{2;3;5;7;11;13;17;19;103}
1
1
1
3
17 19
103
103
4 Conclusion
In this research, several parameters on Goldbach partitions have been investigated. Goldbach equivalence
classes, consisting of those even numbers with the same number of Goldbach partitions, were determined. For
each equivalence class, class size, minimum and maximum elements, sum of elements, maximum elements to
minimum elements ratio and the average element size values has been calculated and investigated for patterns.
Characteristic properties of edge members of equivalence classes were examined by factorization. It was
detected that lowest members prime factors of Goldbach equivalence classes had numerous and small while
largest members prime factors of Goldbach equivalence classes had few and large. The examined parameters in
this work suggest the correlation between G(2N) and the prime factors of the 2N. When number of odd prime
factors of even number gets higher, Goldbach partitions value also gets higher. Besides, the proposition that
G:2N>N surjective is verified for 1<=G(2N)<=4,020,437 interval.
313
References
1. Hartoka, A., (2004), An Investigation on Prime Number Patterns and Goldbach Conjecture, Master
Thesis, Ege University UBE, 146s.
2. Richstein, J., (2000), Computing the number of Goldbach partitions up to 5.108, In Proceeding of the 4th
Symposium on Algoritmic Number Theory, Leiden, 1838:475-490.
3. Richstein, J., (2000), Verifying the Goldbach conjecture up to 4.1014, Mathematics of Computation,
70(236):1745-1749.
4. Fliegel, H.F. and Robertson, D.S., (1989), Goldbachs Comet: the numbers related to Goldbachs
Conjecture, J.Recreational Math., 21(1):1-7.
5. Oliveira e Silva, T., (2008), Goldbach Conjecture Verification, http://www.ieeta.pt/~tos/goldbach.html
(31.08.2008 14:00).
6. Y. Saouter, (1998),Checking the odd Goldbach conjecture up to 1020, Math. Comp. 67, 863{866. MR
98g:11115}
7. J.R.Chen, (1989), Y.Wang, On the odd Goldbach problem, Acta Math. Sinica, 32, 702{718.MR
91e:11108}
8. J. M. Deshouillers, H. J. J. te Riele, Y. Saouter, (1998), New experimental results concerning the
Goldbach conjecture, Proc. 3rd Int Symp. on Algorithmic Number Theory,
.LNCS 1423 204{215. CMP 2000:05}
Biographies
Derya zkan Derya OZKAN was born in zmir. She graduated from Istanbul University Engineering Faculty
(BS, 2001) and got MS degree from Ege University International Computer Institute (Computer Science, 2008).
Currently, she is PhD. Candidate at Istanbul University Department of Computer Engineering.
Bahar lgen Bahar LGEN was born in zmir. She graduated from Dokuz Eylul University Engineering
Faculty (BS, 2000) and got MS degree from Ege University International Computer Institute (Computer Science,
2006). Currently, she is PhD. Candidate at Istanbul Technical University Department of Computer Engineering
and research assistant in the same department at Istanbul Kltr University. Her research areas consist of Natural
Language Processing, Artificial Intelligence, Pattern Recognition, Data and Text Mining.
314
fsen@turgutozal.edu.tr, 2moe@hs-furtwangen.de
Abstract. Besides its life saving features, mechanical ventilators also carry the risk of damaging the lungs of patients. To
prevent this ventilator induced lung injury (VILI), the patient's lung mechanics should be known and respected when making
therapeutical decisions, like settingthe positive endexpiratory pressure (PEEP)or tidal volume. Relatively simple
physiological models are supposed to predict individual reactions to changes in ventilator settings thus allowing to improve
current treatment strategies and then try to improve patients life quality. The aim of this study was to analyze datasets which
were obtained from a group of ARDS (Adult Respiratory Distress Syndrome) patients and to capture individual mechanical
properties of the patients lung.
Introduction
1.1 Background
Artificial respiratory devices are an indispensible therapy in lung failures. It is life-saving on one side.But on the
other hand, may threaten the lives of patients [1]. The most common reason for this problem is non-suitable
energy transfer from the respiratory device to the patients lung leading to baro-, or volume trauma. To improve
thesituation, lung mechanics should be known. Suitable ventilator settings, i.e. PEEP level, tidal volume, should
be adjusted according to the individual properties of the patients lung. Due to thisrelevance research focuseson
the analysis of lung mechanics with the help of physiological models. Better ventilation strategies will be
hopefully derived to better the situation for patients with severe lung disease.
1.2 Objective of the Investigation
The purpose of this study was to analyze pressure-dependent characteristics of ARDS lungs. The investigated
inspiratory range was constrained by a maximal plateau pressure of 45 mbar. Datasets were obtained during a
SCASS (Static Compliance by Automated Single Volume Steps) maneuver [2]. Data analysis was done with the
help of a viscoelastic lung model [3]. Multiple Linear Regression (MLR) analysis is applied for the parameter
estimation [4].The investigation comprised dynamic effects (due to applied MLR analysis) e.g. during
inspiration phase and also static effects e.g. relaxation state of the lung parenchyma at the end of the endinspiratory pause (EIP).
315
Methods
2.1 Datasets
In a multi-center study (the university hospitals of Aachen, Berlin, Bonn, Freiburg i. Br., Gttingen, Hamburg,
Mannheim-Heidelberg and Munich) the datasets of a standardized SCASS maneuver for this study were
obtained.In this SCASS maneuver the performed tidal volumes are selected randomly in a specified range. For
every breath cycle, the plateau pressure is limited to maximal 45 mbar thereby restricting the volume range. The
gas inflation is followed by a5 second pause called endinspiratory pause (EIP).Pressure and flow measurements
are recorded continuously during the ventilation. After each applied SCASS maneuver cycle the patient is
provided with 5 normal breaths. In this way, datasets over the entire lung capacity are obtained and static
compliance can be computed [2].
2.2 Data Analysis
Viscoelastic Model. In this study a viscoelastic model is selected to analyze also viscoelastic properties of the
lung parenchyma. The viscoelastic model was first introduced by Mount in 1955 [3].Viscoelasticity is described
by temporary modifications in stress subsequent to instantaneous alteration in strain [12, 13]. This model belongs
to the class of second order pulmonary models.
Fig. 1.Representation of lung models with the help of electrical analogs. A resistor represents airway resistance and a
capacitor lung compliance. Viscoelasticity is represented (in green) by a parallel combination of resistor R ve and capacitor
Cve.
Parameter Estimation.Theparametersof the viscoelastic model, i.e. Newtonian Resistance (Raw), Viscoelastic
Resistance (Rve), Static Compliance (Cst) and Viscoelastic Compliance (Cve), were calculated by a fitting process
to the patient data.. This process was applied for every volume cycle of the SCASS maneuvers in 17 patients
which summed up to 318 cycles for the fitting process the Viscoelastic Lung Model, as introduced in Figure 1,
was converted to the Laplace domain and represented as a transfer function:
316
In transfer function notation, the numerator represents the Laplacian of respiratory system pressure and the
denominatorthe respiratory system flow. The fitting process was performed using this transfer function to
evaluate the quality of fit for any intermediate parameter constellation. The Matlab (Mathworks, Natick MA,
USA) function fminsearch was used for the optimization steps. [14]
Results& Discussion
Fig. 2.Estimated parameters for all datasets in the way of lower quartiles, medians and upper quartiles are depicted against
plateau pressure. Overall median values are defined on the right side of graphs.
317
collapsed alveoli infunctional dead space of the lung under high pressures. These reopened alveolimay partially
explain the found increases in compliance.
In a SCASS maneuver, the maneuver cycles all start at Functional Residual Capacity (FRC), i.e. at Zero End
Expiratory Pressure (ZEEP).,. To ameliorate the functionality, effects of Positive End Expiratory Pressure
(PEEP) should also be analyzed. In this context, Italian researchers investigated respiratory resistance of some
datasets which are obtained from 16 mechanically ventilated patients who suffer from acute lung injury (ALI).
Cycles start from a predefined PEEP level. Also this study showed that resistance increases non-linearly in
ARDS patients accordingly with increasing pressure. [17,18].
3.3 Method Validation
Datasets were obtained with the following characteristics:
- Rapid flow interruption for analyzing relaxation effects in the lung parenchyma,
- Comprehensive pressure range,
For these purposes, SCASS maneuver was selected. With SCASS maneuver just one P/V dataset is specified
while a normal breath cycle intermitted. The one evaluation takes just 6 seconds. In the midst of every
measurement patients are allowed to take normal breath. Unlike conventional methods, as in super syringe
maneuver which uses a non-physiological tremendous breath, the changed lung history impact is inconsiderable
in the SCASS maneuver. Due to reliability of measurements, gas leakage of the maneuver has to be totally
removed. In SCASS method this handicap can be eliminated via an elegant pressure control in the course of
occlusion. Maneuver settings allow applying plateau pressures up to 45 mbar due to security reasons [19].In
contrast to conventional methods, during data analysis MLR uses pressure, flow and volume data series of the
entire breath [4]. In this way found results can also represent dynamic parameter of the lung mechanics. So they
can be used for continuous tidal ventilation.
3.4 Limitations
Catching sight of low plateau pressure rates were seldom in ARDS. There is huge amount of collapsed alveolus,
dead spaces, in the ARDS lungs and the lung becomes stiff. Therefore opening pressure rates of the alveolus are
higher than normal lungs. For this reason, parameter estimations were scarcely performed in low pressure rates.
A second adversity in data analysis is restricted capability of the model. In some datasets, the model could not
capture all lung dynamics and so could not reach any estimation. To eliminate this handicap, I tried to change
initial numbers. But this did not remedy. Further rectifications should necessary to better the fitting process. The
chosen programming language, Matlab, for the modeling is compatible for future improvements and additions.
In past many different mathematical models were tried to enhance the ventilation strategies. But the most
promising ones of them were the models which are generated accepting the effect of the other physiologic
systems of the human body on the respiratory system.[20] This approach is in fact more realistic. These new
models seem to open a new era in ventilation strategies.
4
Conclusion
Viscoelastic model may supply elucidatory knowledge about the changes in the state of ARDS patients
respiratory system [9, 15]. MLR analysis with the flow interrupter technique provides preferable results than
conventional analysis methods. Its fits of obtained datasets are more robust and less noise sensitive than
previously used techniques, such as two-point analysis [15, 7]. In future, the combination of other physiological
models (i.e. respiratory mechanics, gas exchange and cardiovascular dynamics or other any relevant model)may
ensure a more complete model which will help to find more beneficial ventilation strategies for the patients [20].
Moreover, with the help of this reliable models, therapies can be applied more easily via worldwide
communication opportunities, i.e. GSM, internet.
318
References
1. Dreyfuss D, Saumon G.January (1998) "Ventilator induced lung injury; lessons from experimental studies." Am J
RespirCrit Care Med: 294-323.
2. Sydow M, Burchardi H, Zinserling J, Ische H, Crozier TA, Weyland.(1991) "Improved determination of static compliance
by automated single volume steps in ventilated patients." Intensive Care Med: 108-114.
3. Mount, LE.(1955) "The ventilation flow-resistance and compliance of rat lungs." J Physiol.: 157-67.
4. Guttmann J, Eberhard L, Wolff G, Bertschmann W, Zeravik J and Adolph M. (1992) "Maneuver-free determination of
compliance and resistance in ventilated ARDS patients." CHEST: 1235-1242.
5. Fishman, Alfred P. 2005 "One Hundred Years of Chronic Obstructive Pulmonary Disease." Am. J. Respir. Crit. Care
Med.: 171: 941 - 948.
6. MICHAEL P. MORTELLITI, M.D., and HAROLD L. MANNING, M.D.(2002) "Acute Respiratory Distress Syndrome."
Am FamPhysician: 65(9):1823-1831.
7. Guttmann, J. (1999) "Grundlagen der Lungenmechanik unter Beatmung." Intensivmedizin und Notfallmedizin:
Vol.36:Number 9:S001-S008.
8. C. Schranz, J. Guttmann, and K. Mller, (2010) "Fitting respiratory mechanics in ARDS by a nonlinear recruitment model
with viscoelastic component," Biomed Tech., vol. 55 (Suppl. 1)
9. Ewart R. Carson, Claudio Cobelli, 2001.Modelling Methodology for Physiology and Medicine.Academic Press.
10. Khoo, Michael C K. 2000. Physiological Control Systems: Analysis, Simulation and Estimation. New York: IEEE.
11. Lagarias, J.C., J. A. Reeds, M. H. Wright, and P. E. Wright. (1998) "Convergence Properties of the Nelder-Mead Simplex
Method in Low Dimensions." SIAM Journal of Optimization: Vol. 9 Number 1, pp. 112-147.
12. Bayliss LE, Robertson GW. (1939) "The visco-elastic properties of the lungs." Q J Exp Physiol Cogn Med Sci: 29: 27
47.
13. Zin, De Bora S. Faffe and Walter A. (2009) "Lung Parenchymal Mechanics in Health and Disease." Physiol Rev: 89:
759770
14. SEN, I.F., (2010) Model Based Analysis of Lung Mechanics via Respiratory Maneuvers Hochschule Furtwangen
University Thesis Archive. World Wide Web: http://www.item.hs-furtwangen.de/HFU/
15.Ganzert S, Mller K, Steinmann D, Schumann S and Guttmann J. (2009)"Pressure-dependent stress relaxation in acute
respiratory distress syndrome and healthy lungs: an investigation based on a viscoelastic model." Critical Care: 13:R199.
16. Fredberg JJ, Stamenovic D. (1989) "On the imperfect elasticity of lung tissue." Journal of Applied Physiology: Vol 67,
Issue 6 2408-2419.
17. Pelosi P, Cereda M, Foti G, Giacomini M, Pesenti A. (1995) "Alterations of lung and chest wall mechanics in patients
with acute lung injury: effects of positive end-expiratory pressure." Am J RespCrit Care Med: 152:531-537.
18. Pesenti A, Pelosi P, Rossi N, Virtuani A, Brazzi L, Rossi A. (1991) "The effects of positive end-expiratory pressure on
respiratory resistance in patients with the adult respiratory distress syndrome and in normal anesthetized subjects." Am Rev
Resp Dis: 144:101-107.
19. Ganzert S, Guttmann J, Kersting K, Kuhlen R, Putensen C, Sydow M, Kramer S. (2002) "Analysis of respiratory
pressure-volume curves in intensive care medicine using inductive machine learning." ArtifIntell Med: 26:69-86.
20. Kretschmer J, Wahl A,Guttmann J and Moeller K. (2010) "Dynamic Generation of Physiological Model Systems." In
Proceedings of the World Congress on Medical Physics and Biomedical Engineering 7-12 September: Volume 25/4, 334337.
319
Biographies
smet Ferdi en Mr. en holds a BSc. Degree (2006) in Industrial Engineering form Gaziantep University
(Turkey), MSc. (2010) degree in Biomedical Engineering from Furtwangen University (HFU, Germany).
His research focuses on physiologic modeling, simulation. He also works on forensics and pattern recognition.
He recently joined the Department of Computer Engineering at Turgut Ozal University (Turkey.)
Knut Mller Prof.Mller received his Diploma (1986) in Computer Science with the minor field of study in
Operations Research, Ph.D. (1991) degree in Computer Sciences and an MD degree in Medicine (1992) from
Bonn University, Germany.
Prof. Mller has published over 250 articles, reports and books. His research focuses on medical computer
science, telemedicine and automation, physiological modeling, intelligent monitoring and artificial neural
networks in signal analysis and pattern recognition.
Since 1998, he has been professor in the field of Medical Computer Science, since 2001 he has been director of
the Master Program of Biomedical Engineering, and since 2010 he has been Director of the Institute of
Technical Medicine (ITeM) and Vice Dean of the Faculty of Mechanical and Process Engineering at Furtwangen
University (HFU), Germany.
320
Abstract. In this study we propose a design of genetic algorithm based clustering technique for
wireless sensor networks. The aim of our paper is to create energy efficient and optimal number of
clusters with cluster heads. We used a mobile agent that runs Genetic Algorithm (GA) technique
which is an evolutionary algorithm, to form energy efficient clusters. We analyzed and compared
our algorithm with one of the best known clustering technique, namely, LEACH.
Keywords: wireless sensor networks, clustering, genetic algorithm
Introduction
Wireless Sensor Networks are composed of number of energy constrained sensor nodes which have
ability to sense, message transmission and information processing. Because of the limited energy resource,
several techniques are proposed to decrease energy consumption, such as clustering which is one of the
fundamental technique used in many different types of sensor network applications [1]. In clusters, nodes are
clasiffied as either head or member nodes. While members of clusters are responsible of sending their sensed
data to their cluster heads, cluster heads are responsible of processing these collected sensed data and
transmitting this data to the sink node. The reason why clustering technique is important for efficient energy
consumption is, member nodes send their sensed data only to their cluster head instead of sending directly to
the sink node which is much more far away from the member nodes than the cluster heads.
In this study we propose a genetic algorithm based clustering technique for wireless sensor networks. The
aim of our paper is to create energy efficient and optimal number of clusters with cluster heads. In order to
achive this aim, we assumed a mobile agent that moves to each node and gathers data to use in partitioning the
network into clusters. We used Genetic Algorithm that includes a fitness function which is determined in
accordance with the parameters such as uniformity of the clusters, energy consumption, cluster distance and
direct distance to sink. The rest of the paper is organized as follows: In section 2 background information of
the paper is given. Section 3 describes the details of the proposed algorithm and lastly analysis are given in
Section 4.
321
2 Background
In this section, best known clustering algorithms in wireless sensor networks are summarized. Following this,
the main steps of the genetic algorithm are described.
Clustering is a very important optimization problem for increasing network scalability and lifetime [2]. With
this technique, energy consumption of the sensor nodes are decreased by making them able to send their data to
the cluster head node that locates close to them. There are many clustering techniques in literarute. In [3],
LEACH protocol, a distributed hierarchical self-organized cluster based algorithm, is introduced. In this
technique, sensor network is divided into several clusters where number of clusters are determined previously.
Cluster heads are responsible of collecting data from the member nodes and sending this data to the sink node.
As the higher energy consumption of cluster heads, the role of the cluster heads rotates periodically. In this
technique, each node becomes a cluster head at least once. In [4,5], HEED protocol is proposed which is a
distributed clustering approach that aims to increase network lifetime by distributing energy consumption. This
technique periodically selects cluster heads according to a hybrid of the node residual energy and does not make
any assumptions about the presence of infrastructure. In [6], EEMC protocol is proposed which is an energy
efficient multi level clustering algorithm that aims to decrease energy consumption. This protocol has two main
phases: cluster set-up phase in which the multi-level clustering topology is formed and the data transmission
phase in which the nodes send their data to the sink node via this clustered topology. In [7], HCR protocol is
proposed in which nodes self-organize into clusters that are identified using heuristic based approach. In this
technique GA is used to generate energy efficient hierarchical clusters. In [8], HCR protocol is improved by
using GA to determine the number of clusters, the heads, the members and the transmission schedules. In [9],
GA is used for optimization energy usage and to form number of pre-determined clusters. [10] extends the
algorithm proposed in [9] by improving the GA fitness function.
Genetic Algorithm technique is one of the evolutionary algorithms that mimics the process of natural
evolution. Population, Fitness Function, Selection, Crossover and Mutation are the main steps of genetic
algorithms. Populations are set of solutions and include chromosomes that are composed of series of numbers.
Fitness function measures the quality and the performance of a solution. This step should include the most
important factors that affect the performance of the system. In Selection step, chromosomes with larger fitness
values are selected to be participated in reproduction step. In crossover step, new generation is developed by
passing information from one generation to next. Newly generated chromosomes include portion of information
from their parents. Lastly in mutation step, parts of the chromosomes are changed randomly and the aim of this
step is to avoid much similiarity in population [11].
3
Genetic Algorithm Based Energy Efficient Clustering for Wireless Sensor Networks
In this study, we used genetic algorithm technique to form energy efficient clustered sensor network. We
represent each of the population member in a binary format and each bit represents a node, called gene. While an
ordinary node is represented with the value of 0, a head node is represented with 1. Because of the efficient
energy usage, instead of making sink node to partion the network into clusters, we used a mobile agent that
moves to each sensor node and runs genetic algorithm according to the data it collected from the sensor nodes.
We define the iterations of the agent as forward and backward iteration. In forward iteration, agent collects data
from the sensor nodes and in the backward iteration agent forms the clusters. As the mobile agent enters the
network area and starts its forward iteration, it broadcasts an ADVERTISEMENT message in order to invite
322
nodes to send their node_id, location_info and the battery_status. When the ordinary and the head nodes receive
ADVERTISEMENT message, they send REPLY_ADVERT message including requested information to the
mobile agent. After moving to each node, agent runs the genetic algorithm and determines which of the ordinary
node should be belong to the which of the head nodes cluster. Following this, agent starts its backward iteration
to form clusters.
In order to achive this, it brodcasts an ALERT message to head nodes. When a head node receives ALERT
message, it sends REPLY message including its id number to the agent. Agent looks up its table and sends
INFORMATION message back to the head node including the ids of ordinary nodes belonging to its cluster. As
soon as the head node receives INFORMATION message, it sends CLUSTER message to the nodes whose ids are
included in the INFORMATION message. When the ordinary nodes receive CLUSTER message from the cluster
head, they become members of the cluster of that cluster head. In order to keep on cluster member ships, mobile
agent makes the forward and backward tours periodically.Battery alert, node depletions and deads trigger the
agent to run the genetic algorithm.
Initial chromosomes are generated randomly using random number generator. Population uses the gene
stucture in which the ordinary and head nodes are represented with the 0 and 1 binary values respectively.
Fitness function is determined in accordance with the parameters such as uniformity of the clusters, energy
consumption, cluster distance, direct distance to sink and cluster distance [13]. We used the fitness functions
proposed in [12] and [13].
To prevent cluster head overhead, nodes connected to each cluster head should be uniformly distributed. For
calculations of uniformity of the cluster, we used following equations [12]:
where n is the cluster head number, N is the number of cluster heads, pn is the number of nodes attached to this
cluster-head, and p is the average number of nodes per cluster in a system calculated as [12]:
= Total Sensor Nodes/Total Cluster Heads
Direct distance to the sink, which is sum of all distances from nodes to sink, is an another factor that affects the
fitness function. This distance is defined as follows [13]:
Another factor that affects the fitness fuction is cluster distance which is the sum of the distances from the nodes
to the cluster head and the distance from the head to the sink, is calculated as follows [13]:
where k is the number of member nodes, dih is the distance from node ito the cluster head hand dhsis the distance
from the cluster head hto the sink node s [13].
323
In our algorithm, with the usage of mobile agent, we prevent very high number of message transmission from
cluster heads to sink node. Cluster heads only send their calculated data such as a vehicles calculated location to
the sink node instead of sending states of the nodes. By this way, we decrease the number of message
transmission and gain from energy consumption.
For a cluster with kmember nodes, energy consumption for transmitting the messages from member nodes to the
cluster head, for cluster head to receivemessages from the member nodes and for transfering the aggregated
message from the cluster to the sink are determined respectively as follows [13]:
Last factor that affects the fitness function is remaining energy that represents the total energy available in all the
sensor nodes of a cluster, is calculated for k member nodes as follows [13]:
Chromosomes with the larger fitness value are selected to partipate in reproduction step in which new
chromosomes are generated from these selected chromosomes. For crossover step, we used multi-point crossover
technique in which the crossover points and locations are calculated with random number generator. An example
for reproduction of new chromosomes from two chromosomes having two random cross-over points is given
below:
Before Crossover:
Parent1:
1101000111010100
Parent2: 1 0 0 0 1 1 0 1 0 1 0 0 1 0 1 1
After Crossover:
Child1:
110111010100 0100
Child2:
1101110101000100
Following crossover step, newly reproduced chromosomes are transferred to the mutation step in which one part
of the chromosome is changed randomly according to the mutation probability.
324
References
1. Akyildiz, I. F., Su, W., Sankarasubramaniam, Y., Cayirci, E. (2002). Wireless sensor networks: A survey. Computer
Networks, vol. 38, no. 4, pp. 393-422.
2. Erciyes, K., Dagdeviren, O., Cokuslu, D., Ozsoyeller, D. (2007). A Survey of Graph Theoretic Clustering
Algorithms in Mobile Ad hoc Networks and Wireless Sensor Networks. Journal of Applied and Computational
Mathematics, pp. 162-180.
3. Heinzelman, W., Chandrakasan, A., Balakrishnan, H. (2000). Energy-Efficient Communication Protocol for
Wireless Microsensor Networks. HICSS 2000, vol. 2.
4. Younis, O., Fahmy, S. (2003). Distributed Clustering for Scalable, Long-Lived Sensor Networks, MOBICOM 03.
325
Biographies
Dr. Aysegul ALAYBEYOGLU was born in Manisa in 1983. She received the BSc. degree in Computer
Engineering. from the University of Sakarya in 2005 and a PhD degree in Computer Engineering from Ege
University in 2009. She currently works as an assitant professor in the Computer Engineering Department of
Celal Bayar University. Her research interests are broadly in parallel and distributed systems and computer
networks. More precisely, she works on distributed algorithms for middleware functions in wireless sensor
networks.
Email: aysegul.alaybeyoglu@bayar.edu.tr.
326
Abstract. In this study, a unique model is generated for implementing information security risk analysis by
the aid of machine learning algorithms and this model is applied to a case, where some real data collected
from a qualitative information security risk assessment survey in an organization. In the first part of the study,
a unique survey was carried out for some specific non-quantifiable risks in the organization that were directly
related with their major and critical business processes. The second part of this study is to implement a unique
machine learning risk classification model that deduces the risk level in a binary format with a proper data set
from the results obtained in this survey. In the final phase, several binary classifier algorithms and their
training and test performances are observed by learning curves, Kappa statistics, F-measures and ROC
curves.
Keywords: qualitative risk analysis, information security survey, machine learning, binary classifiers
1 Introduction
In todays business world and even in our daily life, proper and accurate assessment and management of the
information security risks has become a crucial issue. There are several models, methodologies and standards to
assess and analyze information security risks [1], [2], [3], [4], [5]. However, information security risks cannot
always be estimated reliably because each company or organization might be facing different risks or the same
risks with different levels due to divergent environments, cultures, processes and organizations [6]. Some
researches have been made which implement information security risk assessments using machine learning and
similar computational intelligence and reasoning models such as fuzzy logic and belief functions [7], [8], [9],
[10], [11]. Most of these studies either focus on multi-classifier machine learning models for quantitative risks or
technological aspects of information security such as intrusion detection systems, e-mail filter systems and so on
[12], [13]. However, some information security risks are non-quantifiable due to lack of sufficient statistical data
or due to the intangibility of some assets. In such situations, there is always a need for reliable and accurate
automated qualitative risk assessment models that can be used easily by senior management without depending
on the knowledge of information security experts. In addition, such new models must be implemented to
minimize the drawbacks of qualitative risk methodologies such as subjectivity, uncertainty and false predictions
[1], [2].
Machine learning is simply defined as a provider of the technical basis of data mining [14]. It is a new
technology for mining knowledge from data in a manner that the system is trained so that it can establish
327
improved and reliable performance in new situations with new data [14]. Data mining can be defined as the
extraction of useful information from large data sets. It is accepted as an intersection of statistics, machine
learning, data management and databases, pattern recognition, artificial intelligence [15]. Data mining is also
accepted as a stage of a larger process known as Knowledge Discovery in Databases (KDD). Knowledge
discovery is defined as a process for identification of new valid understandable and potentially useful patterns
from large data sets. Hence, machine learning is a discipline that is used in cooperation with data mining where
both disciplines have intersection points, even that some of the algorithms and models are commonly used for
both implementations such as BayesNet, decision trees, neural networks [14], [15], [16]. Successful
implementations and projects have been accomplished in the context of machine learning such as; optimizing
industrial systems, minimizing business risks, fraud detection, prediction and assessment of new investments,
etc. [16], [17], [18].
2 Case Study
In this study, a qualitative information security risk assessment model based on machine learning is implemented
for a governmental institution in Izmir, Turkey. First, an information security survey was conducted. Then, a
machine learning model was generated using the results obtained from the survey. The model was tested among
different selected binary classifier algorithms and the results were analyzed using some standard measures in
order to obtain their comparative performances. It should be noted that due to privacy and confidentiality
considerations of the governmental policies, the name of the organization is not mentioned in this paper.
2.1 Information Security Risk Survey
The information security surveys respondents were the data entry operators and their department manager in the
institution. We held several meetings with the department manager for the design of the survey. Regarding the
major business processes in the data entry department and the data entry personnel, some of the specific
information security risk issues of the organization were identified. During these meetings, the information
security issues that that cannot be quantified and the ones that were the most crucial were included in the
surveys scope. Hence, only some specific assets, vulnerabilities and threats were taken into consideration. For
instance, some of the specified assets were department personnel, data entry application, equipment for formal
endorsement, etc. Some of the threats covered in the survey: disease / illness, natural disasters, intruders inside
the organization (malevolent employee), logical damages to data, etc. Some of the vulnerabilities covered in the
survey were; lack of / insufficient business continuity plans, personnels tendency to hustle or to be careless,
personnels tendency to be disgruntled due to negative work conditions, insufficient job orientation training after
recruitment, and so on. In this manner, 6 assets, 10 threats and 9 vulnerabilities were included in the scope of the
study. Thus, for each of the six assets, each of the 10 threats might impose a possible risk exploiting each of the
nine possible vulnerabilities. Regarding the many-to-many relationship, this would make up (6 x 10 x 9) = 540
possible combinations which would imply 540 distinct risks. However, in real life situations, most of these
possible combinations and relations are neither relevant nor sensible and their probability is zero. These were
automatically discarded from the survey. Some of the other possible combinations were not also taken into scope
of the survey due to institutions managerial primary requirements and strategic decisions. Consequently, the
scope of the survey was limited to 30 distinct topics, or in other words, 30 possible combinations of assets,
threats and vulnerabilities.
The questionnaire forms were given to the survey respondents in printout form where they have to fill in 270
answers amongst 30 risks and 9 evaluation criteria. These questions were either to be answered on a (Yes / No)
or (1 / 2 / 3 / 4 / 5) or (0 / 1 / 2 / 3 / 4) qualitatively ranked scale basis. Some of these survey questions are
denoted in Table 1.
328
Question
Scale Rank
Employees in the data entry department including their supervisor were grouped in several sessions and they
all answered the survey by giving scores to these nine questions for each of the 30 risk relations and filled in the
answer forms in paper format. Sixty-four people attended the survey. After collecting the answer sheets, all of
the qualitative survey scores were entered into an MS Excel file specially designed for the questionnaire, which
also provides some crosschecks and necessary corrections.
2.2 Modeling and Implementation with Binary Classifiers
The qualitative scores obtained from the respondents were analyzed and the threshold value for overall risk was
decided to be 4. The overall risk scores evaluated as 1, 2 or 3 by the survey respondents were defined as
acceptable risks and are marked as Risk = No where all the other scores (4 or 5) are marked as Risk =
Yes. Thus, the basic model was to estimate whether an instance was risky or not. This method has made it
feasible for the binary classifiers in machine learning models where each instance coming from the data set is to
be identified in any one of the two possible classes [14]. After this classification, the results are re-organized as a
proper data set to be used as input for the machine learning classifiers. The whole data set was made of 12
attributes (9 survey questions plus asset, threat, vulnerability parameters and one resultant binary Risk class) and
1920 instances (64 x 30, where 64 survey respondents each with 30 distinct risk topics). In addition, all of the
attributes were set to non-numeric (nominal) type. In the research, all the experiments are carried out by using
Weka software (version 3.6.0). Four different built-in classifier algorithms within Weka are chosen and this data
set is used for observing the learning performances (error rates and some other recommended metrics) of each of
these classifiers with several data sample sets (train and test sets) using cross-validation methodology [14], [19],
[20]. The names and short definitions of the classifier algorithms are denoted in Table 2.
For each of the experiments among these four classifiers (BayesNet, J48, Nave Bayes Tree, VFI), there were
two phases. First phase was the training phase where the whole data set was used for training the classifier
model. In the second phase, the same data set was used as test set by the aid of cross-validation methodology
[14]. 10-folds stratified cross-validation was chosen as a best practice option [14]. This was a crucial point in this
study because no classifier algorithm can be evaluated reliably only by observing its performance values for
training [19]. Some classifier models algorithms might have the danger of over-fitting that could be overcome by
using test sets as well as train sets [14].
329
The primary success or accuracy criterion for any of the classifier algorithms was to achieve a maximum of
10% error rate from the test set. In other words, the sum of TP (True Positive) and TN (True Negative)
classifications should be at least 90% of the whole test set. This can be simply formulated as follows:
Let
kC
TP TN
TP TN FP FN
(1)
* 100
Name
Type
BayesNet
J48
2.3 Results
The results obtained for the binary classifier algorithms using the whole data set for train and test phases where
10-folds stratified cross-validation are denoted in Table 3. It can be seen that J48 algorithm was hardly sufficient
for the training phase (correct classification rate: 90.854%), it was not accurate after the test phase (correct
classification rate: 84.912%) which is due to over-fitting problem. The other three algorithms were not overfitting and their learning performances were observed to be sufficient enough for the 90% correct classification
benchmark criteria. When the results of test phases are compared, BayesNet classifier had slightly the most
accurate values regarding error rates, Kappa Statistics, F-measures and ROC areas.
330
Classifier
name /
Phase
Correctly Kappa
TP
TN
F-measure ROC Area
classified Statisti Rate
Rate (weighted (weighte
(Risk = (Risk = average)
d
instances
c
Yes)
No)
average)
BayesNet /
Train
94.221 %
0.8671
0.94
0.94
0.942
0.989
BayesNet /
Test
92.135 %
0.8341
0.915
0.93
0.921
0.976
J48 / Train
90.854%
0.798
0.913
0.878
0.901
0.951
J48 / Test
84.912 %
0.705
0.86
0.851
0.86
0.934
Nave Bayes
Tree / Train
95.624 %
0.907
0.963
0.946
0.956
0.987
Nave Bayes
Tree / Test
91.302 %
0.82
0.913
0.921
0.915
0.967
VFI / Train
92.105 %
0.8342
0.915
0.93
0.922
0.975
VFI / Test
91.108 %
0.8153
0.911
0.915
0.912
0.965
Some additional experiments were also done to observe whether these three classifiers (J48 was excluded due
to its non-accurate results) provide a feasible learning curve among different test set sizes. For each of the four
classifiers, random test data were generated with set sizes of 153, 238, 321, 494, 900, 1170, 1496, 1920 are used
and their correct classification rates and F-measure rates are analyzed. The results showed that all of the three
classifiers seemed to reach their maximum limits of their learning capacities when the test set size was increased
up to 1920 instances. Some of these results are denoted within their learning curve progression in figures Figure
1 and Figure 2.
Correctly Classified
Instances
100%
90%
80%
70%
60%
50%
40%
30%
20%
10%
0%
0
200
400
600
800
1000
1200
1400
1600
1800
2000
331
Correctly Classified
Instances
100%
90%
80%
70%
60%
50%
40%
30%
20%
10%
0%
0
200
400
600
800
1000
1200
1400
1600
1800
2000
Fig. 2. Learning performance of Nave Bayes Tree within specified test set sizes
3 Conclusion
In this study, a qualitative information security risk assessment model was developed by the aid of machine
learning classifier algorithms. Since the risk deduction is based on two parameters (risky and non-risky), binary
classifier algorithms were shown to be a suitable model. Similar models and promising implementations can also
be derived for other companies. In addition, based on this model, new models can be generated for other
information security domains if risks are to be predicted by qualitative assessments. It was also observed that the
data set size and sample test data sizes were sufficient for the model that enabled us to observe over-fitting issues
and to deduce whether learning curve has reached to its maximum level.
Another important aspect in this study is the parameter selection and usage within the classifier algorithms.
Some of the parameters might be changed and additional results could be observed within the same data set. By
this way, enhanced performance values for the learning capabilities of these algorithms might be obtained.
Similarly, additional experiments could be made using some other suitable classifier algorithms.
It should also be mentioned that a subjectivity problem could exist in this study (and similar studies as well)
due to the values / scores provided by employees in qualitative risk assessments and relevant surveys. This
problem might be mitigated by means of additional new mechanisms and methods in the risk assessment
process.
References
1.
2.
Peltier, T.R. (2001). Information Security Risk Analysis. USA: Auerbach Publications.
Landoll, D.J. (2006). The Security Risk Assessment Handbook - A Complete Guide for Performing
Security Risk Assessments. USA: Auerbach Publications.
3. ISO. (2005). Information Security Management Systems - Requirements ISO/IEC 27001:2005. ISO pub.
4. Tipton, H.F. and Krause, M. (2007). Information Security Management Handbook. USA: Auerbach
Publications.
5. ISO. (2008). Information Security Risk Management ISO/IEC 27005:2008. ISO pub.
6. Pfleeger, S.L. and Cunningham, R.K. (2010). Why measuring security is hard. IEEE Security & Privacy
8(4): 46-54.
7. Qu, W. and Zhang, D.Z. (2007). Security metrics, models and application with SVM in information
security management. In Proceedings of the Sixth International Conference on Machine Learning and
Cybernetics. Hong Kong, China, 19-22 August. IEEE pub.
8. Lv, J.J., Qiu, W.H., Wang, Y.Z. and Zou, N. (2006). A study on information security optimization
based on MFDSM. In Proceedings of the Fifth International Conference on Machine Learning and
Cybernetics. Dalian, China, 13-16 August. IEEE pub.
9. Sun, L., Srivastava, R.P. and Mock, T.J. (2006). An information systems security risk assessment model
under Dempster-Shafer Theory of belief functions. Journal of Management Information Systems 22(4):
109-142.
10. Long, X., Yong, Q. and Qianmu, L. (2008). Information security risk assessment based on analytic
hierarchy process and fuzzy comprehensive. In Proceedings of the 2008 International Conference on
332
11.
12.
13.
14.
15.
16.
17.
18.
19.
20.
Risk Management & Engineering Management. Beijing, China, 4-6 November. IEEE Computer
Society.
Zhao, D-M., Wang, J-H., Wu, J. and Ma, J-F. (2005). Using fuzzy logic and entropy theory to risk
assessment of the information security. In Proceedings of the Fourth International Conference on
Machine Learning and Cybernetics. Guangzhou, China, 18-21 August. IEEE pub.
Pietraszek, T. and Tanner, A. (2005). Data mining and machine learning - towards reducing false
positives in intrusion detection. Information Security Technical Report 10(3): 169-183.
Giacinto, G., Roli, F. and Didaci, L. (2003). Fusion of multiple classifiers for intrusion detection in
computer networks. Pattern Recognition Letters 24(12): 1795-1803.
Witten, I.H., Frank, E. and Hall, M.A. (2011). Data Mining: Practical Machine Learning Tools and
Techniques 3rd ed. Burlington, MA: Elsevier Inc.
Hand, C., Mannila, H. and Smyth, P. (2001). Principles of Data Mining. London: The MIT Press.
Nedjah, N., Luiza, M.M. and Kacprzyk, J. (2009). Innovative Applications in Data Mining. Berlin:
Springer-Verlag.
Hastie, T., Tibshirani, R. and Friedman, J. (2009). The Elements of Statistical Learning, Data Mining,
Inference and Prediction, 2nd edition. New York: Springer.
Delen, D., Sharda, R. and Kumar, P. (2007). Movie forecast Guru: A Web-based DSS for Hollywood
managers. Decision Support Systems 43(4): 1151-1170.
Weiss, G.M. and Provost, F. (2003). Learning when training data are costly: the effect of class
distribution on tree induction. Journal of Artificial Intelligence Research 19: 315-354.
Vuk, M. and Curk, T. (2006). ROC curve, lift chart and calibration plot. Metodoloski Zvezki- Advances
in Methodology and Statistics 3(1): 89-108.
Biographies
Mete Eminaaolu Mete Eminaaolu graduated from the department of computer engineering, Ege
University, Turkey in 1996. He received his M.Sc. degree in 1999 at the department of computer engineering,
Izmir Institute of Technology, Turkey. Mete Eminaaolu continues his PhD education at the department of
computer engineering, Trakya University in Turkey.
He has worked in private sector for fourteen years as manager, auditor and consultant in several areas of
information security and IT governance. He currently works as lecturer at department of computer programming
in Yaar University, Turkey. His main research areas are information security risk assessment, machine learning,
data mining, information security management and cryptology in which he has several international publications.
Mr. Eminaaolu has some international certifications such as CISSP (Certified Information Systems Security
Professional) and CISA (Certified Information Systems Auditor).
Prof. Dr. aban Eren aban Eren graduated from the department of statistics, Bradford University, UK in
1970. He received his M.Sc. degree at the department of statistics, Aston University, UK in 1971 and another
M.Sc. degree at the department of computer science, Bradford University, UK in 1973. aban Eren also received
MS degree in business administration at Hacettepe University, Turkey in 1977. He got his PhD degree in
computer engineering at Ege University, Turkey in 1979.
He is the Dean of Faculty of Science and Letters at Yaar University, Turkey. His current research areas
include IT security, statistics and software reliability. He has authored some national and international articles
and published several books in different areas such as statistics, linear algebra, differential equations, general
mathematics, office applications and computer programming.
Assistant Prof. Dr. Erdem Uar Erdem Uar graduated from department of physics, Trakya University,
Turkey in 1990. Dr. Erdem Uar received his M.Sc. degree in 1994 and his PhD degree in 1996 at the
department of computer engineering, Trakya University.
He currently works as Assistant Professor in the department of computer engineering, Trakya University. He
is also the manager of Trakya University computer center. His main research areas are machine learning,
information security management, medical informatics and mobile applications and he has authored some
national and international publications in these areas.
333
Abstract The basic goal of software projects is completely to meet the requirements of the software that will be developed
and the biggest difficulty while achieving this goal is the errors which occur during and after the development of software.
One of the most important reasons for the errors is that the analysis and the design processes are not performed in the way
that is anticipated. One other problem is that software products are not tested enough. Especially in small and medium-sized
software vendors projects, software testing is ignored. For these reasons, software delivered to customers causes financial
losses and the waste of resources. In this study a hybrid test tool was developed to reduce errors in software projects, and the
application of this hybrid test tool to a sample project developed in accordance with Extreme Programming methodology is
explained.
Keywords: Software testing, agile methods, extreme programming
1 Introduction
Software products have become fundamental and permanent components of many systems with the spread of
technological improvements. This situation has brought plenty of good results as well as some problems. The
reasons for these problems are the developed software products that dont meet the customers requirements,
generating errors or exceeding the budget.
The most important of these reasons are errors occurring in software systems. 59.5 billion dollars were wasted
in the USA because software test quality is low according to a study conducted between 2000 and 2005. The
thought-provoking issue is that 22.5 billion USD out of the 59.5 billion USD could be saved by using
fundamental software test techniques [1]. Software testing is the process of assessment and examination of the
features related to the software project, and this process should be carried out diligently. Test activities should be
determined and a test plan should be formed for the test process.
In the second part of the study the concepts of agile software development and Extreme Programming, and in the
third part test based software development are explained. And finally the last part offers conclusion and
suggestions.
334
2 Agile Methods
Agile methods in software development have become more important in recent years. Customer requirements
can change continually during application development. As a result 40-90% of the project budget is spent on
software maintenance [2]. The changes which occur during the project development cause overruns in time and
budget. Unlike classical software development methods, agile methods ensure implementation of changes to the
project in a shorter time with the practices and life cycle they used. The fundamental principles, notions and
knowledge related to agile methods are published in the Agile Software Development Manifesto. In this
manifesto differences between the classical and the agile methods are emphasized.
Numerousagile methods have been developed since the 1990s. Extreme Programming and Scrum are the best
known and most commonly used agile methods [5]. Other methods includeCrystal Methods [6], Lean Software
Development [7], Feature Driven Development (FDD) [8], Adaptive Software Development [9] andthe Dynamic
System Development Method (DSDM) [10]. Important successes have been achieved in software projects
developed by universities, public offices, small and medium sized businesses using these methods [11, 12, 13,
14].
2.1 Extreme Programming
Extreme Programming (XP) is a software development methodology that meets the customer requirements
and aims to refine the quality of software products. XP was developed by Kent Beck in 1999 and uses some
practices and applications in order to increase software development quality and software production efficiency.
XP is based on 4 distinct values. These are communication, courage, simplicity and feedback. XP is an agile
software development methodology that gives importance to in-group communication [4, 15]. Project
development is done within the project lifecycle with XP. Project lifecycle is dependent on some procedures. To
provide improvement and flexibility in software, XP envisages 12 different procedures. These are: 40-hour work
per week, planning game, customer involvement, testing, simple design, pair programming, continuous
integration, small releases, refactoring, collective ownership, metaphor and coding standards.
In this section a project management tool developed using XP methodology in order to reduce errors in
software projects is explained.
Furthermore, the test-driven software development method and hybrid test software are also explained. And
this method and the corresponding test tool are applied to a sample project developed using XP rules. Errors
occurring in the course of the project and information about the development of the project are discussed.
3.1 Test-Driven Software Development
In this section, there will be a discussion of the ways in which software testing, which is generally overlooked by
small and medium-sized organizations, should be carried out at all stages of the software life cycle.Software
testing which is mostly postponed until the coding or after coding phase, can be applied to the analysis, design
and implementation phases in accordance with XP by using test-driven software development method.
The Phases of the test-driven software development method are:
Analysis Phase:Black-box unit testing is used in the analysis phase of projects. How the requirements
in the specification are met is determined and reduction of error costs is attempted by using this test
technique. The boundaries and constraints of the software that is going to be developed are determined,
and furthermore, how the technological infrastructure is going to be constructed is also determined.
Cooperation between the project team and the customer is provided by using XPs planning game
practice.The information obtained from customer cards is transferred to the hybrid, test based
application, and a calculation of person/hours is performed analysis of the information obtained.
June 1 -4, Kusadasi, Aydin, Turkey http://iscse2011.gediz.edu.tr
335
Design Phase: In this phase, information on the customer cards that are generated as a result of the
planning game is examined in detail. Database structure, main and sub processes are defined.
Information gathered from this definition is transferred to the technical cards section taking place in the
hybrid test based application, and the project deadline is revised in accordance with the information that
is entered into the application.
Coding Phase:The White-box unit test technique is used in this phase. Basic test code is written in the
first place and then functional code is written in accordance with XPs testing practice. When the test is
successful a new test is written and applied to the functional code, and the functional code is rearranged
to succeed the new test code. This iteration continues until all the tests determinedby the developer are
successful. These steps are repeated until the tests specified by the software programmer are successful.
The major advantage of writing test code is that a test application is prepared along with the
development of software project.
Test Phase:This phase contains the confirmation and the acceptance tests that should be performed in
the project.After the acceptance test is completed the maintenance phase begins.
Maintenance Phase:After this phase new requests which are submitted by the customer are considered
and tested for the analysis, design, coding and test phases again.
System Entrance Module: This module enables users to enter the system and to be directed to the
related window with their authorization.
Project Management Module:User profiles and parameters can be defined in this module. Furthermore,
basic definitions related to the application can be entered. This module can be accessed only by users
who have administrator privileges. The basic definitions related to the projects that are to be developed
are formed here. After the project selection related test windows can be listed.
Customer Cards Management Module: In the customer cards management module, which records the
information which comes from the customer, requirements and requests related to a project can be
monitored.
Technical Cards Management Module: This module allows the transferring of technical cards to the
program. Developers can monitor their workload using this module. Information about tests performed
also can be accessed by means of this module.
Reporting Module: This module represents the information related to the customer cards and the
technical cards.
Project Planning:After determining the boundaries of the software to be developed, the functional
requirements that are going to be in the system should be decided. In this phase the number of questions
and the weights associated with the multiple choice questions are determined, and a form design
appropriate for the optic reader was prepared. The connection ports, data types and database schema of
336
the optical reader were defined at the same time. A project boundary determination study was
performed according to the XPs planning game practice. Furthermore, an attempt was made to
determine requirements. These requirements were examined by the team leader and transferred to the
customer cards and the technical cards.
2.
Determining Main and Sub Processes:Eight main processes and 24 sub- processes related to the main
processes were defined for the optic reader evaluation software. The relations between the main and the
sub processes and their generation times were defined, and this information was transferred to the
hybrid test tool.
3.
Coding:In this process, the software development study was begun in compliance with the coding
standards of the university. The requirements of the technical cards that were ascertained by considering
the customer cards were converted to program codes by the developers. XPs simple design practice
was used in the software project. Thus, the goal was to understand, manage, and change the project in a
short time and develop simple software. Based on the customers changeable requirements importance
was given to simple design and feedback as much as possible in the team. Besides, code development
was done by using the test-driven software development method in all phases of the project. During the
coding process, system changes and new components were integrated into the system rapidly, so
detection of errors that could occur in early phases was ensured. New versions were released every
three weeks using continuous integration practice of XP.
4.
Testing:Testing of the software developed was handled separately in each phase beginning from the
analysis phase. Testing was performed in three phases: In the first phase of the project black-box unit
testing was used. In the second phase, the codes developed by the developers and the team leader in
accordance with XPs collective ownership practice were subjected to the white-box unit testing.
Besides, the customer tested the software according to XPs on-site customer practice. Errors were
fixed and new versions were prepared by the team according to the feedback from the customer. The
last phase of the test was the confirmation and acceptance tests. Acceptance tests were performed by the
customer by using real data. With the approval of the customer, the delivery process was completed.
The hybrid test software was used to follow up the operations performed in the test phase. The
Walkthrough method was used in addition to the test methods mentioned above. Thus, it was ensured
that the errors that were missed by the developer of the code were detected and fixed by other
developers.
4 Conclusions
In the last 20 years various models have been developed in order to evaluate software systems and processes,
provide quality certification in software, refine processes and ascertain the capabilities of software products. No
matter how many capabilities and features different models have acquired, software errors have remained a
significant problem.
The findings obtained as a result of this study are listed below:
Software projects can be managed efficiently by using the test-driven software development method
and the hybrid test tool.
The proposed test-driven software development method can be applied to different small and mediumsized software projects.
By means of the hybrid test tool developed, the levels and current error status of the projects can be
monitored.
337
References
1.
2.
3.
4.
5.
6.
7.
8.
9.
10.
11.
12.
13.
14.
15.
Everett, G. D. and McLeod, R. (2007). Software Testing: Testing Across the Entire Software
Development Life Cycle. John Wiley & Sons Inc.
Kajko-Mattsson, M., Westblom, U., Forssander, S., Andersson, G., Medin, M., Ebarasi, S., Fahlgren,
T., Johansson, S.-E., Trnquist, S. and Holmgren, M. (2001). Taxonomy of problem management
activities. In Proceedings of the Fifth European Conference on Software Maintenance and
Reengineering, Pages 1-10.
Beck, K. (1999). Extreme Programming Explained: Embrace Change. Addison-Wesley, 1st Edition.
Beck, K. and C., Andres (2004). Extreme Programming Explained: Embrace Change. Addison-Wesley,
2nd Edition.
Schwaber, K. and Beedle, M. (2001). Agile Software Development With Scrum. Prentice Hall.
Cockburn, A. (2004). Crystal Clear: A Human-Powered Methodology for Small Teams. AddisonWesley Professional.
Poppendieck, M. and Poppendieck, T. (2003). Lean Software Development: An Agile Toolkit for
Software Development Managers. Addison-Wesley.
Palmer, S. R. and Felsing, J. M. (2002). A Practical Guide to Feature Driven Development. Prentice
Hall.
Highsmith, J. (1997). Messy, exciting and anxiety-ridden: Adaptive Software Development. In
American Programmer, Volume10.
Stapleton, J. (1997). DSDM: A framework for business centered development. Addison-Wesley
Professional.
Bossi, P. (2003). Extreme Programming Applied: A case in the private banking domain. In Proceedings
of OOP. Munich.
Silva, A. F., Kon, F. and Torteli, C. (2005). XP south of the equator: An experience implementing XP in
Brazil. In Proceedings of the 6th International Conference on Extreme Programming and Agile
Processes in Software Engineering (XP'2005), pages 10-18.
Lui, K. M. and Chan, K. C. C. (2004). Test driven development and software process improvement in
China. In Proceedings of the 5th International Conference on Extreme Programming and Agile
Processes in Software Engineering (XP2004), volume 3092 of Lecture Notes on Computer Science,
pages 219-222.
Mann, C. and Maurer, F. (2005). A case study on the impact of scrum on overtime and customer
satisfaction. In Agile 2005 Conference, pages 70-79.
Manifesto for Agile Software Development (February, 2011). World Wide Web:
http://agilemanifesto.org/
Biographies
Emin Boranda He is a Lecturer in the Department of Informatics at Maltepe University, Istanbul, Turkey.
He received B.Sc. and M.Sc. degrees in Computer Engineering from Maltepe University in 2003 and 2006,
respectively. He is a PhD student in the Department of Computer Engineering at Trakya University since 2007.
His research interests are in the area of software engineering and software quality.
Fatih Yucalar He is a Lecturer in the Department of Computer Engineering at Maltepe University, Istanbul,
Turkey. He received B.Sc. and M.Sc. degrees in Computer Engineering from Maltepe University in 2002 and
2006, respectively. He is a PhD student in the Department of Computer Engineering at Trakya University since
2006. His research interests are in the area of software engineering and software quality.
aban Eren Professor aban Eren is the Dean of Faculty of Science and Letters of Yaar University, Izmir,
Turkey. He worked in the Computer Engineering Department of Ege University, Izmir, Turkey as a lecturer and
he also served as the chairman of the same department for many years. His current research interests include IT
security, statistics and software reliability. He has authored some national and international publications in the
field of programming, office automation and statistics.
338
Abstract. This paper presents a novel lightweight and flexible wireless protocol based on IEEE 802.11
standard above MAC layer for motion planning to operate a telerobotics system in four directions. The
protocol was tested on a robot car. A complete wireless robot car system by using the proposed custom
wireless protocol is illustrated, performance tests and communication delay measurements are done.
Advantages of the proposed custom wireless protocol are experimentally proven.
1 Introduction
In our modern world of fast communications, applications and devices operating on Wi-Fi technology play an
important role in our daily lives. Wi-Fi operates on unlicensed frequency range of 2.4 GHz allowing free access
to the wireless medium and making it a popular choice for public networking applications. With increasing
number of Wi-Fi users, wireless networks and the medium becomes increasingly crowded affecting the speed
and consistency of network traffic. It is therefore significantly important to develop protocols of data exchange
among devices that are lightweight and flexible to accomplish the necessary task get done. An embedded system
programmed with lightweight software enabling all common networking protocols i.e. UDP, TCP/IP, ARP,
ICMP, etc. operating on 802.11 Wi-Fi standard can constitute an attractive platform for development of different
practical applications.
Motion planning is one of the main areas of concern for developing robotics applications since all robotic
mechanisms need to move physically to accomplish its objectives [1]. This paper presents a novel lightweight
andflexible wireless protocol based on IEEE 802.11 standard above MAC layer for motion planning to operate a
telerobotic system in 4 directions. The protocol was tested on a robot car.
339
2 Related Work
Teleoperation and teleoperated system design received substantial research attention as it can be applied to
many areas [2 - 5]. Systems operated remotely depend mostly on communication protocols and their associated
communication delays because they are considered to be the key factors determining the performance of
teleoperated systems. Research focusing on teleoperation with both wired and wireless communication protocols
can be found in the literature.
In telerobotics, wireless communication gained significant popularity due to its convenience in real
applications. Several wireless communication standards can be implemented for both embedded systems and PC
based systems. In the literature, several researchers preferred to work with a single board computer (SBC)
running an operating system and equipped with an IEEE 802.11b/g wireless network interface card (NIC). In
terms of user friendliness and practical development environment this kind of setup can be the preferred choice
for many researchers. For example, in [6], implementing TCP/IP protocol for wireless communication, a Wi-Fi
based controlled robot arm is designed on PCM-3375 SBC, which runs on embedded Windows operating system
and a IEEE 802.11b NIC. Representing a similar set-up, another wireless setting based on a SBC and IEEE
802.11g with Real Time Streaming Protocol (RTSP) can be found in [7]. In [8], a mobile robot which is
controlled by a PC104 SBC with embedded Windows is presented. There are several other PC based teleorobotics systems with Windows or Linux, and using standard data communication protocols such as TCP/IP and
UDP [9-10].
In data communications, throughput plays the main role which is defined as the average data transfer rate.
However, in telerobotics the focus is on communication delays. In control systems, it is a well known fact that
sensor/measurement and actuator delays can cause the closed loop system to become unstable. In [11] and [12],
effects of communication delays on stability and performance of a teleoperated system is discussed in detail.
TCP/IP and UDP protocols are optimized for throughput and not designed for real-time telerobotics systems.
A teleoperation system requires extremely low communication delays and the data rates are much smaller.
Nevertheless, TCP/IP and UDP protocols are used in [6] and [10]. On the other hand, certain custom protocols
are also reported in the literature. Ulas at al evaluated alternatives of 802.11 based communications protocols for
embedded systems comparing their speed against PCs where the embedded systems do not operate on system
with overhead and PCs are working on an operating system like Microsoft Windows or Linux variants [13].
340
The flashcard placed on the edge of the board and connected to the PIC microcontroller serves as a NIC. The low
level communication with the 802.11b controller card is performed with the NIC drivers which form the first
part of the Airdrop-P system software. The second part of the system software is based on a minimal TCP/IP
stack. This part is mainly based on cooperative multitasking, and does not require a multitasking operating
system environment. It is the software designers responsibility to make sure that all tasks are performed and all
timing requirements are met. This makes software development more difficult, but because of the elimination of
the operating system overhead, system response times may be further optimized.
A similar approach was pursued in this study for development of custom delay sensitive wireless protocol on the
Airdrop-P system for an embedded telerobotics system. The protocol was lightweight, optimized for telerobotics
applications only, and implemented by using the cooperative multitasking approach.
3.2 Robot Car Equipped with an AirDrop-P Card as Slave Arm
The wireless protocol was tested on a robot car equipped with an AirDrop-P card as a slave arm as show in Fig
2 and a standard PC as a master arm. The commands were transmitted from the PC to the robot car for its
operation. Easy configuration options and 802.11b standard reaching up to 11Mbits per second create a flexible
system environment. The model robot car possesses two servo motors on the front and back.
Fig. 2. The wireless robot car equipped with an AirDrop-P Card Slave arm
The PC operating system directly interacts with the NIC sending signals and commands to the slave arm via
the wireless medium. On the slave-arm side the microcontroller receiving the signals by the NIC parses and
transmits them electronically to the mechanical parts of the automobile physically generating necessary motion
to move.
The software on the master arm side (PC) is using the C# programming language. It is an event driven
application. When a key is pressed down on the PC side, the event triggers and a related event matrix is build,
packed trough a socket application and transmitted to the operating system. As long as the key is kept pressed,
no other events trigger. Once the key is released, a key-up event steps in sending stop command to the slave arm.
A global flag is activated when a directional command is pressed. Since it is useless to transmit only directional
commands, when directional commands are transmitted based on the status of the flag.
341
To graphically illustrate the motion Y Axis can be assigned for motions forward and backward while X Axis
can be assigned for motions to the sides, i.e. Left and Right. This would lead us to the fact that at any given point
in time the car needs to be assigned 4 values represented on a 2x2 matrix as given in (1). The first column
represents related axis whereas the second column represents the specific direction on the axis. A typical real
world sample of the motion axis is shown in (2).
[
(1)
(2)
A more complicated model could include a third column for the impact of the specific command based either
on duration or length. Duration would reflect the time related aspect whereas length would reflect special. Such
a model would be assigned lets say a third variable stating arbitrarily 5 seconds. The real world application
would be keep moving forward for 5 seconds.
[
] (3)
Velocity is a variable to be included in the model by definition since every motion requires some degree of
velocity. Motion can be defined as a change of position of a physical matter in space. The velocity can be
defined as in (4):
D=Vxt
(4)
342
(5)
Each member of the matrix represents 8 bits of data. Therefore all members of the matrix are declared as a
character (char) in the programming software.For cases related only to velocity there is no need to transmit
information for axis resulting in the use of a 1x1 matrix as shown in Table 2.
Table 2. The first member (1x1) of the matrix can be assigned one of the 3 following values
Value
0
1
2
Option
Velocity to brake the car
Instructions related to Y axis
Instructions related to X axis
Table 2. The second member (1x2) of the matrix can be assigned one of the following values.
Y Axis (1x1 = 1)
Value
Option
f
Forward instruction on Y axis
r
Backward instruction Y axis
-
Value
l
r
0
X Axis (1x1 = 2)
Option
Left direction on X axis
Right direction on X axis
Stop motion on X axis
The computer keyboard is chosen to be the input device.The letter commands as shown in Table 3 are
assigned to 5 different keys on the keyboard.The slave arm receives the commands from the medium via the
NIC. The microcontroller then parses the signals obtained from the NIC and sendscommands to the responding
pins of the robot car for movement.
Table 3. Upon key press event the matrices to be build
Key
W
S
A
D
F
Option
Forward
Backward
Left
Right
Stop / Brake
Axis Info
Y axis
Y axis
X axis
X axis
Velocity info
Matrix
[1, f]
[1, r]
[2, l]
[2, r]
[0]
Table 4. The probabilities and states are as follows for key up events
Y Axis Keys
(if a f. or b. motion key is released)
X Axis Keys
(if a sideways directional key is released)
[0]
[2,0]
While the car is in a forward or backward directional motion and the corresponding key (W for forward, S for
backward) is released, a key up event is triggered. Such a state would require the transmission of only the stop
command which is represented by a 0 as shown in Table 4. Accordingly a 1 by 1 matrix, a command is created
and transmitted to the slave arm. If a sideways directional key is released (A for left, D for right) this would
require the transmission of more information, since the car would be travelling in either forward or backward
direction before the sideways directional key was pressed down. This time first column of the 2 by 1 matrix
would inform the slave arm that the corresponding piece of data is regarding a sideways motion. Then the second
column of the matrix will be a 0 again as it was the case for forward or backward motion to stop the slave arm
from cutting electrical connection from corresponding pins of the microcontroller.
343
Time (ms)
Ping (ms)
Min
8.435
8.345
Median
9.3065
Max
10.239
190.092
Average
9.26879
13.007
Statistics
6 Conclusion
In this paper we have presented a wireless protocol for telerobotics systems on top of 802.11 standard and
physically put it on test with a real world model robot car.A complete system by using the proposed custom
wireless protocol is illustrated; performance tests and communication delay measurements are done. Advantages
of the proposed custom wireless protocol are experimentally proven.
At this stage of our research the outcomes are positive and encouraging to develop the protocol further in more
complex ways. At further stages the protocol can be put in test in terms of delay sensitivity versus other media
and also Artificial Intelligence or Neural Network models can be worked around it to bring it to industrial level.
Acknowledgements
This work is supported by TUBITAK under the research grant EEEAG-108E091.
344
References
1.
2.
3.
4.
5.
6.
7.
8.
9.
10.
11.
12.
13.
14.
Biographies
Bora Tamer Ylmaz He received a B.B.A. degree in management from University of North Alabama,
Florence, AL, USA, in 2004. He is currently pursuing a M.S. degree in Management Information Systems at
Halic University, Istanbul, Turkey.
He has been working as an investment banker since 2007 professionally. As an M&A Analyst with
UniCredit, Tamer has actively participated in international project groups advising large scale multinational
corporations on privatizations and cross-border acquisitions. His specific sectors were Energy, TMT
(Technology, Media, & Telecoms), and Financial Institutions. Currently he is working as a Research Analyst
with Halk Yatirim tracking financial markets with a specific focus on real estate and IT sectors.
Haluk Gmkaya He received a B.S. degree in electrical and electronics engineering from Middle East
Technical University, Ankara, Turkey, in 1986, a M.S. degree in electrical and computer engineering from
University of Wisconsin-Madison, USA, in 1989, and a Ph. D. degree in computer engineering from stanbul
Technical University, stanbul, Turkey in 1995.
He worked as an Instructor in the Department of Electronics Engineering at Uluda University, Bursa,
between 1990 and 1997. He worked 5 years as a Chief Researcher at the Scientific and Technical Research
345
346
Abstract.The main objectives of a successful software project are developing it to meet the expectations of
the customer's needs, completing it within a planned time and within the planned budget. Unfortunately, it is
impossible to fulfill all of these three objectives in most cases. To ensure the success criteria, software
estimation and measurement methodologies have been incorporated into software projects, making software
size and effort estimation, key aspects of successful project management. There are plenty of size and effort
estimation methods existing in the software industry. Use-Case Points (UCP) method is one of them, which is
based on use-case diagrams. It was developed by Gustav Karner in 1993. In this paper, first the UCP method
is described, and then its application to two software development projects is presented. The results are
analyzed and discussed. Based on the analysis of the straight forward UCP method, modifications are
proposed to improve the method. When the effort is estimated again by using our improved method, almost
identical values are obtained as compared to the actual effort.
Keywords: software effort estimation, size estimation, use-case points.
1 Introduction
It is necessary to estimate the effort needed to determine the cost and the time of a software project before it
is completed. Software effort estimation is one of key elements of project management. Project size should be
estimated to determine the effort. Unfortunately, project managers are not very successful at predicting the
related effort. It is hard to estimate the effort in the beginning of the development phase since detailed
information about the system is not available yet.
There are many of methods to estimate project size and effort. The most common and known ones are
Function Point Analysis [1], MKII Function Point Analysis[2], and COSMIC Full Function Point [3]. But these
approaches have some serious limitations. Counting function points requires experts. Besides these methods
there is Use-Case Points [4] method.
Use-Case Points method is an extension of Function Point Analysis and MKII Function Point Analysis
methods and is based on the same philosophy as these methods. The Use-Case Points (UCP) method can be
347
used to estimate software development effort based on a use-case model and two sets of adjustment factors
related to the environmental and technical complexity of a project [5].
The most widely used concept for modeling and designing of software products is UML (Unified Modeling
Language). UML helps to present different aspects of system through various diagrams [6]. Thus, it is aimed at
imposing a well-constructed architecture on the system. Size and complexity of software systems can be
measured by the use of UML diagrams [7]. Organizations which develop object-oriented software can apply
use-case approach to determine the size of software. A use-case model is an interaction between the software
product itself and the users of that software product [6]. In object-oriented software production, use-cases
describe functional requirements. Use-cases are often used as input for estimating software development effort
[8]. The use-case model may therefore be used to predict the size of the future software system at an early
development stage.
UCP method is explained in some detail in the second part of the paper. In the third part, a case study is
presented wherea UCPmethod is applied to two software project developed.The forth part contains enhancement
suggestions to the UCP method and applications of those suggestions to the same projects. In the last part, the
results obtained and conclusions are given.
Actor Type
Weighting Factor
Simple
Average
Complex
The total Unadjusted Actor Weights (UAW) is calculated by counting how many actors there are of each
category, multiplying each total by its weighting factor (WF), and adding up the products. UAW is determined
with the following equation:
348
(1)
Weighting Factor
Simple
x3
Average
4x<7
10
Complex
x7
15
The Unadjusted Use-Case Weights (UUCW) is calculated counting the number of use-cases in each category,
multiplying each category of use-case with its weighting factor (WF) and adding the products. UUCW is
determined with the following equation:
(2)
(3)
Step 4: Calculation of the Technical and Environmental Factors
The UCP method includes 21 adjustment factors, which concern the technical complexity of the developed
system (13 technical complexity factor), and the environment in which it is developed (8 environmental factors).
All factors and weightings are presented in Table 3 and Table 4.
349
Description
Weight
T1
Distributed system
T2
Performance objectives
T3
End-user efficiency
T4
Complex processing
T5
Reusable code
T6
Easy to install
0.5
T7
Easy to use
0.5
T8
Portable
T9
Easy to change
T10
Concurrent
T11
Security features
T12
T13
Description
Weight
E1
1.5
E2
Application experience
0.5
E3
Object-oriented experience
E4
E5
Motivation
E6
Stable requirements
E7
Part-time workers
-1
E8
-1
1
0.5
The influence of Technical Complexity Factors (TCF) is assessed by assigning a value from 0 to 5 to each of
them. This value is multiplied by a weight of a factor and totaled. TCF is calculated from the following formula:
350
(4)
The level of environmental complexity is directly related to developers experience involved in the software
project. The influence of Environmental Factors (EF) is assessed in a similar way as in the case of TCF. EF is
calculated by the following formula:
(5)
(6)
In order to obtain effort estimation in person-hours one has to multiply UCP by the Productivity Factor (PF). The
following formula is applied:
(7)
The PF is a ratio of the number of person-hours per use-case point based on past projects. Karner originally
proposed a ratio of 20 hours per use-case point [4].If no historical data has been collected, a figure between 15
and 30 is suggested by industry experts. In fact, the most accurate Productivity Factor can be found by recording
the effort spent for designing and implementing the use-cases in time.
3 Case Study
In this study, two software projects developed at our university are addressed as two case studies. Projects are
named as A and B. Project A corresponds to the Student Information System developed for Student Affairs
Department and project B corresponds to the Accountancy Automation System developed for Administrative
and Financial Affairs Department.
351
Table 5 shows some characteristics of the two software development projects used in ourcase studies.
Table 7. Characteristics of two software development projects
Characteristic
Project A
Project B
~ 9600 person-hours
~ 5700 person-hours
Five-layer structure,
5 developers with 2 to 18
years experience, 1
consult
4 developers with 0 to
Education
Finance
Size
Software Architecture
Programming Environment
Project Members
Application Domain
13 years experience
At first, data related to the projects A and B were collected from the project documents, the use-case model
including a requirements specification with brief descriptions of each use-case, iteration plan and several
interviews with project members.
Afterwards, UCP method is applied to the projects A and B. Then enhancements are made to the method
according to the results taken. And lastly, the enhanced method is applied to the same projects. In the case study
productivity factor was assumed as 20 person-hours for each project.
3.1 Application of UCP Method to the Projects
UCP method is applied to the projects A and B. During the application of the method, included and extending
use-cases are considered as Karner suggested while use-cases are classified. Results are listed in Table 6.
Table 8. Implementation of the UCP method to projects
Project
Project Details
A
Simple
Average
Complex
Number of Actors
352
23
Simple
Average
21
Complex
13
415
200
T1
T2
T3
T4
T5
T6
T7
T8
T9
T10
T11
T12
T13
1,005
1,04
E1
E2
E3
E4
E5
E6
E7
E8
UAW:
Number of Use-Cases
UUCW:
Technical Factors
TCF:
Environmental Factors
353
0,875
Use-Case Points
321
203
Productivity Factor
20
20
6440
4040
Actual Effort
9470
5680
0,32
0,29
EF:
Simple
Average
Complex
32
30
Simple
Average
27
11
Complex
17
535
255
T1
T2
T3
T4
T5
T6
T7
T8
T9
T10
Number of Actors
UAW:
Number of Use-Cases
UUCW:
Technical Factors
354
T12
T13
1,005
1,04
E1
E2
E3
E4
E5
E6
E7
E8
0,725
0,875
Use-Case Points
413
259
Productivity Factor
20
20
8260
5180
Actual Effort
9470
5680
0,13
0,09
TCF:
Environmental Factors
EF:
When the results given in Table 6 are examined, it can be seen that the effort estimated by using UCP method
is lower than the real effort. The efforts estimated by using UCP method for Project A and B are 32% and 29%
lower than the real efforts respectively. It is realized that the effort estimation errors originate from ignoring the
details related to the actors and use-cases take place in project.
Some enhancements are made on the method in this respect. Initially ignored included and extending use-cases
are handled as separate use-cases since they include functionality related to the project and require additional
effort. Furthermore, it is realized that alternate scenarios related to the use-cases should be considered. These
scenarios require effort as base scenario. Therefore, alternate scenarios are handled as separate use-cases and
included in the classification.
3.2 Application of the Improved Method
Results in Table 7 are obtained when the enhancements made to the UCP method are applied to projects A and
B. It can be seen that the efforts estimated by using UCP method for Project A and B are 13% and 9% lower than
the real efforts respectively. This approximation of the estimated efforts for projects A and B to the real efforts
are achieved by the increase of the actors and the use-cases. Two and three actors are added for the projects A
and B respectively. Additionally, totally 10 use-cases including 6 average level and 4 complex level use-cases
355
for project A and totally 5 use-cases including 1 simple level, 2 average level and 2 complex level use-cases for
project B are added.
4 Conclusions
In this study first the UCP method is explained in some detail. The method is applied to two software projects
developed at the University and results are analyzed. When we apply this method as Karner suggested, a very
low estimated effort with respect to the real effort is obtained. By means of the enhancements made to the UCP
method, better estimates are obtained. While making improvements to the method, each viewpoint of the usecase model such as relations between use-cases and actors, relations between use-cases and detailed scenarios of
each use-case are handled. It is concluded that it is important to handle each use-case and detailed scenarios of
that use-case in these viewpoints. Technical and environmental factors need to be updated as the number of
developed project increases. Thus, more accurate efforts estimates can be made.
References
1.
2.
3.
4.
5.
6.
7.
8.
9.
Fetcke, T., Abran, A., Dumke, R. (2001). A Generalized Representation for Selected Functional Size
Measurement Methods, 11th International Workshop on Software Measurement, Montreal, Canada.
Symons, C. R., (1991). Software Sizing and Estimating: MK II FPA, John Wiley & Sons.
Abran, A., Fagg, P., Meli, R., Symons, C. (2002). ISO Transposition and Clarification of the COSMIC
FFP Method of Functional Sizing, In Proceedings of the 12th International Workshop on Software
Measurement (IWSM), Magdeburg, Shaker Publ., Aachen, pp. 33-42, October 7-9.
Karner, G. (1993). Metrics for Objectory. Diploma thesis, University of Linkping, Sweden. No.
LiTHIDA-Ex-9344:21.
Ochodek, M. and Nawrocki, J. and Kwarciak, K. (2011). Simplifying effort estimation based on UseCase Points, Journal Information and Software Technology, Volume 53, Issue 3.
Schach, S., R. (2007). Object-Oriented & Classical Software Engineering, 7th Ed., McGraw Hill.
Mohagheghi, P., Anda, B., Conradi, R. (2005). Effort Estimation of Use Cases for Incremental LargeScale Software Development, ICSE '05: Proceedings of the 27th International Conference on Software
Engineering.
Anda, B., Benestad, H. C., Hove, E. (2005). A Multiple-Case Study of Effort Estimation based on Use
Case Points, International Symposium on Empirical Software Engineering.
Braz, M.R., Vergilio, S.R. (2006). Software Effort Estimation Based on Use Cases, Proceedings of the
30th Annual International Computer Software and Applications Conference (COMPSAC'06).
Biographies
Fatih Yucalar He is a Lecturer in the Department of Computer Engineering at Maltepe University,
Istanbul, Turkey. He received B.Sc. and M.Sc. degrees in Computer Engineering from Maltepe University in
2002 and 2006, respectively. He is a PhD student in the Department of Computer Engineering at Trakya
University since 2006. His research interests are in the area of software engineering and software quality.
Emin Boranda He is a Lecturer in the Department of Informatics at Maltepe University, Istanbul, Turkey.
He received B.Sc. and M.Sc. degrees in Computer Engineering from Maltepe University in 2003 and 2006,
respectively. He is a PhD student in the Department of Computer Engineering at Trakya University since 2007.
His research interests are in the area of software engineering and software quality.
Fuat Ince Professor Fuat Ince teaches in Turkish Air Force Academy and he also works as a consultant in
Luna Technology. He received his B.Sc. degree from Bogazici University, Istanbul, Turkey, his M.S. degree and
his PhD. from University of Illinois at Urbana-Champaign. He has teaching experience in foreign universities
and worked in private sector. He was the founder director of TUBITAK Information Technologies Institute.
356
Chapter 2
ELECTRIC
AND
ELECTRONICS
ENGINEERING
357
msengul@khas.edu.tr, 2aozmen@khas.edu.tr
Abstract.In this paper, a genetic algorithm based real frequency technique has been proposed to solve
broadband impedance matching problems. In the techniques present in the literature (for instance simplified
real frequency technique, SRFT), it is necessary for the designer to supply the initials of the free optimization
parameters to the algorithm. But it is not necessary for the proposed technique to start with user defined initial
parameters. An example is given to illustrate the utilization of the method, and the result has been compared
with the result obtained via simplified real frequency technique.
1 Introduction
The matching problem involves the design of a lossless two-port network between a generator and a complex
load impedance, such that the transfer of power from source to load is maximized over a prescribed frequency
band. The power transfer capability of the lossless equalizer is best measured with the transducer power gain
which is defined as the ratio of power delivered to the load to the available power from the generator.
The analytic matching theory requires an explicit expression or, equivalently, a circuit realization for the load
and generator impedances. In practice however, one is usually encountered with experimental real frequency
data for the terminating impedances to be matched. In such cases, in order to be able to use analytic theory, these
data should be approximated with a proper equivalent circuit or an analytic realizable function, which is by no
means easy and satisfactory.
On the other side, matching networks can be attempted to be designed by employing commercially available
computer aided microwave design packages [1-3]. These programs utilize purely numerical methods, and they
are actually devised for the analysis and optimization of given networks based on the circuit elements. That is,
the topology of the network and a good estimate of the element values should be supplied to the programs. Here,
the major difficulty concerning the matching network design is the determination of optimum topology which is
usually unclear. Moreover, the performance function is in general highly nonlinear in terms of the unknown
element values to be optimized. Therefore, it is essential to start with element values which are close enough to
the final solution for ensuring the convergence to a global optimum.
In 1977, Carlin has proposed a new method called real frequency technique which removes the major defects
from which the analytic theory suffers [4]. The method utilizes the experimental real frequency data for the
generator and the load directly, and it is not necessary to assume neither the analytic form of the transfer function
nor the equalizer topology. It has been also shown in various matching problems that in compared to a design
obtained with analytic theory, the real frequency technique results in better performance with simpler structures
358
[4-11]. Because of these advantages, the real frequency technique has turn out to be the most feasible approach
to solve broadband matching problems encountered in practical applications.
In all real frequency techniques developed in literature, the designer must supply the initials of some parameters
which are going to be optimized. But in this work, there is no need to supply a set of user dependent initial
parameters. Broadband matching problem has been solved by using genetic algorithm.
RG
E
ZL
Z2
S
(
p
)
S
(
p
)
g
(
p ) f ( p ) h( p )
22
21
(1)
where g ( p) is a strictly Hurwitz polynomial, f ( p) is a real monic polynomial and is a unimodular constant
( 1 ). If the two-port N is reciprocal as well, then f ( p) is either even or odd and f ( p) / f ( p) .
The polynomials { f ( p), g ( p), h( p) } are related by the losslessness equation
g ( p) g ( p) h( p)h( p) f ( p) f ( p) .
(2)
In almost all applications, one has an idea about the location of transmission zeros of the equalizer network.
Hence the polynomial f ( p) which is constructed on the zeros of transmission of the matching network two-port
is usually defined by the user.
The transducer power gain for the network seen in Fig. 1 can be defined as
TPG
where Z 2 R2 jX 2
4 R2 RL
( R2 RL ) 2 ( X 2 X L ) 2
(3)
1 S11 g h
and Z L RL jX L .
1 S11 g h
In our algorithm, the coefficients of the polynomial h( p) are selected as the free optimization parameters. The
polynomial f ( p) is formed by the user, then from (2) the polynomial g ( p) is obtained. These polynomials are
used to calculate the transducer power gain of the network via (3). Free parameters are optimized via genetic
algorithm to be able to obtain TPG 1 in the frequency band of operation and TPG 0 for other frequencies.
359
In the example, the load impedance ( Z L ) is formed by parallel connection of a capacitor and a resistor. The
normalized values for capacitor and resistor are c 4 and r 1 , respectively. Theoretically maximum TPG is
0.7921 for the selected load. Number of elements used in the matching network is selected as five.
For genetic based approach, Matlab Genetic Toolbox is used. During the solution population size and iteration
number are selected as 50 and 250, respectively. Also there is only one restriction: In the passband ( 0 1 ),
TPG is maximum ( TPG 1 ), and in the stopband ( 1 ), TPG is zero. After applying the proposed genetic
algorithm based approach the following polynomials are obtained;
RG
L1
L2
L3
ZL
C2
C1
C2 4.3593, RG 0.3556 ).
Transducer power gain curves for both methods are given in Fig. 3 for comparison. As can be seen from Fig.
3, a TPG curve which is nearly the same as the TPG curve via SRFT is obtained without supplying any initials.
It is also noted that both curves fluctuate around the maximum available TPG level (0.7921). Namely both
methods converge to the maximum TPG level without any restriction.
360
used as a front-end for the commercially available CAD packages to design practical matching networks for
wireless or in general microwave communication systems.
References
1.
2.
3.
4.
5.
6.
7.
8.
9.
10.
11.
12.
13.
AWR: Microwave Office of Applied Wave Research Inc., AWR, ElSegundo, CA, 2007, from the World Wide
Web: www.appwave.com.
EDL/Ansoft Designer of Ansoft Corp, Ansoft, Pittsburgh, PA, 2007, from the World Wide Web:
http://www.ansoft.com/contact.cfm.
ADS of Agilent Techologies, Agilent, Santa Clara, CA, 2007, from the World Wide Web:
www.home.agilent.com.
Carlin, H.J. and Amstutz, P., (1981). On optimum broadband matching. IEEE Trans. Circuits and Systems 28: 401405.
Yarman, B.S., (1982). Real frequency broadband matching using linear programming. RCA Review 43: 626-654.
Yarman, B.S., (1982). Broadband matching a complex generator to a complex load. PhD Thesis, Cornell
University.
Yarman, B.S., engl, M., and Kln, A., (2007). Design of Practical Matching Networks with Lumped-Elements
via Modeling. IEEE Trans. on Circuits and Systems I: Regular Papers 54(8): 1829-1837.
engl, M. (2008). Modeling Based Real Frequency Technique. AE International Journal of Electronics and
Communications 62(2): 77-80.
engl, M., (2009). Design of Broadband Single Matching Networks. AE International Journal of Electronics
and Communications 63(3): 153-157.
engl, M. and Yarman, B.S., (2008). Broadband Equalizer Design with Commensurate Transmission Lines via
Reflectance Modeling. IEICE The Institute of Electronics, Information and Communication Engineers E91-A(12):
3763-3771.
Volmer, C., engl, M., Weber, J., Stephan, R. and Hein, M.A., (2008). Broadband Matching of a Superdirective
Two-Port Antenna Array. IEEE Antennas and Wireless Propagation Letters 7: 613-616.
Yarman, B.S., (1982). A simplified real frequency technique for broadband matching complex generator to
complex loads. RCA Review 43: 529-541.
Yarman, B.S. and Carlin, H.J., (1982). A simplified real frequency technique applied to broadband multi-stage
microwave amplifiers. IEEE Trans. Microwave Theory and Tech. 30: 2216-2222.
Biographies
Metin engl was born in Bartn, Turkey in 1971. He received B.Sc. and M.Sc. degrees in Electronics Engineering from
stanbul University, Turkey in 1996 and 1999, respectively. He completed his Ph.D. in 2006 at Ik University, stanbul,
Turkey.
He worked as a technician at stanbul University from 1990 to 1997. He was a circuit design engineer at R&D Labs of the
Prime Ministry Office of Turkey between 1997 and 2000. He worked as a lecturer and assistant professor at Kadir Has
University, stanbul, Turkey between 2000 and 2010. Dr. engl was a visiting researcher at Institute for Information
Technology, Technische Universitt Ilmenau, Ilmenau, Germany in 2006 for six months. Currently, he is an associate
professor at Kadir Has University, stanbul, Turkey and working on microwave matching networks/amplifiers, device
modeling, circuit design via modeling and network synthesis.
Atilla zmen was born in Batman, Turkey in 1971. He received the B.Sc., M.Sc. and Ph.D. degrees, all in electrical
engineering, from Istanbul University, Istanbul, Turkey in 1993, 1996 and 2001 respectively.
He served as a research assistant at the Department of Electrical and Electronics Engineering, Istanbul University from
1994 to 2001. Currently he is an assistant professor at the Electronics Engineering Department, Kadir Has University. His
research interests are neural networks, image processing, signal processing and genetic algorithms.
361
Electrical Engineering Department, Iran University of Science and Technology, Tehran, Iran.
2
Abstract. In this paper, we consider discrete-time piecewise affine (PWA) model of chaotic systems. PWA model, as one of
the main classes of hybrid systems, represents a powerful modeling tool for many nonlinear systems including chaotic
systems with absolute-value nonlinearity such as, e.g. the original Chuas circuit and the modified L system. Several
promising methods have recently developed to analyze piecewise linear and affine systems, and the control problem for such
systems have attracted much attention. One of the main approaches control in hybrid systems is optimal control problem.
Moreover, for hybrid systems modeled in PWA, there is a chance of systematically solving model predictive control (MPC).
In order to control the chaotic system, an explicit model predictive control (EMPC) has been used which calculates the
control law as an affine function of system states. In this method, the computation of MPC is moved off-line by means of
MPT toolbox[1]. The off-line control law is easier to implement reducing to a look-up table in comparison with the on-line
approach. Numerical simulations on the chaotic Chua circuit will be presented to show their agreement with the results.
1 Introduction
Recently, hybrid systems have been largely noticed in different communities such as control and computer
science societies [2]. These systems refer to those that involve both continuous and discrete dynamics. In many
physical and industrial systems, a large number of interactions are available between continuous and discrete
parts. Thus the hybrid description and analysis of such systems help us understand these systems behavior in a
better way[3]. Several modeling frameworks have been proposed for hybrid systems among which the two
classes mixed logical dynamical models (MLD) and piecewise affine models (PWA) have been largely used in
control and identification theories respectively. The reason is that they are capable of describing different types
of systems [3],[4]. As stated in [5], these two groups of models are equivalent. Another reason is that systematic
approaches such as model predictive control can be applied to control these systems [6]. From a practical point
of view, the main advantage of MPC is that it is able to handle multivariable systems with constraints in a
systematic manner [7].
As a special case of hybrid systems, PWA systems have been a subject of research in the systems and control
community for some time due to their wide scopes of applications (see [8], [9] for more examples). In fact, PWA
systems are good models equivalent to many typical classes of systems, such as the mixed logic dynamical
systems and the extended linear complementary systems. Many chaotic systems, including Chuas circuit,
cellular neural networks, and the systems with n-scroll attractors or Lorenz-type attractors based on piecewise
362
affine nonlinearities, can be described by PWA models [10]. In particular, the most elementary piecewise affine
functional dependence, the modulus of the dependent variable, |x (t)|, seems to be a reasonable candidate for a
nonlinearity that might lead to a potentially chaotic jerky dynamics such as the so-called L system [11].
2 Problem formulation
Consider Chua circuit which is described by
x1 x1 x2 g1 ( x) u,
x2 x1 x2 x3 ,
(1)
x3 x2 .
where 0, 0, M 1 N 0 and g ( x1 ) Nx1 0.5(M N ) x1 1 x1 1 .
6
the autonomous chaotic behavior is shown in Fig.1 with the
7
8
6
4
2
0
-2
-4
-6
-8
-40
-30
-20
-10
0
x1
10
20
30
40
(2)
for y (k ) Si
The corresponding three regions are defined by S1 y(k ) 40 y(k ) 1, S2 y(k ) 1 y(k ) 1 , and
S3 y(k ) 1 y(k ) 40 , respectively. In the above system,
363
0
0.100
0
0.9857 0.100
1.200
B 0 , b1 b3 0
0
0
(3)
b2 0 0 0 , C 1 0 0.
T
3 Controller design
Model Predictive Control (MPC) has been traditionally and successfully employed in the process industry and
more recently also in a variety of industrial applications. The major advantage of MPC is its straight forward
design procedure. Given a discrete- time control model of the system, including constraints, one only needs to
set up an objective function that incorporates the control objectives. The control action at each time step is then
obtained by measuring the current state and minimizing the objective function over a finite prediction horizon
subject to the equations and constraints of the model. Depending on the model and on the length of the
prediction horizon used in the objective function, this minimization problem varies considerably in complexity.
The first value stemming from the predicted optimal sequence of control inputs is then applied to the system and
the procedure repeated at the successive sampling instant. Further details about MPC can be found in [13].
The other advantage of MPC lies mainly in its handling of constraints and its relative simplicity.
Commonly, a quadratic performance measure over a finite prediction horizon is employed. Problems of this type
are easily translated into a quadratic program (QP). Many optimization software packages include a QP solver
which is efficient enough to perform these optimizations on-line.
In contrast, one of the disadvantages of MPC is the computation time. It turns out explicit MPC solution is a
relatively easy to implement piecewise affine controller. In the next section, the explicit model predictive
controller is designed to control of chaos in discrete Chua circuit.
3 Simulation result
Using MPT toolbox, which is developed for designing MPC control, we have designed the explicit form of
predictive controller. The simulation run-time greatly depends on choosing sample time, prediction horizon,
inputs weights and number of PWA dynamics. The number of control regions increases exponentially with
prediction horizon. In Figures 1 and 2, regions of control law and closed-loop behavior are depicted respectively.
In Figure 2, the controller is activated at t=0.
364
-1
10
10
10
0.5
0
-0.5
-1
-1
4 Conclusions
In this paper, a hybrid PWA model of Chua circuit system was considered. Then based on such model, a model
predictive controller in explicit form was designed. It is shown that this type of co troll is easy to implement
reducing to a look-up table in closed-loop. The control law is an affine function of system states defined over
several regions.
References
[1] M. K. a. P. G. a. M. Baoti. (2004, Multi-Parametric Toolbox (MPT). Available: http://control.ee.ethz.ch/~mpt/
[2] M. S. Branicky, et al., "A unified framework for hybrid control: Model and optimal control theory," Automatic
Control on IEEE Transactions, vol. 43, pp. 31-45, 1998.
[3]L. T. M. Morari, "Hybrid systems," in Proceeding of 8th International Workshop, Lecture Notes in Computer
Science, Berlin, 2004.
[4]A. Bemporad and M. Morari, "Control of systems integrating logic, dynamics, and constraints," Automatica, vol.
35, pp. 407-427, 1999.
[5]W. P. M. H. Heemels, et al., "Equivalence of hybrid dynamical models," Automatica, vol. 37, pp. 1085-1091, 2001.
[6]M. Morari and M. Baric, "Recent developments in the control of constrained hybrid systems," Computers &
Chemical Engineering, vol. 30, pp. 1619-1631, 2006.
[7]C. A. B. Eduardo F. Camacho, Model Predictive Control in the Process Industry New York: Springer-Verlag, 1997.
[8]L. H. Rodrigues, J. P., "Automated control design for a piecewise-affine approximation of a class of nonlinear
systems," in Proceedings of the American Control Conference, 2001, pp. 3189-3194 vol.4.
[9]L. Rodrigues, "Dynamic output feedback controller synthesis for piecewise-affine systems," Ph.D., Stanford
University, 2002.
[10]A. L. Fradkov and R. J. Evans, "Control of chaos: Methods and applications in engineering," Annual Reviews in
Control, vol. 29, pp. 33-56, 2005.
[11]S. J. Linz and J. C. Sprott, "Elementary chaotic flow," Physics Letters A, vol. 259, pp. 240-245, 1999.
[12]Y.-H. Gao, et al., "Observer-based controller design of discrete-time piecewise affine systems," Asian Journal of
Control, vol. 12, pp. 558-567, 2010.
[13] J. Maciejowski, Predictive Control with Constraints: Prentice Hall, 2002.
365
marabi@elec.iust.ac.ir, 2jposhtan@iust.ac.ir
Abstract. In this paper, adaptive model predictive temperature control of an exothermic hybrid batch reactor with discrete
inputs using genetic algorithm minimization is studied. As this reactor contains binary valves as well as a continuous mixing
valve, it is considered as a hybrid reactor. An adaptive strategy is used to estimate the reactor model because the dynamics of
the reactor changes with time due to the chemical reaction performed inside the reactor core. To find the optimal control input
sequence based on a defined cost function, genetic algorithm minimization is used. Finally, reactor outputs are presented and
some modifications are introduced to achieve a desired control performance.
Keywords: Adaptive Control, Model Predictive Control, Genetic Algorithm
1 Introduction
Due to dynamic nature, ability to handle market changes and flexibility in production of wide range products,
batch processes are still preferred in production of low volume and high-valued products such as pharmaceutical,
biochemical and specialty materials [1]. Temperature control of batch reactors is a key problem in process
industries as in many cases, the temperature profile of the reactor should track a prescribed reference trajectory to
achieve the desired product specifications [2]. In batch processing, a sequence of one or more operations is
performed in a predefined order to obtain the desired quality and quantity of products [3]. Also these reactors
may contain on/off valves and motors and other discrete variables as well as continuous variables and dynamics.
Systems involving continuous and discrete states are called hybrid systems [4]. The great interest in hybrid
systems during the last decade is not only because it is an open field for research issues, but also of its increasing
importance in industrial applications [5].
One of the most developed methods in control of hybrid systems which has been the subject of many industrial
and academic research fields, is model predictive control. In [6] model predictive control methods used for
control of hybrid systems are studied. In [7], different classes of hybrid systems are studied. One of the most used
methods in modeling of hybrid systems is to describe them as mixed logical dynamical (MLD) systems [8]. A
switched system is a hybrid dynamical system consisting of continuous subsystems and a switching logic. The
basic problem of stability of switched systems is studied in [9]. There are many different methods used to model
and control hybrid processes including mixed discrete and continuous dynamics. In [10], an online control
strategy based on hybrid predictive control is used to control a mixed continuous batch process. In [11], hybrid
predictive control employing fuzzy models is studied. In [12], model predictive control methods based on
reachability analysis is employed to control a batch reactor with discrete inputs. In [13], model predictive control
of nonlinear hybrid systems with discrete inputs is studied using a hybrid fuzzy model. In [14,15], , hybrid fuzzy
366
model predictive control of temperature in a batch reactor is studied using branch and bound and genetic
algorithm minimization. A supervisory self tune predictive functional control algorithm is introduced in [16,17]
to control the temperature of a hybrid semi-batch reactor with time-varying dynamics. In [18], this self tune PFC
method is used to control the temperature inside of an exothermic batch reactor which its dynamics change due to
the exothermic reaction performing inside the reactor. In this paper, model predictive control method using
genetic algorithm minimization is employed to control the temperature of this exothermic batch process
employing an adaptive model which is obtained online using an RLS algorithm.
T jin ) is a combination of fresh input water temperature (T in ) and reflux water (with temperature T j ).
The temperature of the fresh input water depends on binary valves as follows:
o
if k C 0 and k H 1 then T in T H 150C
(1)
m jc j
dT j
dt
k M cT in (1 k M )cT j cT j S (T j T ) 0S 0 (T j T 0 )
dT
S (T j T ) Q r
dt
k M T in (1 k M )T j
mc
T jin
(2)
(3)
(4)
A B C , A C D
(5)
The number of moles of the components changes according to the following equations:
dM C
dM A
dM B
dM D
(6) The production rates
R1 R 2 ,
R1 ,
R1 R 2 ,
R2
dt
dt
dt
dt
R 1 and R 2 are changes with core temperature as the following equations:
367
m A M A B M B C M C D M D
c
(9)
c A M A c B M B cC M C c D M D
M A M B MC M D
(10)
Q r H 1R1 H 2 R 2
(11)
Parameters of the reactor and exothermic reaction and their values are represented in [18].
T j (k ) 11 (k )T j (k 1) 12 (k )T (k 1) 13 (k )T jin (k 1)
T (k ) 21 (k )T j (k 1) 22 (k )T (k 1)
(13)
Sampling time used for converting inputs and outputs to discrete signals is
1 (k ) R 31
and
2 (k ) R 21
(12)
1 (k ) R 31
T s 20s . Considering
and
2 (k ) R 21
as the
y 1 (k ) T j (k ) 1T (k )1 (k )
(14)
y 2 (k ) T (k ) T2 (k )2 (k )
(15)
which:
1 (k ) [T j (k 1) T (k 1) T jin (k 1)]T
2 (k ) [T j (k 1) T (k 1)]T
(16)
(17)
1 (k ) [11 (k 1) 12 (k 1) 13 (k 1)]T
(18)
2 (k ) [21 (k 1) 22 (k 1)]T
(19)
Pi (k )
G i (k )iT (k )Pi (k 1)
(21)
i (k ) i (k 1) G i (k ) y i (k ) iT (k )i (k 1)
which
(20)
i 1, 2
(22)
covariance matrices from bursting, covariance matrices and parameter vectors are updated only when the
following condition is satisfied:
if T (k )Pi (k 1)if (k ) k DZ (1 i ), i 1, 2
(23)
368
which
1 1 0 0 1 1
M u 1 1 1 1 1 1
1 1 1 1 1 1
(24)
The first row of this matrix denotes the position of the continuous mixing valve, which k M 1 means that the
valve opening should be decreased toward 0, k M 1 means that it should be increased toward 1 and k M 0
means that it should remains unchanged. The second and third rows show the binary valves positions which
possible individuals.
J (X kk h ,U kk H 1 , k , h ) J (X kk h 1 ,U kk H 2 , k , h 1) T (k h ) w (k h )
(25)
Prediction of the reactor core temperature is performed using the following equations:
T j (k h ) 11 (k h )T j (k h 1) 12 (k h )T (k h 1)
13 (k h )T jin (k h 1)
T (k h ) 21 (k h )T j (k h 1) 22 (k h )T (k h 1)
(26)
(27)
which is in the form of the parametric model used for parameter estimation but it uses regressor vectors
369
Mixing valve
Binary valves
Water
Error
switching (s,%)
switching (s,%)
consumption (kg)
5.28
9340,78%
5460,46%
10739
To achieve a better controller performance, the following modifications are used in the controller structure:
370
J (X kk h ,U kk H 1 , k , h ) J (X kk h 1,U kk H 2 , k , h 1)
T (k h ) w (k h )
J m . k M (k h ) k M (k h 1)
Jw . k M (k h ) J in . k H (k h ) k H (k h 1)
(28)
Integral
Mixing valve
Binary
Water
Absolute
switching (s,%)
valves
consumption
switching
(kg)
Error
Modification
(s,%)
Modification 1
5.23
3420,29%
1640,14%
10394
Modification 2
27.01
8960,75%
5940,50%
4620
Modification 3
16.77
600,5%
6640,55%
4300
6 Conclusions
In this paper, adaptive model predictive control of an exothermic hybrid batch reactor based on genetic algorithm
minimization is studied. An RLS algorithm was employed to estimate an adaptive model to predict the future
behavior of the system to be used by the predictive controller. A genetic algorithm has been used to minimizaed
he defined cost function based on reference trajectory tracking error. Simulation results showed that the reference
trajectory was well tracked by the reactor core temperature and some modifications were introduced to decrease
the control effort. Results showed that the control effort can be decreased to some extent by the cost of some
increase in tracking error.
References
[1] Ekaterini Korovessi, Andreas A. Linninger, Batch Processes, Taylor & Francis 2006
[2] Igor Skrjanc, Self-adaptive supervisory predictive functional control of a hybrid semibatch reactor with constraints,
Chemical Engineering Journal 136 (2008) 312319
[3] Dale E.Seborg, Thomas F.Edgar, Duncan A.Mellichamp, Process dynamics and control, second edition, Wiley
Publications (2004)
[4] Gorazd Karer, Gasper Music, Igor Skrjanc, Borut Zupancic, Hybrid fuzzy model-based predictive control of temperature
in a batch reactor, Computers and Chemical Engineering 31 (2007) 15521564
371
[5] Panagiotis D. Christofides, Nael H. El-Farra, Control of Nonlinear and Hybrid Process Systems: Designs for Uncertainty,
Constraints and Time-Delays, Lecture Notes in Control and Information Sciences 324, Springer
[6] E.F.Camacho, D.R. Ramirez, D. Limon, D. Munoz de la Pena, T. Alamo, Model predictive control techniques for hybrid
systems, Annual Reviews in Control 34 (2010) 2131
[7] A. Bemporad, W.P.M.H. Heemels, and B. De Schutter, On hybrid systems and closed-loop MPC systems, IEEE
Transactions on Automatic Control, vol. 47, no. 5, pp. 863869, May 2002.
[8] Alberto Bemporad, Manfred Morari, Control of systems integrating logic, dynamics, and constraints, Automatica 35
(1999) 407-427
[9] Daniel Liberzon ,A. Stephen Morse, Basic problems in stability and design of switched systems, Lecture notes, IEEE
control systems, October 1999
[10] C. de Prada, I. Grossmann, D. Sarabia, S. Cristea, A strategy for predictive control of a mixed continuous batch process,
Journal of Process Control 19 (2009) 123137
[11] Alfredo Nunez, Doris Saez, Simon Oblak, Igor skrjanc, Fuzzy-model-based hybrid predictive control, ISATransactions
48 (2009) 24_31
[12] Gorazd Karer, Gaper Muic, and Borut Zupancic, Predictive Control of Temperature in a Batch Reactor with Discrete
Inputs, Proceedings of the 13th Mediterranean Conference on Control and Automation, Limassol, Cyprus, June 27-29, 2005
[13] Gorazd Karer, Gasper Music, Igor Skrjanc, Borut Zupancic, Model predictive control of nonlinear hybrid systems with
discrete inputs employing a hybrid fuzzy model, Nonlinear Analysis: Hybrid Systems 2 (2008) 491509
[14] Gorazd Karer, Gasper Music, Igor Skrjanc, Borut Zupancic, Hybrid fuzzy model-based
predictive control of temperature in a batch reactor, Computers and Chemical Engineering 31(2007) 15521564
[15] Javier Causa, Gorazd Karer, Alfredo Nunez, Doris Saez, Igor Skrjanc, Borut Zupancic,Hybrid fuzzy predictive control
based on genetic algorithms for the temperature control of a
batch reactor, Computers and Chemical Engineering 32 (2008) 32543263
[16] Igor Skrjanc, Adaptive Supervisory Predictive Control of a Hybrid Fed-Batch Reactor with Slow Actuator, Ind. Eng.
Chem. Res. 2007, 46, 8050-8057
[17] Dejan Dovzan, Igor Skrjanc, Predictive functional control based on an adaptive fuzzy model of a hybrid semi-batch
reactor, Control Engineering Practice 18 (2010) 979989
[18] Gorazd Karer, Igor Skrjanc, Borut Zupancic, Self-adaptive predictive functional control of the temperature in an
exothermic batch reactor, Chemical Engineering and Processing 47(2008) 23792385
Biographies
Mohsen Arabi was born in Qom, Iran, in 1986.He received the B.Sc. degree in electrical engineering from Amir Kabir
university of technology, Tehran, Iran, in 2008. He is now a M.Sc. student in control engineering in Iran university of science
and technology, Tehran, Iran. His research interests include predictive control and batch process control.
Javad Poshtan was born in Tehran, Iran, in 1961. He received his B.Sc. in electrical engineering from Tehran university,
Tehran, Iran, in 1987, his M.Sc. in electrical engineering from Sharif university of technology, Tehran, Iran, in 1991 and his
Ph.D in electrical engineering department of New Brunswick, Canada, 1997.He works in Iran university of science and
technology as an associate professor. His research interests include system identification, fault detection and model predictive
control.
372
Abstract. This paper presents a new key management scheme in order to increase the security and
performance of the SCADA networks. In this new key management scheme, the master station is an initiator
in communication. The secret key is generated and encrypted by security devises on the master station side
instead of RTUs (Remote Terminal Units) or IEDs (Intelligence Electronic Devices) . So, the requaird
hardware and computational power of the RTUs or IEDs side are decreased. The results show that the traffic
network in the key sharing is reduced at about 30% and the speed of the key sharing is increased at about
30% in compare with Kang et al.s key management scheme. Therefore using this new key management
scheme not only the security but also the performance of the SCADA network increase considerably.
Keywords: security, SCADA networks, key managment, symmetric encryption, cryptography.
1 Introduction
A supervisory control and data acquisition (SCADA) networks play a crucial role in the computer-based units
of the industrial and utility infrastructure [1]. The
conventional SCADA networks initially designed to
maximize functionality, with little attention paid to the security of these networks. So the security of the
conventional SCADA networks is not robust [2]. In recent years, many SCADA networks are shifting from
conventional closed proprietary networks to the networks which are integrated with corporate networks [3-4].
This makes SCADA networks potentially vulnerable to disruption of services which could result in serious
disruption to the industrial and infrastructure [2]. So, in order to prevent the potential of high
consequence
damages, major research effort need to be devoted to secure the SCADA system [3]. The core technology used
for system security is cryptography [5-6]. Data encryption is the common cryptography approach to securing the
SCADA networks.
There are two major cryptography algorithms for encryption and decryption of information: symmetric and
asymmetric encryption algorithms [2]. However, some studies show that, the asymmetric encryption algorithms
such as elliptic curve cryptography (ECC) can be used for SCADA systems [7]. But due to the SCADA systems
deal with massive amount of data in a very short period, the symmetric
encryption algorithms are usually
applied to the SCADA networks. In the symmetric encryption, two parties involved in communication, should
share the same key which should be protected from the accesses by others [2]. There are several ways for this
key distribution, which can be categorized in to two groups: centralized key
distribution and decentralized
key distribution. The communication in SCADA networks is only made between the master station (SCADA)
and each slave station (Remote Terminal Unit (RTU) or Intelligence Electronic Device (IED)). So the
decentralized key distribution is suitable for SCADA networks. In dead, in these networks, the master station
takes the role of key distribution center (KDC) [2].
373
There are many key management schemes in order to increase the security of the SCADA networks. Robert et
al. [8] used a KDC for authentication and access control. Their system pre-distributes different symmetric keys
between each slave station and master station to facilitate secure communication with KDC. Pietre-cambacedes
and Sitbon [4] described some generic characteristics and constraints to choose or design appropriate key
management system (KMS) for SCADA. Johnson [1] described some of the weakness in SCADA based system
security and the effectiveness of current security techniques. Xiao et al. [3] proposed a new key management
scheme in order to increase the security, performance and scalability of SCADA systems. Kang et al. [2]
consider the security problems in the radial SCADA networks of electric power systems. They proposed a
symmetric encryption method. They used key management for encryption and provide a solution to the optimal
key distribution period as well. In addition, some distributed key management schemes which developed for
sensor network such as [9-11] may be used for SCADA system.
This paper presents a new key management scheme for SCADA networks. This new key management scheme
is a modification of the key management scheme in [2]. In the proposed key management scheme, the master
station is initiator of
communication and generator of the secret key. So the secret key generator
devices are removed from RTUs or IEDs side. In addition, the communication between master station and slave
station is reduced from 3 communications to 2 communications. Therefore, using this new key management
scheme, not only the network traffic and required hardware of RTUs or IEDs side are reduced but also the speed
and security of the SCADA network are increased considerably.
The rest of this paper is organized as follows: section 2 describes briefly SCADA network and installation of
security devices in SCADA network. The proposed key management scheme for SCADA network is presented
in section 3. Section 4
evaluates the proposed key management scheme. Finally, conclusion is given in
section 5.
2 Background
2.1 SCADA networks context
A SCADA network is a central master system which used for remote measurements and controls of the
infrastructures such as gas, oil and power generation system [2, 4]. Fig. 1 shows a simplified SCADA network .
As it is shown in Fig.1, the SCADA network contains three parts: master station, communication links and
slave stations. The master station contains HMI (Human Machine Interface), historian and applications. This
station provides information in order to supervisory decisions or control actuators. The slave stations contain
RTUs and IEDs, which usually control actuators and sensors [4]. The communication links are used in order to
exchange the measurement data and control signal between the master station and slave station.
374
As it is shown in Fig. 2, the security devices are installed on inlet or outlet of the master station and slave
station at two end ports of the communication links. Recall that the symmetric encryption algorithm, which is
shown in Fig. 3, is usually applied to the SCADA networks.
In the symmetric encryption algorithm, the decryption key is identical to the encryption key. The symmetric
encryption is also called single-key encryption. The key in this algorithm should be exchange in advance
between SCADA side and RTUs or IEDs side in a secure manner and kept secret [2].
Since the communication in SCADA networks is made only between the master station and each RTUs or
IEDs, the master station takes the role of KDC and just one communication counterpart of RTUs or IEDs
implying that the additional KDC is not required. One of the recent schemes of key management for data
encryption in SCADA networks is [2] which is shown in Fig. 4.
In this key management scheme, the master station is initiator in communication. Master station requests
RTUs or IEDs to send the session key by 1. The RTU or IED responds to master station with the key encrypted
with the master key already shared with master station by 2. Finally, master station confirms the key distribution
process by 3.
375
4. Evaluation
The key distribution period is affected by the security SCADA networks and the traffic load of the SCADA
networks. In dead, if the secret key is used over a large time span, it would increase the possibility of being
disclosed. Besides, it is not easy even to change the key too often in that the frequent key changing causes the
increase of the network traffic and communication failures. This issue is more significant in SCADA networks
due to its unique characteristics of data size and communication frequency. So, the performance and security of
the SCADA networks is affected by the key distribution scheme.
In the proposed key management scheme, the key generation is performed on the master station side instead
of RTUs or IEDs side. In addition, the initiator is the
master station. As a result, we used only two
communications between master station side and RTUs or IEDs side in order to share the secret key instead of
several
communications. The secret key generator in RTUs or IEDs side is not required. Table 1 summarize
the comparison between the proposed key management scheme and key management scheme in [2].
376
Table 1:Comparison the proposed key management scheme with key management scheme
in [2]
Parameter
Key management
scheme in [2]
Master station
Master station
Master station
Number of required
communication in key sharing
process
As it is shown in table 1 in the proposed key management scheme, the secret key generator devices are
removed from RTUs or IEDs side. So the required hardware and computation power of RTUs or IEDs side
reduced considerably. In addition, the required communication decreased from 3 to 2 communications. So the
traffic
network and speed of the key sharing are improved considerably. These improvements compute and
summarize in table 2.
Table 2:Improvements of the proposed key management scheme over key management scheme in [2]
Reference
[2]
30%
30%
As it is shown in table 2, in the key sharing of the proposed key management scheme the traffic network is
decreased at about 30% and the speed of the key sharing is increased at about 30% in compare with [2].
Therefore using the proposed key management scheme in SCADA networks not only the computation power,
required hardware in RTUs and IEDs side and traffic network are reduced but also the security of the SCADA
network is increased.
5. Conclusion
By allowing the data collection, data analysis and the control of equipment from remote locations, SCADA
networks provide great efficiency and are widely used. Due to the important role of the SCADA networks in the
infrastructure, the importance of the security and performance of the SCADA networks increased considerably.
This paper presents a new key management scheme in order to increase the security and
performance of the
SCADA networks. In this new key management scheme, the key management scheme in [2] is improved. The
communication links are decreased from 3 links to 2 links. The secret key generation devices are also exchanged
from RTUs or IEDs side to master station side. So the required hardware and computation power of the RTUs or
IEDs side are decreased considerably.
The results show that the traffic network in the key sharing is reduced at about 30% and the speed of the key
sharing increased at about 30% in compare with [2].
Therefore using this new key management scheme in
SCADA networks not only the security but also the performance of the SCADA network increase considerably.
377
References
1.
2.
3.
4.
Biographies
Abdalhossein Rezai was born in Dezful, Iran in 1977. He received the BS and MS degrees in electronics from Department
of Electrical and Computer Engineering, Isfahan University of Technology (IUT), Isfahan, Iran in 1999 and 2003
respectively. He is currently a Ph.D. student in Electrical and Computer Engineering Faculty, Semnan University, Semnan,
Iran.
He is authors and co-authors of several papers. His research interests include network
security, cryptography
algorithm and its application and neural network implementation in
nanoelectronics.
Parviz Keshavarzi was born in Eilam, Iran in 1959. He received the MS degree in electronics from Department of
Engineering, Tehran University, Tehran, Iran in 1988 and the Ph.D. in electronics from Department of Electrical
Engineering, Manchester University, UK in 1999.
He is authors and co-authors of many papers. He is an assistant professor in Electrical and Computer Engineering Faculty,
Semnan University, Semnan, Iran. His research interests
include network security, cryptography algorithm and its
application and nanoelectronics.
Zahra Moravej was born in Shiraz, Iran in 1960. She received the B.E and M.E degrees in
electrical engineering from
Bangalore University, India in 1985 and 1991, respectively, and Ph.D from Banaras Hindu University, India in 2001.
She is an assistant professor in Electrical and Computer Engineering Faculty, Semnan University, Semnan, Iran. She is
author and co-author of many papers. Her research interests include application of artificial intelligence in power system
protection, power quality
monitoring and Substation Automation System (SAS).
Dr. Moravej is IEEE senior member and Cigre member.
378
kemal.buyukkabasakal@ege.edu.tr
Abstract.In this paper, we present a prototype of flexible multi-node transceiver and monitoring system. The
prototype is designed for time-critical applications and can be also reconfigured for other applications like
event tracking. The processing power of FPGA is combined with a simple communication protocol. The most
important contribution of this work is using the virtual Pico Blaze soft core for controlling the data
transmission through the RF modules. The results for capacity usage are very promising as compared to other
similar research.
1 Introduction
RF communication programs for wireless sensor nodes are generally written on microcontroller units
(MCUs) for UART data communication, but the data processing capacity and the number of I/O ports of MCUs
are limited, so they are not convenient for simultaneous multiple parallel data processing that would be needed in
simultaneous multi-node wireless communication and monitoring.
Nowadays, FPGA based designs present powerful alternative to MCU designs due to their parallel data
processing capacities and software reconfigurable structures. Nowadays, as the single chip capacity of an FPGA
that accommodates desirable properties increases, its power consumption decreases. Because of these features,
high performance FPGA based systems has simple circuits, flexible designs and reconfigurable structures.
Besides, depending on the capacity of the I/O ports of an FPGA selected, the communication can be established
simultaneously with wireless multi-nodes which are connected to a single FPGA chip in parallel. This provides
more processing capacity with lower power consumption.
PicoBlaze is an efficient, compact, capable, and cost-effective fully embedded 8-bit RISC microcontroller
core which can be synthesized in Spartan 3 FPGAs (also in Virtex II and Virtex-4) and can be used for
processing data in the implementation of a flexible multi-node transceiver and monitoring system for wireless
sensor nodes. PicoBlaze is similar to many microcontroller architectures but it is specifically designed and
optimized for Xilinx FPGAs. PicoBlaze provides cost-effective microcontroller based control and simple data
processing and consumes considerably less resources than a comparable 8-bit microcontroller architectures. It
occupies just 96 FPGA slices and delivers 44 to 100 million instructions per second (MIPS) depending on the
target FPGA family and the speed grade which is many times faster than commercially available
microcontroller devices [10].
In this paper, we present a prototype of flexible multi-node transceiver and monitoring system. The prototype
is designed for time-critical applications and can be also reconfigured for other applications like event tracking.
379
From here on, the paper is organized as follows: in Section 2 an overview of the related work is presented;
Section 3 gives details on the architecture of the proposed system; in Section 4 experimental results are discussed
followed by conclusion and future work in Section 5.
2 Previous Work
In recent years the variety and complexity of Wireless Sensor Network (WSN) applications, the nodes and
the functions they are expected to perform have great increase. This poses the question of reducing the time from
initial design of WSN applications to their implementation as a major research topic. In this respect, the power of
FPGA as a system design has become even more important. So there are a number of studies in this direction,
which examine the different aspects of integrating FPGA into WSN applications.
In [3] the authors define an FPGA based system for wireless communication. The main objective is to
provide a detailed control of the nRF2401 RF chip using FPGA. As a result, the design of a low price, simple
and easy to control NIOS II system embedded in FPGA is proposed which can be adapted to monitoring WSN.
In [5] the authors focus on a strategy to design a hardware platform for enabling research in massive wireless
sensor networks for ambient systems. A prototype modular sensor node with 25mm form factor is described.
This miniature module comprises 4 different layers: FPGA layer, communication layer, interface and power
layers.
The work in [7] presents a wireless image processing sensor embedded in FPGA including a NIOS II
processor and a microcontroller. Using the capacity of FPGA for image processing, different experiments have
been carried out and the results are presented in the paper.
In most works so far, the reconfigurability advantage of an FPGA is generally not considered. The
conventional wireless multi-node monitoring platforms include a microcontroller and an FPGA to implement all
the processes [4, 5, 6, 7, 8]. In some studies, instead of using both a microcontroller and an FPGA, a single
FPGA chip has been used [1, 2, 3] and they stress on the benefits and advantages of the 32-bit NIOS II soft core
processor.
The aim of this project is to demonstrate the feasibility of using PicoBlaze soft core instead of NIOS II,
where PicoBlaze provides greater programming simplicity and easiness, and requires much less resources
because it is an 8-bit platform. Furthermore, a virtual PicoBlaze soft core processor is used inside the FPGA in
this project instead of using classical independent microcontrollers used in the works mentioned above.
380
Fig. 1. System Overview Block Diagram; System consists of many Pico Blaze cores that run simultaneously and each one is
connected to a transceiver module while one drives the display.
3.1 RF Modules
In this work 433.920 MHz UHF band, ARX 34-X and ATX 34-S RF (Fig. 2. (1, 3)) receiver and transmitter
chips have been used. They use ASK (Amplitude-Shift Keying) modulation technique in a low data rate for short
range ISM.
The transmission protocol is based on preamble and synchronous data transmission. The receiver adjusts the
receiving baud rate in the range from min. 300 to max. 2400 bps. The frame format is defined as:
preamble + sencron + data1+.....+dataX,
where the preamble is 5 bytes of 0xAA or 0x55. Synchronization is 5 bytes 0x00 or 5 bytes 0xFF. The algorithm
for start/stop has been embedded in software in the PicoBlaze. To start receiving data, an acknowledgement in
the form of 3 bytes of 0x3A is sent by the receiver side. Both the sender and the receiver side include a simple
voltage regulator and logic level max232. (Fig. 2. (2, 4))
Fig. 2. Prototype of Wireless Transceiver Node and Transmission Protocol; ATX 34-S transmitter chip (1) and
transmitter circuit (2), ARX 34-C receiver chip (3) and receiver circuit (4).
381
Fig. 3. Block diagram of one Pico Blaze Soft-Core that includes UART module and LCD interface.
4 Experimental Results
Wireless transceiver circuit has analog devices like voltage regulator and RF module and no microcontroller.
Serial data processing from wireless transceiver nodes is realized in PicoBlaze soft core units embedded in the
FPGA. UART communication standard is used for this application. The data is collected from the multiple
nodes in parallel into a single FPGA chip by means of many PicoBlaze cores embedded into the FPGA.
Multiple nodes can be monitored simultaneously, not serially. If desired, the data collected is processed at one
point with a high speed and less resources. The prototype system is designed to show the feasibility of the
system with an LCD and wireless data transceiver nodes. The proposed system has been evaluated based on
logic utilization.
In Table 1, the result for one RF transceiver modules from the FPGA Pico Blaze core source report are
presented. These include results for number of slice flip flops, number of 4 input LUTs (Look-Up Tables) and
number of bounded IOBs (Input Output Blocks). As it can be seen from these results a max of 62 RF
transceivers can be connected. The available pins are also sufficient. If a higher capacity FPGA chip is selected
an even higher number of RF transceivers can be simultaneously connected. A certain percent of the resources
382
has to be put aside for the DSP processing in the FPGA. That is why, for the XILINX XC3S1200E FPGA that is
used in our work, 50 Pico Blaze core control modules for the RF modules have been embedded in the FPGA thus
providing around 20% of the resources of the FPGA for DSP.
Table 1. Source Code Reports Results for 1 RF module
Logic Utilization
Used
Available
Utilization
171
17,344
1%
278
17,344
1%
250
1%
In a similar way, in Table 2 the results from the source report for 50 Pico Blaze core control modules for the RF
and one LCD screen have been presented.
Table 2. Source Code Reports for 50 Pico Blaze core control modules for RF and LCD Screen
Logic Utilization
Used
Available
Utilization
8559
13921
111
17,344
17,344
250
49%
80%
44%
5 Conclusions
In this paper we have presented the initial results from our work related to the use of powerful FPGA
platform and the Virtual PicoBlaze soft-core processor for wireless sensor nodes and applications. These
preliminary results are very optimistic and in the future we plan to test the software with a greater number of RF
modules and incorporate more complicated wireless communication protocols. It is also interesting to
experiment in real environments, for example in industrial settings for collecting information from a number of
nodes used for monitoring a specific production cycle. In such situations both the advantages of FPGA DSP
processing power and the adverse effects of the environment (number of machines in a closed area) will be better
estimated.
References
1.
2.
3.
4.
5.
6.
K. Arshak, E. Jafer and C. S. Ibala (2007). FPGA Based System design suitable for Wireless Health
Monitoring Employing Intelligent RF module. IEEE Sensors, Octaber 2007: 276279
G. Chalivendra, R. Srinivasan and N. S. Murthy. (2008). FPGA Based Re-Configurable Wireless
Sensor Network Protocol. International Conference on Electronic Design, 2008: 1
M. Zhang, H. Zhang. (2010). Design and Implementation of Wireless Transceiver System Based on
FPGA. International Conference on Innovative Computing and Communication and 2010 Asia-Pacific
Conference on Information Technology and Ocean Engineering: 355
B. O'Flynn, S. J. Bellis, K. Mahmood, M. Morris, G. Duffy, K. Delaney, C. O'Mathuna. (2005). A 3D
miniaturised programmable transceiver. Microelectronics International. 22(2): 812
S. J. Bellis, K. Delaney, B. OFlynn, J. Barton, K. M. Razeeb, C. O'Mathuna. (2005). Development of
field programmable modular wireless sensor network nodes for ambient systems. Computer
Communications. 28(13): 15311544
Y. E. Krasteva, J. Portilla, J. M. Carnicer, E. de la Torre, T. Riesgo. (2008). Remote HW-SW
Reconfigurable Wireless Sensor Nodes. Industrial Electronics, 2008. IECON 2008. 34th Annual
Conference of IEEE: 2483
383
7.
C. H. Zhiyong L. Y. Pan and Z. Zeng, M. Q. H. Meng. (2009). A Novel FPGA-Based Wireless Vision
Sensor Node. IEEE International Conference on Automation and Logistics, 2009. ICAL '09: 841
8. J. Portilla, T. R. A. de Castro. (2007). A Reconfigurable FPGA-Based Architecture For Modular Nodes
In Wireless Sensor Networks. 3rd Southern Conference on Programmable Logic, 2007. SPL '07: 203
9. P. P. Czapski, and A. Sluzek. (2007). Power Optimization Techniques in FPGA Devices: A
Combination of System - And Low Levels. International Journal of Electrical, Computer, and
Systems Engineering 1(3): 148-154
10. Xilinx Company. (2008). PicoBlaze 8-bit Embedded Microcontroller User Guide, for Spartan3,
Virtex-II, and Virtex-II Pro FPGAs. UG129 (v1.1.2
Biographies
zkan Akn He was born in Izmir, Turkey in 1979. He received the B.Sc. degree in electrical engineering from Kocaeli
University, Kocaeli, Turkey in 2002 and MSc. degree in electrical engineering from Ege University, zmir, Turkey in 2006.
He has papers concerning the control of power electronics for electrical machines and microprocessor architecture design
using FPGAs. His research interests include the electrical machine control, digital electronics, power electronics, and realtime digital simulation and control. Currently, he is a Research Assistant at the Ege University, Izmir, Turkey.
R.A. Akn has been a member of The Chamber of Electrical Engineers (EMO) since 2002 and Ege University FPGA society.
He has serious contributions for the establishment of FPGA and Electrical Machines Laboratories at Ege University.
lker Baaran He was born in Bursa, Turkey in 1990. He is currently continuing undergraduate education at the
Department of Electrical and Electronic Engineering and Department of Mechanical Engineering (Double Major
Undergraduate Program), Ege University. He plans to receive B.Sc. degree in 2012.
He is working on embedded systems, industrial automation, robotic systems and solar powered vehicles and their power
electronic systems.
Mr. Baaran is a member of The Chamber of Electrical Engineers (EMO) since 2008, Ege University Renewable Energy
Society and Ege University IEEE.
Radosveta Sokullu received her MSc and PhD degree in Telecommunications Engineering from the Sofia Technical
University in 1985 and 1992 respectively. She is currently an Assistant Prof. In the Department of EEE and Head of the
Chair of Telecommunications, Ege University, Izmir, Turkey. She was invited as a visiting professor for the GRASIUS
Program at the Graduate School of Engineering, Kumamoto University in 2008 and 2011.
Her current research interests include simulation for wireless nodes and wireless sensor networks, performance evaluation
and security issues.
rfan Alan He was born in Istanbul, Turkey in 1962. He received B.E.E. and M.S.E.E. degrees from Istanbul Technical
University (ITU), in 1983 and 1986, and the Ph.D. degree in E.E. from the University of Wisconsin-Madison, in 1993.
During most of his Ph.D. studies he has been a research assistant at UW-Madison working on various NASA projects. He
worked at UW-Madison as a research associate in another NASA project during his post doctorate studies. He has been with
Ege University since 1997, and serving as a full Professor since 2004 teaching various undergrad and grad level courses,
supervising many M.Sc. and Ph.D. thesis. His research interests are in solid state power conversion systems and ac machine
drives in general, and resonant and matrix power converters, and induction heating in particular. He has seven SCI journal
articles and many international conference papers.
Prof. Alan is a member of Chamber of Electrical Engineers (EMO). He is the founder of FPGA and Electrical Machines
and Power Electronics Laboratories at Ege University.
Kemal Bykkabasakal He was born in Stuttgart, Germany in 1984. He received the B.Sc. degree in electrical and
electronics engineering from Dokuz Eylul University, Izmir, Turkey in 2006 and MSc. degree in electronics engineering
from Ege University, Izmir, Turkey in 2010.
He is currently continuing his Ph.D. in Ege University. His research interests include the control systems, digital electronics,
and embedded systems. Currently, he is a Research Assistant at the Ege University, Izmir, Turkey.
384
gokhanbas@iyte.edu.tr, 2 mumin.ozcan@gediz.edu.tr
Abstract. This paper presents a novel application of phase only correlation implemented with parallel
processing for two-dimensional (2-D) set of acquired image array having only translational displacement
with some mutual boundary information. Proposed parallel processing tiling algorithm is applied to 2-D
microscopy images and comparison made with batch processing algorithm.
1 Introduction
The problem of 2-D tiling of acquired several images having mutual information are encountered in several
problems such as in digital astrophotography [1], virtual microscopy [2],[3],[4], in general, flattened panoramic
image stitching. Some of the approaches made by researchers are based only the batch processing of phase-ofcorrelation method (POC) in [5], [6], which is bulky and not sufficiently efficient. Therefore, this can be
improved by means of dividing non-algebraic-looped 2-D tiling algorithm in to the parallel processes. In this
manner, proposed algorithm assumes that the acquired images are known by their relative discrete picture
sequence (neighborhood) to each other, and POC is performed locally at least in one neighborhood of each
image in a parallels. Gathering the coordinates of each image relatively to each, at last, the starting image, is
sufficient to tile all images. In this manner, proposed algorithm is applied to bunch of microscopy images and
made a comparison with the traditional batch process, it is proved that the parallel processing is superior the
sequential process.
385
1.2 Methods
In this paper, POC is used with parallel scheduling algorithm to find the relative coordinates of images with
respect to each other. In the first place, sequentially taken images are placed in an array describing a
neighborhood representation. After that, starting from a middle point within the neighborhood array, finding the
correlation and determining a relative location vector between aforesaid images constructs a path. By means of
extending more than one way among four possible ways makes tiling like a diverging process from the selected
middle point. When the edges of the neighborhood matrix or already POC performed images are encountered in
the neighborhood array, the written down mutual relative vector expressions are composed together to form a
meta data for image to be tiled.
2 Problem Formulation
2.1 Used Notation
In this part of the paper, problem of high resolution 2-D image tiling is formulated by means of mathematical
notion. Let us consider that we have acquired M N = MN images
1 i MN ). As a 2-D notation, an image in rectangular grid M N may also be denoted as I ij , 1 < i < M ,
1 < j < N , besides L = I ij . Before going into tiling procedure, it is suitable to point out following remarks.
Remark 1.It is assumed that all input images are all known to their four neighbors in the way of how they are
sorted in the acquired image set
L . i.e. ,
N 4 ( I ij ) = I (i 1) j , I (i 1) j , I i ( j 1) , I i ( j 1) ,
(1)
(2)
= lk , l p , lm , ln , lk , l p , lm , ln L,
Equations (1-2) say that, for an image I ij , images N 4 ( I ij ) are known to their acquistion sequence in the set
L.
Remark 2.It is assumed that all gathered MN images are acquired such that an image has sufficient mutual
information with its four-neighboring images defined in Equation (1).
Remark 3.Ultimate goal of the 2-D tiling procedure is based on building a meta-data that keeps the relative
neighborhood translation vector(s) (RNTV). More precisely, RNTV is used to indicate the discrete vectoral
distance from image
I ij L to image
composed of v I i j , I ij vectors as a lookup table such that I i j is selected as starting image, and is can be
s s
s s
regarded as a kind of relative zero coordinate (is , i j ) .
386
v I ij , I ij = v I ij , I ij
(inverse element ),
v I ij , I ij = v I ij , I kl v I kl , I ij , I kl L. (RNTV conservation).
(3)
(4)
Notation shown here is sufficient to explain the procedure of 2-D image tiling.
i.e. = 1 and
= 1 [6]. Equation (5) does not hold for 1 or 1 that is why derivation of tree-map of
neighboring images are essential for evaluation of RNTV s, that is, meta-data knowledge.
In [6], tiling is performed in a sequential manner. As a rule of thumb example of this, tiling may be alligned
with the order of the image acquisition. In other words, a tree map may have a one-way sequence such that,
v Ii
s js
(6)
i =1 j =1
, I ij = v I 11 , I ij = F (i, j ),
I ij
(7)
Algebraic-loop free structure of F(i, j) of Equation (6) permits calculation F(i, j) in a parallel way provided that
tiling map and its RNTVs have been determined before, preparation of which takes negligible amount of time
compared to POC operations. Accordingly, proposed algorithm of Algorithm 2 (Appendix), parallel processing
completes tiling process rapidly as shown in Table 1 and this advantageous property is pointed out here.
387
Figure 1: Tree map of images that representing the neighborhood connections and corresponding RNTV s
connecting each image is shown, where, until all of the M N discrete image grid is spanned and not duplicated
with any other image
that
N D (Ii
fk j fk
I ij L
. The term ( i
fk
, jf
are already mapped in the tree so there exist no not included neighbors of
Ii
fk j fk
Figure 2: A rule of thumb tiling sequence, which can be regarded as a one way tree map of Equation (6), is
shown for M = N = 4 image array.
388
Table 1. Comparison of two given algorithms namely, Algorithms (1) vs. (2) in the sense of speed. (This
process is carried out on a intel i3 M 350 four-parallel processor computer. Experimental images had a
bytes
resolution of 1280 960 3 image . Convolution region of interest is entire image Noting that batch tiling makes use
of only single CPU.) Amount of mutual information is ~ %10.
References
1.
2.
3.
4.
5.
6.
7.
Keith Wiley, Steve Chambers, Digital Astrophotography: The State of the Art: Long-Exposure Webcams and
Image Stacking Techniques, Patrick Moore's Practical, Astronomy Series, Springer London, 2005, Section 1, 4560.
Steinberg DM, Ali SZ. Application of virtual microscopy in clinical cytopathology. Diagn Cytopathol. Dec
25(6):389-96, 2005.
Szeliski R., Found. Trends. Comput. Graph. Vis., Vol. 2, No. 1. (2006), pp. 1-104.
Kumar R.K., Velan G.M., Korell S.O., Kandara M., Dee F.R., Wakefield D., Virtual Microscopy for Learning
and Assessment in Pathology,Journal of Pathology, 204, no. 5, 613-18, 2004.
Bas G., Demir G.K., A Sinc Interpolation Based Image Stitching for Virtual Microscopy, BIOMUT 16 Ulusal
Biyomedikal Muhendisleri Bilisimi, ANTALYA, Nisan 2010.
Bas G., Graphical User Interface Development for Camera and 3-D Robot Synchronisation In Virtial
Microscopy by Using Qt Framework and ITK Libraries , Dokuz Eylul University, Final Year Graduation
Projects, June 2010.
B. S Reddy and B. N. Chatterji, An FFT-based technique for translation, rotation, and scale-invariant image
registration, IEEE Transactions on Image Processing 5, no. 8 (1996): 1266-1271.
389
Appendix
Biographies
Gkhan Ba- Gkhan Ba was born in Afyonkarahisar. He received his B.S.in Electrical Enginering at Dokuz Eylul
Univesity with Honors degree in 2010. Currently he is a Rsrc. Asst. at zmir Institure of Technology as a PhD. candidate.
His research interests are Lyapunov based controller design for robotic systems, and biomedical image processing.
Mmin zcan- Mmin zcan was born in Balkesir. He graduated Industrial Engineering department of Sakarya
University in 2003.He received his M.S degrees in Industrial Eng. from the Dokuz Eyll University in 2007. Currently he is
continuing his Ph.D.work in operation Research at the Sakarya University. Between 2009-2011, he worked as a research
assistant in zmir Gediz University. Since March 2011,he has been lecturer in zmir Gediz
University.
390
Abstract. Modeling MTS data effectively is important for many decision-making activities. Separately analyzing
each one of time series in MTS is not accurate because of the interactions among them, therefore for analyzing an MTS
all the time series must be regarded simultaneously. Existence of noise on time series makes this task hard.A lot of
research has gone into the development of ways of analyzing multivariate time series (MTS) data in both the statistical
and artificial intelligence communities. Statistical MTS modeling methods include the Vector Auto-Regressive process,
the Vector Auto-Regressive Moving Average process, and other non-linear and Bayesian approaches, while various AI
methods have been developed for different purposes. These include dependence detection in MTS of categorical data,
knowledge-based temporal abstraction, Bayesian clustering of similar MTSs, and forecasting. Modeling MTS data are
usually performed with VAR models. In this paper, a new AI method, recently introduced in [1], will be used for
determining parameters of VAR. This method (The Charged System Search (CSS)) is the most recent meta-heuristic
algorithm which utilizes the Newtonian law of mechanics in addition to the electrical physics laws to direct the agents
in order to recognize the optimum locations. Results show significant improvement in model accuracy.
Keywords: multivariate time series, Charge System Search, modeling, artificial intelligence
1 Introduction
Recent developments in computing have provided fast access to vast amounts of online data; processing power
and storage devices continue to become cheaper and more powerful, networks are providing more bandwidth
and higher reliability, and online analytic processing allows rapid retrieval of data from data warehouses. This is
especially true for the recording of time series data, for example, in medical and financial sectors. This
necessitates methods for the analysis and modeling of time series which are accurate and fast. A time series is a
sequence of data points, measured typically at successive times spaced at uniform time intervals. Models for
time series data can have many forms and represent different stochastic processes. Time series can be regarded
as univariate or multivariate. A univariate time series which is independent from other time series can be
analyzed or modeled individually. However, multivariate time series is a vector of time series which are
depended to each other with high interaction among them. Multivariate Time Series (MTS) data are common in
many fields such as medicine, finance and science, and innovations in modeling this kind of data are still
welcome.
Modeling MTS data effectively is important for many decision-making activities. Separately analyzing each one
of time series in MTS is not accurate because of the interactions among them, therefore for analyzing an MTS all
the time series must be regarded simultaneously. Existence of noise on time series makes this task hard. Filtering
or distinguishing noise from the main data is a usual solution to this problem. However, in some applications we
391
needs to model noise or disturbance along with main data such as determining performance of the system, model
must be fitted to the both data and disturbance to have the accurate prediction in the minimum variance index. In
this application a model is fitted to the time series and by evaluating and analysis it minimum possible variance
for system can be obtained [2]. For determining the accurate value for minimum variance the model must be
accurate in fitting to systems and disturbances data. In other words, model must be able to follow the vibration
in output. This usually necessitates models with higher orders. A lot of research has gone into the development
of ways of analyzing multivariate time series (MTS) data in both the statistical and artificial intelligence
communities. Statistical MTS modeling methods include the Vector Auto-Regressive process, the Vector AutoRegressive Moving Average process, and other non-linear and Bayesian approaches [3], while various AI
methods have been developed for different purposes [4,5,6]. These include dependence detection in MTS of
categorical data, knowledge-based temporal abstraction, Bayesian clustering of similar MTSs, and forecasting
[7,8,9]. Modeling MTS data are usually performed with VAR models. It has been argued that an advantage is
that it does not need any prior assumptions for a structural model. For reaching higher accuracy, higher order for
VAR and capable parameter estimation methods are needed. Consequently, higher order of VAR leads to higher
number of unknown parameters. Estimating the value for high number of parameters (high dimensional
problems) is hard or even impossible for some methods. Most algorithms cannot be used in such problems, as
they are usually trapped in local minimums. AI methods practically are used for finding best parameters for this
kind of models [10,11,12].
In this paper, a new AI method, recently introduced in [1,13,14], will be used. This method (The Charged
System Search (CSS)) is the most recent meta-heuristic algorithm which utilizes the Newtonian law of
mechanics in addition to the electrical physics laws to direct the agents in order to recognize the optimum
locations. Similar to many other meta-heuristics, this algorithm contains a number of agents, so-called Charged
Particle (CP). Each CP is considered as a charged sphere which can impose an electric force to the other CPs and
therefore, each CP is treated as a source. Some applications have shown that this algorithm outperforms other
well-known algorithms such as Evolutionary Algorithms (EA). This algorithm is fast and accurate in finding the
global minimum and can escape from local minimums; hence it is very appropriate for finding VAR parameters
with high orders.
The organization of this paper is as follows: section 2 explains the CSS algorithm in more detail; section 3
shows the simulation results and compares the CSS with GA, CEP and PSO and section 4 presents the
conclusion and summarizes the simulation results.
392
3 Results:
A multivariate time series with two depended time series has been assigned to compare ability of algorithms in
modeling MTSs. Modeling will be performed on 200 points of multivariate time series. Four algorithms have
been selected for this comparison. All of them are well known in AI domain. These algorithms are continuous
Genetic Algorithm (GA), Particle Swarm Optimization (PSO), Classical Evolutionary Programming (CEP), and
Charged System Search (CSS).
393
parameters
Population
Number of variables
Variable bounds
Mutation factor
Selection rate
Inertia factor
acceleration coefficients
Tournament size q
GA
PSO
CEP
value
100
42
[-100,100]n
0.1
0.5
1
2
10
VAR model have been selected for modeling with order 10, so 42 (
) parameters must be
determined by the algorithms. With increasing order of VAR, dimension of cost function and in consequence
complexity of problem increase exponentially.
VAR is generalized form of AR in which time series are depended to one another (1).
p
X (t ) Ai X (t i ) b
i 1
a
Ai i1
ai 3
ai 2
,
ai 4
(1)
b
b 1
b2
By considering (1), the chromosomes or individuals of the algorithms can be defined as (2).
(2)
Minimum least square of errors, among MTSs points and models outputs, makes cost function for the
algorithms (3). The algorithm can minimize this cost function further and be fast in reaching to minimum value
is capable one for modeling this kind of time series.
2
200
Cost X (i ) X (i ) .
(3)
i 1
Where X (i ) are i-th points of MTS and X (i ) are i-th points of VAR models outputs.
Table 2 compares minimum cost found by each algorithm after 10000 generations. It is clear that CSS is more
accurate in getting close to the global minimum. For comparing
Table 2 : Minimum cost found by the algorithms
Cost
GA
778.22
CEP
346.86
PSO
1839.3
CSS
48.85
the algorithms about their speed some figures have been drawn. Figure 3 shows speed of the algorithm in
decreasing value of the cost function versus generations. Speed of the algorithm is more crucial when the
algorithm is being used in online application in which modeling process is repeating along with the passage of
time. Accuracy is more important when optimization task is performed in offline applications in which the
algorithm has enough time to getting close to the global minimum. In addition to these two main factors
releasing from local minimums is most important factor that decides the speed and accuracy of algorithms.
Figure 2 shows this factor too. The algorithm which stays on a cost during several generations (like PSO) loses
394
its speed and opportunity for searching more accurately. However figure 2 shows that none of the algorithms
could reach the global minimum, CSS has more speed in decreasing value of the cost function.
8
15
time series 1
GA
Time series 2
GA
10
4
5
Cost
Cost
2
0
0
-5
-2
-10
-4
-6
0
50
100
number of iteration
150
-15
0
200
100
number of iteration
150
200
Time series 1
CEP
50
Time series 2
CEP
3
1
2
0
Cost
Cost
1
0
-1
-1
-2
-2
-3
-3
-4
-4
-5
0
50
100
number of iteration
150
-5
0
200
25
50
100
number of iteration
150
200
20
Time series 1
PSO
20
Times series 2
PSO
15
15
10
10
Cost
Cost
5
0
0
-5
-5
-10
-10
-15
-15
-20
-20
-25
0
50
100
number of iteration
150
-25
0
200
50
100
number of iteration
150
200
0.6
1.5
Tim series 1
CSS
Tim series 2
CSS
0.4
0.2
0.5
Cost
Cost
0
0
-0.2
-0.5
-0.4
-1
-1.5
0
-0.6
50
100
number of iteration
150
200
-0.8
0
50
100
number of iteration
150
Figure 3 : Shows differences between models determined by the algorithmsand main data
395
200
For increasing the accuracy of model there are some policies such as increasing order of VAR model, increasing
number of generation and selecting other kind of models like VARMA.
5
10
CEP
GA
PSO
CSS
Cost
10
10
10
10
2000
4000
6000
8000
number of iteration
10000
12000
Figure 3 shows accuracy of models and differences between models determined by the algorithms and main data.
Conclusion
In this paper, a new AI method, have been utilized. This method (The Charged System Search (CSS)) is the most
recent meta-heuristic algorithm which utilizes the Newtonian law of mechanics in addition to the electrical
physics laws to direct the agents in order to recognize the optimum locations. This algorithm is fast and accurate
in finding the global minimum and can escape from local minimums, hence it is very appropriate for finding
VAR parameters with high orders. VAR is an approach to model MTS data. For reaching higher accuracy in
modeling, higher order for VAR and capable parameter estimation methods are needed. Consequently, higher
order of VAR leads to higher number of unknown parameters. Estimating the value for high number of
parameters (high dimensional problems) is hard or even impossible for some methods. As a result there is a need
for a powerful algorithm to adapt these parameters precisely. The CSS has good potential for applications in
such areas. In this study performance of CSS has been compared with other AI algorithms such as EAs and
PSO. Simulation results clear that CSS is more appropriate for modeling VAR than GA, PSO and CEP.
References
1.
2.
3.
4.
5.
6.
7.
Kaveh A, Talatahari S (2010) A novel heuristic optimization method: charged system search. Acta Mech
213(34):267289
M Jelali (2010), Control System Performance Monitoring Assessment, Diagnosis and Improvement of Control
Loop Performance in Industrial Automation, 25 Jan 2010.
Luc Bauwens, Sbastien Laurent, Jeroen V. K. Rombouts, Multivariate GARCH models: a survey, JOURNAL
OF APPLIED ECONOMETRICSJ. Appl. Econ. 21: 79109 (2006)
O. Valenzuelaa, I. Rojasb, F. Rojasb, H. Pomaresb, L.J. Herrerab, A. Guillenb, L. Marqueza, M. Pasadasa,
Hybridization of intelligent techniques and ARIMA models for time series prediction, Fuzzy Sets and Systems
159 (2008) 821 845.
Shinn-Ying Ho, Li-Sun Shu, and Jian-Hung Chen, Intelligent Evolutionary Algorithms for Large Parameter
Optimization Problems, IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, VOL. 8, NO. 6,
DECEMBER 2004.
M. Han , Y. Wang, Analysis and modeling of multivariate chaotic time series basedon neural network Expert
Systems with Applications 36 (2009) 12801290.
A Silvestrini, temporal aggregation of univariate and multivariate time series models: a survey, 10.1111/j.14676419.2007.00538.x.
396
8.
9.
10.
11.
12.
13.
14.
P. Kellam, X. Liu, N. Martin, C. Orengo, S. Swift, and A. Tucker, A Framework for Modeling Virus Gene
Expression Data, PHD dessert.
S. Swift and X. Liu, Predicting Glaucomatous Visual Field Deteriorationthrough Short Multivariate Time Series
Modeling, PHD dessert.
I. Rojas, F. Rojas, H. Pomares, L.J. Herrera, J. Gonzalez, O.Valenzuela, The synergy between classical and softcomputing techniques for time series prediction, in: Advances in Artificial Intelligence, Lecture Notes in Computer
Science, Vol. 2972, 2004, pp. 3039
C.-S. Ong, J.-J. Huang, G.-H. Tzeng, Model identification of ARIMA family using genetic algorithms, Appl. Math.
Comput. 164 (3) (2005) 885912.
Valds, Julio; Barton, Alan, Multivariate Time Series Model Discovery with Similarity BasedNeuro-Fuzzy
Networks and Genetic Algorithms April 2003, published in Proceedings of the IEEE, INNS, IJCNN 2003
International Joint Conference on Neural Networks. IEEE Catalog Number: 03CH37464C,ISBN: 0-7803-7899-7.
Oregon, Portand, USA.
A Kaveh , S Talatahari, An enhanced charged system search for configuration optimization using the concept of
fields of forces, Springer, Struct Multidisc Optim, 2011.
Kaveh A, Talatahari S (2010) Charged system search for optimum grillage systems design using the LRFD-AISC
code. J Constr Steel Res 66(6):767771.
Biographies
Yousef Alipouri received the B.Sc. degree in control engineering from University of Tabriz, Tabriz, Iran, in 2009. He is
currently pursuing his study towards the M.Sc. degree at Iran University of Science and Technology. His research interest
focuses on control system theory and artificial intelligence.
Javad Poshtan received his B.Sc., M.Sc., and Ph.D. degrees in electrical engineering from Tehran University, Tehran, Iran,
in 1987, Sharif University of Technology, Tehran, Iran, in 1991, and the University of New Brunswick, Canada, in 1997,
respectively. Since 1997, he has been with the Department of Electrical Engineering at Iran University of Science and
Technology. He is involved in academic and research activities in areas such as control systems theory, system identification,
and estimation theory.
397
emredone@hotmail.com, 2adnankaya@sdu.edu.tr
Abstract. This study presents a 2.4 GHz RF power amplifier module design with active biasing for ISM
band wireless communication. The PA modul has 3 stage. First this PA was designed with two supply and
then -4V supply was removed and active biasing circuit replaced. These two circuit were obtained according
that result of simulation gain 29 dB ,noise 3 dB.
1 Introduction
The fast growing wireless communication industry is demanding more radio spectrum, and has driven RF
design into higher microwave frequency bands at 2.4 GHz and beyond. Wireless communication products such
as portable phones, require low cost power amplifiers operating at low voltage with high efficiency and high
power for portable handsets.[1]
This paper presents the design of a 2.4 GHZ ISM band power amplifier module with active biasing. The
power module is a 3 stage amplifier consisting of an input amplifier stage, a driver stage, a power output stage.
Also there is a active stage in module. Two bipolar transistor (AT41486 and BFG31) and two GaAs MESFET
(CLY2 and CLY5) power transistors are used in this amplifier. This module fabricated on low cost FR-4
substrate.The main purpose of power amplifier increasingthe power ofinputsignal froma poweroutletto
communicate.
Thisproject
isaimed
to
bethepowergainbetween25-30
dB.
In
addition,2.4GHzoperatingbandlinearamplifieras possibleandis intended to representa stableoperation. In
addition to theseamplifierdesignedforlowdistortionas possible,class Abiasingcircuitsare designedaccording to
the study.
First power amplifier has been designed two supply as +5V and -4V. Then -4V supply has been removed and
replaced BFG31 for active biasing. With the S parameters of the selected transistors by manufacturers AWR
design environment (AWRDE) program is modeled.
398
(2)
R3=(Vdd-VE)/Ids
(2)
VB=VE+VBE
(3)
399
VB=Vdd*[R1/(R1+R2)]
(4)
VDD=IBB*(R1+.R2)
(5)
(6)
R1=Vdd[Ibb[1+[(Vdd-Vb) / Vb]]]
(7)
3 Simulation Results
Determinedbefore starting to designthe poweramplifiertoGain25-30dB, 2-3dBfornoise, S11-10to-15 dB,
S22-20and-25dbdesignwasin line withthetargets set. Thedesign,simulationresultsobtainedare given inTable1. In
addition,the data obtainedbythe activebiascircuitdesignis given intable1.
400
S11 (dB)
- 11.5
+5V&
Biasing
-13
S22 (dB)
- 20.6
- 23.6
S21 (dB)
29.5
29.4
Gain (dB)
29.6
29.5
Noise (dB)
Stability
10
11
Active
VSWR
1.7
1.5
Table 1. Comparison Performans Power Amplifier with -4V Supply and Active Biasing
4 Conclusion
The design fabrication and measurements of a 2.4 GHz RF power amplifier module have been presented. The
first PA has 3 stage amplification which uses a BJT (AT 41486) and two GaAs MESFET (CLY2-CLY5). Then
-4V negative supply was removed and active biasing circuit was replaced. Asthe mostrecentresultsobtained
from thesimulationwere compared.S11andS22parameters ofthe simulationwere improvedas a result ofthe
noisehas not changed. Gain remainedalmostthe same.
401
References
1.
2.
3.
4.
5.
6.
7.
8.
Huey-Ru Chuang,Sin-yue Lin,Sen-you Liu and Cheng-Yung Chiang,(1998) Design of RF Power Amplifier for
ISM-Band Wireless Communication, Applied Microwave&Wireless
Pozar, D. M., 2005. Microwave Engineering Third Edition. John Wiley & Sons, Inc., United States of America.
Smith, J. R.,1998. Modern Communication Circuits, Mcgraw-Hill. New York.
Sayre Cotter W.,2008. Wireless Communication Design , Mcgraw-Hill. Second Edition. New York.
Gilmore R. , Beser L. 2003 Practical RF Circuit Design for Modern Wireless Systems, Artech
House.Boston.London
Hazouard M. 2002 Conception et ralisation d'amplificateurs micro-ondes de puissance l'aide de la mthode
des frquences relles Bordeuaux Luniverste
Ludwig, R. and Bretchko, P., 2000. RF Circuit Design. Prentice-Hall, Inc., United States of America.
ATF 55143 datasheet,www.alldatasheet.com
Biographies
Emre Dne is a master Electronics and Communication Engineering student in at Sleyman Demirel University
and currently working on developing 2.4 GHz power amplifier for Rf modules for ISM-band wireless communications.
Adnan Kaya is a associate professor in at Sleyman Demirel University. He works about electromagnetic field
and microwave technique.
402
Abstract. The Modular multiplication is one of the most important arithmetic operations in modern applied
cryptography. It requires both multiple-precision multiplication and some method for performing modular
reduction. Montgomery reduction technique allows efficient implementation of modular multiplication
without explicitly carrying out the classical modular reduction step. In [1], new techniques for performing
modular multiplication and exponentiation are proposed, namely, Spectral Modular Multiplication (SMM)
and Spectral Modular Exponentiation (SME). As the names imply, these methods work completely in the
frequency domain with some exceptions such as the initial, final, and some partial transformations
calculations. Since the proposal of SMM, as far as authors knowledge, this is the first contribution providing
an actual implementation.
Keywords: Modular Multiplication, Number Theoretical Transform (NTT), RSA, FPGA
1 Introduction
Many public-key encryption and digital signature schemes require computations in Z q (the ring of integers
with multiplication and addition modulo a positive integer q). For example, the RSA, Diffie-Hellman, and
ElGamal schemes require multiplication and exponentiation in Z q . The efficient implementations of modular
arithmetic operations including modular multiplication, inversion, and exponentiation have been at the center of
research activities in cryptographic engineering. Hence, there is an extensive literature on methods for multipleprecision modular arithmetic [1-6].
Spectral techniques for integer multiplication have been known for years. Due to Schnhage and Strassen [4],
large to extremely large sizes of numbers can efficiently be multiplied using the spectral integer multiplication.
In [1], some spectral modular reduction techniques are introduced which yielded new algorithms for performing
modular multiplication and exponentiation, namely Spectral Modular Multiplication (SMM) and Spectral
Modular Exponentiation (SME). These methods work completely in the frequency domain with some exceptions
such as the initial, final, and some partial transformations calculations.
Since the proposal of SMM, as far as authors knowledge, this is the first contribution providing an actual
implementation. The paper is outlined as follows: In Section 2, a formal definition of Discrete Fourier Transform
(DFT) and spectral modular multiplication concepts are given. In Section 3, a hardware implementation of SMM
is described and the algorithm is examined through with case studies. Finally, some conclusions are drawn in
Section 4.
403
2 Modular Multiplication
In this section, we give some preliminary notation and a formal definition of DFT over Z q . We then present
the Spectral Modular Product (SMP) and SMM algorithms. Note that the information given in this section can be
found in [1, 2], they are included here simply for the sake of completeness.
(1)
j 0
for i, j 0,1,..., d 1. We say x(t ) and X (t ) are transform pairs, x(t ) is called a time polynomial and sometimes
SMM can be divided into three sub-procedures: Initialization, spectral modular product, and finalization. Initial
and final stages consist of some data arrangements.
404
N (t ) are transform pairs of x(t ), y(t ) and n(t ) respectively. We assume that x(b) x , y(b) y and n(b) is a
multiple of modulus n with n0 1 for some integers x, y n and b 0 .
INPUT: X (t ) , Y (t ) and N (t ) ; spectral polynomials
OUTPUT: Z (t ) DFT ( z(t )) where z xy 2 db mod n and z (b) z,
1. Z (t ) : X (t )
Y (t )
2. : 0
3. for i 0 to d 1
4.
z0 : d 1. Z0 Z1 Zd 1 mod q
5.
: z0 mod b
6.
: z0 / b
7.
Z t : Z t N (t ) mod q
8.
Z t : Z t z0 (t ) mod q
9.
Z t : X t
t modq
Compute (t ) , bd mod n
2.
3.
X d (t ) : DFTdw ( x d (t ))
4.
Y (t ) : DFTdw ( y(t ))
5.
N (t ) : DFTdw (n(t ))
6.
Z (t ) : SMP( X d (t ), Y (t ), N (t ))
7.
z (t ) : IDFTdw (Z (t ))
8.
return z (b)
Since operations are carried out in a finite ring spectrum, one has to deal with overflows that might occur
during the computations. Due to [1, 2], overflows can be precluded if the parameters satisfy
(2)
2sb2 q
Note that in general, a single classical modular multiplication is faster than a single SMM; however, spectral
methods are very effective when several modular multiplications with respect to the same modulus are needed.
An example is the modular exponentiation.
405
3 Implementation
The primary objective of this study is to provide a hardware implementation of SMP and thereby SMM.
Unlike other architectures in the literature, the modular multiplication operation is performed in the spectral
domain.
SMP hardware accepts three parameters d (DFT length), b (base), and q (size of the ring). The parameters d
and b are crucial in terms of area and speed of the overall SMP architecture. We first select a ring, this comes
with the principal root of unity w. The DFT length d determines s d / 2 . Using the inequality in (2),
allowable b values and hence the parameter u can be found. This informs about the size of the input bits k via
k s u .
3.1 Methodology
The implementation consists of two essential steps: First, the algorithm is implemented and simulated in
MATLAB. Second, an hardware of the SMM algorithm is designed and implemented in VHDL. The architecture
is simulated in Xilinx ISim Simulator to verify the results with MATLAB. Finally, the architecture is
synthesized for the Xilinx FPGA. Figure 1 shows the hardware structure of the algorithm.
SMP
d*v
X(t)
Xt
d*v
Y(t)
(t)
d*v
(t)
d*v
Yt
Teta
Gama
Z(t)
d -1
d*v
Zt
invd
rdy
ready
clk
clock
reset
rst
start
start
(a)
start.iteration
Z(t)
iteration
Initialization
Iteration
d*v
Zt
X(t)
Y(t)
d*v
d*v
Xt
d*v
Yt
Const
clock
clk
reset
rst
start.initial
Zt
Z(t) init
Montgomery
Coefficient
Const
ready.initial
start
d*v
Finalization
Z(t)
iteration
Zt
Zt
Z(t)
final
d*v
d*v
q
rdy
Zt_out
(t)
d*v
(t)
d*v
d -1
alfa
alfa
clk
clock
rdy
rst
reset
Teta
start.final
ready.final
start
Gama
invd
clk
clock
rst
reset
start.iteration
start
rdy
ready.iteration
(b)
Controller
ready.initial
ready.iteration
clock
clk
reset
rst
ready
start
start
start.initial
start.iteration
start.final
ready
ready.final
(c)
Figure 1: The hardware structure of SMM (a) Top-level (b) The inner structure
406
1
2
3
4
5
6
7
8
4
8
12
16
20
24
28
32
2
4
8
16
32
64
128
256
Min.
Period (ns)
6,879
7,315
7,378
7,480
7,531
7,713
7,781
7,882
Max.
Frequency
(MHz)
145,361
136,711
135,535
133,685
132,789
129,653
128,526
126,871
Number of
Slice
Registers
2522
2521
2520
2519
2518
2517
2516
2515
Number of
Slice LUTs
3678
3699
3699
3700
3701
3703
3703
3705
with q 240 1 . The DFT exists in this ring with d 16 and w 32. The maximum value of u is found to be
17. Using these parameters in the SMP architecture, several simulations are performed for different values of u.
The results are summarized in Table 2. Note that, the sizes of the operands range between 8 to 128.
Min.
Period (ns)
1
2
4
8
12
16
8
16
32
64
96
128
2
4
16
256
4096
65536
7,804
8,243
8,408
8,811
9,269
9,271
Max.
Frequency
(MHz)
128,137
121,317
118,928
113,496
107,889
107,864
Number of Slice
Registers
8115
8114
8112
8108
8104
8100
407
Number of
Slice
LUTs
11742
11785
11787
12084
11795
11799
4 Conclusion
The work at hand presents the first documented implementation of recently proposed algorithm for modular
multiplication, SMM [1, 2]. This algorithm is implemented in VHDL, tested for various bit length sizes and
synthesized for the Xilinx Virtex-5 XC5VLX50 FPGA.
A possible application for the presented architecture is public-key primitives such as the RSA, ElGamal, or
elliptic curve operations, since SMP can be used as a building block for modular exponentiation. The value of
spectral techniques will be appreciated in these applications. It is foreseen that spectral algorithms yield efficient
and highly parallel architectures for hardware realizations of public-key cryptosystems.
References
1.
2.
3.
4.
5.
6.
7.
8.
Saldamli, G. (2005). Spectral Modular Arithmetic. Ph.D. Thesis, Department of Electrical and Computer
Engineering, Oregon State University, USA.
Saldamli, G. and Koc, C.K. (2007). Spectral Modular Exponentiation. 18th IEEE Symposium on Computer
Arithmetic, 123--130, IEEE Computer Society Press.
Menezes, A.J., van Oorschot, P.C. (2001). Vanstone, S.A.: Handbook of Applied Cryptography, CRC Press.
Schnhage, A., Strassen, V. (1971). Schnelle Multiplikation Grafer Zahlen. Computing, 281292.
Amanor, D.N., Paar, C., Pelzl, J., Bunimov, V., Schimmler, M. (2005). Efficient Hardware Architectures for
Modular Multiplication on FPGAs. International Conference on Field Programmable Logic and Applications,
539542.
Montgomery, P. (1985). Modular Multiplication without Trial Division. Mathematics of Computation, 519- - 521.
Blahut, R.E. (1985). Fast Algorithms for Digital Signal Processing, Addison-Wesley.
Naussbaumer, H.J. (1982). Fast Fourier Transform and Convolution Algorithms, Springer.
Biographies
smail San was born in Denizli, Turkey, in 1983. He has double major in Electrical-Electronics Engineering and Avionics.
I received my BSc in Electrical-Electronics Engineering from Anadolu University in Eskiehir in Turkey in 2008 in five
years. I received my second BSc in Avionics from Anadolu University in Eskiehir in Turkey in 2008 in five years.
His research interests are in security of digital communication channels, cryptographic embedded systems, secure
hardware design and architectures for communication and computer systems.
PhD student Mr. San is a member of IEEE.
Nuray At was born in Eskisehir, Turkey. She received her B.Sc. degree in Electrical and Electronics Engineering from
Anadolu University; two M.Sc. degrees in Electrical and Computer Engineering from University of California Davis and
Georgia Institute of Technology, respectively; and Ph.D. degree in Electrical and Electronics Engineering from Anadolu
University.
Her research interests include information security, cryptographic engineering, communications systems, estimation, and
signal processing. Dr. At is a member of IEEE.
408
Anadolu University, Faculty of Engineering and Architecture, Electrical and Electronics Engineering Dept.,
Eskisehir, Turkey
1
isan@anadolu.edu.tr, 2nat@anadolu.edu.tr
Abstract. Security protocols such as IPSec, SSL and VPNs employ various cryptographic primitives. In order
to provide data integrity and message authentication in such protocols, special family of cryptographic
algorithms named cryptographic hash functions are used. Keccak hash algorithm is one of the finalists in
SHA-3 competition that aims to develop the next generation hash algorithm. With advancement in
technology, some of these protocols and applications call for higher throughputs. In this work, we present an
embedded System on Chip (SoC) design that can operate at high frequencies yielding high throughputs in
relatively smaller area than previously reported loop unrolled or pipelined architectures for the Keccak hash
algorithm. Hardware/Software codesign approach is used to exploit the benefits of both methodologies. The
proposed SoC architectures were synthesized using VHDL and implemented on Xilinx Virtex5 device.
Performance analysis is presented revealing the superiority of the proposed SoC design in terms of execution
time.
Keywords: Cryptographic hash function, Keccak, SoC Design, Acceleration, FPGA
1 Introduction
With advancement in technology, several applications including networking and streaming ask for more
computational speed as well as data security at the same time. Security protocols such as IPSec, SSL and VPNs
employ many cryptographic primitives to meet security needs of such applications. In order to provide data
integrity and message authentication in such protocols, special family of cryptographic algorithms named
cryptographic hash functions are used.
Cryptographic hash algorithms take a message of arbitrary-length as input and produce an output of fixed-length
referred to as hash-value, or simply hash. The fundamental idea of cryptographic hash functions is to generate a
condensed representation of the input called the message digest. A hash function should have two properties as a
minimum: compression and ease of computation [9].
Recently, the National Institute of Standards and Technology (NIST) organized a public competition to develop
the next generation hash algorithm called SHA-3. Keccak hash algorithm is one of the finalists in the
competition whose hardware performance worth investigation. In this study, we propose a System on Chip (SoC)
design which uses an efficient hardware core named Keccak Coprocessor on Xilinx FPGAs with MicroBlaze
soft processor.
Authentication protocols based on cryptographic hash algorithms are extensively used in todays networking and
409
streaming applications. These protocols authenticate the content of a packet to be transmitted as follows: At the
transmitter, an integrity check value for a packet is computed using a cryptographic hash algorithm, this value is
then added into a specific authentication header with payload, and transmitted over a channel. At the receiver,
the integrity check value is computed for a received packet using the same cryptographic hash algorithm. If none
of the fields differ, it is assumed that the content of a packet is not altered and/or in error. This provides a
mechanism to verify the integrity of a packet transmitted over a channel. However, it does not provide
confidentiality since there is no encryption involved. Thus, the cryptographic hash algorithm provides
authentication but not the privacy. This work focuses on the acceleration of data integrity and message
authentication services using the Keccak hash algorithm.
We adapt Hardware/Software codesign approach in order to exploit the advantages of both methodologies. The
design rationale of the proposed architecture is to accelerate the performance of computational intensive tasks
used by the Keccak cryptographic hash algorithm. We use specially designed Keccak Coprocessor that achieves
high throughputs in a small area. Owing to the serialization exploited in the Keccak Coprocessor, the required
area for the design is relatively smaller than the ones presented in the literature [2,4,11]. Furthermore, the low
latency presented by the coprocessor empowers to operate in higher frequencies [13].
In this study, we attach the Keccak Coprocessor to MicroBlaze processor via both the Processor Local Bus
(PLB) [15] and the Fast Simplex Link (FSL) bus [16]. The latter connection performs relatively better than that
of the former bus attachment, since the FSL directly links the Keccak Coprocessor to the execution stage of the
MicroBlaze processor. The benefit of a microprocessor based system together with a custom hardware is realized
via embedded system approach by Xilinx FPGAs. Results show that a substantial gain in the performance of
computational intensive tasks used by the cryptographic hash algorithm is achieved. Moreover, we emphasize
the fact that this performance improvement is obtained without significantly increasing the required area (silicon
gate count) of the system.
The paper is outlined as follows. Section 2 briefly overviews the literature on cryptographic hash functions
and their realizations on FPGA based devices. Section 3 describes the Keccak hash algorithm. The proposed SoC
design is presented in Section 4 including details for the MicroBlaze based embedded system and Keccak
coprocessor details. Section 5 shows performance results of the proposed SoC architectures. Finally, some
conclusions are drawn in Section 6.
2 Literature Review
In recent years, several cryptographic hash functions have been successfully attacked, and serious attacks have
been published against SHA-1, a Secure Hash Algorithm designed by the National Security Agency (NSA) and
published by the NIST as a U.S. Federal Information Processing Standard. Although no successful attacks have
yet been reported on the SHA-2 variants, they are algorithmically similar to SHA-1; hence, NIST has decided to
develop one or more additional hash functions named SHA-3 through a public competition [1,5,6,8,10].
In this study, we present a SoC design that efficiently provides data integrity and message authentication
services using the Keccak hash algorithm, one of the finalists in the SHA-3 competition. Keccak is a family of
hash functions that are based on the sponge construction. The sponge construction operates on a state of b-bits.
In order to attain a certain security level, usually large state sizes are called for. However, the large state sizes
may be cumbersome in the hardware implementation since the operations must be done with large state contents.
In straightforward implementations of the Keccak Hash algorithm [1-2], this resulted in a very large amount of
silicon area utilization that questions the practicality and/or suitability of the algorithm. This work remedies the
above mentioned problem.
Early Keccak hash algorithm implementations seen in the literature are parallel, loop-unrolled or pipelined
architectures using an extensive amount of slice resources on FPGAs. Our custom designed Keccak Coprocessor
uses on-chip memory components of FPGAs, implementing state and summation registers in separate RAM
units on the device. For further details of our custom designed Keccak Coprocessor, please contact authors. Our
contribution can be stated as building a system which efficiently provides data integrity and message
410
authentication services using a dedicated Keccak Coprocessor via Hardware/Software design methodology.
b r c bits, where r is the bitrate and c is the capacity which determines the attainable security level.
In the initialization phase, all the bits of the state are set to zero. The input message is padded and divided
into blocks of r bits. The sponge construction then proceeds with the absorbing phase followed by the
squeezing phase.
In the absorbing phase, the r-bit input message blocks are XORed with the first r-bit of the state, and the
resulting outputs are interleaved with the function f. When all message blocks are processed, the sponge
construction alters to the squeezing phase.
In the squeezing phase, the first r-bit of the state is returned as output blocks and are also interleaved with
the function f. The number of output blocks can be arbitrary and is chosen by the user.
411
412
Multiport
Memory
Controller
IIC
SPI
Cache
Links
DXCL
IXCL
DLMB
Local
Memory
UART
DPLB
GPIO
Microblaze
ILMB
IPLB
BRAM
FSL
Interrupt
Controller
LMB
Buses
Timer
Coprocessor
Keccak Coprocessor
addrB
weB
dinB
PortB PortA
State
Register File
PortB
select
Port D
addrA
Port D Port C
Datapath
Control
Unit
11
addr
addrD
addrC
11
weC
dinC
Data out
C
Data out
State
start
init
ready
ctrl
Register File
C
Port C
dinA
0
PortA
data in
address
64-bit
1-bit
weA
we
Instruction
Memory
413
set; hence, the user can access the message digest from the register file via user interface.
BRAM
SLMB
SLMB
dlmb_cntrlr ilmb_cntrlr
SLMB
dlmb_cntrlr ilmb_cntrlr
DLMB
DLMB
ILMB
ILMB
SFSL
SFSL
MFSL
MFSL
KECCAK
COPROCESSOR
Microblaze
Microblaze
DPLB
IPLB
DPLB
IPLB
mb_plb
mb_plb
SPLB
XPS
Timer
SPLB
SPLB
SPLB
SPLB
UART
KECCAK
COPROCESSOR
XPS
Timer
UART
(a)
(b)
Fig. 4. (a) PLB connection (b) FSL connection for the Keccak Coprocessor.
FIFO
. . .
FSL_S_Clk
FSL_S_Data
FSL_S_Control
FSL_S_Read
FSL_S_Exists
Fig. 5. Block diagram of an FSL with Master and Slave signal interface.
414
5 Performance Results
Several SoC architectures are experimented for different SoC configurations. Table 1 lists the configuration
parameters for the Keccak Software, Keccak PLB, and Keccak FSL architectures.
Keccak Software is an architecture without the Keccak coprocessor, that is, the Keccak hash algorithm is
realized in a pure software domain. The C code for the Keccak hash algorithm is written with the same
configuration parameters as its correponding hardware cores.
Keccak
coprocessor
Barel
shifter
DDR2
sdram
Area (Slices)
5 27 (69120)
5651 (69120)
5837 (69120)
BRAM
blocks
19 (148)
21 (148)
25 (148)
Keccak FSL and Keccak PLB architectures have the Keccak coprocessor as a peripheral unit in the MicroBlaze
based embedded system. In Keccak PLB, the coprocessor is connected to the MicroBlaze via PLB whereas in
Keccak FSL this connection is made via FSL.
Several sets of measurements are taken during experiments revealing the benefits of using hardware based
Keccak hash algorithm core compared to the software based realization of the Keccak hash algorithm in terms of
execution time. Fig. 6 shows execution times of the three different SoC architectures for different cache
configurations without the compiler optimization where the vertical axis has a logarithmic scale, ICache and
DCache denote the instruction and data caches, respectively. Note that different cache configurations affect the
system performance. However, for all cases, the Keccak Software architecture has the highest execution time to
compute hash function among others.
415
416
Fig. 8.Speedups of Keccak FSL and Keccak PLB with respect to Keccak Software
6 Conclusion
In this study, an efficient SoC design is introduced for the systems, which needs to have the services of data
integrity and message authentication. Thanks to the efficient Keccak coprocessor and bus based approach
provided by the Microblaze processor, the resulted SoC design accelerates the system performance for the
applications such as networking and fast authenticated data communication where the performance is important
[7]. The reconfigurable computing [12] provides significant speed-up which is achieved with the Microblaze
processor based platform by the Xilinx Virtex-5 FPGAs [14].
System performance in a pure software design is accelerated by using dedicated hardware. In this study, we
used Xilinx embedded system platform utilizes to accelerate cryptographic hashing algorithms in embedded
domain. MicroBlaze processors of Xilinx embedded platform allow us to connect the custom hardware via some
peripheral connections. Our Keccak Coprocessor is connected via PLB and FSL buses to the MicroBlaze
processor. The experiments show that the speed-up provided by the proposed embedded SoC system is very high
compared to the pure software design without so much increase in the area.
With the proposed SoC design, the area needed is reduced and the efficiency is increased significantly on
FPGAs. This will also decrease the power usage on the system. With the approach presented, one can achieve
high enough performance and the security at the same time.
References
1.
2.
3.
4.
5.
6.
7.
Aumasson, J.P., Henzen, L., Meier, W., and Phan, R.: SHA-3 proposal BLAKE Version 1.3, 2010,
http://www.131002.net/blake/blake.pdf, accessed January 2011
Bertoni , G., Daemen, J., Peeters, M., and Assche, G. V.: KECCAK sponge function family main document, Version
2.1, http://keccak.noekeon.org/Keccak-main-2.1.pdf, accessed January 2011
Beuchat, J.L., Okamoto, E., and Yamazaki T.: Compact implementations of BLAKE-32 and BLAKE-64 on FPGA,
Proc. of the 2010 Int. Conf. on Field-Programmable Technology-FPT 2010, Beijing, China, December 2010, pp. 170177
Akin A., Aysu A., Ulusel O.C., and Savas E;. Efficient hardware implementations of high throughput SHA-3
candidates keccak, luffa and blue midnight wish for single- and multi-message hashing, In Proceedings of the 3rd
international conference on Security of information and networks (SIN '10). ACM, New York, NY, USA, 168-177.
Ferguson, N., Lucks, S., Schneier, B., Whiting, D., Bellare, M., Kohno, T., Callas, J., and Walker, J.: The Skein Hash
Function Family, Version 1.1, http://www.skein-hash.info/sites/default/files/skein1.1.pdf, accessed January 2011
Gauravaram, P., Knudsen, L.R., Matusiewicz, K., Mendel, F., Rechberger, C., Schlffer, M., and Thomsen, S.S.: Grstl
a SHA-3 candidate, http://www.groestl.info/Groestl.pdf, accessed January 2011
Gonzalez, I., and Gomez-Arribas, F.J.: Ciphering algorithms in MicroBlaze-based embedded systems, IEE Proc.,
417
8.
9.
10.
11.
12.
13.
14.
15.
16.
Comput. Digit. Tech., March 2006, Vol. 153, Issue 2, pp. 87-92
Hongjun Wu: The Hash Function JH, September 2009, http://icsd.i2r.a-star.edu.sg/staff/hongjun/jh/jh_round2.pdf,
accessed January 2011
Menezes, A.J., van Oorschot, P.C., and Vanstone, S.A.: Handbook of Applied Cryptography, CRC Press, 1996
NIST: Federal Register Notice, Vol. 72, No.14, 2007
http://csrc.nist.gov/groups/ST/hash/documents/FR_Notice_Jan07.pdf , accessed January 2011
SHA-3 zoo: http://ehash.iaik.tugraz.at/wiki/SHA-3_Hardware_Implementations, accessed January 2011
Todman, T.J., Constantinides, G.A., Wilton, S.J.E., Mencer, O., Luk, W., and Cheung, P.Y.K.: Reconfigurable
computing: architectures and design methods, IEE Proc., Comput. Digit. Tech., March 2005, Vol. 152, Issue 2, pp.
193207
Xilinx, WP245: Achieving Higher System Performance with the Virtex-5 Family of FPGAs,
http://www.origin.xilinx.com/support/documentation/white_papers/wp245.pdf,
accessed January 2011
Xilinx, UG190: Virtex-5 FPGA User Guide,
http://www.xilinx.com/support/documentation/user_guides/ug190.pdf, accessed January
2011
Xilinx, DS531. (2010) LogiCORE Processor Local Bus (PLB) v4.6, September 21.
Xilinx, DS449. (2010) LogiCORE Fast Simplex Link (FSL) v20 Bus, April 19.
Biographies
smail San was born in Denizli, Turkey, in 1983. I have double major in Electrical-Electronics Engineering and Avionics.
I received my BSc in Electrical-Electronics Engineering from Anadolu University in Eskiehir in Turkey in 2008 in five
years. I received my second BSc in Avionics from Anadolu University in Eskiehir in Turkey in 2008 in five years.
His research interests are in security of digital communication channels, cryptographic embedded systems, secure
hardware design and architectures for communication and computer systems.
PhD student Mr. San is a member of IEEE.
Nuray At was born in Eskisehir, Turkey. She received her B.Sc. degree in Electrical and Electronics Engineering from
Anadolu University; two M.Sc. degrees in Electrical and Computer Engineering from University of California Davis and
Georgia Institute of Technology, respectively; and Ph.D. degree in Electrical and Electronics Engineering from Anadolu
University.
Her research interests include information security, cryptographic engineering, communications systems, estimation, and
signal processing.
Dr. At is a member of IEEE.
418
Abstract. Wireless Sensor Networks (WSNs) draw more and more attention due to the increasing
number and variety of their applications. From system design point of view, the well known OSI
protocol model does not reflect the specifics and operation of WSN applications. Considering the
scarce energy and processing resources of WSNs, joint optimization and design of networking
layers, named cross-layer design stands as the most promising solution. Because of different types
of applications have different needs and features, it is important to define the adequate order of
demands. The aim of this paper is to provide an application oriented methodology to defining
optimization parameters in WSNs and apply it to a Cross Layer Design Framework.
1 Introduction
Wireless Sensor Networks (WSNs) draw more and more attention due to the increasing number and variety of
their applications. The nature of WSNs shows differences from Wired Networks and Ad Hoc Networks. From
system design point of view, the well known OSI protocol model does not reflect the specifics and operation of
WSN applications. Considering the scarce energy and processing resources of WSNs, joint optimization and
design of networking layers, i.e., cross-layer (CL) design stands as the most promising alternative to inefficient
traditional layered protocol architectures. Recently, CL design has emerged as a viable approach, however
presented solutions are quite varied and there is no common platform that allows their comparison and
evaluation from system design and implementation point of view. The need to define a suitable protocol design
model that fits their specific requirements and operation has become even more pressing.
The type of WSN application and its specific requirements play an important role in the definition of the system
characteristics and the protocol framework structure. Different types of applications have different needs and
features, so it is important to determine the adequate order of demands. Thus the most important optimization
criteria related to the specific application can be determined and such knowledge can greatly benefit the system
designer. The aim of this paper is to provide an application oriented methodology to defining optimization
parameters in WSNs and apply it to a Cross Layer Design Framework.
419
Despite the large variety, from engineering and conceptual point of view, the existing and envisaged
applications for WSNs, can be organized in groups based on their different needs and characteristic features. The
classifications of WSN applications can show the most important optimization criteria, besides that it can show
the relevance and order of importance of these criteria and their requirements.
The approach for defining optimization parameters that is suggested in this paper includes the following steps:
1. Definition of WSN application groups based both on their application characteristics and implementation
limitations. 2. Applying a multi-criteria decision method (MCDM) for clearing defining the relevance and order
of importance of the selected criteria. 3. Application of the selected parameters to the CL framework and final
definition of the specific design goals.
WSNs applications can be organized in groups based on their different requirements. We have utilized the
classification presented in [7] and have tried to extend it further by considering the possible optimizations that
these groups of WSN applications might be subjected to. In our work we are considering the following 3 larger
groups: Habitat&Environmental, Public&Industrial and Military&Health applications. In the following subsection these groups of applications is explained.
420
be used. The sensors might be expected to perform distributed collaboration and overcome obstacles such as
trees and rocks that block the line of sight of a sensor.
The major goal of our work is to define an application based approach and methodology for initial evaluation
of the major optimization criteria, functions and protocols to be designed for a given WSN, using a cross layer
design framework. Such a methodology aims to provide an easy tool for boundary estimation and consistency
checking that will be helpful for the system designer. WSN have become quite complex and often initially
defined optimization criteria, based on the given requirements might contradict. However this comes to light
only at the very final stage of protocol design and its is a long and timeconsuming job to start everything from
the very beginning. Thus providing a possibility to evaluate the interrelation between the requirements and the
best optimization criteria in a simple way for groups of applications can help the designer determine the main
conflicts from the very initial stage. In this section we describe the building blocks (milestones ) of the proposed
methodology.
The milestones are: 4.1 applications and requirements 4.2 AHP consistency analysis 4.3 Major Functions and
Optimization Criteria in the SE Cross Layer Framework. 4.4 Definition of the specific steps.
421
for Habitat&Environment, as explained above the lifetime of the network is the most important parameter for
optimizing the performance, while for Health&Military applications the reliability and delay are more important
than the prolonged lifetime. Thus there is prioritization in the optimization parameters which is related to the
defined application groups. Also this order of significance is unfortunately not always as straight forward as it
might look at first sight.
0,4
0,3
0,2
0,1
0
Energy Consumption
Scalibility
Latency
Reliability
422
role of the above suggested methods within that framework. The framework includes the following major blocks
(Fig. 2): main communication tasks, additional services, application interactions and hardware and operating
systems block.
Public&Industrial
Military&Health
Energy Consumption
Scalability
Latency
Reliability
Energy Consumption
Latency
Reliability
Scalability
Reliability
Latency
Energy Consumption
Scalability
Which
Protocols
in
the
Main
Communication Tasks Block will be most
affected
Wireless Channel Based Adaptive Power
Control
Medium Access Functions
Topology & Coverage Control
Wireless Channel Based Adaptive Power
Control
Multi-hop Routing Functions
Medium Access Functions
Error Detection & Control
Fault tolerance
Topology & Coverage Control
423
5 Conclusion
Although in general all sensor networks have common design goals, grouping the applications as mentioned
above points out to an important specific: for each group the order of significance of satisfying the major
requirements like reducing the energy consumption, providing low latency and high reliability along with
scalability etc. is different. And unfortunately it is not always as straight forward as it might look at first sight.
Sometimes there might even be contradictions in the optimization criteria to be met.
Accordingly we have investigated the optimization criteria for WSNs based on three major groups of
applications. Different from other grouping available in literature, which are primarily used for classification
purposes, our work aims at combining both the design and implementation limitations and requirements. Further
on the specific optimization parameters are selected and using a MCDM they are checked for consistency, before
finally limiting the set in the CLF and defining the specific goals. We have also addressed the functions that have
to be ensured by a CLF in each group. By combining optimization criteria and the functions to be met the
proposed CLF will provide greater efficiency at the system design stage. This will help to design and implement
WSNs applications in a shorter period of time.
References
1.
Akyildiz, I.F., Su, W., Sankarasubramaniam Y. and Cayirci E. (2002). A Survey on Sensor Networks. IEEE
Communications Magazine, Vol. 40, No.8, pp. 102-116, Aug. 2002.
2. Srivastava,V. and Motani, M. (2005). Cross-layer design: A survey and the road ahead. IEEE Communications
Magazine, 43(12): 112--119.
3. Marrn,P. J., Minder, D.l, Lachenmann, A. and Rothermel, K. (2005). TinyCubus: An Adaptive Cross-Layer
Framework for Sensor Networks. it - Information Technology, vol. 47(2), pp. 87-97.
4. Melodia,T., Vuran, M. C. and Pompili, D. (2006). The State-of-the-art in Cross-layer Design for Wireless Sensor
Networks. Springer Lecture Notes in Computer Science (LNCS), vol. 3883, June 2006.
5. Bouabdallah-Othman, F., Bouabdallah, N. and Boutaba,R. (2009). Cross-Layer Design for Energy Conservation in
Wireless Sensor Networks. IEEE ICC 2009, Dresden, Germany, June 2009.
6. Liu,S., Bai,Y., Sha,M., Deng,Q. and Qian,D. (2008). CLEEP: A Novel Cross-Layer Energy-Efficient Protocol for
Wireless Sensor Networks. IEEE International Conference on Wireless Communications, Networking and Mobile
Computing (WiCOM), IEEE Communications Society Press, October 12 - 14, 2008, Dalian, China.
7. Yick,J., Mukherjee, B. and Ghosal, D. (2008). Wireless sensor network survey. Computer Networks, 52(12): 22922330.
8. Karaca, O. and Sokullu, R. (2009). Comparative Study of Cross Layer Frameworks for Wireless Sensor Networks.
Proc. of WirelessVitae 2009, Aalborg, Denmark, May 17-20.
9. Shakkottai,S., Rappaport, T. S. and Karlsson, P. C. (2003). Cross-layer design for wireless networks. IEEE
Commun mag., Vol. 41, No. 10, 2003, pp. 74-80
10. Zhao, N. and Sun, L. (2007). Research on Cross-Layer Frameworks Design in Wireless Sensor Networks. ICWMC
'07: Proceedings of the Third International Conference on Wireless and Mobile Communications, March 2007.
11. Karaca, L. O., and Sokullu, R. (2010). Using AHP for Multi-Criteria Decision in WSN, The Fourth International
Student Conference on Advanced Science and Technology- ICAST-2010, 25-26 May,2010.
12. Sokullu,R. and Karaca,O. (2009). Simple and Efficient Cross Layer Framework Concept for Wireless Sensor
Networks. The 12th International Symposium on Wireless Personal Multimedia Communications, September 7
10, 2009, Sendai, Japan.
424
Biographies
Ozlem Karaca received her Bachelors degree in Electrical and Electronics Engineering from Dokuz Eylul University,
Turkey. After graduation, she became a teaching assistant at the department of Technical Programs at Izmir Vocational
School. She obtained M.S. degree in Electrical and Electronics Engineering from Ege University. She received her Ph.D.
degree from Ege University in 2010. She participated in CRUISE Project. She is currently a lecturer at Dokuz Eylul
Univesity. Her current research interests include wireless sensor networks, cross layer design and cross layer frameworks.
Radosveta Sokullu received her MSc and PhD degree in Telecommunications Engineering from the Sofia Technical
University in 1985 and 1992 respectively. She is currently an Assistant Prof. In the Department of EEE and Head of the
Chair of Telecommunications, Ege University, Izmir, Turkey. She was invited as a visiting professor for the GRASIUS
Program at the Graduate School of Engineering, Kumamoto University in 2008. Her current research interests include
wireless networks and wireless sensor networks, performance evaluation and security issues.
425
Abstract.There is an especial remark for Evolutionary Programming (EP) among EAs. EP was the first introduced algorithm
in EA. In EP family, there are some well-known algorithms such as: Classical EP (CEP), Fast EP (FEP) and Levy distributed
EP (LEP). Levy distribution function will be used in the proposed method in mutation operator as it has long tails, thus giving
offspring big step size to search map quicker which is necessary for some cost functions. Novelty of the proposed method
against former variants like LEP, is defining new operator which can improve performance of LEP. In this paper, our main
objective is to extract and benefit from some useful information which has been ignored by other variants. Location of each
parent has some information which is captured from previous iteration. In short, new method adds new factor to equation
which produces offspring. This factor pulls offspring to gathered point of parents. A set of benchmark cost functions has been
used to compare the results of the proposed method with some other known algorithms.
Keywords: Evolutionary Programming, Levy distribution function, gathered point, weighted mean operator
1 Introduction
Optimization has long been one of the interesting fields of research. It has been linked with many investigations.
Many real-world applications are confronted with determining the values of some unknown parameters or
coefficients, which must be adapted in optimal ways. Defining a cost function is a way to transform this problem
to a minimization (maximization) problem. After choosing a proper cost function, the main question is how to
minimize this cost function. Evolutionary Algorithms (EA) are first candidates for hard optimization tasks. They
can deal with many kinds of problems and cost functions such as: multimodal, discrete, without derivative
function, with bounders on variables or some subjection (restriction on variables) and high dimensional. These
capabilities have made them suitable for many applications.
In the case of evolutionary computation, there are four historical paradigms that have served as the basis for
much of the activity of the field: genetic algorithms (Holland, 1975) [1], genetic programming (Koza, 1992,
1994) [2], evolutionary strategies (Recheuberg, 1973) [3], and evolutionary programming (Forgel et al., 1966)
[4]. The basic differences between the paradigms lie in the nature of the representation schemes, the reproduction
operators and selection methods [5].
These methods have drawn much attention to the research community in conjunction with the parallel and/or
distributed computations. EP, in particular, was studied initially as a method for generating artificial intelligence
[6,7].
Evolutionary programming is stable kind of evolutionary algorithm. It has many advantages in optimization
multimodal problems. Three algorithms, Classical EP (CEP), Fast EP (FEP) and Levy distributed EP (LEP) are
most famous in EP.
This paper proposes a novel EP algorithm, which adds a new factor promoting the performance of EP, such as:
426
speed, accuracy and stability of final answers. 12 test functions are gathered from several references for a better
comparison of the results and to show generality of the method's performance. One advantage of the proposed
method is its capability in working with other EP variants. So this method may promote the performance of
almost all methods introduced above. Most works having been performed for improving EP are in the mutation
operator. Most of the researchers have tried to introduce a better distribution function in place of the Gaussian
distribution function. But in this study an attempt has been made to introduce a new approach that contains all
previous work and is capable of matching with almost all previous researches.
The organization of this paper is as follows: section 2 explains the EP variants which have been used for
comparison in this paper; section 3 introduces the basic idea behind the Inertia Weighted Evolutionary
Programming (IWEP); section 4 introduces some detail of the test functions, shows the simulation results and
compares the new method with CEP, FEP and LEP and section 5 presents the conclusion and summarizes the
simulation results.
(1)
(2)
where xi ( j ), xi ( j ), i ( j ) and i ( j ) denote the j th component of the vectors xi , xi,i and i , respectively.
N (0,1) denotes a normally distributed one-dimensional random number with mean zero and standard deviation
one. N j (0,1) indicates that the random number is generated anew for each value of
commonly set to
( 2 n ) 1 and (2 n )
427
R N (0,1)
parent
Xi
origion
IWEPs offspring
region
1-a
mean point
global minimum
Fig. 1: location of offspring which is produced by CEP and MCEP. Offspring is somewhere inside of the dashed circles.
Levy distribution function will be used in the proposed method in mutation operator in equation (4) as it has
long tails, thus giving offspring big step size to search map quicker which is necessary for some cost functions.
428
4. Results:
12 well known benchmark cost function have been selected for comparing the algorithms. Descriptions of the
cost functions used in this paper are mentioned in Tables I. Some main properties of the functions will be
explained in this part but a more detailed description of these cost functions is available in [9].
The cost functions are parted to 3 subgroups: Unimodal- high-dimensional, multimodal-high-dimensional and
low-dimensional.
Functions f1 f4 are high dimensional-unimodal problems, which have only one global minimum that is also
their only local minimum too.
Functions f5 f8 are high dimensional-multimodal functions where the number of local minima increases
exponentially with the problem dimension. They appear to be the most difficult class of problems for many
optimization algorithms of which CEP has slow convergence on these functions [9].
Functions f9 f12 are low-dimensional functions, which have only a few local minima but a number are hard
for minimization by evolutionary algorithms.
Table 1. The 12 benchmark functions used in our experimental study, second column introduces name of
functions
f1 ( x)
Sphere Model
f 2 ( x)
f 3 ( x)
f 4 ( x)
f 5 ( x)
i 1
xi2
xi i1 xi
n
i 1
n
( j 1 x j )
i
i 1
max i xi ,1 i n
i 1
xi sin
x
i
f 6 ( x)
f 7 ( x)
Ackley's function
f 8 ( x)
f 9 ( x)
f10 ( x)
1
4 x12 2.1x14 x16 x1 x2 4 x22 4 x 42
3
f11x
Hartman's Family2
f12 ( x)
Shekel's Family3
i 1
20 exp( 0.2
1 n 2
1 n
xi ) exp( n i1 cos 2xi ) 20 exp(1)
n i 1
n
n
x
1
xi2 i1 cos( ii ) 1
4000 i 1
25
1
1
j 1
]
2
500
j i 1 ( xi aij ) 6
i1[( x ai )( x ai )T ci ]
10
The ranges of the variables and dimensions of the cost functions are chosen like references [9]. For highdimension functions, the number of the variable is D=30 (if it is not limited by the cost function).
All parameters of algorithms mentioned in Table II, is similar for the algorithms in all simulation results.
In remain the best algorithm among IWEP, CEP, FEP and LEP is going to be selected on 12 test functions. For
avoiding any concurrence in the results, all of the algorithms were run 50 times and the averages of the obtained
results are presented in Table III.
All of the algorithms are run until a pre-specified generation is reached. The number of generations is shown in
the second column of Table III. In this table, the best result
429
10
Population size
Range bound of variables
Number of repetition
Number of generation
Value of coefficients a in equation 3
100
Mentioned in last column of table V
50
Mentioned in second column of tables V
Normal random number N (0,1)
f1
f2
f3
f4
f5
f6
f7
f8
f9
f10
f11
f12
Number of
generation
1500
2000
5000
5000
3000
5000
1500
2000
100
100
200
100
CEP
FEP
LEP
IWEP
FMIN
Range bound
155.67
0.418
1161
0.607
-8373
44.275
8.43
2.373
3.432
-1.031
-3.322
-8.578
25.194
0.596
4438
34.38
-12542
1.864
3.773
0.971
1.047
-1.031
-3.201
-8.311
26.685
0.707
650
1.07
-11167
13.435
5.36
0.694
1.458
-1.031
-3.164
-9.145
3.50e-8
2.56e-6
6.72
0.016
-5748
3.55e-15
0.0016
0.0049
2.884
-1.031
-3.322
-10.52
30
30
30
30
30
30
30
30
2
2
6
4
0
0
0
0
-12569.5
0
0
0
1
-1.031
-3.32
-10.5
[-100,100]n
[-10,10]n
[-100,100]n
[-100.100]n
[-500,500]n
[-5.12,5.12]n
[-32,32]n
[-600,600]n
[-65.53,65.53]n
[-5,5]n
[0,1]n
[0,10]n
obtained for each cost function is highlighted. These results show that the new proposed method in most cases
can approach the global minimum better than other mentioned methods. All of the parameters have been
considered to be the same for three methods.
As Table III shows accuracy of the methods to find the global minimums, in order to compare the speed of
methods, the 12 figures have been drawn. All results of cost functions specified in Table III are drawn in Fig 213.
All methods have begun from the same initial point (first generation).
Figures 2-6 are for unimodal cost functions. IWEP has the best answer for all functions in this group. These
results were expected, since IWEP has a factor pulling offspring toward the gathered point (mean value) of
parents. Unimodal cost functions have only one local minimum which is also their only global minimum. Thus,
the mean value can steer offspring below the hole in a short time.
Figures 6-10 are results obtained for multimodal cost functions. These functions make it difficult for EP
families to find global minimum as they have many local and several global minimums. IWEP has the best
answer in this group. Good results in this group are very important as there are few methods that have acceptable
performance.
The third group's results are drawn in Figures 11-14.
430
5. Conclusion:
The results on low dimensional, unimodal, and multimodal cost functions show accuracy, speed of convergence
and ability of escaping from local minimums, respectively. Results show the proposed algorithm is more accurate
and fast in finding the value and location of the global minimum in all three groups of the cost functions. The
results clear benefits of using gathered point in determining the location of offspring. Consequently, the method
benefits more information which have been caught by parents in breeding offspring have more accuracy and
speed in convergence to global minimum. One advantages of the proposed method is its capability to adapt with
other variants of EP.
431
f1
10
F2
15
F3
10
10
FEP
LEP
CEP
IWEP
10
10
10
FEP
LEP
CEP
IWEP
10
10
10
-2
10
Cost
Cost
Cost
10
10
10
10
-4
10
-8
10
1500
F4
500
10
2000
2000
3000
number of iteration
4000
5000
F6
10
-10
-5
10
-10
-1
10
-2
1000
10
10
10
FEP
LEP
CEP
IWEP
Cost
10
1500
-10
FEP
LEP
CEP
IWEP
10
1000
number of iteration
F5
10
Cost
10
-10
10
500
1000
number of iteration
-5
10
FEP
LEP
CEP
IWEP
Cost
-6
10
1000
2000
3000
number of iteration
4000
-10
5000
F7
-15
500
1000
1500
2000
number of iteration
2500
10
3000
F8
4000
5000
F9
10
10
2000
3000
number of iteration
FEP
LEP
CEP
IWEP
10
1000
10
10
FEP
LEP
CEP
IWEP
FEP
LEP
CEP
IWEP
10
0
Cost
Cost
Cost
10
-2
10
10
-1
10
-4
10
-2
10
-3
10
FEP
LEP
CEP
IWEP
-6
10
-8
500
1000
number of iteration
10
1500
500
1000
number of iteration
1500
10
2000
F11
F10
40
60
number of iteration
80
100
F12
-10
-10
FEP
LEP
CEP
IWEP
20
0.3
-0.4
-10
FEP
LEP
CEP
IWEP
FEP
LEP
CEP
IWEP
-0.3
-10
-10
Cost
Cost
Cost
0.4
-0.2
-10
-10
-0.1
-10
0.5
-10
-10
20
40
60
number of iteration
80
100
50
100
number of iteration
150
200
-10
20
432
40
60
number of iteration
80
100
References
1.
2.
J.H. Holland, Adaptation in Natural and Artificial Systems, University of Michigan Press, Ann Harbor, 1975.
Koza JR. Genetic Programming: on the Programming of Computers by Means of Natural Selection, MIT Press,
1992.
3. H.-P. Schwefel, Numerical Optimization of Computer Models, Wiley, Great Britain, 1981.
4. L.J. Fogel, Artificial Intelligence through Simulated Evolution, Wiley, New York, 1966.
5. S. N. Sivanandam, S .N. Deepa, Introduction to Genetic Algorithms, Springer, pp. 2-5, 2008.
6. C Y Lee, and X Yao, Evolutionary Programming Using Mutations based on the Levy Probability Distribution, IEEE
Trans. on Evolutionary Programming, Vol. 8, No. 1, February 2004.
7. L. J. Fogel, A. J. Owens, and M. J.Walsh, Artificial Intelligence through Simulated Evolution, New York, Wiley,
1966.
8. Evolutionary computation with extinction: Experiments and analysis. G. B. Fogel, G. W. Greenwood and K.
Chellapilla. Piscataway, US : Proc. of 2000 Congress on Evolutionary Computation, 2000.
9. X. Yao, Y. Liu, and G. Lin, Evolutionary Programming Made Faster, IEEE Trans. on Evolutionary Computation,
vol. 3,no. 2, 1999.
10. Li. Xiaoan, K. Jichang, An Improved Fast Evolutionary Programming, IEEE Tran. of Northwestern Ploy technical
University. 2001. Vol. 19 No.1
Biographies
Yousef Alipouri was born in Tabriz, Iran, 1987. He received the B.Sc. degree in control engineering from University of
Tabriz, Tabriz, Iran, in 2009. He is currently pursuing his study towards the M.Sc. degree at Iran University of Science and
Technology. His research interest focuses on control system theory and artificial intelligence.
Javad Poshtan received his B.SC., M.SC and PHD degrees in Electrical Engineering from Tehran University, Tehran, Iran,
in 1987, Sharif University of Technology, Tehran, Iran, 1991, and University of New Brunswick, Canada, in 1997,
respectively. Since 1997 he has been with the Department of Electrical Engineering at Iran University of Science and
Technology. He is involved in academic and research activity in areas such as control system theory, system identification,
and estimation theory.
433
Abstract: Following the discovery of the new fundamental circuit element called memristor, memcapacitive and
meminductive systems similar to memristive systems have also been described. It has been speculated that two new circuit
elements must also exist, namely the memcapacitor and the meminductor. Memcapacitor, which is a special case of
memcapacitive systems, is a circuit element characterized with a certain type of memory and its (mem)capacitance is fluxdependent. Two-capacitor problem, which is the behavior of a circuit consisting of two capacitors, has been investigated
extensively in the literature. However, there is no such analysis in the current literature for a capacitor and a memcapacitor.
In this work, a circuit consisting of a memcapacitor and a linear time-invariant capacitor is analyzed. The time-domain
solution of the system is given. An ongoing transient has been found in the circuit. It is shown that the memcapacitorcapacitor circuits may have interesting properties and behave rather different from linear time-invariant capacitor circuits.
1. Introduction
The claim by Dr. Chua in 1971 that a fundamental circuit element called memristor must have existed paved
the way for Him and Kang to come up with yet another claim in 1976 that systems called memristive systems
with similar properties to memristors exist and a memristive system should have a zero-crossing current-voltage
hysteresis curve [1-2]. Almost 37 years later, an HP research team has announced that they have found the
missing circuit element [3]. A review on the brand new circuit element named memristor can be found in [4]. In
the year 2009, Ventra, Pershin and Chua has written a paper called Circuit elements with memory and claimed
that there must be other circuit elements besides memristors, namely inductors and capacitors with memories,
those with similar properties to memristors and they called these elements the meminductor and the
memcapacitor [5]. The equations describing a memcapacitive system and charge-voltage hysteresis curve are
also given in [5]. A commonly-used variable capacitor is actually a memcapacitor according to the equations
describing a memcapacitive system. Although a meminductor has not been realized in nano dimensions yet,
research on ionic and solid-state memcapacitors exist in literature [6-7]. Some recent papers on memcapacitor
are about its modeling by menas of some simulation programs and its usage in chaos circuits [8-12]. Another
interesting review on the memcapacitors can be found in [13].
Parallel connection of two capacitors is found in all circuit theory books. However, to the best of our
knowledge, a memcapacitor connected in parallel with a capacitor has not been studied in the literature yet. In
this paper, to the best of our knowledge, for the first time in literature, a circuit consisting of a memcapacitor and
a capacitor without a supply is studied. This paper is arranged in the following way. In the second section, a
434
memcapacitive system and the Flux-dependent memcapacitive system, the Memcapacitor, are shortly explained.
In the third section, the memcapacitor-capacitor circuit and its analytical model are given. The voltage, current
and flux of memcapacitor are given. The results are discussed in the Conclusion section.
q t C X ,V .V t
(1)
where
X is a space-state equation.
V is the memcapacitive system voltage.
C is the memcapacitive system memcapacitance.
A memcapacitor is a special case of a memcapacitive system. Its memcapacitance is a function of the memcapacitor flux,
which is desribed as the integration of memcapacitor voltage with respect to time. Memcapacitor terminal equation then reads
q(t ) C .V t
(2) or
(3)
Where
V (t)
q (t)
435
In this study a linear-flux dependent memcapacitance function is used. Memcapacitor memcapacitance can be
written as a linear function similar to the HP memristor memristance. Then, its memcapacitance is given as
C Co K t
(4)
t V t .dt o
(5)
436
x 10
-3
Memcapacitance (F)
1.5
0.5
0.1
0.2
0.3
Memcapacitor Flux (Wb)
0.4
0.5
q
d
VMC
Cp
dt
(6)
iC i MC 0 Cp
dV dq .
dt dt
(7)
Cp
dV d
(C( ).V) 0
dt dt
(8)
Cp
dV d
Co K V 0
dt dt
(9)
d
Co Cp K V 0
dt
(10)
A Co Cp K V
(11)
A Co Cp K (0).V(0)
(12)
C Cp K (0).V(0) (13)
A
o
Co Cp K
o
p
The memcapacitor current is calculated by taking the derivative of the memcapacitor charge:
i(t )
dq(t )
,
dt
i(t) Cp
(14)
Cp .AKV
dV
(15) C dV d C K V 0
p
o
dt dt
dt Co Cp K 2
(16)
q(t) A CpV
(17)
437
d
dt
(18)
At B C o C p K2 / 2
(19)
B C o C p (0) K (0) 2 / 2 .
(20)
rearranging (12)
0 At B C o C p K2 / 2
(21)
C o C p
C p 2At BK
2
(22)
dt
(23)
C p 2At BK
2
The system is simulated for the parameters given in Table 1. Memcapacitance, Voltage, Current, Flux and
Charge of the Memcapacitor are shown in Figures 6-10. The voltage of the memcapacitor increases during the
transient. The charge of the memcapacitor-capacitor system is conserved. The memcapacitor flux and therefore
the memcapacitance varies as a function of time as well.
Table 1: The circuit Parameters.
Parameter
Value
0.5 mF/Wb
K
Cp
1 mF
CO
1.5 mF
VC (0)
5V
(0)
0.25 Wb
438
0.03
0.025
0.02
0.015
0.01
0.005
0.2
0.4
0.6
0.8
Time(s)
0.0115
0.011
0.0105
0.01
0.0095
0.009
0.0085
0.008
0.2
0.4
0.6
0.8
Time(s)
x 10
-5
0.8
0.6
0.4
0.2
0.2
0.4
0.6
0.8
Time(s)
439
60
50
40
30
20
10
0.2
0.4
0.6
0.8
Time(s)
0.2
0.4
0.6
0.8
Time(s)
4. Conclusion
In this paper, a circuit consisting of a capacitor and a memcapacitor with a linear flux-dependent
memcapacitance and without an external power supply is inspected. It is shown that a memcapacitor and a
capacitor are not in steady-state even though they have the same voltage in sharp contrast to two conventional
capacitors in steady-state connected in parallel. The time-domain solution of the emecapacitor-capacitor system
connected in parallel is given. It is found that the memcapacitance of the memcapacitor varies even when the
voltage of the memcapacitor and capacitor happens to be the same. As time progresses, the system voltage
decreases, whereas the memcapacitor charge increases due to the fact that total charge and energy is conserved
in the system. However, transient of the circuit is ever-continuing if there is no saturation mechanism of the
memcapacitor. Considering the case that a saturation mechanism does exist, the transient ends in a finite time.
The Memcapacitor-capacitor problem should be investigated further for a better understanding of the circuit
element memcapacitor so that it can be used more effectively. Also, new circuit elements such as memcapacitor
and their combinations with other circuit elements should be studied in detail so that their behavior pattern can
be understood more and be used more effectively in practice.
440
References
[1] L. O. Chua, Memristor - The Missing Circuit Element, IEEE Trans. Circuit Theory, vol. 18, pp. 507-519, 1971.
[2] L. O. Chua and S. M. Kang, Memristive devices and systems, Proc.IEEE, vol. 64, pp. 209-223, 1976.
[3] D. B. Strukov, G. S. Snider, D. R. Stewart, and R. S. Williams, The missing memristor found, Nature (London), vol.
453, pp. 80-83, 2008.
[4] O. Kavehei, A. Iqbal, Y. S. Kim, K. Eshraghian, S. F. Al-Sarawi and D. Abbott, The fourth element: characteristics,
modelling and electromagnetic theory of the memristor, Proc. R. Soc. A 466, pp. 2175-2202, 2010.
[5] M. Di Ventra, Yu. V. Pershin and L. O. Chua Circuit Elements With Memory: Memristors, memcapacitors and
meminductors, Proc. IEEE, vol. 97, pp. 17171724, 2009.
[6] J Martinez, M Di Ventra. Pershin, Yu. V., Solid-state memcapacitive system with negative and diverging capacitance,
Physical Review B, 81, 195430, 2010.
[7] Matt Krems, Yuriy V. Pershin and Massimiliano Di Ventra, Ionic Memcapacitive Effects in Nanopores Nano letters, ,
10, pp 26742678, ACS Publications, 2010.
[8] D. Biolek, Z. Biolek and V. Biolkova SPICE Modeling of Memristive, Memcapacitative and Meminductive Systems
[9] Blaise Mouttet A Memadmittance Systems Model for Thin Film Memory Materials arXiv:1003.2842v1 [condmat.mes-hall].
[10] Pershin, Y.V., Di Ventra, M., Memristive circuits simulate memcapacitors and meminductors Electronics letters,
2010.
[11] Zhiheng Hu, Yingxiang Li, Li Jia, and Juebang Yu, Chaos in a charge-controlled memcapacitor circuit, International
Conference on Communications, Circuits and Systems (ICCCAS), Chengdu, pages: 828831, July 2010.
[12] Zhiheng Hu, Yingxiang Li, Li Jia, and Juebang Yu, ,Chaotic oscillator based on voltage-controlled memcapacitor,
International Conference on Communications, Circuits and Systems (ICCCAS),
pages: 824-827, July 2010.
[13] Yuriy V. Pershin and Massimiliano Di Ventra,Memory effects in complex materials and nanoscale systems, Advances
in Physics, v. 60, pages 145-227, 2011.
441
Abstract. In this paper daynamic model of quadrotor helicopter is presented. Because of parametric
uncertainties and nonparametric uncertainties in the model and also existence of external disturbances, an
advanced control strategy is necessary to obtain good performance in various conditions. An adaptive
feedback linearizing controller is proposed to slove the tarjectroy tracking problem. A robust adaption law is
used to update the controller parameters. Simulation results illustrate the good tracking performance of
proposed control scheme.
Keywords: Adaptive control, feedback linearization, tracking control, quarotor
1 Introduction
Nowadays unmanned aerial vehicles are good solution to search and rescue operation, surveillance and
civilian applications. Hence automatic control of UAVs has attracted the attention of many researchers. Due to
the mechanical simplicity, high payload and high maneuverability, quadrotor configuration is highlighted in last
few years. Quadrotor is underactuated nonlinear MIMO system with strong coupling between states and
therefore it is a good test bench for implantation of different control techniques.
Many researchers have solved trajectory tracking problem for UAVs in different ways. Feedback
linearization method in combined with other techniques, is used to solve the tracking problem in [1, 2, 4, 6, 7]. A
vision-based stabilization and output tracking control method is proposed in [1]. In [4], the stabilizing control
problem for a quadrotor is solved by designing a nonlinear control system, which decomposes system into an
inner attitude and an outer velocity control loop. In [6] Dynamic inversion is applied to the inner loops, which
yields internal dynamics that are not necessarily stable. A robust control approach is used to stabilize the internal
dynamics. Two types of nonlinear controllers for an autonomous Quadrotor helicopter are presented in [7]. A
feedback linearization controller involves high-order derivative terms and the second type involves a new
approach to an adaptive sliding mode controller. An integral predictive and nonlinear robust control strategy is
used to solve the path following problem for a quadrotor in [10]. Backstepping approach is used to control the
quadrotor in [3, 5, 9]. Backstepping technique is combined with neural networks in [3]. In [5] Backstepping
approach is used for Controlling a UAV while Aerodynamic forces and moments are estimated by neural
networks. A nonlinear adaptive controller is proposed, which is using backstepping technique mixed with neural
networks, is presented in [9]. An L1 adaptive output feedback controller design process is presented in [8]. In
[11] an Output Feedback Control of a Quadrotor UAV Using Neural Networks is performed.
In this paper tracking control problem is solved through inner and outer control loop, which are control the
attitude and altitude of vehicle, and longitudinal and latitudinal movements, respectively. Feedback linearizing
442
controller is applied to inner control loop. Parameters of system, which are used by controller, are estimated by
robust parameter identifier. Outer control loop generated instantaneous desired angle for inner controller
according to the error of position.
The rest of this paper is organized as follows: in section 2 dynamic model of quadrotor is presented, proposed
control scheme is exposed in section 3, section 4 allocated to the simulation results. The major conclusions of the
paper are drawn in section5.
Pitch
Roll
Rear
Yaw
X
Y
Z
T2
z
Right
y
443
U1
x (S S C S C ) m
Y
(C S S S C ) U1
U1
Z g (C C )
m
a b c U
1
1
1 2
a 2 b 2 c 2 U 3
a 3 b 3 U 4
)2(
U1 b(12 22 32 24 )
2
2
U 2 lb ( 2 4 )
2
2
U 3 lb (1 3 )
2
2
2
2
U 4 b ( 1 2 3 4 )
1
2
3
4
(3)
I yy Izz
I xx
, b1
J tp
I xx
, c1
l
I xx
, a2
J tp
I xx I yy
J tp
Izz I xx
1
1
, b2 , c2 , a3
, b3 , c3 , Where Ixx, Iyy and Izz are
I yy
I yy
I yy
I zz
I zz
I zz
quadrotor moment of inertia around x, y and z axis of body fixed frame respectively. l is distance between center
of rotor and center of vehicle. Jtpis inertia of rotor.
3 Controller design
The control strategy has been split into two stages. Attitude and altitude are controlled in inner loop while
longitudinal and latitudinal motions are controlled in outer loop. Initially we select output vector that makes the
dynamic inverse easy to find. To stabilize the zero dynamics, we select some desired state variable such that
reminded state variable is bounded [6].Since the system has some unknown parameters such as variable payload
and moment of inertia tensor, there are parametric uncertainties in model; an adaptive control law is used to
update the control parameters. A leakage modification is applied to modify the integral action of adaptation law,
to avoid the parameters drift due to external disturbances and nonparametric uncertainties [12].According to
difference between reference trajectories of X and Y and path travelled by UAV the control signals are generated
by outer loop. Outputs of the outer loop are given to inner loop as desired angle. Reference trajectory is produced
so that the tracking objectives are satisfied, while the velocities and the accelerations at the end point of
trajectory are zero. By using this trajectory, longitudinal and latitudinal variable become bounded [5]. The
control scheme is shown in Fig.2
Z
,,
Aero dynamic forces
and moments
Adaption law
Estimated parameter U1,2,3,4
Zd
Trajectory
generator
Xd
Yd
Outer
controller
d
d
Feedback
linearization
controller
U1
transitional
XY
U2
U3
rotational
U4
quadrotor
444
(4)
i 1
As we seen derivative of output is not related to the control inputs. By differentiating again, it yields
y L2f h j L f L h j Lr L f h j L2r h j Lr L h j L Lr h j L2 h j
4
i 1
i 1
i 1
(7)
(8)
Lg L f h 1 ... Lg L f h 1
4
1
.
.
.
D( x)
.
.
.
L
L
h
...
L
L
h
4
4
g 1 f
g 4 f
Assuming
(9)
the system is input-output linearizable [12]. And the following nonlinear feedback:
(10)
y3 3 L L f h 3
y4 4 L L f h 4
(11)
i yri 1i ( y ri y i ) 2i ( y ri y i )
(12)
445
(13)
W T E1
Where
is a regression matrix,
is a vector of filtered error signal, is a symmetric and positive definite
matrix as adaption gain and is a constant positive definite diagonal matrix. By using the estimated parameter
can be rewritten as follows:
(15)
u D 1 ( x)( L2 L L h)
a
Where
and
are estimation of
and
respectively.
(16)
(17)
U1
Y ( - S )
m
and
(18)
(19)
(20)
(21)
di
3x
4x
5x
a 3 4t
f
4x
a5 x 6 12t f
5t
20t
2
f
2
f
dx
df
di
(24)
t 3f
The reference trajectories for the y and z directions are given by similar relationship. [5]
446
4 Simulation result
In this section, some computer simulations are performed to show the tracking performance and robustness
of the proposed controllers in presence of parametric and nonparametric uncertainties. The UAV parameters are
listed as follows: m = 2 kg, Ixx = 7.110-3 kgm2, Iyy=7.110-3 kgm2, Izz=14.210-3 kgm2, l = 0.24 m, Jtp= 1.42102
Nms2, g = 9.81 m/s2.
Feedback Linearization coefficients and Controller Coefficients for outer loop are chosen as follow:
Adaption law gain and
1 2 1 2 10, 1 2 5 z1 z 2 1, C x1 C x 2 C y1 C y 2 2
This simulation shows the performance of adaptive tracking controller in presence of parametric uncertainties,
while external sinusoidal disturbance is applied to system in all directions. We want to quadrotor moves from
initial point (0,0,0) with initial orientations (0,0,0) to final position (0.5m,0.5m,0.5m) with final orientations
(0,0,0.5rad) in 2 seconds. Fig.3 shows the six state variables in this scenario.
Fig. 3. Position and attitude of quadrotor using the adaptive feedback linearization
controller
As shown in Fig.3 output variables track the trajectory with negligible error and reach to final desired value at
the specified time. The trajectory is produced by outer loop such that the initial and final values of velocity and
acceleration along X and Y axis become zero. As obviously seen in fig.3 by using adaptive controller in inner
loop system track the commands, which is generated by outer loop, very well in presence of external
disturbances. Figure 4 shows estimated parameters of system according to time.
447
5 Conclusions
In this paper, an adaptive feedback linearizing controller is designed to tracking control of quadrotor helicopter.
The control strategy has been split into inner control loop, which is control the attitude and the altitude, and outer
control loop, which is control the longitudinal and the latitudinal motions. An adaptive control law is used to
update the control parameters to cope with parametric uncertainties in model. The simulation results shown at
the end demonstrates the effectiveness of adaptive feedback linearization controller in presence external
disturbances and model uncertainties in different conditions. Future work will involve the implementation of this
control strategy in a real quadrotor helicopter.
References
1.
Altug E.; Ostrowski JP.; Taylor CJ. (2005). Control of a quadrotor helicopter using dual camera visual feedback,
International journal of robotics research, vol.24:329-341.
2. Mian AA.; Daobo W.(2008). Modeling and backstepping-based nonlinear control strategy for a 6 DOF quadrotor
helicopter, Chinese journal of aeronautics, vol.21: 261-268.
3. Madani T.; Benallegue A.(2008). Adaptive Control via Backstepping Technique and Neural Networks of a
Quadrotor Helicopter, In Proceedings of the IFAC international conference., Seoul, Korea.
4. Voos H. (2009). Control Systems Design for a Quadrotor UAV, At-Automatisierungstechnik (57): 423-431.
5. Das A.; Lewis F.; Subbaro K. (2009). Backstepping Approach for Controlling a Quadrotor Using Lagrange Form
Dynamics, J. Intell. Robot Syst(56):127151.
6. Das A.; Lewis F.; Subbaro K.(2009). Dynamic inversion with zero-dynamics stabilisation for Quadrotor control.
IET Control Theory Application, Vol.3: 303314.
7. Lee D.; Kim H.; Sastry S.(2009). Feedback Linearization vs. Adaptive Sliding Mode Control for a Quadrotor
Helicopter, J. Control, Automat., and Systems 7(3): 419-428.
8. Michini B.; How J. (2009). L1 Adaptive Control for Indoor Autonomous Vehicles: Design Process and Flight
Testing, In Proceedings of the AIAA Guidance, Navigation, and Control Conference August, Chicago, Illinois.
9. Huang M.; Xian B.; Diao C.; Yang K.; Feng Y. (2010). Adaptive Tracking Control of Underactuated Quadrotor
Unmanned Aerial Vehicles via Backstepping, In Proceedings of the. AACC. Marriott Waterfront, Baltimore, MD,
USA.
10. Raffo G.; Ortega M.; Rubio F. (2010), An integral predictive/nonlinear control structure for a quadrotor helicopter,
Automatica(46): 29-39.
11. Dierks T.; Jagannathan S. (2010). Output Feedback Control of a Quadrotor UAV Using Neural Networks, IEEE
TRANS. Neural Networks 21(1):50-66.
12. Shojaei K.; Shahri AM.; Tarakameh A. (2011). Adaptive feedback linearizing control of nonholonomic wheeled
mobile robots in presence of parametric and nonparametric uncertainties, Journal of Robotics and ComputerIntegrated Manufacturing (27):194-204.
Biographies
Zahra Boroujeni born in Tehran, Iran in 1989. She received a B.S. degree degree in Electrical engineering (control)
from Iran university science and technology, Tehran (IUST) in 2010. Now she is student of M.Sc. Control
Engineering in IUST.
Her research interests are: Nonlinear control; System identification; Adaptive control; Robotic; Mechatronic.
Mostafa Mohammadi born in Sardasht, Iran in 1986. He received a B.S. degree in Electrical engineering (control) in
2008. Now he is student of M.Sc. Control Engineering in IUST.
His research interests are: Intelligent and adaptive control, System identification, Adaptive control, robotic, aerial
robots
448
Sahra Boroujeni born in Boroujen, Iran in 1984. She received a B.S. degree in Mechanical engineering from
Shahrekord University, Iran in 2007. Now she is student of M.Sc. mechanical Engineering UTM Skudai, Johor,
Malaysia.
Her research interests are: solid mechanic, bio medical engineer, fatigue and material science aspects, intelligent and
adaptive control,
Alireza Mohammad Shahri received a B.S. degree in Electrical engineeringfrom K.N.Toosi University of
Technology, Iran in 1985. He then receivedhis M.Eng and Ph.D. degrees in Electrical Engineering from University of
Wollongong, Australia in 1994 and 1998, respectively.
His research interests are: Instrumentation and measurement systems,Industrial computer networks, Mobile Robots,
and Mechatronic systems.Dr. Shahri is currently Faculty Member at the department of ElectricalEngineering at Iran
University of Science & Technology.
449
Abstract.The aim of this research is to develop an advanced driver assistance system characterized with the function of
adaptive front-lighting system (AFS). This paper presents a new method to adjust the amount and direction of light of
headlamp car in different road conditions. It calculates the angle curve and real distances to it and to front car and then
changes the density and radiation angle of front light car. The paper proposes a novel Road curve detection algorithm and
new method to measure the distance to the front car. The most prominent contribution of this paper is: Measurement angle of
the road curve is done with recognition only one of lane markings. Using the same algorithm to detect lane markings and
front car,both of them is detected at the same time at the image.This makes the analysis time greatly reduced and System
performance especially in high speed to improve.
Keywords: lane mark detection, adaptive front-lighting system (AFS), intelligent vehicles
1 Introduction
Progress of science and technology in recent years has led many advanced countries actively invest in making a
smart car. Road Accident Statistics Annual show Factors such as vision loss or did not have good visibility to the
road, driver fatigue resulting from driving long-time , non-driver's reaction time, etc is a major cause of
accidents. In fact, all these factors are the errors and limitations performance car driver.Therefore, as part of the
Smart car project, automated guided vehicles as one of the main goals in the automotive industry is followed.
This paper proposes a new method of Automatic control front-lighting car, with the aim of designing automated
guided vehicles. Despite the lower percentage of trips are done at night, but statistics show a significant
percentage of accidents happen at night. In automatic control front-lighting car the Light intensity and angle of
radiation of car headlights according to road conditions will be set automatically. To adjust the intensity and
direction angle of the light two factors is checked: the first is existence vehicle moving on the road in front of the
host car, to prevent vision loss driver must reduce the amount of light, and the second factor is the curve road.
Set the angle of light, according to the curved angle of the road done. As shown in the figure 1, rotating light
degree pitch angle according to the transverse and longitudinal road increased visibility to the driver.
In order to adjust the angle of light, the road curve identified, and then according to the angle of the curve road,
car headlights are set. Many methods are used for determine the angle of Road curves. Most of the proposed
450
methods based on measuring the angle of rotation of vehicle wheels, using sensors and angle which is embedded
on the wheels.
Fig 1. adaptive front-lighting system (AFS)-using ADS increase driver visibility range
The main problem is dependence onperformanceof driver thatcauseserrors of it be transmittedto automatic
control front-lighting car systemandaccuracy ofits performance influenced bydriverperformance. To prevent this
and increase accuracy, using a CCD/CMOS camera in front of car has been suggested.
The camera sends the image taken from road conditions to a central microcontroller. The microcontroller by
processing the image can determine present of the road curve and angle of it.
This method allows adding other automatic controls to the car without the need for new circuit and only using
processing the image of the road condition, such as obstacle detection, pursuit of a subject's mind, etc In addition
to increased accuracy.
Many methods are used for lane markings recognition and each method has its own advantages and
disadvantages according to its concept. As described in [1] [2], the inverse perspective mapping (IPM) approach
is adopted to generate the birds eye image of the road plane so as to remove the perspective effect and extract
lane markings through some constraints of road geometry. This method is time-consuming due to its complex
computation. Other research combines road model and Hough Transformation to estimate the initial lane
boundary and improve the availability of lane detection without the ability of curve lane recognition [3].
A feature-based approach [4] is applied to detect the preceding vehicle. The vehicle features, including texture,
edge, symmetry and shadow of vehicle images, are used for vehicle segmentation and recognition. In generally,
the single feature is not satisfied to detect a vehicle resulted from the variation of illumination or environment.
The template-matching is the other solution to detect a vehicle by means of preparing numerous vehicular
templates (such as: the characteristics of edge and wavelet, etc.) in advance to perform a match with the current
images so as to locate the matching region with the maximum correlation as the vehicle-existing region [5].
Normally this approach can obtain a better reliability than the approach of feature-based detection. Recently,
using a learning algorithm [6] for vehicular detection has become prevailing; therein, a neural network or other
training mechanisms are taken to go through training from a large volume of vehicle images and then carry out
the classification through a classifier.
This paper presents a novel algorithm for detect the road curve and angle of it. It uses only one of lane markings.
It is important because both of them are not always available.Especially when the car is moving at highway or
when more than one vehicle is moving on the road and when the car is overtaking.
The research also proposed new algorithm for determined the distance front car from host car if it exit. It is
necessary because High intensity light reduces visibility other car driver, therefore the light should be reduced.
Next section introduces the lane detection algorithm. The road curve detection algorithm and calculation the real
distance is presented in section 3 and 4. Section 5 specifies results. Finally, section 6 concludes the whole paper
and looks forward the future work.
451
(a)
(b)
Fig 2. The vertical lines in the image dont cut lines direct road (a) but cut the curved road in more one point (b).
The Distance between two intersection points defines the curved location.
This algorithm is based on Number of intersection points of vertical lines in the image with the road curves.If the
road is curved, one of vertical lines cuts it in more than one point; at least, otherwise the road lane intersects all
the vertical lines at just one point. In this method after the detection the lane markings, vertical lines at the half
image is drawn and the number of intersection points with the curve of the road is checked. If the number of
intersection points was just one point the process is finished else the process is repeated for line Vertical
Location fourth image. This process continues until reaching a desired accuracy. If the road curve cuts the
vertical line in more one point means the curve is located at the Distance between two
intersections.Formoreaccuratevertical linesshouldbe checkedmore veridical lines. For example if the road curve
cuts the line vertical location fourth image. It means the curve is located at the Distance between two
intersections but for more accurate it is necessary lines eighth image and if needed 1/16 image is also checked.
In order to determine the angle of the curve, it is approximatedbytwo crossover lines. Angle between two lines is
the angle of the road curve with reasonable accuracy. To draw a line is needed at least two points so if the
vertical line cuts the curve in two point, save the points and then the next vertical line is drawn and the new
points is save, now we have the two points that is needed for drawn the approximated lines. But To have more
accurate can be used more points.In this paper, the best line is approximated using minimum square error
method.
In the next algorithm, the image is divided into two parts: the land area and the sky area. In Land area which the
road surface is visible and is usually image, the image is divided into n area.Based on the available points in
each area of the n area, first degrees line is obtained. If theroadis not curved, the second line is along thefirst line
inanyconsecutivetwo areas therefore both of two lines have nearly the same slopes but the difference between
slopes of these two lines indicates there is a curve in this area and Difference slope between the two lines
452
represent the angle the road curve. The number areas (n) Defines the accuracy of the algorithm. If the number of
regions (n) is more algorithm accuracy is higher.
Fig3. In the next algorithm, the image is divided into n parts, If theroadis not curved, the second line is along
thefirst line inanyconsecutivetwo areas therefore both of two lines have nearly the same slopes but the difference
between slopes of these two lines indicates there is a curve in this area and Difference slope between the two
lines represent the angle the road curve.
Assuming the camera which is used, were able to image the road to a maximum distance of 100 meters with a
sufficient resolution,since the average maximum speed limit on highways is 100 kilometers per hour, so
Maximum time available for image processing and conclusion is:
Tmax = 100 m / 100 km/h = 3.6 s
So time is very important factor in selecting algorithms for image processing and Since the second algorithm,
due to lower computing operations require less time to process therefore, the second algorithm is preferred.
3.2 The real distance the curve to the front wheels of car
For finding the distance of the curve, we used the Two-Point Perspective 3D Model from a Single Image method
[8]. It calculates the actual distance between two points using their distance in image and perspective geometry
principles.
If we take a side view of the scene, we will get something similar to the following diagram:
453
The headway distance estimation model is used to calculate the distance between the host and preceding vehicle.
The headway distance estimation model as shown in Figure 7 mainly bases on the image of preceding vehicle in
the image plane and the relation of equivalent triangle equation (3) to calculate the distance between the host and
preceding vehicle.
(3)
where H is the height of mounted camera from the ground; f is the focus length of camera; D1 is the distance
from thecamera to the front tip of host vehicle; Pr is the size of everypixel in row direction of the monitor; R is
the size of image inrow direction, few parameters as listed below need to beobtained through calculating the
collecting data from images; D is the distance between the host camera and the tail ofpreceding vehicle; D2 is
the headway between two cars; and Z is the coordinate of image in row direction of the precedingvehicle bottom
end point.
454
5 Results
To test the proposed algorithm, we used the C3088 Camera. It is a color camera module with digital output. It
uses a CMOS image sensor OV66201 from Omnivision. It has a digital video port that supplies a continuous
8/16 bit-wide image data stream. All the camera functions, such as exposure, gamma, gain, white balance,
windowing, etc. is programmable.
The format used in the project is the simplest one: Zoom Video Port Format. In this format 8 bits represent the
intensity (Y channel) of one pixel, the other 8 bits represent the U and V channels, but are not used.
The motion control of the front-lighting system is done by a step motor. It has found that the stepper motor is
preferable than DC motor because of the dynamic performance, cost and the control [9].
Figure 8 shows Flowchart overall system performance.
6 Conclusions
This paper describes a new method for the adaptive front lighting system base on image processing.
The method for several images that was taken at different conditions of various roads (forest roads, mountain
roads, desert roads, etc.) was tested. It performs well, even if only a lane marking is seen, and image processing
algorithms can calculate angle and place of the curved with any accuracy.
1 CMOS VGA (640x480) CAMERACHIP with Omni Pixel Technology
455
With using the camera could add other intelligent control without the new circuit and only analyzing and
processing images taken by cameras.
In the future, we will devote to some other fields such as the situation, there isnt or may not recognize the lanes
marking.
References
1.
2.
3.
4.
5.
6.
7.
8.
9.
Yong Zhou, Rong Xu, Xiaofeng Hu and Qingtai Ye, A robust lane detection and tracking method based on
computer vision, 2006 IOP Publishing Ltd, vol. 7, pp. 62-81, February 2006
A. Broggi, M. Bertozzi, A. Fascioli, C. Guarino Lo Bianco, and A. Piazzi, Visual Perception of Obstacles and
Vehicles for Platooning, IEEE Trans. on Intelligent Transportation systems, vol. 1, pp. 164-176, September 2000.
Suzuki A.,Yasui N.,Kaneko M., Lane Recognition System for Guiding of Autonomous Vehicle, Intelligent
Vehicle '92, pp. 196-201, September 2000.
C. Hoffmann, T. Dang, C. Stiller, Vehicle Detection Fusing 2D Visual Features, IEEE Intelligent Vehicles
Symposium, pp. 280-285, 2004.
M. Betke, E. Haritaglu and L. Davis, Multiple Vehicle Detection Andand Tracking in Hard Real Time, IEEE
Intelligent VehiclesSymposium, pp. 351356, 1996.
Ayoub Khammari, Fawzi Nashashibi, Yotam Abramson, Claude Laurgeau, Vehicle detection combining gradient
analysis and AdaBoost classification, IEEE Conference on Intelligent Transportation Systems, pp1084-1089 ,
September 2005.
Wenhong Zhu, Fuqiang Liu, Zhipeng Li, Xinhong Wang, Shanshan Zhang, A Vision Based Lane Detection and
Tracking Algorithm in Automatic Drive, 2008 IEEE Pacific-Asia Workshop on Computational Intelligence and
Industrial Application.
Faustinus Kevin Gozali, Two-Point Perspective 3D Modeling from a Single Image, A Thesis Submitted to
Department of Electrical & Computer Engineering Carnegie Mellon University.
C K CHAN, K.W.E.CHENG, S.L.HO ,T .M. FUNG, Development of Electric Vehicle with Advanced Lighting
System and All Electric Drive, 2009 3rd International Conference on Power Electronics Systems and
Applications.
Appendix
The algorithm for several images of roads with different conditions was tested. The results for sample desert road, mountain
road and highway is shown. We test the detection front car program in desert road, detect curved in mountain road image and
both of them in highway image. The results are shown in follows.
(a)
(b)
(c)
Fig 1. Desert road, there is a car in front of the host car.(a) is the output camera.(b) is output the detection front car
program.(c) is detection lines Rear bumper and the car roof.
456
(a)
(b)
Fig 2.mountain road, the road is curved. (a) is the output camera.(b) shows output of detect curved programs.
(a)
(b)
(c)
input image
50
100
150
200
(d)
Fig 3.highway image, the road is curved and there is a car in front of the host car.(a) is the output camera.(b) is output the
detection front car program.(c) is detection lines Rear bumper and the car roof. (d) shows output of detect curved programs.
Biographies
Maryam Sakhdari- Maryam sakhdari born in neyshaboor/iran in desember 1987. She received a B.S. degree in Electrical
engineering (electronic) from Iran university science and technology, Tehran (IUST) in 2010. Now she is student of MS
electronic Engineering in Tarbiat Modares University.
Her research interests are: image processing & machine vision; smart control; optical Communication; metamaterial.
457
Abstract. Differential evolution (DE) is an effective and robust evolutionary optimization method.It employs
simple mutation and crossover operators to generate new candidate solutions, and utilizes a one-to-one
competition scheme to deterministically decide whether the offspring will replace their parents in the next
generation. This paper presents a proportionalintegralderivative (PID) controller design based on DE for
DC motor control. The control parameters such as maximum overshot, rise time, settling time and steadystate errors were affected by changing PID constants in a system controlled with PID controller. Optimum
PID constants tuning is required to achieve expected control performance. In this study, optimum PID
constants have been estimated by using DE algorithm. The unit step response of the system was observed and
compared with that of PID-PSO (PID-Particle Swarm Optimization). Obtained results show that the PID
controller based on DE is having better response than PID-PSO.
Keywords: PID tuning, differential evolution, particle swarm optimization
1 Introduction
Today, there are various control methods used in control applications such as PID, fuzzy control and neural
network. According to statistics nearly 90% industrial objects are controlled by PID controllers: [1], [2]. PID
control is called as a classical control method. It is basic and easily implemented to many different areas: [3].
Chang has proposed a systematic method to select gains of a discrete PID controller: [4]. Huang and Yasunobu
presented a practical design method for fuzzy PID control and designed some examples both linear and nonlinear
plants: [5].
In the PID control design, there are three basic parameters directly affecting the system control performance.
Therefore, researchers focus on the strategies to tune these three parameters.Among the well-known approaches
are the integral of squared time weighted error rule (ISE), integral of absolute error rule (IAE): [6], and the
Ziegler-Nichols (Z-N): [7]. However, these traditional methods cannot meet the expected control performance.
PID tuning can be considered as an optimization problem because the aim of PID tuning is to determine the
optimal PID parameters.Considering it as an optimization problem many optimization algorithms have been
introduced into PID control such as genetic algorithm (GA): [8] and particle swarm optimization (PSO): [9]. In
reference [10], PID control parameters are determined by a modified binary PSO algorithm. In other study, PSOPID has proposed for real-time control systems: [11]. Proposed PSO-PID control system has the salient merits of
458
u (t ) K p e(t ) Ki e(t ) d (t ) K d
0
de(t )
d (t )
(4)
where K p is proportional gain, K i is the magnitude of the error plus the integral gain, and K d is the integral of
error plus the derivative gain. Input signal u (t ) will be sent to the plant (system to be controlled), and the new
output y out (t ) will be obtained. This new output y out (t ) will be sent back to the sensor again to find the new
error signal e(t ) . The PID controller takes this new error signal and computes its derivative and its integral
again. This process goes on and on.
In this paper, a PID controller using the DE was developed to improve the step response of DC motor. The DE
algorithm was mainly utilized to determine three optimal controller parameters K p , K i and K d , such that
controlled system could obtain a good step response. Fig. 1 shows block diagram of optimal PID control for DC
motor.
Kt
0.01Kg-m/A
459
In the proposed the DE method, each individual contains three members K p , K i and K d . These members are
assigned as real values. The DE algorithm is detailed in section 3.
3.1 Initialization
DE algorithm starts with initializing a population of PN (population number) solution vector, which has target
individuals PG X 1,G , X 2,G ,....X PN ,G . Where X i,G X i,G1 , X i,G 2 ,....X i,G n , i 1,2....PN is an n-dimensional
vector with parameter values determined randomly and uniformly between predefined search ranges [Xmin, Xmax].
3.2 Mutation
The mutation operation of DE applies the vector differentials between the existing population members. For each
target vector X i,G X i,G1, X i,G 2 ,....X i,G n , i 1,2....PN , a mutant vector is generated by
(2)
where a, b, c 1, 2,...PN are random indexes and must be different from each other. F is a mutation factor
0,2 which affects the differential variation between two individuals: [1].
3.3 Crossover
In order to increase the population diversity, a crossover operator is applied to each mutant vector and its
corresponding target individual. The crossover operator is represented as follows,
r j CR
vi ,G 1 if
u ,G 1
xi ,G otherwise
(3)
where CR is a crossover rate in the range 0,1 , and r j 0,1 is a uniform random number. As a result of
crossover, a trial vector, U ,G u i,G1 , u i,G 2 ,.....u i,G n , is yielded for each mutant vector.
3.4 Selection
In order to decide whether or not it should become a member of next generation, the trial vector u i ,G 1 is
compared to the target vector xi ,G , using the selection criterion. The selection in DE is deterministic and simple.
460
The resulting trial vector u i ,G 1 will only replace the original parent xi ,G if it has a lower fitness function value.
The selection criterion of DE can be expressed as follows,
U ,G if f U ,G f X ,G
X ,G 1
otherwise
X ,G
(4)
where f V ,G and f X ,G are the fitness function of U ,G and X i ,G , respectively. In order to decide whether or
not it should become a member of generation G 1 , the trial vector u ,G 1 is compared to the target vector x ,G
using the greedy criterion. The resulting trial vector u ,G 1 will replace the original parent x ,G if it has a lower
cost function value; otherwise, the old value x ,G is retained.
3.5 Termination
Above mentioned algorithm steps are repeated until the fitness is acceptable or the evolution time
completed. Algorithm is ended, once the reach to termination criteria. The individual which has best
fitness function value has been admitted solution.
4 Simulation Studies
In this section, control performance of PID controller based on DE, PSO and Z-N method are compared. In
order to carry out the comparison of the DE and PSO algorithms in similar conditions, the values of similar
control parameters of the algorithms were chosen; for example, population size is set to 20 and number of
iterations is set to 50. For DE algorithm, mutant factor is F 0.5 , and crossover rate is CR 0.8 . For PSO
algorithm, c1 c 2 2 and inertia weight, wt , is decreasing linearly from 0.5 to 0.1.
The cost function is very important for PID tuning based on DE since it directly determines the goal of
optimization. In this paper, a cost function, J w , is defined stated as follows:
J w
1
N
en
(5)
n 1
where en is absolute error, w is PID parameter vector, and N is number of samples. The aim is to minimize
J w , by adjusting w . Initial population was randomly generated. Search space is defined in the range
0 K p 50 , 0 K i 50 , and 0 K d 1 .
Fig. 2 and Table 2 show the values of the PID parameter obtained by DE, PSO and Z-N. Whereas the DE and
PSO approximately determine the same values for PID parameters, the Z-N method determines the different
values then others. The averaged absolute error of PID controller based on DE lower than that of PID-PSO. The
control performances of tuning methods are drawn in Fig. 4.
Table 2. PID parameters and averaged absolute error obtained by DE, PSO and Z-N
Algorithms
Kp
Ki
Kd
DE
PSO
Z-N
25.1413
29.2032
38.2353
26.9509
26.9704
63.8978
0.1375
0.1504
0.0039
Averaged
absolute error
4.6350
4.6353
6.6952
Settling time
(second)
0.1
0.1
0.45
461
Fig. 3 shows that the DE and PSO provide similar performance and superior to that of Z-N method. Both
algorithms have no overshoot and small settling time. Maximum overshoot occurred 7% in the control
performance by Z-N method.
Fig. 3. The control performances of PID controller based DE, PSO and Z-N
5 Conclusions
PID is a control method which is widely used in control applications. It is simple and powerful tool. It is
essential that the PID control parameters have estimated in system control. Henceforth, this paper focuses on
PID parameter tuning as an optimization problem. The PID controller based on DE was designed for DC motor
control. Also, PID-PSO and Z-N controller were designed for comparison. The unit step response of the system
462
was observed. By a comparative study, it is shown that the optimization methods were utilized in improving the
control performance of PID. Proposed PID controller based on DE reduced the averaged absolute error for step
response of DC motor control. It can be efficiently used for PID tuning problem.
References
1.
2.
3.
4.
5.
6.
7.
8.
9.
10.
11.
12.
13.
14.
15.
16.
17.
18.
19.
Zheng, T.G.; and Quan, D.H. (2007). ACS algorithm-based adaptive fuzzy PID controller and its application to
CIP-I intelligent leg. J. Cent. South University Technology 14 (4): 528-536.
Lin, Y.; Ang, K.H.; and Chong, G.C.Y. (2006). Patents, software, and hardware for PID control. IEEE Control
System Magazine 26 (1): 42-54.
PID 2006, IEEE Control System Magazine 26 (1): Feb. 2006.
Chang, P.H. (2009). A systematic method for gain selection of robust PID control for nonlinear plants of secondorder controller canonical form. IEEE Transactions on Control Systems Technology 17 (2): 473-483.
Huang, Y.; and Yasunobu, S. (2000). A general practical design method for fuzzy PID control from conventional
PID control. In Proceedings of the Ninth IEEE International Conference on Fuzzy Systems (FUZZY IEEE 2000),
San Antonio, Texas, USA, 7-10 May 2000, vol. 2, 969-972.
Zhen, L.L.; and Wang, L. (2008). Probability-based binary particle swarm optimization algorithm and its
application to WFGD control. In Proceedings of International Conference on Computer Science and Software
Engineering, Wuhan, China, 12-14 December 2000, 443-446.
Ziegler, J.G.; and Nichols, N.B. (1942). Optimum settings for automatic controllers. Trans. ASME 64: 759-768.
Zhang, J.H.; and Zhuang, J. (2009). Self-organizing genetic algorithm based tuning of PID controllers. Information
Sciences 179: 1007-1018.
Panda, S.; and Padhy, N.P. (2008). Comparison of particle swarm optimization and genetic algorithm for FACTSbased controller design. Applied Soft Computing 8: 1418-1427.
Ma, C.; Qian, L.; Wang, L.; Menhas, M.I.; and Fei, M. (2010). Determination of the PID controller parameters by
modified binary particle swarm optimization algorithm. In Proceedings of Control and Decision Conference
Chinese (CDCC), Xuzhou, China, 26-28 May 2010, 2689-2694.
Wai, R.J.; Lee, J.D.; and Chuang, K.L. (2011). Real-time PID control strategy for maglev transportation system via
particle swarm optimization. IEEE Transactions on Industrial Electronics 58 (2): 629-646.
Cheng, S.L.; and Hwang, C. (2001). Optimal approximation of linear systems by a differential evolution algorithm.
IEEE Transactions on System Man and Cybernetics, 31 (6): 698707.
Karaboga, N. (2005). Digital IIR filter design using differential evolution algorithm.EURASIP Journal on Applied
Signal Processing 8: 1269-1276.
Yu, B.; and He, X.S. (2006). Training radial basis function networks with differential evolution. Transactions on
Engineering Computer and Technology 11 (2): 157160.
Liao, T.W. (2010). Two hybrid differential evolution algorithms for engineering design optimization. Applied Soft
Computing 10(4): 1188-1199.
Storn, R.; and Price, K. (1995). Differential evolution-a simple and efficient adaptive scheme for global
optimization over continuous spaces. (Technical Report No. TR-95-012, International Computer Science Institute
(ICSI), Berkeley, California, USA.
http://www.engin.umich.edu/group/ctm/examples/motor/PID2.html
Cheng, S.L.; and Hwang, C. (2001). Optimal approximation of linear systems by a differential evolution algorithm.
IEEE Transactions on System Man and Cybernetics 31 (6): 698707.
Yu, B.; and He, X.S. (2006). Training radial basis function networks with differential evolution. Transactions on
Engineering Computer and Technology 11 (2), 157160.
Biographies
Ayhan Gn was born in Ktahya, Turkey, in 1973. He received his B.Sc. degree in Electrical-Electronics Engineering at
Near EastUniversity, M.Sc. degree in Electrical-Electronics Engineering at DumlupnarUniversity and Ph.D. degree in
Electrical-Electronics Engineering at OsmangaziUniversity, in 1996, 2001, and 2007, respectively. He has substantial
experience of control systems, pendulum control, neural network, and fuzzy control.
Currently, he is Assis. Prof. at DumlupnarUniversity, Department of Electrical-Electronics Engineering. He is interesting
463
464
Abstract. In this paper, a new CMOS analog neuron circuit for hardware implementation of artificial neural networks is
presented. The neuron circuit consists of two main parts: the synapse circuit which uses a new low-power four-quadrant
analog multiplier circuit and a sigmoid function generator circuit. The multiplier circuit uses differential-mode input voltages
and produces a single-ended current output. The proposed circuit is implemented and simulated in HSPICE using 0.18m
CMOS technology parameters. In order to justify the performance of the proposed neuron circuit, a neural network for
solving XOR problem is designed and successfully simulated in HSPICE.
1 Introduction
In the recent years, neural networks have earned more popularity in different applications specially those which
need real-time processing. Since, most of signal processing applications need a large amount of real-time data
processing and because neural networks are inherently parallel processors, so, neural networks have found many
applications these days. However, most of the time, artificial neural networks are implemented using computers
which have not a parallel nature but hardware implementation of neural networks can be made parallel easily.
There are different approaches to hardware implement neural networks which are: analog, digital and mixedsignal implementations [4-8]. However, analog implementation of neural networks has many advantages such as
small size and higher speed with reduced power consumption because analog synapse circuits can be
implemented by analog multipliers which use fewer transistors in comparison with digital parallel multipliers
which use at least several thousand transistors [1-3]. However, artificial neural networks are systems which
consist of a large number of simple processing units, called neurons.
A neuron has generally a highdimensional input vector and a simple output signal which is usually a non-linear function of the input vector
and the weight vector. The function to be performed on the input vectors is hence defined by the non-linear
function and the weight vector of neuron. A simple neuron structure is shown in fig.1 while the neuron function
is expressed in (1).
(1)
Vout = F ( Xi.Wi)
465
Where, F is the activation function of neuron and Xi is the input signal of the ith synapse and Wi is the connection
weight of the ith synapse. Depending on the application, the neurons could be connected to each other by using a
special construction. However, the operation of each neuron is controlled by setting weight values in each
synapse in order to recognize different patterns which are expected from the neural network. These days,
researches in the field of neural networks are highly focused on designing neuron circuits which are adjustable
[4]. So, designing different circuits and implementing different activation functions is needed [5, 6]. In the
following, a new low-power multiplayer circuit and a sigmoid activation function generator circuit is proposed
and used in designing a neuron structure. Finally, in order to justify the neurons performance, a MLP neural
network using the proposed neuron is presented to solve the XOR problem.
466
Vdd
M13
M15
M14
Vx+
M3
M4
Vx-
M16
M17
M18
Iout
M2
Vy+
M7
M1
M5 M6
Vy-
M10
M8 M9
M12
M11
Vss
Fig. 2. The multiplayer circuit used in synapse [1]
(2)
Where, VX = VX+-VX- , VY = VY+-VYGm is MOSFETs transconductance which can be defined by:
Gm =(Cox.W/L.ID)0.5
(3)
However, the multiplayer circuit which is shown in fig. 2, is simulated using HSPICE in 0.18m CMOS
technology. Fig. 3, shows the simulation result.
467
As it can be seen in fig. 3, the analog multiplier circuit shows a four-quadrant behavior and has a linear
characteristic for the input voltages in the range of -200mv to +200mv.
468
VDD
M2
M1
M3
M4
M6
M7
M8
M10
Vin
M9
M11
M12
Vbias
M5
M13
M14 M15
M16
Vss
The sigmoid circuit is also simulated using 0.18m CMOS technology parameters in HSPICE. Simulation
results are shown in fig. 5. As shown in fig.4, the circuit works in the voltage mode and both the inputs and
outputs of the sigmoid circuit are in the form of voltage signals.
469
4 Conclusions
In this paper, a new low-power CMOS analog neural network is presented and simulated using 0.18m
technology parameters in HSPICE. Because of using a low-power four-quadrant analog multiplier with singleended current output in the neuron circuit, there is no need for extra full-adder circuit in the synapse circuit. So,
the total power consumption of the circuit is minimized. The simulation results for the XOR problem is offered
in the paper, which shows that the proposed circuit can be successfully used in other applications.
470
References
[1] JACEK JASIELSKY, WOJCIECH KOLODZIEEJSKI, Four-Quadrant CMOS Transconductance Multiplier Operating
at Low Voltage and High-speed, Mixed design of integrated circuits and systems (MIXDES 2010), June 24-26, wroclaw,
Poland, 2010.
[2] M. Dei, N. Nizza, G. M. Lazzerini, P. Bruschi, M. Piotto, A four-quadrant CMOS analog multiplier based on the non
ideal mosfet I-V characteristics, IEEE PRIME - Research in Microelectronics and Electronics Conference, pp. 33-36, 2008.
[3] M. Dei, N. Nizza, G. M. Lazzerini, P. Bruschi, M. Piotto, A four-quadrant analog multiplier based on a novel CMOS
linear current divider, IEEE PRIME - Research in Microelectronics and Electronics Conference, pp.128-131, 2009.
[4] LEE, B.W., and SHEU, B.J. General-purpose neural chips with electrically programmable synapses and gain-adjustable
neurons, IEEE J. Solid-state Circuits, 27, pp. 1299-1303, 1992.
[5] AL-RUWAIHI, K.M., NORAS, J.M. A programmable CMOS current comparator circuit for analog VLSI neural
network using identical small-dimension MOS transistors, Int. J.Electron ., 78, pp. 347-358, 1995.
[6] HOLLIS, P.W., and PAULOS, J.J, Artificial neural networks using MOS analog multipliers, IEEE J. Solid-"State
Circuits, 25, pp. 849-855, 1990.
[7] GOWDA, S.M., SHEU, B.J., CHOI, J., HWANG, C., and CABLE, J.S. "Design and characterization of analog VLSI
neural Network modules, IEEE J. Solid-State Circuits, 28, pp.302-313, 1993.
[8] B. Gilbert, A precise four-quadrant multiplier with sub-nanosecond response, IEEE J. Solid-State Circuits, pp. 365-373,
1968.
471
Abstract. Power Amplifier, raising the antenna used to transmit the RF signal at wireless transceiver systems. Battery life of
portable devices is very important. Therefore, low power consumption should be considered when designing PA. Thus low
power consumption of the circuit elements are taken into account when choosing. AGC has been added to improve the
stability and available dynamic range of the PA. In addition, despite to changing conditions AGC allows us to achieve
constant output power of PA. This study, 2.45 GHz for the transmitter added AGC PA has been designing and the results of
the measurement circuit are examined.
Keywords: AGC, PA, ISM band, transceiver systems
1 Introduction
In this study, low-power, high gain and added the AGC of PA were design and analysis for 2.45 GHz wireless
communication systems.
RF PAs are radio frequency working, transmitter provide the necessary gain and power, usually the antenna
feeding circuit. High-power for running the most current drawing system and the most heat-generating part of.
especially on portable devices such as mobile phone efficiency that directly affects the battery life is a very
important role in.
2.
Working frequencies of 2.45GHz, the A-class power amplifier analysis and design were performed, and
performance improvement for wireless transmitter systems. A class reason for the choice of amplifier high
linearity and minimum distortion. Despite that the bias of this type amplifiers are similar to small-signal
amplifiers, the basic differences are to permit the instantanes output voltage and current oscillations. In study
two floors of the A-class power amplifier ATF 55143 (Avago) used low-noise FET. Radial stub was used for
destroy parasites of supply voltage. Two stubs used to make matching at RF signal amplifier circuit. Directional
coupler the output of the circuit and amplifier feedback used for AGC system.
In the circuit applications 55143 FET, LTC5509 detector and the LM393 dual comparator used.
472
473
Figure4.PACircuitGraphicS11 andS21
Figure5.PANFCircuitGraphic
474
voltage value of the directional coupler was given equation 1. Reflected voltage value of the directional coupler
was given equation 2. Depend on these, the ratio of returned voltage with respect to incoming voltage is given at
the 3. equation. The return loss equation is given in the 6th expression
Vi C
C j
e
D
(1)
Vr C
C j
e
D
(2)
1
Vr
D
Vi 1 1
D
(3)
RL(dB) 20log(
Pr
)
Pi (4)
475
Figure 9. PACircuitMeasurementResults.
476
Conclusion
The reason for the addition of the AGC circuit to the PA system is to increase the usable dynamic range and
stability.
However, due to feedback to the AGC system will cause earnings to fall. In the simulations, the supply voltage is
applied as a 3.6V, but in application circuit it is applied as a 3.9V, because of the better working conditions. In
the both simulation and application circuits, when AGC system was added it has seen that S21 decreased.
Thanks
References
[1] Belen, M.A.,Kaya, A., SD Teknik Bilimler Dergisi, 2011, Cilt1 Say 1, Pasif Elemanlar Eklenerek 2.4 GHz
ISM Band Uygulamalar in Mikroerit Band Geiren Filtre Tasarm
[2] Kahriman,M., Kaya, A., SD Teknik Bilimler Dergisi, 2011, Cilt1 Say 1, RF Devreler in G lme Ve
Adaptif Kontrol Uygulamas
[3] Pozar, M., 1998. John Wiley&Wiley, Microwave Engineering.
[4]Cotter W. Sayre, Copyright 2008, 2001 by The McGraw-Hill Companies, Complete Wireless Design,
Second Edition.
[5] Bowick C., Blyler J., Ajluni C., 2008, Newnes, RF Circuit Design, Second Edition.
[6]Hsieh M., Kim J., Harjani R., A European ISM Band Power Amplifier Module, Department of Electrical and
Computer Engineering, University of Minnesota, Minneapolis, MN 55455, USA.
[7] Multanen M., Salonen P., Kivikoski M., 2003, Variable-Gain Amplifier For Adaptive Antenna Control At
The 2.4-Ghz Ism-Band, Microwave And Optical Technology Letters / Vol. 41, No. 2, April 20 2004 , Tampere
University of Technology Institute of Electronics
[8] 55143, website, http:// www.agilent.com/semiconductors/, Contact date: 13.02.2011
Biographies
Hseyin ARK is a master Electronics and Communication Engineering student in at Sleyman Demirel University and
currently working on developing Added AGC PA Design For 2.45 Ghz ISM Band Low Power Transceiver Systems.
Adnan Kaya is a associate professor in at Sleyman Demirel University. He works about electromagnetic field and
microwave technique.
477
Abstract.This paper presents a design tool for automated design of analog CMOS operational amplifiers
(OPAMPs). In order to fit the circuit performance into desired specifications, a multi-objective optimization
approach based on genetic algorithms (GA) is proposed. The transistor sizes and bias voltages as well as the
values of passive components are calculated based on analytical equations describing the behavior of the
circuit. The optimization algorithm is developed in MATLAB and the performance of the designed circuit is
verified using SPICE simulations based on EKV 0.5m CMOS technology parameters. Different analog
amplifiers such as Folded-Cascode and Cascode OPAMPs were successfully designed and verified using the
proposed design tool. It is also shown in this paper that the design results obtained from the proposed
algorithm in MATLAB, have a very good agreement with the obtained circuit simulation results in SPICE for
all considered circuit topologies.
Keywords:Multi-objective optimization, analog, OPAMP, CMOS
1 Introduction
With the increasing demand for analog and mixed-signal integrated circuits, the optimal design of analog
amplifiers such as operational amplifiers (OPAMPs) is becoming more challenging and time consuming
[1,2]. Since, some performance parameters of analog amplifiers are most likely conflicting with each other,
analog circuit design suffers from long design time, high complexity, high cost and requires highly skilled analog
designers [1-5]. Soft computing methods such as fuzzy optimization approach [6] and Genetic Algorithms (GA)
can be used to decrease the design duration and time-to-market of analog integrated circuits [2-5]. Moreover, in
the recent years, the new EKV transistor model is used for circuit simulations because it is strongly based on
device physics and it links the design equations with small number of model parameters. So, by using EKV
model, it is possible to find solutions of design parameters without solving complex and time-consuming design
equations [7]. However, in an analog design problem design goals involve multiple design specifications. So, in
the OPAMP design problem, it is desirable to achieve maximum gain as well as maximum bandwidth while the
power consumption and noise are kept in the minimum value. In many cases, the above performances are most
likely conflicting with each other. So, in these cases the designer can use multi-objective optimization in order to
optimize both objectives simultaneously to yield a better solution. In this paper a multi-objective optimization
using genetic algorithm is used to find the optimum solution to the design of cascode and folded-cascode
OPAMPs.
478
(a)
(b)
Fig. 1.OPAMP structures considered in this paper
(a): Cascode (b): Folded-Cascode
The chromosome contains transistor dimensions W, L and bias voltages or currents of transistors. Of course,
some of the widths and lengths in the OPAMP circuits are equal for example (W 1=W2, L1=L2). The automated
design and optimization algorithm is coded in MATLAB using the analytical equations of OPAMPs shown in
figure 1.
In this section, as an example, the automated design algorithm for the Folded-Cascode OPAMP is reviewed.
The analytical equations of a Folded-Cascode OPAMP used in this paper are as follows:
(1)
(2)
479
(3)
GBW=gm1/CL
(4)
SR=I3/CL
(5)
(6)
As it is obvious from the circuit topologies, some widths and lengths are equal, so, the unknown design
parameters which are considered as the chromosome vector are: W 1, W3, W4, W6, W8, W10, L6, L8, Vb1, Vb2, Vb3.
For all OPAMP structures considered in this paper it is assumed that, V DD= 1.5V, VSS=0V andCL=2pF. The
GAprogram determines this vector of the parameters such that the fitness function is maximized. Fitness is a
numerical quantity describing how well a solution (individual) meets predefined design objectives and
constraints. On the other hand, the total performance of a circuit is evaluated by fitness function. For fitness
evaluation in a multi-objective problem, several methods can be used. Here, the fitness function (ff) is defined as:
Maximizing gain
Maximizing bandwidth
Minimizing noise
480
As mentioned before, the four main objectives are conflicting and thus require trade-off decision. The multiobjective optimization can be formulated as the product of single-objectives as follows:
Minimize
(7)
3 Simulation results
In this paper, two different types of widely used operational amplifiers (Cascode and Folded-Cascode OPAMPs)
are designed and optimized using multi-objective optimization approach implemented in MATLAB using GA as
a multi-objective optimization tool. SPICE software and EKV transistor model are also used as a verification tool
for circuit simulations. Furthermore, optimization is done in both frequency and time-domain characteristics. The
simulation results are summarized in table 1, where they are compared with the desired values. Finally, the design
parameters (transistor dimensions and bias voltages) which are resulted from the optimization algorithm in
MATLAB are used for SPICE circuit simulations in order to verify the performance of the proposed algorithm.
Figures 2, 3 show the SPICE simulation results for AC performance of the considered OPAMP circuits.
Desired
Cascode
Noise(nV/Hz)
<400
Gain(dB)
>70
GBW(MHz)
>5
PM(deg)
>60
SR(V/s)
>10
offset(mv)
<6
Vomax(V)
>1.2
Vomin(V)
<0.25
Swing(V)
>1
P(mW)
<2
Vdd
1.5
Vss
324.8
95.4
16.23
87.8
9.65
4.8
1.21
0.28
0.93
0.517
1.5
0
FoldedCascode
3 0.3
86.4
10.8
88.3
9.4
0.71
1.32
0.21
1.11
1.23
1.5
0
481
Conclusions
In this paper, the multi-objective optimization algorithm is developed in MATLAB and the performance of the
designed circuits are verified using SPICE simulations based on EKV 0.5m CMOS technology parameters.
Different analog amplifiers such as Folded-Cascode and Cascode OPAMPs were successfully designed and
verified using the proposed design tool. It is also shown that the design results obtained from the proposed
algorithm in MATLAB have a very good agreement with those obtained from circuit simulation results in SPICE
for all considered circuit topologies.
References
1.
2.
3.
4.
5.
6.
7.
482
Abstract. Firearms leave characteristic striations on the bullet while the bullet travels through the barrel. The
number, orientation, width and depth of these striations show similarities for the bullets fired from the same
gun. In this work, 1D profiles calculated from 2D images are used to match bullets for automated firearm
identification. For that purpose a profile clustering algorithm is proposed to exclude the deformed land
impression regions in the profile extraction process.
Keywords: firearm identification, bullet matching, profile extraction
1 Introduction
Firearms leave characteristic striations (land impressions and groove impressions) on the bullet while the bullet
travels through the barrel. The number, orientation, width and depth of these striations show similarities for the
bullets fired from the same gun. Using the similarity in these features, a bullet collected from a crime scene can
be matched to another bullet in the criminology laboratory archive. The matching, and firearm identification
process is performed by a comparison microscope by the firearm examiners traditionally. However, this
traditional method is subjective and extremely time consuming due to the large number of evidences in the
criminology laboratory archive. Therefore automated firearm identification systems have been popular over the
last decades.
There are several studies in the literature for automated bullet matching [1-5]. In one of the early studies, [1], the
position, amplitude, and width of the striations measured by scanning electron microscope are used for bullet
matching. In [2], 3D images are used to compare the positions of the striations. In [3], Chu et al. proposed an
edge detection based method for locating striation marks and selecting valid correlation areas. In [4], SIFT
features are extracted from 2D images for bullet matching. In another study, [5], Lenfused multiple images of
the bullet recorded under different illumination conditions to obtain a high quality image. Then, he projected the
2D image in the striation direction to extract 1D profile. These 1D profiles are used for bullet matching after
applying a top-hat filter. The bullet matching methodology followed in this study is parallel with the Lens
methodology. The contributions of this study mainly focus on the image fusion and profile extraction steps of the
methodology proposed by Len.
Image fusion plays an important role for obtaining high quality images in different fields of forensic science. In
[6], Wen & Chen proposed a wavelet-based image fusion technique (Wen & Chens Image Fusion Algorithm)
for different forensic science applications. The experiments conducted in this study show that the application of
this fusion algorithm (proposed by Wen & Chen) to bullet matching also give good results compared to the
Lens Image Fusion Algorithm [7].
483
There can be deformed regions which are not characteristic for the bullet matching operation on the bullet
surfaces. These regions corrupt the 1D profile extracted from the projection of the 2D image in the striation
direction. In order to exclude these deformed regions from the profile extraction step, a profile clustering
algorithm is proposed in this study.
2 Method
The profile extraction methodology followed in this study is summarized in Figure 2. First the images taken
under different illumination conditions are fused. For that image fusion operation two different algorithms are
used in this study. These are Len [7] and Wen&Chens [6] image fusion algorithms. In the second step, the
fusion image is rotated in a way that the striations are parallel to the horizontal axis of the image coordinate
system using a Hough transform based line detection algorithm [8]. Finally the profile clustering algorithm is
utilized to extract the 1D image profile.
(1)
484
where
is
th
the n column of the
ith window.
(2)
Step 3: Finally the Teager function is calculated from the difference function as it is given below.
(3)
The Teager-Kaiser Energy Function calculated for the bullet shown in Figure 3 is given in Figure 4. The location
of the peak values (greater than an adaptive threshold) of that function gives the deformation region boundaries
over the bullet surface. If the distance between two marginal peaks is smaller than a certain distance, the region
between these peaks is excluded from the profile extraction process. The average of the columns of the
remaining parts of the image is calculated to yield the final 1D profile feature.
Figure 4. The Teager-Kaiser Energy Function calculated for the bullet land impression shown above. There are
4 marginal peaks (deformation region boundaries) which are greater than the adaptive threshold level
485
486
Figure 7.(a-b) Two bullet land impression image samples (c) Fusion result according to Lens Algorithm [7]. (d) Fusion
result according to Wen&Chens Algorithm [6]
In order to evaluate the performance of the algorithms, the concentration curve is used [5]. The x-axis of the
curve represents the archive percentage and y-axis represents the probability of finding an element (land
impression) in the corresponding initial percentage of the archive list. In Table 4, values from the concentration
curve for different image fusion and profile clustering conditions are given. It is seen in the table that the
proposed profile clustering algorithm and the Wen&Chens image fusion algorithm both increase the bullet
matching performance.
Table 4. Concentration curve values for different image fusion and profile clustering conditions
Image
Method
Fusion
Profile
Clustering
%1
%10
%20
Len [7]
NO
%19
%44
%69
%79
Wen&Chen [6]
NO
%19
%46
%85
%87
Len [7]
YES
%23
%31
%65
%79
Wen&Chen [6]
YES
%19
%58
%88
%88
Acknowledgement
This work is supported by TBTAK 107G194 Research Grant.
487
References
1.
2.
3.
4.
5.
6.
7.
8.
9.
Gardner, G.Y. (1978). Computer identification of bullets. IEEE Trans. Syst. Man Cybern. Vol. 8 pp. 6976.
Leloglu, U. M., Deviren, M., Sermen, C., Balci, K. (2000). Mermi ekirdekleri zerindeki Set zlerinin Otomatik
Elenmesi. 8. Sinyal Isleme ve Uygulamalari Kurultayi.
Chu, A., Song, J., Vorburger, T., Thompson, R., Renegar, T., and Silver, R. (2010). Optimal compression and
binarization of signature profiles for automated bullet identification systems. Proc. SPIE 7729, 77291C.
Sayar, A., Tetiker, F., Acar, E., Oskay Acar, B., Sakarya, U. (2011). Bullet matching using SIFT feature. In Proc.
IEEE 19th Signal Processing, Communication and Applications Conference, Antalya, Turkey, April 20-22.
Len, F.P. (2006). Automated comparison of firearm bullets. Forensic Science International, pp. 40-50.
Wen, C., & Chen, J. (2004). Multi-resolution image fusion technique and its application to forensic science.
Forensic Science International, vol. 140, pp. 217-232.
Len, F. P. (1997). Enhanced imaging by fusion of illumination series. Proceedings of SPIE, vol. 3100, Sensors,
Sensor Systems, and Sensor Data Processing, Otmar Loffeld, Editor, pp. 297-308.
Duda, Richard O., and Peter E. Hart. (1972). Use of the Hough transformation to detect lines and curves in pictures.
Communications of the ACM 15, no. 1, pp. 11-15, January 1.
Kaiser, J.F. (1990). On a simple algorithm to calculate the energy of a signal. Proc. IEEE ICASSP-90,
Albuquerque, NM, pp.381-384.
Biographies
Erman Acar Erman Acar received his B.Sc. and M.Sc. degrees in electrical and electronics engineering from Middle
East Technical University in 2008 and 2011 respectively. He currently works at TBTAK UZAY as a researcher. His main
research topics are computer vision, pattern recognition and biomedical signal processing.
Fatih Tetiker Fatih Tetiker received his B.Sc. degree in computer science from Bilkent University in 2008. He currently
works at TBTAK UZAY as a researcher. His main research topics are computer vision and image processing.
Banu Oskay Acar - Banu Oskay Acar received his B.Sc. and M.Sc. degrees in electrical and electronics engineering from
Middle East Technical University in 1999 and 2002 respectively. She currently works at TBTAK UZAY as a senior
researcher. His main research topics are pattern recognition, audio and speech processing.
Ufuk Sakarya - Ufuk Sakarya received his B.Sc. and M.Sc. degrees in electrical and electronics engineering from Middle
East Technical University in 1998 and 2002 respectively. He received his Ph.D. degree in electronics engineering from
Ankara University in 2009. He currently works at TBTAK UZAY as a chief researcher. His main research topics are
pattern recognition, computer vision, image and video processing.
488
Abstract.In this study a glider tracking system and monitoring is developed to establish glider flight
performance and it is aimed to develop a technique for separating hook with the flight performance data
obtained from the model. The timer existing in the glider monitors time-dependent execution of predetermined flight movements of the model glider. GPS and microcontroller based circuitry is designed to
track the model flight and help to find where it would land. The variables such as altitude, direction, speed
and distance are accurately measured for the model glider. The modeler can watch flight cruising over Google
Earth software in three-dimension by transferring flight data to a computer environment.
Keywords: glider, timer, hand unit, thermal soaring, tracking system
1 Introduction
With the purpose of making aviation activities more attractive, increasing flight information and improving the
skills and encouraging sporting activities in various aviation clubs and federations has been established around
the world. Among the aviation activities ballooning, parachuting, model aircraft, and glider can be considered.
Developing a glider model and its flight techniques hasshown continuous improvements in parallel with the
developing technology. In general, the gliding movement, without any driving force propeller, fan, etc., just
using air currents (thermal) is called free flight. A glider is a kind of aircraft that does not depend on engine or
motor-based power to achieve flights [1].Another definition according to Federal Aviation Administration
(FAA), a glider is an aircraft heavier than air that is forced to move up in flight by the dynamic reaction of the air
in opposition to its lifting surfaces, and whose free flight is not dependent on an engine [2].
The most fair for modelers is to lose their models. If the tail winglet does not open because of a malfunction
arising from the timer or mechanical problems the model continues to rise or fly until scattering the hot air
masses ascending from the ground to the air, which is called thermal. The model, with a large thermal can go 2-3
km away in 3 minutes. It has been made possible follow it up with the addition of GPS so that such a distance
may cause the loss of model glider.
A literature review shows that tracking processes are studied using various methods and algorithms [4], [5], [6].
Our contribution, which we describe in this paper, is to develop and demonstrate a tracking system and
monitoring for a model glider.
489
The purpose of the modeler is to successfully achieve the model free-flight motion over the specified time under
the specified rules at a desirable performance. For this reason, the flight modeler observes the weather conditions
(thermals) carefully and whenever captures an appropriate thermal, the modeler should release the model
immediately. Thermals arise when the air near the ground becomes less intense than the surrounding air due to
heating or humidity changes at the earths surface [3]. The model glider is moved into the air with the aid of a
rope length of 50 m. The model is left with the aid of rope through the hook mechanism and the model passes
through the free flight. Referees run the clock as soon as the model is left from the rope. The model should be
disconnected from the hook at the moment of hot air beginning to rise. As soon as the model is left from the
glider hook, the timer starts. The timer's task enables the model to implement its time-dependent flight control
mechanisms (rudder, aileron tail, etc.) within the rules in a specified time period. These movements are
predetermined and cannot be changed by the modeler during the flight. For example, we can predetermine the
movements such as once the model is left the hook, it lifts its tail wingletup to 2 degrees and the rudder returns to
the right of 30 degrees after 30 seconds with the start of flight movements. Leaving the model at the right
moment and regulating these movements in respect to time completely depends on the experience of modeler.
Thermal Soaring
The principles of thermal soaring with a glider are very simple, and most medium/large size gliders are capable
of riding the thermals'.
Thermals are columns of air that are warmer than the air surrounding them. Warm air is less dense than cool air,
and so it rises up - rather like helium filled balloon does. The rate of rise depends on the temperature of air; the
higher the temperature, the less dense the air and so the faster and higher it will rise.
Thermals appear because the sun warms different surfaces at different rates; for example, water absorbs the sun's
rays so thermals won't be present over the water but the roof of a house, or a road, will be warmed by the sun
quickly and so strong thermals will be generated over these types of surface.
490
Basic function of a timer shown in Fig. 2 is to improve performance of a model during the flight by having made
different angles to the tail winglet, helm, etc. The model can realize these actions with respect to time only
because of the rules. It is forbidden to interference the model while it flights by any algorithm or external as the
rules necessitate.
Fig.3. Arising of thermal disturbance: a) the tail winglet parallel to the earth b) the tail winglet by 45 o (1: gum
elastic, 2: cord)
The winglet at tail that remains parallel to the earth is called de-thermalizer (thermal disturbance). The
winglet takes the position as shown in Fig. 3a during the flight.
As the one end of cord is connected to the rear side of de-thermalizer, shown by 2, in Fig. 3b the other end of
491
cord is connected to the timer available in the glider body. When the flight time is over the timer leaves the cord
free. Meanwhile, the gum elastic, shown by 1, achieves to leave the rear of the de-thermalizer up with an angle
of 45 degrees. Therefore, the de-thermalizer takes the second position as shown in Fig. 3b. The model in this
position wants to lift its nose up but, it cant have this action because of no driving force. Finally it will begin to
lose altitude and will lend.
The sensitive measurement of variables for the model glider such as altitude, direction, speed, distance is
successfully implemented via GPS. The system basically consists of a glider, hand unit and interface program.
GPS receiver in the glider sends its geographical data to the hand unit.
Fig.4. a) The interface program and b) hand unit displaying position and direction
The data for the position and the direction of the glider are received after the selection serial port and pressing
the start button in the interface program shown in Fig. 4a. Flight and position of the model can be shown on
world map by Google Earth Software in real time.The coordinateand altitude information incoming fromthe
model can be observed at the left side of the interface program. The altitude for the glider can be sketched in real
time over the flight duration.
A typical GPS based tracking system is depicted in Fig. 5a. Rapidly locating of direction and distance to reach a
glider is vital and cause of pleasure for a modeler. Direction location can be found by calculating the angle
difference between the direction that modeler should go and the direction that the modeler goes on. The modeler
is positioned at the origin of coordinate system and the angle is calculated between model and modeler. North
line is used as a reference (zero degree).
Fig. 5. a) A typical GPS based tracking system and b) direction location accessing on a model
The indicated numbers in Fig. 5b represent 1: model position, 2: modeler (follower), 3: direction that modeler
goes on, 4: direction that modeler should go, 5: angle difference between the direction that modeler should go
492
and the direction that modeler goes on, and 6: the north line respectively.
The hand unit is designed to calculate the direction angle regarding with the Fig. 5b and to display on its
screen. The direction is found as the angle difference between the direction angle that modeler should go and the
direction angle that modeler goes on regarding with the north line.
Fig. 6. a) Distance calculation between two points and b) their representation in analytic plane
Position knowledge is normally received from GPS in degree and in minute with their latitudes and longitudes as
shown in Fig. 6a. For the calculations to be easy latitudes and longitudes are expressed in degrees. Assume that
the sample points incoming from GPS are
degrees north latitude,
degrees west
longitude for point 1 and
degrees north latitude,
degrees west longitude for the
point. These two points are demonstrated in Fig. 6b.
The difference between the point 1 and the point 2 gives the distances in parallel to the latitude and the longitude
respectively.
, the distance in parallel to the latitude
, the distance in parallel to the longitude
The distance between the all latitudes is the same and is equal to
km all over the world. That means, if one
degree is
in km then the distance between the points is
km.
Since the longitudes narrow from the equator towards the poles, the distance between the two longitudes is
in equator only. An approach, (1), can be utilized in calculating the distance between longitudes at a
latitude point.
(1)
The distance between longitudes at the latitude of the second point is calculated by
km. The distance between the longitudes is
in
km. The two distances in latitude and the longitude between the two points are the horizontal and vertical
components of a perpendicular triangle. The hypotenuse of the triangle presents the actual distance between the
points.
(2)
493
The same points are entered to the Google Earth Software and the result is found to be of 888.4898 m., which
is much closed by the algorithm result.
Experimental Results
The first flight shown in Fig.7 is realized using some standard setup adjustments. Then, the second flight
designates more improvement because of choosing the first flight as a reference. The flight angle of dethermalizier is increased by 2 degrees after the first flight. At the instant of leaving the cord end from the model
the de-thermalizier angle is lifted up 3 degrees from the beginning. After 0.5 seconds, it is reduced by -3 degrees.
A lightweight and highly accurate tracking system is designed and the measuring performance of flight is
successfully achieved for a model glider. With the help of GPS-based hand unit developed, the data related with
the glider such as the height from the earth, the distance to the location and the geographical position are
transferred in real-time to the modeler. Even if the model glider is far away from the modeler, it can be easily
located by the help of the tracking system. The hand unit can transfer the coming data to the computer over USB
port. The flight cruising can be monitored in three dimension in real time using the Google Earth software. The
flight cruising data can be stored for later use via the logging capability of the system.
A discussion of approaches to the design of a tracking system regarding with soaring has raised several areas
for future work. These approaches are computationally expensive, and will benefit from insight drawn from
simpler studies of soaring.
494
References
1.
2.
3.
4.
5.
6.
Kagabo, W. B. (2010). Optimal Trajectory Planing for a UAV Glider Using Atmospheric Thermals, Master Thesis,
Faculty of the Department of Mechanical Engineering in the Kate Gleason College of Engineering at Rochester
Institute of Technology, New York.
Administration (2003). Federal Aviation Glider Flying Handbook, Flight Standards Service.
Allen, M. J. (2007). Guidance and Control of an Autonomous Soaring UAV, NASA Dryden Flight Research
Center Edwards, California.
Parrott, G. C. (1970). Aerodynamics of gliding flight of a black vulture Coragyps atratus. J. Exp. Biol. 53, 363374
Tucker, V. A. & Parrott, G. C. (1970). Aerodynamics of gliding flight in a falcon and other birds. J. Exp. Biol. 52,
345367.
Anderson, J.D. (2000). Introduction to Flight (Fourth Ed.), Boston, Massachusetts, McGraw Hill,
Biographies
Mert Gner He was born in Istanbul, Turkey, in 1986. He received B. S. degree in electronics and computer education
from Marmara University, Technical Education Faculty, Computer-Control Education Department in 2010.
Mr. Gner is still working as a R&D engineer in ENTES Electronics in Istanbul.
Mustafa Onat He was born in Trabzon, Turkey, in 1963. He received B. S. degree in electrical and electronics engineering
from Middle East Technical University, Engineering Faculty, Electrical and Electronics Department, Turkey, in 1988 and the
M.S. and Ph.D. degrees in computer and control education from Marmara University, The Institute for Graduate Studies in
Pure and Applied Sciences, Computer & Control Education Department, Istanbul, Turkey, in 1996 and 2001, respectively. .
Major Field of study: Internet and GSM based control, real time control, developing industrial product.
Cihan Ylmaz He was born in Istanbul, Turkey, in 1986. He received B. S. degree in electronics and computer education
from Marmara University, Technical Education Faculty, Computer -Control Education Department in 2010.
495
Abstract. This paper presents an adaptive trajectory tracking control scheme for a wheeled mobile robot
with an inverted pendulum (WMRIP) in presence of parametric uncertainties and disturbances. First feedback
linearization technique is applied to the dynamic of heading and pendulum angles subsystems. Then
disturbances and parametric uncertainties are compensated by use of adaption law. For overcoming
underactuated property of the system, the integrated-error of displacement is added to control signal of
inverted pendulum angle as a virtual disturbance so error of displacement is converged to zero by adaptive
controller. The computer simulations illustrate the performance of proposed controller.
Keywords:Adaptive trajectory tracking, Inverted pendulum wheeled mobile robot, Disturbance observer,
underactuated system
1 Introduction
The inverted pendulum system has the unique characteristic that has been considered as a prototype nonlinear
systems whose control is known to be not easy. Plan of WMRIP first has come from The Industrial Electronics
Laboratory at the Swiss Federal Institute of Technology, Lausanne, Switzerland [1]. WMRIP has evoked a lot
of interested recently and some commercial product is available like Segway and P.U.M.A car. Such vehicles
are of interest because they have a small footprint and can spin on the spot. From the theoretical point of view,
the WMRIP models have attracted much attention in the field of control theory and engineering, because they
are nonlinear and underactuated with inherent unstable dynamics. A wheeled inverted pendulum mobile robot is
a selfbalancing vehicle with two wheels attached on the sides of its body. Similar systems like the cart and
pendulum, and the Pendubot have been studied in the several literatures. Unlike these systems, however, the
pendulums motion in the present system is not planar and the motors driving the wheels are directly mounted
on the pendulum body.
Several works have been done in recent years to control of motion WMRIP. In [1] dynamic equations of system
is achieved by Newtons approach. Then by linearizing the results about operating point, the linear controller is
designed. In [2] velocity and position of system are controlled by partial feedback linearization. In [3] Several
coordinate transformations and partial-feedback linearization techniques are done then derived to transform the
vehicles dynamics into an upper-triangular form. Next, nested p-times differentiable saturation and
backstepping techniques are combined to design a control law for the transformed system. Also a disturbance
observer is designed to estimate external disturbance. [4] used two sliding layers and uncertainty defined as
slope of surface that robot move over that. Adaptive robust controller is designed in [5]. The proposed
controller mechanisms using physical properties of wheeled inverted pendulums ensure that the outputs of the
system track the given bounded reference signals within a small neighborhood of zero. Intelligent control is
496
used widely in control of WMRIP. In [8] WMRIP used as a boxer robot that control by RBF neural network
controller. Also [7] used linearizing model for design of Takagi-Sugeno fuzzy controller. In [13] adaptive
backstepping controller is proposed. In [6] and [15] neural network adaptive controller proposed. Direct
adaptive fuzzy basic function network controller is designed in [14]. In this work classic adaptive controller is
suggested for control of WMRIP. This controller has simple structure and easy to implementation and also has
sufficient performance.
The rest of the paper is outlined as follow. Section 2 briefly describes the nonlinear modeling of the robot.
Section 3 develops proposed adaptive controller. The performance of proposed controller is illustrated in
section 4 by computer simulation and finally section 5 concludes the paper.
2 Problem Statements
As shows in figure 1 the dynamic of WMRIP system consist of three degree of freedom and two control inputs
consist of right and left wheels torque. The equation of motion of WMRIP from [3] is:
D
2 RJ
(T L T R )
1
1
[ml cos( )(ml sin( ) 2 (T L T R )) M x mgl sin( ) cos( )]
0
R
x
(1)
1
1
[J (ml sin( ) 2 (T L T R )) m 2l 2 sin( ) cos( )]
0
R
Where:
J
D2
(M w w2 )
2
R
J
M x M m 2(M w w2 )
R
J Jv
J Jc J p
0
M x J m 2l 2 cos( ) 2
Where is heading angle of robot, x is the vehicles displacement along x -axis and is angle of pendulum.
The parameters that used in equation (1) are introduced in table 1. By linear transform in control inputs the
dynamic of WMRIP can be decoupled to two independent subsystems. First subsystem is nonlinear SISO
system where heading angle is Its output and T1 is its control input. Another subsystem is underactuated
nonlinear SIMO system where displacement and pendulum angle are its outputs and T 2 is its control input. T 1
and T 2 defined as follow:
T1 1 1 T L
T 1 1 T
R
2
(2)
497
Torque provided by wheel actuators acting on the left and right wheels
Mw
Jw
Gravity acceleration
Distance between the left and right wheels along the y-axis
Jc
Jv
, of the pendulum
z
Pendulum
CG
Right wheel
TR
x
mg
Left wheel
chassis
O
TL
g 2T 2
(3)
498
3 Control Design
As showed in previous section, the control task is to design two control input
and
for two subsystems
and . T 2 and T 2 can transform to input right and left wheels torque by following equation:
T L 0.5 0.5 T1
T 0.5 0.5 T
2
R
(4)
1
(u1 des )
g1
T2
1
(u 2 f 2 )
g2
(5)
and
are designed as
k [e
1
and
e ]
(6)
des
where
k [ ]
2
where
s k
and
12
des
11
and
0
des
[ k11 k12 ]
22
s k
21
and
[ k 21 k 22 ]
are chosen
f 2 f 2
And g 2
g 2 g 2 g 2
Where f
and
f 2 and
g 2 . Here the
uncertainty and the external disturbance are combined and treated as a total disturbance. The plant (3.b) is
rewritten as:
f f 2 ( g 2 g 2 )T d 2
2
where d 2 is bounded external disturbance. Summing up unknown terms into a one unknown term the plant
(3.b) can be rewritten as
(7)
f g 2T b2 ( , , d 2 )
2
where
b ( , , d ) f 2 g 2T 2 d 2 is total disturbance and includes both the internal uncertainty and the external
2
2
u2
(8)
Adaptation law is employed to estimate b 2 . If the new variables are defined as follow:
499
z u2
(9)
b2
(10)
Then
can be estimated as
Where
by:
(11)
z
n s2
(12)
d T
0T 0
So by having 1
, is PE signal. Since regression signal is PE, so estimated disturbance converges to real
value exponentially. Thus is used to monitor the total disturbance . The feedback linearization controller
(5.b) using disturbance observer is given as:
(15)
1
T2
g2
(u 2 f 2 b 2 )
This method similarly is employed for omit effect of parametric uncertainty and external disturbance of
subsystem and results as below:
(16)
1
T1
g2
(u 2 des b1 )
where is total estimated disturbance and includes both the internal uncertainty and the external disturbance
of
subsystem.
of MWRIP,
integrated-error of displacement can added to pendulum angle dynamic system as a virtual disturbance.
u 3 k 31 ( x des x ) k 32 ( x des x )
(17)
Where
is integrated-error of displacement and
and
show desire displacement and its first derivate
computed from trajectory generator.
and
are positive constants.Control law (17) for subsystem
is
modified as:
(18)
1
T2
g2
(u 2 f 2 b2
u3 )
500
4 Simulation Results
In this section, computer simulation is performed to evaluate the performance of the proposed controller. In this
simulation, parameters are chosen to match with a real world robot. In order to show the performance of the
adaptive tracking controller, the results of nonadaptive controller that designed based on feedback linearization
technique in presence of unknown parameters and disturbance are bring in this section.
2
The plant parameters are taken as m 1.5kg , M 5kg , J 1.5kgm 2 , M 1kg , , J c 2.5kgm , D 0.15m ,
W
w
J
1.5kgm , l 1.2m , R 0.25m . The initial conditions are taken as M 8kg , (0) 0.2 rad s ,
x (0) 1m , x (0) 0 m s ,
(0)
10
rad
and
k
12
(0) 0.01rad s .
k
11
4, k
21
25 , k
22
10 , k 0.25 , k 32 0.1 ,
31
10 and 0 .
A bounded smooth desired trajectory is chosen as 1 meter radius circle which 1M s linear velocity.
Nonadaptive controller that design in section 3.1 have good results where there is no external disturbance and
without parametric uncertainty. But control objective failed when there exist parametric uncertainty and
external disturbance and system became unstable. Also in this case there is no control on displacement. If
external disturbances are chosen as d1 0.5 and d 2 0.6 from 4 second of simulation until 5 second,
nonadaptive controller became unstable. The results of simulation of adaptive controller shows in figure 2.
(a)
(b)
(c)
Fig. 3. Robot variables without parametric uncertainty in present of external disturbance where (a) shows
displacement along x-axis, (b) is pendulum angle and (c) is robot orientation
In presence of parametric uncertainty the results of one parameter estimator adaptive controller are shown in
figure 3. In this case some parameters are different in model form what are used is feedback linearization such
as m 4.5kg , M 8kg , J 3kgm 2 , J 7.5kgm 2 , J 5kgm 2 and l 2.6m .
w
(a)
(b)
(c)
Fig. 4. Robot variables with parametric uncertainty in present of external disturbance where (a) shows
displacement along x-axis, (b) is pendulum angle and (c) is robot orientation
501
5. Conclusion
In this paper, an adaptive feedback linearization controller was proposed to solve trajectory tracking problem of
MWRIP. Proposed controller has simple structure and easy to implement and also has sufficient performance
in presence of external disturbances and parametric uncertainties and adaptive control is capable at tracking
reference signals satisfactorily, with all of the closed-loop signals uniformly smooth and bounded. Computer
simulations illustrate the performance of the proposed controller.
References
1.
2.
3.
4.
5.
6.
7.
8.
9.
10.
11.
12.
13.
14.
15.
Felix Grasser, Aldo DArrigo, Silvio Colombi, Member, IEEE, and Alfred C. Rufer, Senior Member,
IEEE.(2002). JOE: A Mobile, Inverted Pendulum. IEEE TRANSACTIONS ON INDUSTRIAL ELECTRONICS.
49(1), FEBRUARY: 107-114
Kaustubh Pathak, JaumeFranch, and Sunil K. Agrawal. (2005). Velocity and Position Control of a Wheeled
Inverted Pendulum by Partial Feedback Linearization. IEEE TRANSACTIONS ON ROBOTICS. 21(3): 505-515
Khac Duc Do and Gerald Seet. (2010). Motion Control of a Two-Wheeled Mobile Vehicle with an Inverted
Pendulum. J Intell Robot Syst .Springer. 60: 577-605
Jian Huang, Zhi-Hong Guan, Takayuki Matsuno, Toshio Fukuda, and Kosuke Sekiyama. (2010). Sliding-Mode
Velocity Control of Mobile-Wheeled Inverted-Pendulum Systems. IEEE TRANSACTIONS ON ROBOTICS,
26(4): 750-758
Zhijun Li, Member, IEEE, and Jun Luo. (2009). Adaptive Robust Dynamic Balance and Motion Controls of
Mobile Wheeled Inverted Pendulums. IEEE TRANSACTIONS ON CONTROL SYSTEMS TECHNOLOGY,
17(1):233-241
Ching-Chih Tsai, Senior Member, IEEE, Hsu-Chih Huang, Member, IEEE, and Shui-Chun Lin, Member, IEEE.
(2010). Adaptive Neural Network Control of a Self-Balancing Two-Wheeled Scooter. IEEE TRANSACTIONS
ON INDUSTRIAL ELECTRONICS, 57(4):1420-1428
H. Azizan*, M. Jafarinasab, S. Behbahani, M.Danesh. (2010). Fuzzy Control Based on LMI Approach and Fuzzy
Interpretation of the Rider Input For Two Wheeled Balancing Human Transporter" , IEEE International
Conference on Control and Automation Xiamen, China, 9-11 June.
Jin-Seok Noh, Geun-Hyung Lee, and Seul Jung. (2008). Control of a Mobile Pendulum System for a Boxing
Robot Game" , International Conference on Smart Manufacturing Application in KINTEX, Gyeonggi-do,
Korea,9-11 April.
Zhijun Li*, Kun Yang and Yipeng Yang. (2009). Support Vector Machine Based Optimal Control for Mobile
Wheeled Inverted Pendulums with Dynamics Uncertainties. Joint 48th IEEE Conference on Decision and Control
and 28th Chinese Control Conference Shanghai, P.R. China,16-18 December.
Ryoko Imamuraand Toshinobu Takei,Shinichi Yuta.(2008). Sensor Drift Compensation and Control of a
Wheeled Inverted Pendulum Mobile Robot. 978-1-4244-1703-2/08/$25.00 2008 IEEE Conference.
Hyung-Jik Lee and Seul Jung.(2008). Control of a Mobile Inverted Pendulum Robot System. International
Conference on Control, Automation and Systems, Seoul, Korea, Oct. 14-17.
P.K.W. Abeygunawardhana, Michael Defoort and Toshiyuki Murakami. (2010). Self-Sustaining Control of TwoWheel Mobile Manipulator using Sliding Mode Control. The 11th IEEE International Workshop on Advanced
Motion Control March 21-24, Nagaoka, Japan
Tatsuya Nomura, Yuta Kitsuka, Haruo Suemitsu, and Takami Matsuo. (2009). Adaptive Backstepping Control for
a Two-Wheeled Autonomous Robot. ICROS-SICE International Joint Conference 2009. August 18-21, Fukuoka
International Congress Center, Japan
Ching-Chih Tsai, Shui-Chun Lin and Bor-Chih Lin. (2010). Intelligent Adaptive Motion Control Using Fuzzy
Basis Function Networks for Self-Balancing Two-Wheeled Transporters. 978-1-4244-6588-0/10/$25.00 201 0
IEEE.
Zhijun Li and Yang Li and Chenguang Yang and Nan Ding. (2010).Motion Control of an Autonomous Vehicle
Based on Wheeled Inverted Pendulum Using Neural-adaptive Implicit Control. The 2010 IEEE/RSJ International
Conference on Intelligent Robots and Systems October 18-22, Taipei, Taiwan.
502
Biographies
Amirkhosro Vosughi was born in Shiraz, Iran, in 1986. He received the B.S. degree in electronic engineering from
Amirkabir University of Technology (Tehran Polytechnic), Tehran, Iran, in 2009 andhe is currently a M.S student in
Electronic Faculty Control branch in IranUniversity of Science & Technology, Tehran, Iran.
Hes interested fields consist of Robotic and Instrument and Control Theory.
Alireza Mohammad shahri received a B.s. degree in Electrical engineering from K.N.Toosi University of Technology,
Iran in 1985. He then received his M.Eug and Ph.D degrees in Electrical Engineering from University if
WollongongAustralia in 1994 and 1888, respectively. His research interests are: Instrumentation and measurement systems,
Industrial computer networks, Mobile Robots and Mechatronic systems. Dr. shahri is currently Faculty Member at the
department of Electrical Engineering at Iran University of Science & Technology.
503
Abstract. Electromagnetic levitation systems (maglev) inherently show unstable behavior in case of there
is no active stabilizing controller in closed loop. Therefore, this paper outlines and describes design
principles of stabilizing I-PD controller for a 4-pole hybrid electromagnet by using Coefficient Diagram
Method (CDM) providing a good balance of stability, response and robustness. Effectiveness of the
proposed design technique is verified by numerical and also experimental results.
1 Introduction
The 4-pole hybrid electromagnet can solely produce levitation with full redundancy and can lead to couple of
being linear motor mover structure.[3] A possible industrial application is a freight conveyance system, used
in clean rooms, such as in a semiconductor manufactory. In this kind of applications, the electromagnet is
suspended below a track core. The controller either keeps the gap clearance at a constant distance, or changes
the gap length according to some rules with active stabilization.[1-2]
The well-known PID controller structure stabilize the system, but there is a zero-point near the origin, which
results in a large overshoot appeared at the output. This problem can be resolved by introducing the I-PD
configuration of PID structure. The key of I-PD controller design is how to decide the controller gains,
proportional, integral and derivative.Furthermore, since pure derivative is avoided from the point view of
practical realization, one significant issue rises up by questioning what should be definitely the limits of the
time constant of a pseudo derivative filter.
In this study, we apply Coefficient Diagram Method (CDM) to decide the gains of I-PD controllers for each
motion axes and stability limits of the time constant of admissible pseudo derivative filter. CDM is an
algebraic settlement applied to polynomial loop of the system on the parameter space, where a special
diagram is invoked to present and interpret the necessary information and also good design criteria of
predefined specifications. The basic idea of this approach is to use the stability indexes and the equivalent
time constant derived from the characteristic polynomial as the design basis. The design methodology is quite
simple and provides a good balance of stability, response and robustness.[4]
504
(a)
(b)
Fig. 1. (a) Fundamental structure of 4-pole hybrid electromagnet. (b) Test bench.
4-pole electromagnet can control 3-degree of freedom motion by appropriate excitation of each one of the
pole coil via an axes transformation matrix.[6-7] Due to the page limitation and simplification of the
discussions, the governing dynamical equations of the system will be taken into consideration only for vertical
direction, z. Moreover, for inclination (roll, yaw) dynamics, same equations can be developed as well.
Transfer function for vertical direction is described as following;
G( s)
K iz
Z ( s)
Vz ( s) m L s 3 m R s 2 R K z
(1)
The fundamental idea of controlling 4-pole hybrid electromagnet is described as follows; it is assumed that
there exists three virtual winding currents, (iz, i, i) which are employed to control motions of vertical
direction z, and inclinations , correspondingly [6-7]. Hence, the feedback controllers are designed each one
of the motion freedom independently. The relationships between virtual currents of the each degree of
freedom and actual winding currents are obtained from current transformation matrix as shown in Fig. 2.
i1 1 1 1
i
iz
2 1 1 1 i
i3 1 1 1
i
i4 1 1 1
(a)
(b)
Fig. 2. (a) Control Method of 4-pole hybrid electromagnet. (b) Current transformation
matrix
505
Z ref
KI
s
KP
VZ
K iz
m Ls m Rs R Kz
3
s KD
f s 1
I-PD
Controller
n i 1 1
P( s) a0 j ( s) s 1
i 2 j 1 j 1
(2)
where the design parameters; the equivalent time constant (), the stability indices (i) and the stability limits
(i*) are defined as;
506
(3)
ai2
, i 1,, n 1
ai 1 ai 1
0 n ,
a1 a0 , *i
1
i 1
(4)
1
i 1
The equivalent time constant specifies the speed of the time response. The stability indices and stability limits
indicate stability margin and waveform of the time response [8-10-11]. The variation of the stability indices
due to the plant parameter change points out the robustness. Stability conditions for Manabes canonical
polynomial are described as;
(5)
i 2.5,2,2, i 1,, n 1
KP
(6)
3 2 2 L2 RK z R 3 m
R4m
,
K
I
3 2 2 L2 K iz
33 2 2 1 L3 K iz
KD
(7)
a
R2m
L
, 1 1 2 3
1 Lz K iz
a0
R
To
set
the
previous
discussions on a
concrete base a test
bench was designed
as photographed in
Value
10
0.3
6.8410-6
13.44
1.50
Unit
z0 [mm]
iz0[A]
Kz [N/m]
Kiz [N/A]
Lz,, [H]
Value
4.3
0.0
20991
14.87
0.016
Unit
0, 0 [rad]
i0, i0 [A]
K, K [Nm/rad]
Ki, Ki [Nm/A]
Epm [AT]
Value
0.0
0.0
106.43
3.13
2689
Determination of the proper I-PD gains is fulfilled by using (7). For the ideal case, degree of the
characteristic polynomial is 4th and corresponding coefficient diagram is depicted in Fig. 4.
507
10
ai
i
1,12.*i
*
i
a , , 1.12 ,
10
10
10
10
-2
2
indices
Fig. 4. Coefficient Diagram for ideal case. (without pseudo derivative filter)
As we see from Fig. 4., stability conditions are met, namely, there is no cross section between i (stability
indices) and i* (stability limits or margins). However, practical realization dictates that a low pass filter
should be inserted through derivative term, the pseudo derivative filter. At that point, one significant issue
comes out; "What is the proper value of the filter time constant?" In selection of the filter time constant, two
dominant factors should be taken into account; noise emission of the gap length sensors and stability margins.
Hence, there should be a compromise between them. Smaller time constant will result in much sensor noise
entrance to the system and have adverse effect on the performance. On the other hand, bigger time constant
will threaten stability of the system and deteriorate the performance as well. CDM resolves that issue by
providing a visual tool. Consistency of the controller parameters (7) with a given filter time constant can be
tracked and tested by CD. For example, if the filter time is set to f =0.017sec., stability margin of the system
is to be passed over as illustrated in Fig. 5.
10
ai
f 0.017 sec
10
*i
2
a, ,
*
i
10
10
10
10
-2
Stability margins
are exceeded
-4
2
3
indices
Fig. 5. Coefficient Diagram for the pseudo derivative filter case. (f =0.017sec)
It should be noted that with addition of the pseudo derivative filter, the closed loop characteristic equation
becomes 5th order polynomial as seen in Fig.5. Furthermore,
CDM has following unique properties; the degree of convexity of the diagram for the coefficients of the
characteristic polynomial provides a general measure for stability, while the general inclination of this curve
gives a measure for the speed of response. The change of the shape of the coefficient curve due to the
parameter variation is a measure of robustness.
508
-4
experiment
reference
simulation
4
2
0
-2
-4
-6
x 10
-3
experiment
reference
simulation
4
Alpha Axis Inclanation (rad)
2
0
-2
-4
0.5
1.0
1.5
2.0
2.5
3.0
-6
Time (s)
0.5
1.0
1.5
2.0
2.5
3.0
Time (s)
(a)
(b)
Fig. 6. Position tracking response (a) z-vertical axis (b) -inclination axes
0.8
1.5
0.6
1
Winding Currents (A)
0.4
0.2
0
-0.2
-0.4
-0.6
-0.8
simulation
experiment
0
0.5
1.0
1.5
Time (s)
2.0
2.5
3.0
0.5
0
-0.5
-1
pole1
pole2
pole3
pole4
-1.5
-2
0.0
0.5
1.0
1.5
2.0
2.5
3.0
Time (s)
(a)
(b)
509
GAP
SENSOR
x 10
-4
4
LOAD
Q8
POWER
AMP
TARGET
PC
DAC
ADC
-2
CPU
HOST
PC
-4
0
-3
1
Virtual Axes (A, Alpha & Beta) Currents (A)
CPU
x 10
alpha axis
beta axis
1.5
1
0.5
0
-0.5
-1
-1.5
-2
0
Time (s)
MATLAB
XPC-TARGET
Time (s)
z axis current
alpha axis current
beta axis current
0.5
-0.5
-1
0
Time (s)
5 Conclusions
The discussed and outlined I-PD control design scheme based on CDM is a suitable and relatively very
easy approach for maglev systems. The technique has potential to apply any kind of motion control systems
without sacrificing a good balance point among robustness, stability and speed of response via visual design
support. The experimental and numerical calculation results validate effectiveness of the proposed design
method.
References
1.
2.
3.
4.
5.
6.
7.
8.
9.
Sinha, P. K., Electromagnetic Suspension, Dynamics and Control, Peter Peregrinus Ltd., London, U. K., (1987).
Astrm, K.J., T. Hagglund : PID Controllers: Theory, Design and Tuning. 2nd Edition.Research Triangle Park, NC:
Instrument Sociaty of America, (1995)
Liu, L. "Semi-Zero-Power Maglev Control and its Coordination with Linear Motor for Flexible Conveyance
System", PhD. Thesis, The University of Tokyo, June 2002
Manabe, S. "Coefficient Diagram Method", 14th IFAC Symp. on Automatic Control in Aerospace, Seoul(1998),
B. Taworn, B. Jongkol, N., Noriyuki, K., Yasuzumi, O. "CDM Based Two-Degree of Freedom PID Controllers
Tuning Formulas", Proc Sch Engineering Sch Inf Technol Electron Tokai Univ, Vol:30, pp.53-58, 2005
Bittar, Moura Sales A. R., PID Controller Design for Magnetic Leviataion Model, International Conference on
Cybernetics and Infotrmatics, Feb. 10-13, 2010
Ho, W.K., C.C. Hang, L. Cao, "Tuning of PID controllers based on gain and phase margin specifications",
Automatica, 31, pp. 497-502, 1995
A.I. Al-Odienat, A. A. Al-Lawama,"The Advantages of PID Fuzzy Controllers Over The Conventional Types",
American Journal of Applied Sciences, PP. 653-658, 2008
Li, T-H. S., Shieh, M-T. ,"Design of a GA-based fuzzy PID controller for non-minimum phase systems", Fuzzy Sets
and Systems, Volume 111, Issue 2, pp. 183-197, 16 Ap 2000
510
Chapter 3
INDUSTRIAL
ENGINEERING
511
Abstract.Outsourcing is the contractual delegation of a service or an activity that is normally, but not always, performed
in-house to an outside supplier. The classification of hybrid formed companies based on outsourcing arrangements is a
complex process which can be done by deductive reasoning. The purpose of this study is to propose a fuzzy inference system
which consists of fuzzy if-then rules and produces output values for given input values for the determination of outsourcing
arrangements according to the degree of outsourcing and ownership of a company. In order to test the applicability of the
proposed framework, a hypothetical case study is demonstrated.
1 Introduction
The make-or-buy decision which is fundamental to industrial organizations [1], is the act of making a
strategic choice between producing an item internally (in-house) or buying it externally (from an outside
supplier). The buy side of the decision is also referred to as outsourcing. Outsourcing is defined as the
contractual delegation of a service or an activity that is normally, but not always, performed in-house to an
outside supplier (vendor) [2]. The advantage of outside purchase (outsourcing) lies in the "high powered
incentives" of the market and in economy of scale, which means a specialized producer supplying multiple users
produces the input more cheaply [3].
Two alternative sourcing decisions as insourcing (make) or outsourcing (buy) are parts of make-or-buy a
certain decision. However there are other outsourcing arrangements for the hybrid forms. The classification of
hybrid formed companies based on outsourcing arrangements is a complex process which can be done by
deductive reasoning in which we wish to deduce or infer a conclusion, given a body of facts and knowledge. A
fuzzy inference system (FIS) employing fuzzy if-then rules can model the qualitative aspects of human
knowledge and reasoning processes without employing precise quantitative analyses [4].
The purpose of this study is to propose a fuzzy inference system which consists of fuzzy if-then rules and
produces output values for given input values for the determination of outsourcing arrangements according to the
degree of outsourcing and ownership of a company.
The following section includes the literature review of outsourcing arrangements. In the next section the
proposed methodology is explained by defining the fuzzy rules. Then a case study demonstrated the applicability
of the method. In the last section the study is concluded.
512
2 Literature Review
There are several classifications of outsourcing arrangements types in the literature. Loh and Venkatraman
(1991) classified the insourcing and outsourcing types according to the dimensions of the internalization of
human and technical resources [5]. The highest internalization of human resources and technological resources
represent complete in-house operations and the lowest internalization of human resources and technological
resources mean the complete outsourcing. The other outsourcing types in between are the facility management
outsourcing, systems integration outsourcing, and the time-sharing outsourcing. Currie and Willcocks, analyzed
(1998) four types of IT sourcing decisions in the context of scale, client/supplier, interdependency and risk
mitigation [6]. These types are total outsourcing multiple-supplier sourcing, joint venture/strategic alliance
sourcing, and insourcing. Elam (1988), suggested the possible use of three forms of cooperative arrangement;
full-equity ownership, partial ownership, and no ownership [7]. Dibbern et. al. (2004) developed a classification
of six outsourcing arrangements the kind of which are based on two parameters; degree of outsourcing and
ownership [8].
The classified outsourcing arrangements are briefly explained below [8]:
Spin-off (SPO): Outsourced activities could be spun off into a separate services unit or company a situation
where the ownership is still internal but the function is either totally or selectively outsourced [9]-[10].
Joint ventures (JV): When the spin off is jointly owned between the client and vendor organizations. Such
joint ventures are based on a strategic partnership [11]-[12].
Traditional outsourcing (TO): The ownership of the activities is handed over to the outsourcing vendor and
outsourcing is of the total variety [13].
Selective outsourcing (SLO): Outsourcing is selective when the ownership of the activities is handed over to
the outsourcing vendor [14].
Insourcing / Backsourcing (I/B): A firm has the activities entirely inside and has total ownership of it and the
activities had not previously been outsourced [15], backsourcing refers to the same situation but the unit has
been brought back in-house after being outsourced [16].
Facilities sharing (FS): The firm chooses to share ownership of activities with either a vendor or others in the
same industry [17].
3 Proposed Methodology
This paper proposes a fuzzy inference system for the determination of outsourcing arrangements according to
the degree of outsourcing and ownership of a company. Fuzzy inference system (FIS) is the process of
formalizing the reasoning process of human language using fuzzy logic. This process involves three major
concepts which are membership functions, fuzzy set operations, and inference rules [18]. A fuzzy rule is a
conditional statement, expressed in the form if-then. The deduction of the rule is called the inference and
requires the definition of a membership function characterizing this inference. FIS uses fuzzy if-then rules to
determine a mapping from fuzzy sets in the input universe of discourse to fuzzy sets in the output universe of
discourse.
In the literature, three types of FISs, Mamdani FIS, Sugeno FIS, and Tsukamoto FIS, have been widely used in
various applications. The differences between these FISs lie in the consequents of their fuzzy rules, and thus
their aggregation and defuzzication procedures differ accordingly [19].
A typical FIS consists of four steps; fuzzification, fuzzy rules, fuzzy inference and defuzzification.
Step 1: (Fuzzification) Fuzzification process involves the definition of the membership functions of
input/output variables by linguistic variables.
Step 2: (Fuzzy rules)A FIS with i-input variables has r=pi rules, where p is the number of linguistic terms per
input variable. As the dimension and complexity of a system increase, the size of the rule base increases
exponentially.
513
Step 3: (Fuzzy inference) Fuzzy inference is an inference procedure to derive a conclusion based on a set of
if then rules. In this study, Mamdani method is used for fuzzy inference. The Mamdani inference method is
manually constructed on the basis of expert knowledge and the final model is neither trained nor optimized. The
mechanism of Mamdani inference method is as follows: (1) If there is more than one input in the rule, fuzzy set
operations should be applied to achieve a single membership value; (2) then implication method (min) is applied
to reach each rules conclusion; (3) the outputs obtained for each rule are combined into a single fuzzy set, using
a fuzzy aggregation operator (max).
Step 4: (Defuzzification) Through fuzzy inference, the result can be converted into a definite value. In this
study the center of area defuzzification method is used in order to get the advantage of its continuity and
disambiguity. Center of area (also called centroid method or center of gravity method) is the most prevalent and
physically appealing of all the available defuzzification methods [20]. It is given by the algebraic expression as
follows;
(1)
4 Case Study
As a case study, three different hybrid formed company is classified based on outsourcing arrangements by
using a Mamdani type FIS which is developed by using MATLAB Fuzzy Logic ToolboxTM .(Fig. 1).
514
RULE 6: IF degree of outsourcing is Medium AND ownership is Low THEN Selective Sourcing is High AND
Joint-Venture is Medium AND Traditional Outsourcing is Medium AND Facilities sharing is Low.
RULE 7: IF degree of outsourcing is Low AND ownership is High THEN Insourcing/Backsourcing is High
AND Spin-off is Medium AND Facilities sharing is Medium AND Joint-Venture is Low.
RULE 8: IF degree of outsourcing is Low AND ownership is Medium THEN Facilities sharing is High AND
Insourcing/Backsourcing is Medium AND Joint-Venture is Medium AND Spin-off is Low AND Selective
Sourcing is Low.
RULE 9: IF degree of outsourcing is Low AND ownership is Low THEN Selective Sourcing is Medium AND
Facilities sharing is Medium AND Joint-Venture is Low.
After the construction of fuzzy rules, the fuzzy inference process is performed to derive a conclusion based on
a set of if-then rules. Figure 2 represents the fuzzy rules for the inference system which is formed by using
MATLAB Fuzzy Logic ToolboxTM.
I1
65
90
50
I2
85
10
50
SPO
0.73
0.00
0.50
JV
0.52
0.50
0.87
TO
0.44
0.86
0.13
SLO
0.44
0.5
0.50
I/B
0.48
0.00
0.13
FS
0.27
0.00
0.50
When we analyze the Table 1, for the first hybrid form company which has a point of 65 for degree of
outsourcing and a point of 85 for ownership, we can say that with 0.73 possibility it is a Spin-off. The second
company which has 90 point for degree of outsourcing and 10 point for ownership, is possibly be a Traditional
Outsourcing with 0.86 membership degree. Lastly, the third company having 50 point for degree of outsourcing
and 50 point for ownership, has 0.87 possibility to be a Joint Venture.
515
5 Conclusion
This study presents a framework for the classification of hybrid formed companies based on outsourcing
arrangements. These arrangements are classified by a FIS employing fuzzy if-then rules. Fuzzy if-then rules
produces output values for given input values for the determination of outsourcing arrangements according to the
degree of outsourcing and ownership of a company. For the reasoning process, Mamdani type FIS is utilized in
which input and output parameters are represented with the linguistic variables of low, medium and high by
using nine if-then rules. In order to test the applicability of the proposed framework, a hypothetical case study is
demonstrated. The scores based on input parameters of three hybrid formed companies are evaluated by experts
and the reasoning of these companies is obtained by using the constructed inference system. The hypothetical
case study demonstrated the applicability of the FIS in classification of outsourcing arrangements under
uncertain environment.
References
1.
2.
3.
4.
5.
6.
7.
8.
9.
10.
11.
12.
13.
14.
15.
16.
17.
18.
Grossman, G. M. and Helpman, E., (2002). Integration versus outsourcing in industry equilibrium*,
The Quarterly Journal of Economics 117(1): 85-120
Nicholson, B.; Jones, J.; and Espenlaub S. (2006). Transaction costs and control of outsourced
accounting: Case evidence from India, Management Accounting Research 17: 238258.
Nooteboom, B. (1993) Firm size effects on transaction costs. Small Business Economics5: 283295.
Jang, J.S. R. (1993). ANFIS: Adaptive-Network-based Fuzzy Inference Systems, IEEE Transactions
on Systems, Man, and Cybernetics, 23(3): 665-685.
Loh, L. and Venkatraman, N., (1991). Outsourcing as a Mechanism of Information Technology
Governance: A Test of Alternative Diffusion Models, Working Paper No. BPS3271-91, Massachusetts
Institute of Technology, Alfred P. Sloan School of Management, Cambridge, MA
Currie, W. L. and Willcocks L. P., (1998). Analysing four types of IT sourcing decisions in the context
of scale, client/supplier interdependency and risk mitigation, Information Systems Journal, 8(2), 119
143.
Elam, J.J., (1988). Establishing Cooperative External Relationships, Transforming the IS Organization,
in Elam, J.J., Ginzberg, M.J., Keen, P.G.W., and Zmud R.W. (Eds.), ICFT Press, Washington, DC.
Dibbern, J.; Goles, T; Hirschheim, R; and Jayatilaka, B. (2004). Information systems outsourcing: A
survey and analysis of the literature. The DATA BASE for Advances in Information Systems 35(4): 6
102.
Heinzl, A., 1993. Outsourcing the Information Systems Function Within the Company: An Empirical
Survey, In Proceedings of the International Conference of Outsourcing of Information Services,
University of Twente, The Netherlands.
Reponen, T. (1993). Outsourcing or Insourcing?, In Proceedings of the 14 International Conference on
Information Systems, Orlando, Florida, 103-113.
Fitzgerald, G. and Willcocks, L. P. (1994). Contracts and Partnerships in the Outsourcing of IT, In
Proceedings of the 15th International Conference on Information Systems, Vancouver, Canada, 91- 98.
Marcolin, B. L. and McLellan, K. L. (1998). Effective IT Outsourcing Arrangements, In Proceedings of
the Annual Hawaii International Conference on System Sciences, 654-665.
Earl, M. J., 1996. The Risks of Outsourcing IT, Sloan Management Review 37(3): 26-32.
Lacity, M. C.; Willcocks, L. P. and Feeny, D. F. (1996). The Value of Selective IT Sourcing, Sloan
Management Review 37(3): 13-25.
Lacity, M. C. and Hirschheim, R. A. (1995). Beyond the information systems outsourcing bandwagon :
the insourcing response, Chichester ; New York: Wiley.
Hirschheim, R. A. and Lacity, M. C. (1998). Reducing Information Systems Costs through Insourcing:
Experiences from the Field, In Proceedings of the 31st Annual Hawaii International Conference on
System Sciences, Hawaii, 644-653.
Currie, W. L. (1998). Using Multiple Suppliers to Mitigate the Risk of IT Outsourcing at ICI and
Wessex Water, Journal of Information Technology 13: 169-180.
Duque, W. O., Huguet, N. F., Domingo J. L. and Schuhmacher, M. (2006). Assessing water quality in
rivers with fuzzy inference systems: A case study. Environment International, 32: 733742.
516
19. Jang, J. S. R., C. T. Sun, and E. Mizutani, Neuro-fuzzy and Soft Computing: A Computational
Approach to Learning and Machine Intelligence, Prentice-Hall, Upper Saddle River, NJ, 1997.
20. Ross, T. J., Fuzzy Logic with Engineering Applications, McGraw-Hill, New York, 1995.
21. Mamdani, E. H.(1974). Application of fuzzy algorithms for control of simple dynamic plant. IEE
Proceeding 121(12), 15851588.
Biographies
Ahmet Can Kutlu was born in Istanbul / Turkey in 1981 and graduated in electrical engineering at Istanbul
Technical University (ITU) in 2004. After the graduation in engineering management master program at the
institute of science and technology at ITU in 2007, he began his PhD studies in industrial engineering program at
the same institute at the same university in Istanbul / Turkey.
His research interest is about the management issues and methods concerning the fields of strategy, quality,
energy and human resources.
Mr. Kutlu has been working as a research assistant in the department of Industrial Engineering at the faculty
of Management at Istanbul Technical University since 2005
Hlya Behret was born in Kayseri / Turkey in 1979 and graduated in industrial engineering at Erciyes
University (Kayseri/Turkey) in 2001. After the graduation in industrial engineering master program at the
institute of science and technology at Bogazici University (Istanbul/Turkey) in 2006, she began her PhD studies
in industrial engineering program at institute of science and technology at Istanbul Technical University (ITU) in
Istanbul / Turkey.
Her research interest is about production planning and control and methods concerning the optimization of
complex systems under uncertainty.
Miss. Behret has been working as a research assistant in the department of Industrial Engineering at the
faculty of Management at Istanbul Technical University since 2006.
517
Abstract. Nowadays every firm uses telecommunication networks to continue their existence and complete
their daily operations. In this paper, we investigate an optimization problem that a firm faces when acquiring
network capacity from a market in which there exist several network providers offering different pricing and
quality of service schemes. In the mathematical model, quality of service level guaranteed by network
providers and minimum quality of service level which is needed to accomplishing operations are denoted as
fuzzy numbers. Finally, a heuristic solution procedure is proposed to solve resulting nonlinear mixed integer
programming model with fuzzy coefficients.
Keywords: telecommunication, fuzzy quality of service, heuristic algorithm
1 Introduction
Usage of data networks encompasses not only the traditional data applications but also newer
applications such as real-time audio/video streaming, voice over TCP/IP, real-time transactions and other batch
transactions over digital networks [1]. Firms in other words customers can lease or buy required amount of
network resource from telecommunication network suppliers to perform their tasks. Each of these mentioned
operations are assigned one of the acquired resources. Every task requires distinct amount of network capacity
and different quality of service (QoS) requirements. As a consequence, not only cost but also QoS levels that are
guarantied by network providers have to be taken into account by firms while selecting telecommunication
network suppliers.
The QoS requirements for an operation include a wide range of parameters such as bandwidth, loss
ratio, maximum delay and delay variation (jitter) [2]. Due to the non-deterministic nature of telecommunication
network environment, it is not always possible to correctly estimate these parameters beforehand. Therefore, the
main contribution and innovation of proposed work is the development of a mechanism for solving complex
selection and task allocation problems in communication networks. Based on the fact that accurate information
needed for making these decisions is not available. In the QoS related problems, QoS requirements by the users
can be regarded as fuzzy terms, since there exist the imprecision of fuzziness inherent in human judgments.
Network managers often avoid specifying crisp values for their goal [3]. Moreover, the imprecision of network
state seems to be always present and our best bet is to live with it [3].
518
In this paper, we investigate an optimization problem that a firm faces when acquiring network capacity
from a market in which there exist several network providers offering different pricing and quality of service
schemes. Quality of service level guaranteed by network providers and minimum quality of service level which
is needed to accomplishing operations are denoted as fuzzy numbers. Moreover, proposed model has ability of
coping with different price discount policies (volume discount pricing) that are offered by network providers in
order to encourage customers to acquire large scale capacity. The problem is modeled from the firms
(customers) point of view and assumes customer is rational decision maker. Although there exist some studies
in current literature related to routing and infrastructure by taking into account fuzzy quality of service, best to
our knowledge there is almost no study about how fuzzy quality of service constraints effect network capacity
acquiring and task allocation policies. To sum up, the aim of this research is suggesting a novel mathematical
model and developing efficient solution algorithm to the capacity acquisition and task allocation problem which
is encountered by firms using network capacity.
2 Literature Review
Fuzzy logic has been applied to the several networking area during the last decade, and we believe that
our study can extend the use of fuzzy logic in this area and particularly in task allocation. It is possible to divide
current literature into two clusters that are related to application of fuzzy QoS parameters in telecommunication
industry. One of these clusters might be considered as routing related researches under the assumption of fuzzy
delay, jitter and package loss. Chen et al. [3] argues that it is unreasonable and inefficient to express the QoS
constraints with deterministic crisp values. They presents new fuzzy genetic algorithm for QoS multicast routing
and simulation experiments to demonstrate the algorithms efficiency. Cohen et al. [2] also tries to solve routing
problem (path selection problem) under uncertainty. They present a fuzzy-based method for path selection under
additive quality of service constraints, where the information available for making routing decisions is
inaccurate. Guerin et al. [4] and Lorenz et al. [5] argue that the QoS constraints usually represented in the formof
crisp limits. However, in actual network environment thiskind of information is quite imprecision because of
non-negligiblepropagation delay of state message, relative longupdate period due to overhead concern, and
hierarchicalstate aggregation. Rai et al. [6] proposes a multi-objective Pareto-optimal technique using Genetic
Algorithm (GA) for group communication, which determines a min-cost multicast tree satisfying end-to-end
delay, jitter, packet loss rate and blocking probability constraints. Their model incorporates a fuzzy-based
selection technique for initialization of QoS parameter values at each instance of multicasting. Another existing
set of literature about fuzzy QoS parameters mostly concentrates on selection or ranking of telecom suppliers or
services under inaccurate or vague data. Robak et al. [7] introduce modeling methodology of the QoS
requirements for e-Business transactions between the service requestorsand the service providers by usage of
fuzzy methods. Wang [8] argues that service selection based on QoS-aware can bring great benefits to service
consumers, as it is able to reduce redundancy in search.
This section of literature review is devoted to articles that are concerned with supplier selection under
existence of fuzzy information. nt et al. [9] develop a supplier evaluation approach based on the analytic
network process (ANP) and the technique for order performance by similarity to ideal solution (TOPSIS)
methods to help a telecommunication company in the GSM sector in Turkey. Kumar et al. [10] presents fuzzy
Multi-objective Integer Programming Vendor Selection Problem formulation that incorporates the three
important goals: cost-minimization, quality-maximization and maximization of on-time-delivery-with the
realistic constraints such as meeting the buyers demand, vendors capacity, vendors quota flexibility, etc. Amid
et al. [11] develops a fuzzy multi objective model is formulated in such away as to simultaneously consider the
imprecision of information and determine the order quantities to each supplier based on price breaks. Amid et al.
[12] suggests a weighted maxmin fuzzy model to handle effectively the vagueness of input data and different
weights of criteria.
519
3
Problem Definition and Proposed Mixed Non-Linear Integer Mathematical
Programming Model with Fuzzy QoS Constraints
In the model, it is assumed that all tasks and operations that have to be performed by firm can be
divided into two discrete sets namely size fixed and time fixed tasks without loss of generality. It is also assumed
that all mentioned tasks independently require different transmission time, bandwidth and quality of service
level. In the mathematical model, QoS level guaranteed by network providers and minimum QoS level which is
needed to accomplishing operations are denoted as triangular fuzzy numbers (TFN). Two important QoS factors,
delay and jitter, are modeled by fuzzy membership functions. Figure 1 shows delay membership functions for
two different telecommunication suppliers.
Objective function of presented model includes two cost terms, cost of acquiring capacity that reflects
different pricing and discount schemes of suppliers and the potential opportunity costs such asthe cost of
degradation in quality (which is more relevant for real time streaming applications), and the QoS switching cost
when the decision maker intentionally downgrades the QoS level or switch to another supplier. The parameters
define network providers are bandwidth, duration and QoS level.
ik :
ik :
cik :
Li :
Ci :
i th supplier
i th supplier
c oj :
wj :
Size of task
r jU :
r jL :
j ( j AS ) in terms of bits
520
j th , respectively
Decision variables:
0, otherwise
0, otherwise
Mathematical Model:
mi
Min
c
iI k 1
mi
x ik (rjU rj ik )c oj yij
ik ik
(1)
iI jAT k 1
Subject to :
w j yij
jAS
yij ~ j ~i
~
y ~
ij
mi
xik ik ik Li
t
r
y
j j ij ik
jAT
k 1
t j yij Li
i I
(2)
i I , j J
(3)
i I , j J
(4)
i I , j AT
(5)
521
mi
xik ik ik 0
r
y
j ijt
j J
k 1
y 1
y t
ij
iI
ijt
iI t Li
r y
j
iI t Li
mi
x
k 1
ik
ijt
wj
1
mi
yij xik 0
i I , t 1,2,, Li
(6)
j J
(7)
j AT
(8)
j AS
(9)
i I
(10)
i I , j J
(11)
k 1
i I , j J , t Li
yijt yij
rj ik y r
L
ij j
rjik yij r
U
j
(12)
i I , j AT , k 1,.., mi (13)
i I , j AT , k 1,.., mi (14)
xik ik Li Ci
k 1
i I
(15)
rj , t j 0
j J
(16)
mi
xik 0,1
yij 0,1
yijt 0,1
i I , k 1,, mi
i I , j J
i I , j J , t Li
(17)
(18)
(19)
Above mathematical models objective function represents cost minimization problem for firms that
buy capacity from telecommunication market. Objective function includes two cost terms. First expenditure term
consists of total capacity acquiring cost by taking into account discounts. Second term reflects opportunity cost
because of not meeting target transmission rate. Constraint set (2) ensures that we can use only up to available
capacity. Fuzzy constraint sets (3) and (4) guarantee that task can be assigned to only resources (suppliers) which
provide lower or equal delay and jitter values, respectively. Constraint (5) ensures that all time fixed tasks
assigned to a resource are completed when the resource is available. Constraint set (6) prevents using more
bandwidth than available at any time. Constraint (7) together with (18) ensures that a task is allocated just only
one resource and all tasks are allocated. (8) Numbered constraint set guarantees that all time fixed jobs actually
placed into required amount of time slice. On the other hand, constraint (9) forces to reverse enough capacity for
size fixed tasks. Constraint set (10) prevents buying more than one bandwidth level from any supplier. Next
constraint (11) enforces assigning tasks only to resources that is already bought. Constraint set (12) ensures that
a task is allocated a network resource only if it is occupies a time slice on it. Constraints (13) and (14) put
limitation on possible minimum and maximum transmission rate of each time fixed task, respectively.
Constraints set (15) ensure not acquiring capacity more than sold by network providers. Rest of the constraints is
used for sign and binary restrictions
522
somehow similar to bin packing formulation. 2-Dimensional Bin packing problems are considered as NP Hard
[13]. Therefore, it can be concluded that suggested cost optimization model for firms is also NP Hard in strong
sense. To sum up, existing commercial optimization software packages do not have ability of solving our model
in reasonable amount of time. Hence, it is essential to develop fuzzy heuristic algorithm in order to solve existing
problem. In the context of possibility theory, there are different approaches to model the coefficients of the
objective function and/or the constraints as fuzzy members. Here the transformation method proposed by Gen et
al. [14] and used by [15] is adopted and extended to transform fuzzy right hand side numbers and fuzzy
technological coefficients numbers in greater than or equal type constraints to crisp constraints.
n
~
(20)
i 1,, m
a~ij x j bi
j 1
~
bi is the fuzzy right hand side term of i-th constraint. Crisp formulation of (20) is given
below.
n
((1 )a
j 1
ij1
aij 2 ) x j (1 )bi1 bi 2
i 1,, m
(21)
where 0 1 is a cutoff value that has to be established parametrically. Therefore, the transformation of the
fuzzy QoS constraints sets (3) and (4) in Section 3 into crisp inequalities are performed by using (21). In order
to solve the problem is settled down parametrically to obtain the value of objective function ( z ) for each one
of those 0,1 . The result is, however, a fuzzy set planner has to decide which pair ( , z ) considers the
optimal if he wants to obtain a crisp solution.
by using (21). Afterwards, sort all tasks first increasing order of j then increasing order of j .
Step 2: Sort the list by decreasing order of sizes of tasks. Call this list J .
Step 3: For each supplier i I calculate delay equivalent of resource i and jitter equivalent of resource i for
given a given value of
by using (21).
Step 5: Sort the list obtained at Step 4 by increase order of effective unit capacity cost cik . Where cik cik / Liik
. Call this list I .
Step 6: Let rj
rjU , j AT . Starting from the first element of J assign tasks to suppliers (bins) again starting
Step 7: After allocation of all tasks, find least used most expensive supplier. Try to move all tasks in this bin into
other already chosen bins by taking opportunity cost and swapping tasks among resources. Calculate the current
total cost by Eq. (1)
Step 8: Repeat Step 7, if it is feasible and economical, otherwise stop.
Table 1 shows the results of a test case where firm has to complete 100 tasks and also there exists 40
different suppliers in the market. Test scenario analyzes indicates that changing value of does not effect the
number of supplier selected. Yet, it effects which supplier is selected. Optimal Solution for decision maker is
observed when is chosen 0.5.
523
# of
Supp
-liers
#
of
Tasks
40
40
40
40
40
40
40
40
40
40
40
100
100
100
100
100
100
100
100
100
100
100
0
0.1
0.2
0.3
0.4
0.5
0.6
0.7
0.8
0.9
1
41.7375
42.2456
42.0878
42.5576
55.1896
39.4625
39.9097
40.9396
40.9205
42.0530
52.7371
246.637
246.637
246.637
246.637
235.015
246.637
246.637
246.637
246.637
246.637
235.015
5
5
5
5
5
5
5
5
5
5
5
Unused
Capacity
(%)
0.12
0.26
0,19
0,26
0,05
0,03
0,16
0,25
0,36
0,39
0,002
Run
Time
(sec.)
3.939
3.699
3.483
3.858
4.965
3.912
3.919
4.190
3.929
3.945
4.934
Acknowledgements
This research is supported by the Scientific and Technological Research Council of Turkey (TBTAK), ShortTerm R&D Funding Program (Grant No. 110 K 609).
References
1.
2.
3.
4.
5.
6.
7.
8.
9.
10.
11.
12.
13.
14.
15.
Kasap, N., Aytu, H., and Ereng, S. S. (2007). Provider Selection and Task Allocation Issues in Networks with
Different QoS Levels and All You Can Send Pricining, Decision Support Systems, 43(2), pp.375-389.
Cohen, A., Korach, E., Last, M., Ohayan, R. (2005). A fuzzy-based path ordering algorithm for QoS routing in
non-deterministic communication networks, Fuzzy Sets and Systems, 150, pp.401-417.
Chen, P., Dong, T. (2003). A fuzzy genetic algorithm for QoS multicast routing, Computer Communications, 26,
pp.506-512.
Guerin, R., Orda, A. (1999). QoS-based routing in networks with inaccurate information: theory and algorithms,
IEEE/ACM Trans. Networking, 7 (3), pp.350364.
Lorenz, D., Orda, A. (1998). QoS-based routing in networks with uncertain parameters, IEEE/ACM Trans.
Networking, 6 (6), pp.768778.
Rai, S., C., Misra, B., B., Nayak, A., K., Mall, R., Pradhan, S., K. (2010). A Multi-Objective QoS optimization
with fuzzy based parameter setting for real-time multicasting, Int. J. Communications, Network and System
Sciences, 3, pp.530-539.
Robak, S., Pieczynski, A. (2008). Fuzzy modeling of QoS for e-business transactions realized by web services,
Journal of Applied Computer Science, 16(1), pp.69-79.
Wang, P. (2009). QoS-aware web services selection with intuitionistic fuzzy set under consumers vague
perception, Expert Systems with Applications, 36, pp.4460-4466.
nt, S., Kara, S. S., Ik, E. (2009). Long term supplier selection using a combined fuzzy MCDM approach: A
case study for a telecommunication company, Expert Systems with Applications, 36, pp.3887-3895.
Kumar, M., Vrat, P., Shankar, R. (2006). A fuzzy programming approach for vendor selection problem in a
supply chain, Int. J. Production Economics, 101, pp.273-285.
Amid, A., Ghodsypour, S., H., Q'Brein, C. (2009), A weighted additive fuzzy multi objective model for the
supplier selection problem under price breaks in a supply Chain, Int. J. Production Economics, 121, pp.323-332.
Amid, A., Ghodsypour, S., H., Q'Brein (2011). A weighted maxmin model for fuzzy multi-objective supplier
selection in a supply chain, Int. J. Production Economics, 131, pp.139-145.
Martello, S., Vigo, D., (1998). Exact solution of the two-dimensional finite bin packing problem, Management
Science, 44(3), pp.388-399.
Gen, M., Tsujimura,Y., Ida, K., (1992). Method for solving multi objective aggregate production planning
problem with fuzzy parameters. Computers and Industrial Engineering, 23, pp.117120.
Mula, J., Peidro, D., Poler, R., (2010). The effectiveness of a fuzzy mathematical programming approach for
supplychain production planning with fuzzy demand. . International Journal of Production Economics, 128,
pp.136143.
524
Abstract. SWOT (Strengths, Weaknesses, Opportunities and Threats) analysis is a commonly used and an
important technique for analyzing internal and external environments in order to provide a systematic
approach and support for a decision making. SWOT is criticized mostly for considering only qualitative
examination of environmental factors, no priority for various factors and strategies, and no vagueness of the
factors under fuzziness. In this paper, fuzzy TOPSIS (Technique for Order Preference by Similarity to Ideal
Solution) integrated with fuzzy analytical hierarchy process (AHP) is used to develop fuzzy multi-criteria
SWOT analysis in order to overcome these shortcomings.
1 Introduction
In SWOT analysis is a commonly used and important technique for analyzing internal and external
environments in order to provide a systematic approach and support for a decision making [1], [2], [3], [4], [5].
This approach involves systematic thinking and comprehensive diagnosis of factors related to a new product,
technology, management, or planning [6]. SWOT, commonly known in strategic management area, analyzes the
opportunities and threats identified by an internal environment appraisal as well as the strengths and weaknesses
by an external environment appraisal [7], [8]. The external and internal environments include variables which are
outside and inside the organization, respectively. The top management of the organization cannot affect either
type of the variable for a short-term [9].
SWOT, an early stage of the strategic planning process, has the goal of developing and adopting a strategy
resulting in a good fit between internal and external factors [4]. SWOT maximizes strengths and opportunities,
and minimizes threats and weaknesses. In other words, it transforms weaknesses into strengths, and threats into
opportunities [10], [11]. SWOT can also be used when strategy alternative suddenly arises and the decision
context relevant to it has to be analyzed [11].
In a conventional SWOT analysis, the magnitude of the factors is not quantified to determine the effect of
each factor on the proposed plan or strategy [12]. In other words, SWOT analysis does not provide an analytical
means to determine the relative importance of the factors, or the ability to assess the appropriateness of decision
alternatives based on these factors [13]. A new hybrid method was presented in [4] that eliminates the
weaknesses in the measurement and evaluation steps of the SWOT analysis and this technique involves the
utilization of the analytic hierarchy process (AHP) in the SWOT analysis and has been referred to as AWOT in
subsequent studies [13], [14]. The quantified SWOT analytical method adapted to the Multi-Attribute Decision
Making concept is used in [15] and AHP and a multi-layer scheme for the simplification of the complex
525
problems are utilized [7]. Furthermore analytical network process (ANP) in a SWOT analysis was suggested in
[16].
While the proper use of SWOT is a good basis for strategy formulation [13], it still has certain structural
problems. The majority of papers assume that the demand is deterministic, but in reality this assumption is not
true [17]. An important problem is the lack of considering uncertain and two sided factors and there is a lack of
prioritization of the factors and strategies and there are too many extractable strategies.1 Despite its wide
applications, in [18] mention seven problems of the SWOT method and most important ones are as follows:
Usually only qualitative examination of environmental factors is considered [19];
It considers no priority for various factors and strategies;
It does not consider the vagueness of the factors.
In this paper, a model is proposed for quantified SWOT under fuzziness. After internal and external criteria
are determined, Changs fuzzy AHP [20] is utilized to determine the weight vector of each type of criteria. Later,
by using the linguistic scores of each strategic alternative for both internal and external criteria, and the weight
vector of each type of criteria, Chens fuzzy TOPSIS [21] is utilized. According to the ranking results the degree
of internal dimension scores from weakness to strength and the degree of external dimension scores from threats
to opportunities are obtained. At last, the strategic alternative is selected with the largest strength and opportunity
scores.
The rest of the paper is organized as follows. In Section 2, Literature Reviews of Fuzzy SWOT, Fuzzy AHP
and Fuzzy TOPSIS are expressed. In Section 3, fuzzy multi-criteria SWOT analysis is modeled. Finally,
conclusions are given.
2 Literature Review
2.1 Fuzzy SWOT
In SWOT analysis some uncertainties can be encountered in the evaluation of an organizations internal and
external environment. It can be stated that it is impossible for the opportunities and threats arising from the
external environment of the firm to be always definite in any condition. Consequently, it is also impossible to
measure the numerical values precisely. Similarly, it is not always possible to measure or evaluate the strengths
and weaknesses which are the consequences of the assets and abilities of the firm exactly [1]. Besides, in SWOT
analysis, it is difficult to evaluate the factors in any case and at anytime in a binary based Aristotle logic. For
instance, it may not be realistic to represent the image about the goods and services produced by a firm in two
points as sufficient (1) or insufficient (0). In real life, there may be different evaluations between these two.
Therefore, it will not be possible to understand the actual situation with an evaluation which does not assume
fuzziness [1].
In [1] it is emphasized that in most cases the internal and external factors cannot be fully recognized as
positive or negative, as their impact on the organization could be observed within a wide spread which may
include both positive and negative effects and in [1] a fuzzy SWOT algorithm considering uncertain and twosided factors, and prioritization of that factors and strategies are presented. Strategies to reduce the dangers of
transporting hazardous material in Iran are made in [22] by using the method of fuzzy SWOT analysis. In [23] a
fuzzy quantified SWOT analytical procedure that integrates the MCDM concept and fuzzy method to help
decision makers assess the competitive position and degree of locations developing internal distribution centers
(IDC) is proposed. By using this method the environmental position of locations developing IDCs can be judged
according to their environmental position and competitive degree in four quadrants, and competitive strategies
further developed to strengthen competitive advantage in accordance with the grand strategy matrix (GSM)
model. A new fuzzy quantified SWOT procedure that integrates the fuzzy ANP method to help decision makers
assess the competitive position and degree of each transportation industry developing in Taiwans transportation
market is presented in [24]. The proposed model can handle the criteria characteristics, which may be
independent, dependent and interdependent. A two phased decision model for supplier selection is proposed in
[7]. In first phase, fuzzy quantified SWOT analysis is applied for evaluating suppliers. The output of this stage is
526
the weight of each supplier. In second phase a fuzzy linear programming model is applied to determine the order
quantity.
2.2 Fuzzy AHP
AHP is one of the well-known multi-criteria decision making techniques that was first proposed by Saaty
(1980) in [25]. The classical AHP takes into consideration the definite judgments of decision makers [26].
Although the classical AHP includes the opinions of experts and makes a multiple criteria evaluation, it is not
capable of reflecting humans vague thoughts [27].
As the uncertainty of information and the vagueness of human feeling and recognition, it is difficult to provide
exact numerical values for the criteria and to make evaluations which exactly convey the feeling and recognition
of objects for decision makers. Therefore, most of the selection parameters cannot be given precisely. Thus
experts may prefer intermediate judgments rather than certain judgments. So the fuzzy set theory makes the
comparison process more flexible and capable to explain experts preferences [28].
Different methods for the fuzzification of AHP have been proposed in the literature. AHP is firstly fuzzified in
[29], and in this study, fuzzy ratios which were defined by triangular membership functions were compared. The
comparison ratios based on trapezoidal membership functions are used in [30]. A new approach was introduced
in [20] for handling fuzzy AHP, with the use of triangular fuzzy numbers for pair-wise comparison scale of
fuzzy AHP, and the use of the extent analysis method for the synthetic extent values of the pair-wise
comparisons. A fuzzy objective and subjective method based on fuzzy AHP was proposed in [31]. A selection
among the transportation companies is analyzed in [32] by using fuzzy axiomatic design and fuzzy AHP. They
developed a fuzzy multi-attribute axiomatic design approach and compared it with fuzzy AHP.
2.3 Fuzzy TOPSIS
TOPSIS, one of the classical Multi-criteria decision making methods, was developed by Hwang and Yong
(1981) in [33]. It is based on the concept that the chosen alternative should have the shortest distance from the
positive ideal solution (PIS) and the farthest from the negative ideal solution (NIS). TOPSIS also provides an
easily understandable and programmable calculation procedure. It has the ability of taking various criteria with
different units into account simultaneously [34].
A number of fuzzy TOPSIS methods have been developed in recent years. Fuzzy numbers to establish fuzzy
TOPSIS was first applied in [35]. A fuzzy TOPSIS method developed in [36] where relative closeness for each
alternative is evaluated based on fuzzy arithmetic operations. TOPSIS method is extended to fuzzy group
decision making situations by considering triangular fuzzy numbers and defining crisp Euclidean distance
between two fuzzy numbers in [21]. The methodology proposed by Chen (2000) is further improved in some
studies [37], [38]. In addition the fuzzy TOPSIS method is extended based on alpha level sets with interval
arithmetic [39], [40].
Fuzzy TOPSIS has been introduced for various multi-attribute decision-making problems. Fuzzy TOPSIS is
used for plant location selection in [41] and for supplier selection in [42]. Fuzzy TOPSIS also is utilized for
industrial robotic system selection [43]. A modified fuzzy TOPSIS to select municipal solid waste disposal
method and site is used in [34]. Furthermore fuzzy TOPSIS integrated with fuzzy AHP is used to propose a new
FMEA (failure modes and effects analysis) which overcomes the shortcomings of traditional FMEA [44].
Another modified fuzzy TOPSIS for selection of the best energy technology alternative is proposed in [45].
Fuzzy TOPSIS is used for modeling consumers product adoption process [46].
3 Methodology
To overcome the shortcomings of SWOT analysis, a fuzzy multi-criteria SWOT approach is proposed in this
paper. Firstly, internal and external criteria are determined as it was in traditional SWOT analysis. Then, after
appropriate linguistic scores of each criterion are determined, Changs fuzzy AHP is utilized to determine the
527
weight vector of each type of criteria. Later, by using the linguistic scores of each strategic alternative for both
internal and external criteria, and the weight vector of each type of criteria, Chens fuzzy TOPSIS is utilized.
According to the ranking results the degree of internal dimension scores from weakness to strength and the
degree of external dimension scores from threats to opportunities are obtained. At last, the strategic alternative is
selected with the largest strength and opportunity scores.
The proposed fuzzy SWOT model is illustrated by Fig. 1.
Best strategic
alternative
selection
Experts
Weights
obtained by
Fuzzy AHP I
Experts
Internal
Criteria
External
Criteria
Fuzzy
TOPSIS I
Fuzzy
TOPSIS II
SW
Dimension
OT
Dimension
Weights
obtained by
Fuzzy AHP II
SWOT
Analysis
528
Opportunity
Alternative II
Alternative I
Strength
Weakness
Alternative III
Alternative I
Threat
4 Conclusion
SWOT, an early stage of the strategic planning process, has the goal of developing and adopting a strategy
resulting in a good fit between internal and external factors. It is a commonly used and important technique for
analyzing internal and external environments in order to provide a systematic approach and support for a
decision making. In other words, it aims at transforming weaknesses into strengths, and threats into
opportunities.
As SWOT is criticized mostly for considering only qualitative examination of environmental factors, no
priority for various factors and strategies, and no vagueness of the factors, in this paper in order to overcome
these shortcomings a fuzzy multi-criteria SWOT is proposed. The proposed model consists of three parts: first,
utilizing Changs fuzzy AHP to determine the weight vector of internal and external criteria, second, utilizing
fuzzy TOPSIS to obtain the ranking results the degree of internal dimension scores from weakness to strength
and the degree of external dimension scores from threats to opportunities, and third, determining the best
strategic alternative by evaluating the internal and external dimensions.
For further research, we suggest other multi-criteria methods like ELECTRE, VIKOR, or Utility Models to
be used and the results of different methodologies to be compared.
References
1.
2.
3.
4.
5.
6.
7.
8.
Ghazinoory, S.; Zadeh, A.E.; and Memariani, A. (2007). Fuzzy SWOT Analysis. Journal of Intelligent &
Fuzzy Systems 18: 99108.
Kangas, J.; Kurtila, M.; Kajanus, M.; and Kangas, A. (2003). Evaluating the management strategies of a
forestland estate-the S-O-S approach. Journal of Environmental Management 69: 349358.
Kotler, P. (1998). Marketing Management: Analysis, Planning, Implementation and New Jersey:
Control Prentice-Hall.
Kurtila, M.; Pesonen, M.; Kangas, J.; and Kajanus, M. (2000). Utilizing the analytic hierarchy process
(AHP) in SWOT analysis-a hybrid method and its application to a forest-certification case. Forest
Policy and Economics 1: 4152.
Stewart, R.; Moamed, S.; and R. Daet (2002). Strategic implementation of IT/IS projects in
construction: A case study. Automation in Construction 11: 681694.
Weihrich, H., (1982). The TOWS matrixa tool for situation analysis. Long Range Planning 15(2): 5466.
Amin, S.H.; Razmi, J.; and Zhang, G. (2011). Supplier selection and order allocation based on fuzzy
SWOT analysis and fuzzy linear programming. Expert Systems with Applications 38: 334342.
Dyson, R. G. (2004). Strategic development and SWOT analysis at the University of Warwick,
European Journal of Operational Research 152: 631640.
529
9.
10.
11.
12.
13.
14.
15.
16.
17.
18.
19.
20.
21.
22.
23.
24.
25.
26.
27.
28.
29.
30.
31.
32.
33.
34.
35.
Houben, G.; Lenie, K.; and Vanhoof K. (1999). A knowledge-based SWOT-analysis system as an
instrument for strategic planning in small and medium sized enterprises, Decision Support Systems 26:
125135.
Arslan, O.; and Deha, I.E. (2008) SWOT analysis for safer carriage of bulk liquid chemicals in tankers,
Journal of Hazardous Materials 154: 901913.
Christensen, R.; Berg, N.; and Salter, M. (1976) Policy formulation and administration (Homewood:
Richard D. Irwin).
Masozera, M.K.; Alavalapati, J.R.R.; Jacobson, S.K.; and Shresta, R.K. (2006). Assessing the
suitability of community-based management for the Nyungwe Forest Reserve, Rwanda Forest Policy
and Economics 8: 206216.
Kajanus, M.; Kangas, J.; and Kurtila, M. (2004). The use of value focused thinking and the AWOT
hybrid method in tourism management, Tourism Management 25: 499506.
Leskinen, L. A.; Leskinen, P.; Kurttila, M.; Kangas, J.; and Kajanus, M. (2006) Adapting modern
strategic decision support tools in the participatory strategy process - A case study of a forest research
station, Forest Policy and Economics 8: 267278.
Chang, H. H.; and Huang, W.C. (2006). Application of a quantification SWOT analytical method,
Mathematical and Computer Modeling, 43: 158169.
Yksel, .; and Dadeviren, M. (2007). Using the analytic network process (ANP) in a SWOT analysis
A case study for a textile firm. Information Sciences 177: 33643382.
Snyder, L. V. (2006). Facility location under uncertainty: A review, IIE Transactions 38: 537554.
Hill, T.; and Westbrook, R. (1997). SWOT Analysis: Its Time for a Product Recall. Long Range
Planning 30(1): 4652.
Buyukozkan, G.; and Feyzioglu, O. (2002). A fuzzy logic based decision making approach for new
product development. International Journal of Production Economics 90: 27-45.
Chang, D.Y. (1996). Applications of the extent analysis method on fuzzy AHP. European Journal of
Operational Research 95(3): 649655.
Chen, C. (2000). Extensions of the TOPSIS for group decision-making under fuzzy environment. Fuzzy
Sets and Systems 114: 1-9.
Kheirkhah, A. S.; Esmailzadeh, A.; and Ghazinoory, S. (2009). Developing strategies to reduce the Risk
of hazardous materials transportation in Iran using the method of fuzzy SWOT analysis, Transport 24:
325-332.
Lee, K.; and Lin, S. (2008). A fuzzy quantified SWOT procedure for environmental evaluation of an
international distribution center. Information Sciences 178: 531549.
Kuo, M. S.; Tzeng, G. H.; and Chen, S. Y. (2009). A new fuzzy quantified SWOT procedure for
environmental evaluation of transportation markets, in Proceedings on 17th Symposium on Fuzzy
Theory and Its Applications.
Saaty, T. L. (1980).The Analytic Hierarchy Process. New York: McGraw-Hill.
Wang, T. C.; and Chen, Y. H. (2007). Applying consistent fuzzy preference relations to partnership
selection, International Journal of Management Science 35: 384388.
Seme, N. Y., Bayrakdarolu, A.; and Kahraman, C. (2009). Fuzzy performance evaluation in Turkish
Banking Sector using Analytic Hierarchy Process and TOPSIS, Experts Systems with Applications 36:
11699-11709.
Kahraman, C.; Cebeci, U.; and Ulukan, Z. (2003). Multi-criteria supplier selection using fuzzy AHP,
Logistics Information Management 16: 382-394.
Laarhoven, P. M. J.; and Pedrycz, W. (1983). A fuzzy extension of Saatys priority theory, Fuzzy Sets
and Systems 11: 229-241.
Buckley, J. J. (1985). Fuzzy hierarchical analysis, Fuzzy Sets and Systems 17: 233247.
Kahraman, C.; Ulukan, Z.; and Tolga, E. (1998). A fuzzy weighted evaluation method using objective
and subjective measures, in Proceedings of the international ICSC symposium on engineering of
intelligent systems 5763.
Kulak, O. and Kahraman, C. (2005). Fuzzy multi attribute selection among transportation companies
using axiomatic design and analytic hierarchy process, Information Sciences 170: 191210.
Hwang, C. L.; and Yoon, K. (1981). Multiple attributes decision making methods and applications
Berlin: Springer.
Ekmekiolu, M.; Kaya, T.; and Kahraman, C. (2010). Fuzzy Multi-criteria Disposal Method and Site
Selection for Municipal Solid Waste, Waste Management 30: 1729-1736.
Chen, S. J.; and Hwang, C. L. (1992). Fuzzy multi attribute decision making (lecture notes in
economics and mathematical system series) New York: Springer-Verlag.
530
36. Triantaphyllou, and Lin, C. L. (1996). Development and evaluation of five fuzzy multi- attribute
decision making methods, International Journal of Approximate Reasoning 14: 281310.
37. Chu, T. (2002). Selecting plant location via a fuzzy TOPSIS approach, International Journal of
Advanced Manufacturing Technology 20: 859864.
38. Chu, T; and Lin, Y. (2002). Improved extensions of the TOPSIS for group decision making under fuzzy
environment, Journal of Information and Optimization Sciences 23: 273286.
39. Jahanshahloo, G. R.; Hosseinzadeh, L.; and Izadikhah, M. (2006). Extension of the TOPSIS method for
decision-making problems with fuzzy data, Applied Mathematics and Computation 181: 15441551.
40. Chu, T.C.; and Lin, Y. C. (2009). An interval arithmetic based fuzzy TOPSIS model. Expert Systems
with Applications 36: 1087010876.
41. Yong, D. (2006). Plant Location Selection Based on Fuzzy TOPSIS, International Journal of Advanced
Manufacturing Technologies 28: 839-844.
42. Chena, C.-T.; Lin, C.-T.; and Huangb, S.-F. (2006). A fuzzy approach for supplier evaluation and
selection in supply chain management, International Journal of Production Economics 289301.
43. Kahraman, C.; evik, S.; Ate, N. Y.; and Glbay, M. (2007). Fuzzy multi-criteria evaluation of
industrial robotic systems, Computers & Industrial Engineering 52: 414433.
44. Kutlu A. C.; and Ekmekiolu, M. (2010) Fuzzy Failure Modes & Effects Analysis by Using Fuzzy
TOPSIS Integrated with Fuzzy AHP, in 1th International Symposium on Computing in
Science&Engineering, Kuadas, Aydn, Turkey.
45. Kaya, T.; and Kahraman, C. (2011). Multi-criteria decision making in energy planning using a
modified fuzzy TOPSIS methodology, Expert Systems with Applications 38: 65776585.
46. Kim, S.; Lee, K.; Cho, J. K.; and Kim, C. O. (2011). Agent-based diffusion model for an automobile
market with fuzzy TOPSIS-based product adoption process, Expert Systems with Applications 38:
72707276.
Biographies
Mehmet Ekmekiolu was born in Istanbul in 1984 and graduated from Istanbul Technical University,
Mechanical Engineering Department at 2006. He completed the Master Program of Industrial Engineering at
ITU at 2009. Since then, he has been studied at Industrial Engineering PhD Program of ITU.
His research interests are engineering economy, multi criteria decision making, production planning, quality
processes, and strategic planning.
Mr. Ekmekiolu worked as a research assistant at Management Engineering Department at the faculty of
Management at ITU between May 2007 and March 2011. He has been working as a production planning
engineer at FORD OTOSAN since April 2011.
Ahmet Can Kutlu was born in Istanbul / Turkey in 1981 and graduated in electrical engineering at Istanbul
Technical University (ITU) in 2004. After the graduation in engineering management master program at the
institute of science and technology at ITU in 2007, he began his PhD studies in industrial engineering program at
the same institute at the same university in Istanbul / Turkey.
His research interest is about the management issues and methods concerning the fields of strategy, quality,
energy and human resources.
Mr. Kutlu has been working as a research assistant in the department of Industrial Engineering at the faculty
of Management at Istanbul Technical University since 2005
Cengiz Kahraman was born in Istanbul / Turkey in 1965 and graduated in industrial engineering at Istanbul
Technical University (ITU) in 1988. Then he graduated from industrial engineering master and PhD program at
the institute of science and technology at ITU in 1990 and 1996 respectively. He became assistant professor in
1996 and associate professor in 1997. He has been a full time professor at the department of Industrial
Engineering of Istanbul Technical University since 2003.
His research areas are engineering economics, quality control, statistical decision making, multicriteria
decison making, and fuzzy applications in industrial engineering. He published about 100 papers in
international journals and four edited Springer books. He is the editorial board member of nine international
journals. He has guest-edited many special issues of international journals.
531
Fatih University, Faculty of Engineering, Department of Industrial Engineering, 34500, Istanbul, Turkey
1
aliiseri@fatih.edu.tr, 2rkizilaslan@fatih.edu.tr
Abstract. The aim of this study is to evaluate the effectiveness of fuzzy logic and artificial neural networks
(ANN) in predicting movements in Istanbul Stock Exchange (ISE) 100 index (Symbol: XU100) for
03.01.200525.06.2010 period. The performance of both techniques is evaluated using different metrics. Also
performances are compared with some basic techniques such as moving average and regression analysis. A
comprehensive evaluation procedure is described, involving the use of trading simulations to assess the
practical value of predictive models. In the experiments presented here, trading decisions based on fuzzy
logic resulted in a hit ratio of 61.1% and an average return on investment of 2.12% per month, while trading
decisions based on ANN resulted in a hit ratio of 61.4% and an average return on investment of 1.88% per
month during a period when the ISE 100 index grew by 1.22% per month.
Keywords: Stock market prediction, Istanbul Stock Exchange (ISE), Forecasting, Fuzzy Logic, Artificial
Neural Network (ANN)
1 Introduction
It is a well-known fact that forecasting stock returns is difficult because of market volatility that needs be
captured in used and implemented models. These financial time series behave like a random walk and several
studies have shown that their serial correlation is economically and statistically insignificant [1]. So, most of the
traditional approaches such as regression and time series analysis are not sufficient to predict their movement.
Fuzzy decision support system is a special rule-based system that uses fuzzy logic instead of Boolean logic in its
knowledge-base and derives conclusions from user inputs and fuzzy inference process [2] while fuzzy rules and
the membership functions make up the knowledge-base of the system. On the other hand, artificial neural
networks (ANN) incorporate a large number of parameters which allows learning the intrinsic non-linear
relationship presented in time-series, increasing their forecasting capabilities [3], [4]. Numerous studies have
shown the successes and failures of expert systems such as fuzzy logic and ANN in forecasting financial data.
Studies generally showed that these systems outperform the traditional forecasting techniques, such as
regression, moving average, or auto-regressive models [5], [6], [7]. An excellent survey of the literature of soft
computing methods for predicting stock market is provided by Atsalakis & Valavanis [8].
Numerous studies applied expert decision systems to stock price returns in the developed markets of the United
States or Europe. There is, however, little research in the literature on the value of expert systems to predict
stock price movements in emerging markets. And there is evidence that emerging markets behave differently
than the developed markets [9].
532
Research on the Istanbul Stock Exchange based on expert systems is also limited. Some important studies are by
Yumlu et al. [10] and Bildirici & Ersin [11]. Decisions on these two studies are based on ANN. Unfortunately,
there is no study based on fuzzy logic for predicting ISE 100 index.
The Istanbul Stock Exchange was established on December 26, 1985 for the purpose of ensuring that securities
are traded in a secure and stable environment, and commenced to operate on January 3, 1986. The ISE has
contributed to the development of Turkish capital markets and Turkish economy since the date of its
establishment. ISE National 100 index (XU100) is an overall composite index that has been calculated on daily
basis since October, 1987. Although there are many other indices that are calculated by ISE, this index is the
most widely used by many to capture the overall market performance.
The basic aim of this study is to predict movements in XU100 for 03.01.200525.06.2010 period using fuzzy
logic and (ANN). However, we also used other traditional techniques such as regression analysis to compare the
performance of expert systems with the traditional methods. So performance comparisons are done both with
traditional methods and other expert systems. The study is organized as dataset compilation, performance
criteria, model construction, results and conclusion.
2 Dataset Compilation
For the purposes of this work, a dataset has been compiled containing daily minimum, maximum, opening and
closing values of XU100, daily transaction volume of ISE, Dollar/TL and Euro/TL rates, gold price (TL/kg),
daily bond return, daily closing value of NasDaq composite index, an indicator that captures the trend in XU100
and day factor. The dataset covers a period from 03.01.200525.06.2010. From this dataset daily percentage
returns are calculated and used in the constructed models as inputs and target. Trend factor is used for capturing
the history of the index. It tracks the up and down trends in the XU100. Day factor is another factor that is used
to investigate the influence of days. We used 1 for Monday, 2 for Tuesday, etc.
3 Performance Criteria
In this study, three performance criteria are used: Root mean squared error (RMSE), hit rate (directional
success), return on investment (ROI).
RMSE is the optimized criteria in ANN applications, although from an end-users point of view it is quite
abstract especially for predicting stock movements.
Instead, two other practical performance criteria are used: Hit rate and ROI.
Hit rate is the directional success of the proposed model. It is the ratio of the correct predictions of upward and
downward movements. It is an important metric as buy/sell decisions will be based on predictions that the index
will rise or fall.
ROI is computed by basing trading decisions on the output of the model. A simple trading strategy is proposed
here: No Threshold All In/Out. All trades are conducted at the start of business (i.e., at the opening value of
XU100), and the models predict its closing value. If the market is predicted to rise by any amount (no threshold)
this signals a buy, while if it is predicted to fall this signals a sell. We start with an initial capital amount of 1000
TL. Then, the first day that a buy signal is received from the model, an investment is made in the idealized
XU100 fund using the full capital amount. Buy signals on subsequent days are treated as hold instructions until a
sell signal is received. When all investment units currently held are sold at the current price after a sell signal, we
yield a capital amount. Subsequent days sell signals are again treated as non-trade instructions, until another buy
is received, and the process is repeated. The overall ROI over the investment period is the percentage gain in
investment capital. We also account for transaction charges. We assume a charge of 0.02% which is the rate of
533
one of the reputable online trading services.1 Similar trading strategies are used in many papers such as
OConnor & Madden (2006) [12].
4 Model Construction
While constructing our models, we used daily percentage returns instead of absolute values. For instance, instead
of taking the closing value of NasDaq composite index, we take the daily percentage return of NasDaq
composite index. This is the same for dollar, euro, and gold. Also we add the daily returns of dollar and euro
together making them one factor.
(1)
http://www.sankomenkul.com/
534
Nas Daq
Daily
Return
Trend ISE
Next Days
Daily
Return of
ISE
0,47
Stationary
Stationary
Negative
0,21
0,34
Stationary
Positive
Positive
0,22
Negative
0,19
Stationary
Positive
Stationary
0,57
Stationary
Positive
0,43
Stationary
Positive
Negative
0,21
Positive
Stationary
Stationary
0,49
Negative
Negative
Positive
0,12
Positive
Stationary
Negative
0,08
Negative
Negative
Stationary
0,43
Positive
Positive
Positive
0,4
Negative
Negative
Negative
0,45
Positive
Positive
Stationary
0,47
Negative
Stationary
Positive
0,15
Positive
Positive
Negative
0,13
Negative
Stationary
Stationary
0,37
Stationary
Negative
Positive
0,36
Negative
Stationary
Negative
0,48
Stationary
Negative
Stationary
0,39
Negative
Positive
Positive
0,11
Stationary
Negative
Negative
0,25
Negative
Positive
Stationary
0,43
Stationary
Stationary
Positive
0,33
Negative
Positive
Negative
0,46
Stationary
Stationary
Stationary
0,46
Nas Daq
Daily
Return
Trend ISE
Positive
Negative
Positive
Positive
Negative
Stationary
Positive
Negative
Positive
Weight
535
Weight
To develop a knowledge based rule construction system most of the multiple experts are used but it is very
difficult to integrate knowledge from these experts. So in construction of fuzzy rules we have followed a
different method. As mentioned before in fuzzification step, by using the input and output variables
membership functions, number of incidence of all possible cases are determined. For example, there were
potentially 33=27 rules because each rule is consisted of three variables, and all variables are evaluated with
three linguistic variables: Positive, Stationary and Negative. All of the possible cases were observed one by one
and then weights of the rules determined according to the percentage occurrence of rules. List of the all possible
cases and weights of these rules are shown in Table 1.
Compositions of the input values are guided by the fuzzy rules that formed before. Standard max-min
inference method was used in this process which is most commonly used strategy in fuzzy composition step.
The process of computing a single number which represents the best outcome of fuzzy set evaluation is called
defuzzification [14]. We choose the center of gravity method which is most frequently used in defuzzification
step.
5 Results
In this paper we have used different performance criteria like RMSE, hit rate and average ROI. However the
most importing criteria are the hit rate which shows the directional success of our decisions and ROI which
shows the return of investment based on our trading strategy.
Table 2 shows the performance of each method in these 3 criteria based on our proposed No Threshold All
In/Out. As seen in Table 2, five days moving average model is not suitable for forecasting of stock exchange
values. Not only hit rate but also average ROI and RMSE values are worse than the other techniques. While we
look at the performance of other techniques it could be said that hit ratios and ROI values are approximately
same, but it is not. Because while the hit rates are approximately same, frequency of these up or down values
may differ.
536
No action
-----
3.05
49.9
0.36
1268
Regression
1.87
60.9
1.94
3555
Fuzzy Logic
2.43
61.1
2.12
3995
ANN
1.89
61.4
1.88
3418
100
16.9
30,215,979
Oracle
So it is also important to consider that the loss when predicting an upward trend as a downward one is not the
same with the loss when predicting a downward trend as an upward one. In the first case the investor suffers a
loss; in the second case the investor, if he has allocated assets to risk-free government bonds, will still have some
positive returns. Further, correct prediction of the stock price direction does not refer to the magnitude of the
stock price movement. This means that gains from correct predictions and losses from incorrect predictions may
vary [15]. From this point of view for example an indicative amount of 1000 TL (Turkish Lira) is allocated as
an initial investment in 03.01.2005. At the end of the 25.06.2010 date the net gain asset will be 1268 TL in
moving average, 3555 TL in regression analysis, 3418 TL in ANN and 3995 TL in fuzzy logic.
6 Conclusion
It is very complicated decision to buy or sell a stock since there are too many factors that affects the stock
price values. Different forecasting techniques were put in practice in this paper and the study shows that the
performance of the stock price forecasting can be significantly done by using ANN and fuzzy logic. The best and
most important difference of this paper that only two factors, NasDaq and trend, have been taken as an input to
forecast the ISE stock exchange values while most of the previous in general, the average number of input
variables are between four and ten. It is very important to decrease the number of input variables in order not to
increase the complexity of the model and decrease the time of calculation. And also while increasing the number
of variables it will also be difficult to reach the variables. But this paper enables us to not only forecast ISE stock
exchange values with only two input variables but also better forecasting performance from the previous studies.
References
1.
2.
3.
4.
5.
6.
7.
G. Hawawini, D.B. Keim, On the predictability of common stock returns: world-wide evidence, in:
R.A. Jarrow, V. Maksimovic, W.T. Ziemba (Eds.), Handbook in Operations Research and Management
Science, vol. 9, Finance, North-Holland, Amsterdam, 1995, pp. 497544, Chapter 17.
Kandel,A.(1992). Fuzzy expert systems. BocaRaton,FL:CRCPress.
Haykin, S. (2001). Neural networks, principles and practice.
Specht D., Probabilistic neural networks. Neural Networks 1990;3:10918.
Refenes, A. N., Zapranis, A., & Francis, G. (1994). Stock performance modeling using neural
networks: A comparative study with regression models. Neural Networks, 7(2), 375388
Kuan, C., & Lim, T. (1994). Forecasting exchange rates using feed forward and recurrent neural
networks. Journal of Applied Econometrics, 10, 347364.
Haefke, C., & Helmenstein, C. (1996). Forecasting Austrian IPOs: an application of linear and neural
network error-correction models. Journal of Forecasting, 15, 237251.
537
8.
9.
10.
11.
12.
13.
14.
15.
Atsalakis, G.S., Valavanis, K.P. (2009), Surveying stock market forecasting techniques Part II: Soft
computing methods. Expert Systems with Applications 36, 59325941
Bhattacharya U, Daouk H, Jorgenson B, Kehr C-H. When an event is not an event: the curious case of
an emerging market. Journal of Financial Economics 2000;55:69101.
Yumlu, S., Gurgen, F. G., & Okay, N. (2005). A Comparison of global, recurrent and smoothedpiecewise neural models for Istanbul Stock Exchange prediction. Pattern Recognition Letters, 26, 2093
2103
Bildirici, M., Ersin, .., (2009) Improving forecasts of GARCH family models with the articial
neural networks:An application to the daily returns in Istanbul Stock Exchange, Expert Systems with
Applications 36, 73557362
OConnor, N., Madden, M.G., (2006) A neural network approach to predicting stock exchange
movements using external factors. Knowledge-Based Systems 19, 371378
Fasanghari M., Mantezar G.(2010). Design and Implementation of Fuzzy Expert System for Tehran
Stock Exchange Portfolio Recommendation, Expert System with Applications 37 (2010) 6138-6147
Ngai,E.W.T.(2003).Design and development of a fuzzy expert system for hotel selection. Omega(The
International Journal of Management Science), 31, 275286.
Atsalakis, G.S., Valavanis, K.P. (2009), Forecasting stock market short-term trends using a neuro-fuzzy
based methodology. Expert Systems with Applications 36, 1069610707
538
Abstract. Forecasting of intermittent demand is one of the most important problem in spare parts inventory
systems. Demand pattern of spare parts is far out from fast moving items. Intermittent demand series occurs
when a demand series has a significant portion of the periods with no demand and demand size variability is
very high. This makes it difficult to forecast accurately. However, many of the traditional forecasting methods
perform poorly when demand pattern is intermittent. Because of this it can be reasonable to apply Artificial
Neural Network(ANN) in terms of considering properties and nonlinear relationships of data.
In this study, several ANN forecasting techniques were used by defining different types of inputs based on
the characteristic of spare parts demand patterns. Our study compares the performance of ANN to those
traditional
methods
like;
exponential
smoothing,
Crostons(Syntetos&Boylan)
and
Crostons
(Leven&Segersted) methods.
The models were applied to different types of real data sets of a major airline company in Turkey. After
evaluating the performance of models based on MSE criteria, it shows that our proposed method is better than
other.
Keywords: Forecasting,Intermittent demand, Neural network modelling
1 Introduction
In most companies, customers are demanding faster service in competitive environment so planning activites
based on forecasts becoming more important to enable the availability of spare parts in manufacturing and
service systems. Demand for spare parts show irregular pattern with a high degree of uncertainty as to when and
at what level demand will occur. Managing uncertainty when the demand is non smooth is a complex issue as
well as an unexplored one. Forecasting of non-smooth demand items is very difficult because demand has
various numbers of periods with no demand and the variability of demand size is high. For this reason,
traditional forecasting methods are insufficient for non smooth demands and forecasting this type of demand
requires special techniques since the assumptions for continuity and normal distribution do not hold.
Intermittend demand patterns are very common, particularly in organizations that hold many spare parts. In the
aerospace and automotive sectors, organizations may have thousands or tens of thousands of stock keeping units
(SKUs) classified as intermittend [1]. If interarrival time between nonzero demands is very high, keeping these
items in inventory consistently could cause holding costs. So managing spare parts is a critical operational issue
in manufacturing and service industry because of small improvements may be translated to substantial cost
savings.
539
When demand is non smooth, traditional forecasting methods may perform inadequately in capturing the
nonlinear pattern in data. It would be better to apply ANN models which exhibit good performance of modeling
non-linear and complex systems.
In literature, there are several methods such as exponential smoothing, regression etc. methods have been used
several decades in forecasting applications. These methods give good results when the demand pattern is smooth.
However, many of these methods perform poorly when demand pattern is non-smooth.
Exponential smoothing method is most commonly used method in forecasting applications but Croston (1972)
noted that this method is not appropriate when demand pattern of items is intermittend. He defined a bias that
gives most weight on the most recent demand date, proposed a method that uses average size of nonzero
demand occurances and average interval between these demand values. Syntetos and Boylan (2005) introduced a
new correction factor (1-/2) and applied it to the Crostons method, where is smoothing constant used for
updating interdemand intervals [9]. A different and independently developed method was define by LevenSegerstedt (2004) [10]. Detailed informations about these methods introduced in section 2.
In forecasting applications of non-smooth demand values, different types of ANN models like: Multi-Layered
Perceptron (MLP) methods with back-propagation (BP) algorithm which is most commonly used methods in
literature was used[4]. Besides this, Recurrent Neural Network (RNN) [5], Generalized Regression Neural
Network (GRNN) [6], and some different hybrid methods [7] were used.
In MLP application, Gutierrez et al., they used two variables for input which are; the number of periods
separating the last two nonzero demand transactions as of; i: the end of the immediately preceding period ii: the
number of periods separating the last two nonzero demand transactions as of the end of the immediately
preceeding period. Output was the predicted value of the demand transaction for the current period. They choose
3 hidden units and use Back Propagation algorithm. For learning rate value of 0.1 and for momentum factor
value of 0.9 was selected.
In GRNN application, A. N. Pour et al., they used two more different inputs such as:
1. The number of consecutive period with no demand transaction immediately preceding target period.
2. The mean of demand for four period immediately preceding target periods.
As opposed to back propagation algorithm, in GRNN there are no training parameters, such as learning rate and
momentum. However, there is a smoothing factor that is used when the network is applied to new data.
In RNN application, A. N. Pour et al., their network consists of four layers, namely: input layer, hidden layer, a
context layer and an output layer and used four more input values are such as:
1. The number of consecutive period with no demand transaction, immediately preceding target period.
2. The number of periods separating the last two nonzero demand transactions as of the end of the
immediately preceding target period.
3. The number of period(s) between target period and first zero demand immediately preceding target
period.
4. The maximum demand among six periods immediately preceding target period.
2 Forecasting Methods
In this study real data sets of 20 types of spare parts demand in an airline inventory items have been used.
These spare parts were chosen among others for the diversity of their ADI which is greater than 1.32 and CV 2
which is lower than 0.49 and is classified into intermittent demand category. The data were handled for 64
monthly periods from 2005 to 2010. The data series were divided into three sets, namely: training, validation,
and test sets. From 62 monthly observations, 12 observations have been used in initialization, 38 observations
are used in validation while the last 12 observations are used as the test set. In order to evaluate the performance
of the neural networks, their forecasts are compared to the three conventional best forecasting methods which are
exponential smoothing, Croston (Syntetos&Boylan), and Croston (Leven&Segersted) explained below.
540
Demand Size
Demand Interval
Single Exponential
Smoothing
---
---
Croston (Original)
---
---
Croston
(Syntetos&Boylan)
Croston
(Leven&Segersted)
---
---
Initialization: First i periods of data is used to initialize purpose with value is 0,2 (i=12)
Calibration: The value of that minimizes the current performance measure can be updated by SOLVER
in Excel using the c period of data (cinitial=38)
Dynamic updating of performance block: optimum which is found in calibration is used to forecast of
(i+c+1). data. Then set c=c+1 and optimum is updated using c periods of data to forecast (i+c+1). data. This
optimum value is updated to minimize based on current performance measure by SOLVER considering
range of c period which is increased by one in each step of updating data to forecast. Thats to say
SOLVER optimizes the value of (p-1) times
Performance measure block: In this step according to current performance measure performance of last
(p) data is computed and until this process is repeated for selected intermittent demand forecasting
methods. Finally all the process is repeated for each performance measure.
The results showed that using dynamic value gives better results than previously defined constant value.
541
(1)
8. ; refers to the forecast of the number of periods the (t+1) th lump of demand spans;
(2)
9. Lt ; refers to the current level estimate using Browns exponential smoothing.
(3)
542
10. Tt refers to the current trend estimate using Browns exponential smoothing.
(4)
Where
refers to the
non-zero demand for period t;
refers to the forecast for the (
non-zero demand;
refers to the number of periods of the
lump of demand spans;
refers to the interval between the (
and
lump of demand;
MATLAB based ANN program is used for all ANN analysis. Primarily, the data set divided into three part;
train dataset (first %60 of all data), validation dataset ( %18 of all data), test dataset (last %22 of all data). In
the last column of Table 2, best ANN models are listed for all twenty dataset. Explanation of the coded ANN
models are such as; the name of ANN algorithm (TLRN, MLP etc.), number of hidden layer used (1 or used
hidden layer), backpropagation algorithm (B) and Levenberq-Marquert (L) or Momentum (M) respectively. All
of these alternativies are applied in all ANN learning algorithms and beat ANN models arre as shown in Table
2.
Table 2. MSE Results of Conventional Methods and ANN Methods
Table 2 shows performance of the methods considering MSE error measure. As can be seen in this table simple
average of MSE for our proposed models with new defined inputs is 2.852. While this measure for method
ANN method with eight input is 3.62, for Croston (Syntetos&Boylan) is 3.578, for Croston (Leven&Segersted)
is 3.93 and for Exponential Smoothing is 4.121 which indicate the superiority of our proposed models with new
defined inputs.
Results indicate that ANN forecasting have superior performance in comparison to other forecasting methods
considering new defined input values.
The findings suggest the potential value of ANN for producing accurate forecasts and strengthening the decision
making process. As mentioned before, ANN algorithms display promise for use in multivariate forecasting but in
forecasting this type of non-smooth demand data any specified ANN model can not be defined as a best model
for the following reasons: First, in a company there are thousands of different types of SKUs that demand pattern
is non-smooth. Second, demand size variability of these data is very high. Third, there are some different
543
classification studies on this type of data (categorized as lumpy, erratic, intermittent etc) are done but there is no
absolutely specified borders. Because of this reason it is difficult to define a specific model for this type of data.
Four, there are many different types of ANN algorithms which use different types of parameters, have
advantages and disadvantages on their working styles.
The reasons of these can be increased more and more. Bu in this study we showed that all different types of
ANN models used in this study gives best results in forecasting of non-smooth items demand. But if we analyse
the performance of the different ANN algorithms, as shown in Table 3 MLP is best, RBF and GFF is second.
RN and TLRN gives best results in two data values and TDNN in only one data value. Besides this, in this study
different types of ANN models applied but best networks are showed in this paper.
Table 3. ANN models that gives best results
3 Conclusion
Managing a service parts inventory can be a tricky business. One wants to simultaneously keep stocks low to
contain inventory costs and keep stocks high to insure availability. This classic inventory management dilemma
can be eased by accurate demand forecasts. Unfortunately, while demand forecasting is difficult in general, it can
be especially difficult to forecast demand for service parts. This is because service parts demand is
intermittent. Intermittent demand data contain a large percentage of zero values, with random non-zero
demands mixed in at random times. Classical time series forecasting methods usually perform poorly when
applied to intermittent demand.
The purpose of this research was to estimate the best forecasting technique used to handle spare parts demand
problems. Initially, historical demand data of spare parts is classified depending on its properties such as average
interdemand interval and coefficient of variation based on Syntetos scheme. Then, spare parts demand
forecasting models are used considering the optimum alpha model has been presented in 2.2. The smoothing
parameter of the forecasting methods is optimized dynamically via Excel SOLVER to give minimum error
instead of using a fixed smoothing factor values as in earlier studies (i.e 0.05). This approach is used in three
conventional forecasting methods. The models are performed on 20 real intermittent demand data series from the
airline industry. The forecasting techniques are evaluated on the basis of Mean Squared Error (MSE) to reveal
the efficiency of the forecasting techniques.
Although there are various forecasting methods developed for intermittent demand series in literature, using
neural networks in spare parts forecasting can be reasonable in terms of capturing non linear relations in data set.
For this reason we handled ANN in spare parts forecasting. Our study compares the performance of ANN
forecasts to those from three traditional time-series methods showing good performance. We concluded that
ANN models are mostly superior to three traditional forecasting models even under a relatively simple network
topology. However, in cases where a significant decrease in the average of the nonzero demand sizes arises
between the training sample and the test sample, we observed that forecasts from traditional time-series methods
tend to improve in performance relative to ANN forecasts with optimum values of the smoothing constant in
these traditional methods. Overall results related to comparisons are given as the following table considering
different performance measures.
544
References
J. E. Boylan, Intermittend and Lumpy Demand: A Forecasting Challenge, Foresight, International
Journal of Applied Forecasting 1(1), 2005, pp. 499-510
2. Vinh, DQ., Forecasting irregular demand for spare parts inventory, Department of Industrial
Engineering, Pusan National University, Busan 609-735, Korea, 2005
3. Neuralyst User s Guide, Chapter 3
4. R. S. Gutierrez, A. O. Solis, S. Mukhopadhyay, Lumpy Demand Forecasting Using Neural Networks,
Int. Journal of Production Economics 111, 2008, 409-420
5. M.R. Amin Naseri, B.R. Tabar, Neural Network Approach to Lumpy Demand Forecasting or Spare Parts
in Process Industries. Int. Conferance on Computer and Communiciation Engineering, 2008
6. M.R. Amin Naseri, B.R. Tabar, B. Ostadi Generalized Regression Neural Network in Modelling Lumpy
Demand. 8th Int. Conferance on Operations and Quantitavie Management Engineering, 2007
7. Z. Hua, B. Zhang, A Hybrid Support vector Machines and Logistic Regression Approach for Forecasting
Intermittend Demand of Spare Part, Applied Mathematics and Computation 181, 2006, 1035-1048
8. Syntetos, AA,, Boylan, JE., The Acuuracy of Intermitted Demand Estimates, Int.Journal of Forecasting,
21: 303-314, 2005.
9. Croston, JF., Forecasting and stock control for intermittent demands, Operational Research Quarterly,
23, 289-304, 1972.
10. Levn, E., Segerstedt, A. Inventory control with a modified Croston procedure and Erlang distribution,
International Journal of Production Economics, 90, 361-367, 2004.
11. Neurosolutions User s Guide, Chapter 3
1.
545
Abstract. This article includes implementation of simulation method in hospitals. Because of the increasing
complexity of the health care systems day by day, system improvement studies become mandatory. There is a
scarce and unbalanced use of resources. Excessive queues and patient waiting times cause patient
dissatisfaction. In this paper, a literature review is performed for process improvement, organization
development and workforce planning studies with simulation. In view of hospital management, personnel and
patient satisfaction new performance assessment parameters are proposed in Conclusion part.
Key-words:Hospital, simulation, flexible working, process improvement
1 Introduction
Service systems both in the world and Turkey, which is a part of economy, are gradually increasing.
Especially in banking, education, health and tourism service sectors, there are increasing competition and rapidly
changing market conditions.
In addition, because of rising income levels at both developed and developing countries with increasing
average life cycle, health care sector is in technological development and economic growth transformation
process. As a result of this, with the quality of presented service and importance, increasing the productivity
level has become the prerequisite to compete. For these reasons, in hospitals and in other health care institutions,
process improvement, working time models, workforce planning and performance measurement and evaluation
systems need to be more strengthened. As it s been well known, hospitals are first step of health care systems.
The hospital system basically starts with the registration of the patient and continues with clinical services. As
a result of the complex structure of hospitals, different consecutive processes and patient treatment occur. The
treatment process varies depending on the types of the patient. Accordingly, efficient use of hospital resources is
of great importance. Appropriate treatment plans for different patient, efficient use of resources and for patient
satisfaction, these plans should be standardized (ages, chronic illness or). Harper [1] presented a framework by
using simulation for modelling hospital resources, planning and management of beds, operating theatres and
workforce needs.
Scarce and unbalanced health care resources cause patient waiting queues [2]. Excessive queues increase the
waiting times of outpatients in front of the clinics and this situation causes patient dissatisfaction. Compared with
manufacturing process, increasing Work in Process time and high Lead time brings patient dissatisfaction.
In a sense, solutions should be produced for the transition of lean manufacturing in hospitals. Naturally, rising
lead time and work in process cause excessive stress and tension in hospitals. Financial losses consist of
unplanned and uncontrolled study at hospitals. Therefore patients do not want to get involved in this complex
process and delaying treatment diagnostic processes. This situation also makes it compulsory Optimal allocation
546
of scarce health care resources is tackled by using discrete event simulation to reduce waiting times and costs
and to improve patient flow and increase patient satisfaction [3].
Zlch et al. [4] developed an object-oriented simulation procedure (The Simulation Procedure -OSim-GAM)
in their study, for process optimization and flexible working times, personnel capacity utilization in hospitals.
They have tested different working time models.
In this study, we have analyzed the application of simulation methods in health care systems. Identification of
personnel requirement, working time, improving processes, reducing the waiting time, and efficient use of
resources were examined.
2 Literature Review
With the application of health insurance, as a result of the innovations made in health sector, became flood of
demands and inevitable health care expenses. A large share of the hospitals brings efficiency of resources
performance raise [5].
547
548
simulation study to analyze intensive care unit of a public hospital in Hong Kong to improve capacity utilization
and the quality of care. Boginski et al. [36] presented a simulation model, developed in Arena, to analyze
hospital operations and resource utilization. Data used for analyzing is derived from RFID. Patient flow process
and resource utilization is analyzed by using simulation.
Brenner et al. [37] presented a simulation model for staff and resource utilization of emergency department.
Workforce configuration and resource utilization was analyzed and the optimal number of staff and resources
was investigated by using simulation. Coelli et al. [38] concerned about the utilization of equipment and staff
utilization of a mammography clinic by using discrete event simulation. Villamizar et al. [39] presented a
simulation study that handles the utilization of resources and patient scheduling in a physiotherapy clinic.
3 Methods
Simulation method which allows analyzing the systems with different performance parameters and
comparison between different alternative scenarios is an effective method to be used in decision making process,
resource utilization, process improvement,.
In this paper, the applications of simulation methods are examined in service sector, especially in hospitals.
The scope of research is classification of problems experienced in hospitals. At this study identification of
personnel requirement, working time, improving processes, reducing the waiting time, and efficient use of
resources were examined. Among the methods used for solving problems, the applications of simulation
methods are examined. By the solution proposals with traditional methods, highlighted the importance of
simulation methods and tried to determine deficiencies in the literature. Beside this, the utilization of the
simulation software programs and the potential of applications for development of simulation software were
investigated for service systems. The applicability of simulation programs for hospital service systems was
investigated. Which performance measurement parameters are forefront in hospital systems and which methods
are being used in the process improvement were examined In addition, a preparatory work of simulation
software that meets all the performance parameters and in accordance with the hospitals and service systems has
been started.
4 Conclusion
The various assessments made with surveys, include the personal and biased observations of the participants.
Simulation-aided studies of assessments based on more objective and quantitative measurements.
Existing simulation programs were developed accordance with the manufacturing and assembly sector. There
are variety of difficulties in the use of and implementing for the service sector. As a result of the research
literature related to hospital or health sector, there could be said, except a few study, the simulation applications
has been done with existing software packages. Also there is an inadequacy of a comprehensive simulation
software including all components and the parameter features of the hospital.
Further studies may be done by a simulation software with the performance evaluation parameters from
different aspects as the hospital management, personnel and patient satisfaction for the evaluation of
performance measurement and assessment.
The aspect of hospital management performance measurement, the number of patients treated, costs of
treatments of each disease, personnel wages, costs of the equipment used in treatments should be considered.
From the perspective of personnel performance measurement the duration of treatments and the workload as
physical and psychological have become important. None of the patient wants to wait for treatment for a long
time. For the performance measurement in view of patient satisfaction, the waiting times for registration,
equipments, units and treatment should be evaluated. This could be determined to the simulation system as the
rate of waiting times of total time.
549
References
1.
2.
3.
4.
5.
6.
7.
8.
9.
10.
11.
12.
13.
14.
15.
16.
17.
18.
19.
20.
21.
22.
23.
24.
Harper, P.R. (2002). A framework for operational modelling of hospital resources. Health Care
Management Science 5, 165173.
Angelis, V.D., Felici, G., Impelluso, P. (2003). Integrating simulation and optimisation in health care
centre management. European Journal of Operational Research 150 (2003) 101114.
Jacobson, S.H., Hall, S.N., Swisher, J.R. (2006).Discrete-event simulation of health care systems.
Patient Flow: Reducing Delay in Healthcare Delivery. International Series In Opeations Research &
Management Science, Volume 91, 2006, DOI: 10.1007/978-0-387-33636-7.
Zlch, G., Stock, P., Hrdina, J., 2006, Process optimization and efficient personnel employment in
hospitals. Operations Research Proceedings, Selected Papers of the Annual International Conference of
the German Operations Research Society (GOR), Jointly Organized with the Austrian Society of
Operations Research (GOR) and the Swiss Society of Operations Research (SVOR) Karlsruhe,
September 68, 2006, Volume 2006, ISBN 978-3-540-69994-1, Hrsg., Springer Berlin Heidelberg.
Can, A. and biciolu, H. (2008). Ynetim ve yneticilik ynnden niversite hastanelerinin
deerlendirilmesi. Sleyman Demirel niversitesi ktisadi ve dari Bilimler Fakltesi Dergisi, 2008,
C.13, S.3 s.253-275.
Bader, O. (1997). Personalentwicklungsstrategien im Krankenhaus- Unter Bercksichtigung der
Situation in der Freien und Hansestad Hamburg, Uni Disseration, Uni Diss. Hamburg, Uni. Diss.
Gnal, M. and Pidd, M. (2008). DGHPSIM: Supporting smart thinking to improve hospital
performance. Winter Simulation Conference 2008.
Gnal, M. and Pidd, M. (2009). Understanding the effects of waiting time targets using modelling and
simulation. Lancaster University Management School Working Paper 2009/2.
Sinreich, D. and Marmor, Y.N. (2004). A simple and intuitive simulation tool for analyzing emergency
department Operations. Proceedings of the 2004 Winter Simulation Conference.
Hoot, N.R., LeBlanc, L.J., Jones, I., Levin, S.R., Zhou, C., Gadd, C.S., Aronsky, D. (2008). Forecasting
Emergency Department Crowding: A discrete event simulation. Annals of Emergency Medicine.
2008;52:116-125.
Steins, K. (2010). Discrete-event simulation for hospital resource planning possibilities and
requirements. Linkping Studies in Science and Technology Licentiate Thesis No. 1446, LiU-Tryck,
Linkping, Sweden, 2010.
Revetria, R., Giribone, P., Oliva, F., Schenone, M. (2007). Data driven simulation in the health and
hospital field: a model for the G. Gaslini clinical pathology laboratory workout. Internatonal Journal
Of Systems Applcatons, Engneerng & Development. Issue 4, Volume 1, 2007.
Ogulata, S.N., Cetik, M.O., Koyuncu, E., Koyuncu, M. (2009). A simulation approach for scheduling
patients in the department of radiation oncology. Journal of Medical Systems (2009) 33:233-239.
Vissers, J.M.H., Adan, I.J.B.F., Dellaert, N.P. (2007). Developing a platform for comparison of hospital
admission systems: An illustration. European Journal of Operational Research 180 (2007) 12901301.
Werker, G., Saure, A., French, J., Shechter, S. (2009). The use of discrete-event simulation modelling to
improve radiation therapy planning processes. Radiotherapy and Oncology 92 (2009) 7682.
Ahmed, M.A. and Alkhamis, T.M. (2009). Simulation optimization for an emergency department
healthcare unit in Kuwait. European Journal of Operational Research 198 (2009) 936942.
Mans, R.S., Russell, N.C., van der Aalst, W., Bakker, P.J.M., Moleman, A.J. (2010). Simulation to
analyze the impact of a schedule-aware workflow management system. Simulation, Vol. 86, Issue 8-9,
August-September 2010 519541.
Hentze, J., Kammel, A. (1993). Personalcontrolling, Bern,Stuttgart, Wien,: Haupt s.11.
Forsberg, H. et al. (2011). Managing health care decisions and improvement through simulation
modelling. Q Manage Health Care, Vol.20, No.1, pp. 15-29.
Vissers, J. (1998). Patient flow-based allocation of inpatient resources: A case study, in European
Journal of Operational Research, Amsterdam, 105(1998)2, S. 356-370.
Vissers, J.M.H., Bertrand, W., De Vries, G. (2001). A framework for production control in health care
organizations. Production Planning & Control, 12: 6, 591 604.
Brckner, B. (2002). Schwerpunktaktion der hessischen Areitsschutzverwaltung Arbeitszeitgestaltung
im Krankenhaus, in: ArbeitsplatzKrankenhaus, Hrsg.: Hessisches Sozialministerium, Wiesbaden:2002,
s.12-13.
Henning, K., Isenharrd, I., Flock, C., (1998). Kooperation im Krankenhaus, Bern: Verlag Huber, S.74.
Fields, H. (2003). Wake Up, Doc., in US News and World Report, Washington, 135(2003)20, S.70-71.
550
25. Yurtkuran, A., Emel, E. (2008). Smulaton based decson-makng for hosptal pharmacy management.
Proceedings of the 2008 Winter Simulation Conference.
26. Sundaramoorthi, D., Chen, V.C.P., Rosenberger, J.M., Kim, S.B., Buckley-Behan, D.F. (2009). A dataintegrated simulation model to evaluate nursepatient assignments. Health Care Manage Sci (2009)
12:252268.
27. Yi, P., George, S.K., Paul, J.A., Lin, L. (2010). Hospital capacity planning for disaster emergency
management. Socio-Economic Planning Sciences 44 (2010) 151160.
28. Cochran, J.K. and Bharti, A. (2006). A multi-stage stochastic methodology for whole hospital bed
planning under peak loading. International Journal of Industrial and Systems Engineering, Vol. 1, Nos.
1/2, pp.836.
29. Groothuis, S., Hasman, A., Van Pol, P.E.J., Lencer, N.H.M.K., Janssen, J.J.J., Jans, J.D.M.J., Stappers,
J.L.M., Dassen, W.R.M., Doevendans, P.A.F.M., Van Merode, G.G. (2004). Predicting capacities
required in cardiology units for heart failure patients via simulation. Computer Methods and Programs
in Biomedicine (2004) 74, 129141.
30. Couchman, A., Jones, D.I., Griffiths, K.D. (2002). Predicting the future performance of a clinical
biochemistry laboratory by computer simulation. Simulation Modelling Practice and Theory 10 (2002)
473495.
31. Macdonnell, M. and Teehan, S. (2006). Discrete event simulation analysis of the hospital bed
management problem. 15th International Conference on Management of Technology (IAMOT 2006),
China.
32. Altnel, .K. and Ula, E. (1996). Simulation modeling for emergency bed requirement planning. Annals
of Operations Research 67(1996)183-210.
33. Troy, P.M. and Rosenberg, L. (2009). Using simulation to determine the need for ICU beds for surgery
patients. Surgery 2009;146:608-20.
34. Landa, P. (2009). A Montecarlo simulation tool to verify the efficiency of operating room plans.,
Universit degli Studi di Genova, Dipartimento di Economia e Metodi Quantitativi, Discussion Paper
n.36, September 2009.
35. Kim, S.C., Horowitz, I., Young, K.K., Buckley, T.A. (1999). Analysis of capacity management of the
intensive care unit in a hospital. European Journal of Operational Research 115 (1999) 36-46.
36. Boginski, V., Mun, I.K., Wu, Y., Mason, K.P., Zhang, C. (2007). Simulation and analysis of hospital
operations and resource utilization using RFID data. 2007 IEEE International Conference on RFID
Gaylord Texan Resort, Grapevine, TX, USA March 26-28, 2007.
37. Brenner, S., Zeng, Z., Liu, Y., Wang, J., Li, J., Howard, P.K. (2010). Modelng and analyss of the
emergency department at Unversty of Kentucky Chandler Hosptal usng smulatons. Journal of
Emergency Nursing 2010;36:303-10.
38. Coelli, F.C., Ferreira, R.B., Almeida, R.M.V.R., Pereira, W.C.A. (2007). Computer simulation and
discrete-event models in the analysis of a mammography clinic patient flow. Computer Methods and
Programs in Biomedicine 8 7 ( 2 0 0 7 ) 201207.
39. Villamizar, J.R., Coelli, F.C., Pereira, W.C.A., Almeida, R.M.V.R. (2011). Discrete-event computer
simulation methods in the optimisation of a physiotherapy clinic. Physiotherapy 97 (2011) 7177.
551
Abstract.This paper presents a production scheduling approach using Faborg-Sim simulation tool with
priority rules. Faborg-Sim consists of three modules as modelling, simulating and performance evaluationThe
initial values of the system could be modelled in modelling module. Then, simulation and scheduling could
be done with using priority rules that integrated in Fabog-Sim simulation module, for the best performance
value of production schedule, the simulation runs are repeated for each priority rule of scheduling. After
repetitions the performance measurement parameters obtained and could be evaluated. Faborg-Sim was
developed at the Department of Industrial Engineering of the University of Suleyman Demirel. A case study
that using Faborg-Sim performed for modelling and evaluating in a machine parts manufacturing system.
Key words:Production scheduling, simulation, priority rules
1 Introduction
Management of capacity and changing demand are affected by many different factors. Meeting customer
demands and due dates become very important with changing manufacturing systems and product variety [1].
Sustainable competition in the market and entering new markets require strong management and controlling
production parameters and capacity.
On time delivery, short processing times, lower personnel costs, processing costs, service and product quality
provide competition power in market and customer satisfaction. All these conditions occur after correct
management of the production. Production management, especially choosing the appropriate scheduling
approach gets difficult under the conditions with unexpected tasks, events. The system loses its effectiveness
under this uncertainty and complex condition [2]. At this point production scheduling comes into question.
Scheduling that affected by many factors, such as production levels, lot size limits, due dates, job precedence,
priority rules so on, is a very important function for a production system [3]. Scheduling is a kind of assignment
problem which is related to assign tasks over a period under constraints [4]. Generally it is pointed two types of
constraints in scheduling problems, that are resource capacity and technological constraints [5].Different
methods are used for scheduling problems that become a focus in development, application and evaluation of the
systems. The usage of methods changes the structure of system and objective function. In the literature heuristics
and operations research models such as integer programming, dynamic programming and branch and bound
techniques are mostly used to solve the production scheduling problems. These methods need assumptions to
ease the problem and get optimal solutions, but not always the effective assumptions could be developed, then
552
the simulation method becomes an effective method to analyse the problems and evaluating different convincing
results and for practical purposes [6, 7, 8].
In a job shop scheduling system, for production problems as interruption of machines, re-work and work in
process presented expert systems by simulation method or developing a structure to minimize total completion
time considering due dates, precedence or priority rules and maintenance control processes [9, 10, 11, 12, 13].
Job shop scheduling problems are described as NP hard problems. Because of the hardness of getting the
optimal solution, it is accepted to get near optimal solution in practice [14]. Job shop scheduling problems are
categorized in two groups, first, static job shop problems and the second, dynamic job shop problems. In the
static job shop problems, there are jobs to be sequenced on various machines and jobs arrivals are static, but in
dynamic job shop problems there are randomly and continuously job arrivals in a period [15]. Simulation method
is mostly used for the dynamic job shop problems. Performance assessment parameters such as resource
utilization rate, capacity utilization rate, definition of bottlenecks could be obtained. Reduction of lead times and
comparison of alternative scenarios are the possible results of simulation [16].
In this study, we examined a production scheduling problem using simulation software Faborg-Sim with
priority rules. Faborg-Sim allows modelling, simulation and performance evaluation. For the best performance
value of production schedule, the simulation runs are repeated for each priority rule for scheduling. After
repetitions the performance measurement parameters obtained and could be evaluated with a case study by using
Faborg-Sim performed for modelling and evaluating in a machine parts manufacturing system.
2 Literature Review
Priority rules are used for decades as scheduling procedure in industry. It is known that implementation of the
priority rules are easy in practice. Geiger et al [17] proposed a study for production scheduling that uses priority
rules with implementing by using discrete event simulation and comparison of performance of the rules. A
dynamic stochastic job shop scheduling problem has been presented with coordination of priority rules and the
analysis of performance by using simulation [18].
The main purpose of scheduling rules is minimizing the total completion time and meet due dates. Job shop
scheduling problems are studied by using simulation in the literature. Moghaddam and Mehr [19] indicated this
with an example of job scheduling problem using simulation tool Visual SLAM with the aim of minimizing
makespan and evaluation of the system performance. It is studied that for the flexible job shop scheduling
problems on the objectives of the minimizing makespan, total workload of machines and critical machine with
simulation, using Matlab [20].
Minimizing tardiness is an objective of some priority rules used in scheduling problems. A priority rule for
minimizing mean tardiness in a dynamic job shop environment has been presented and compared by using
simulation [21]. Each tardy job causes a tardiness cost. Minimizing tardiness and reducing total tardiness cost is
studied and the performance parameters are compared with simulation [22]. Selecting the right dispatching rules
in production scheduling improves machine utilization. Selection of a product mix and a dispatching rule
development has been studied to gain maximum profit for a job shop scheduling [23].
For the dynamic assembly job shops it is studied scheduling with different earliness, tardiness and holding
costs of each job. The aim of study is to present implementing priority rules and their costs related with earliness,
tardiness [24]. A simulation based assembly scheduling system is presented that aimed to optimize due dates and
optimal utilization of personnel and material resources [25]. It is possible to make an evaluation of performance
by using simulation. An assembly job shop problem has been presented with priority rules to minimize the flow
time and has been done a simulation based evaluation [26].
Setup time changes could have an influence on due dates of jobs. The scheduling problems including setup
times were reviewed [27]. Literature was classified into three groups such as jobs, class and job-and-class setup
cases. Simulation integrated solution method was used by comparing the priority rules performance in a job shop
553
scheduling problem with sequence dependent setup times [28]. About the minimization of the makespan, Balas
et al [29] studied job shop scheduling introducing release dates, deadlines and sequence dependent setup times.
Simulation could be used as a decision support tool that a neuro genetic decision support system integrated
with simulation to achieve the performance parameters such as flow time, number of tardy jobs, total tardiness
and machine utilization rates has been presented in a study [30]. Simulation method is used also for analysing in
flexible manufacturing systems. Sabuncuoglu [31] examined the effects of scheduling rules on the performance
of flexible manufacturing systems including the changes on processing times and breakdown rates.
554
Described personnel, working times, machines, functions, job processing times are the input of the FaborgSim and due to these values the selected priority rule are used, then the evaluation module demonstrates the
performance parameters such as personnel utilization rate, machine utilization rate, the number of delivered jobs,
lead time. ,
The Manufacturing of machine parts or products are modeled as network graphs, where each activity is
assigned to at least one machine, workplace or human resource, with separate setup and execution times if
required.
Gantt charts are seen in production environment for scheduling jobs and provide the jobs to be seen the
starting time and end time clearly. In this study Faborg-Sim uses Gantt charts for scheduling jobs after
simulation as machine and personnel Gantt charts.
Priority Rules
Production
Orders
FCFS
LCFS
SPT
LPT
EDD
LRNOP
MRNOP
SRPT
LRPT
SIRO
Fig. 2. Priority rules for customer orders in Manufacturing Systems (modified from [40])
-
First Come First Served Rule (FCFS): The job arriving earliest at the machine will be the next
processed job on the machine [21].
Last Come First Served .Rule (LCFS): The job arriving latest at the machine will be the next processed
job on the machine.
Shortest Processing Time (SPT): The job with the shortest processing time among waiting jobs will be
processed next on the machine to minimize total flow time[21].
Longest Processing Time (LPT): The rule assigns the job with longest processing time to the machine to
minimize the total completion time.
Earliest Due Date (EDD): The job with the earliest due date will be processed next on the machine [21].
It is aimed to improve customer satisfaction.
Lowest Remaining Number of Operations (LRNOP): The job that has the lowest operations remaining
will be processed next on the machine [41]. It is aimed to maximize the number of delivered customer
orders.
555
Most Remaining Number of Operations (MRNOP): The job that has most operations remaining will be
processed next on the machine [41]. It is aimed to maximize the utilization rate of capacity.
- Shortest Remaining Processing Time (SRPT): The job that has the shortest total remaining processing
time will be processed next on the machine. It is aimed to minimize the total completion time and
minimize the latest job delivery time.
- Longest Remaining Processing Time (LRPT): The job that has the longest total remaining processing
time will be processed next on the machine. It is aimed to maximize the utilization rate of capacity.
- Service in Random Order (SIRO): The job that is selected randomly from waiting jobs will be
processed next on the machine.
Algorithms of priority rules were developed and were coded in C#. The priority rules module was integrated
in to Faborg-Sim. Thus, using appropriate rules production systems can be simulated easily.
Customers
M3_MATKAP
1 Unit
Products
Customer Orders
M5_MONTAJ
1 Unit
P01
P02
P03
Workstation /
Machine
Worker /
Personnel
M2_FREZE
3 Parallel Units
M4_KAYNAK
1 Unit
P04
556
20%
10%
0%
-10%
-20%
-30%
-40%
-50%
-60%
-70%
-80%
-90%
-100%
10%
5%
A1 A2 A1 A2 A1 A2 A1 A2 A1 A2 A1 A2
0%
-5%
FCFS
LCFS
LPT
EDD
LRNOP
LPRT
-10%
-15%
GADLT-Lead Time
GADWIP-Work In Process
GADSOL-Sum of Logistics
4 Conclusion
In this paper, production scheduling is studied by using priority rules as integrated to Faborg-Sim simulation
tool. The simulation runs are repeated for each scheduling priority rule. After repetitions the performance
measurement parameters obtained and could be evaluated. The effects of different set-up times and lot size on
scheduling are investigated for priority rules in the manufacturing systems.
A number of different criteria and parameters exist for evaluation of production systems and effects on
performance parameters in a production environment. By integrating simulation and production scheduling
methods, it is possible to evaluate various performance parameters with given input values. System bottlenecks
could be seen by visual and excessive waiting times could be eliminated. Simulation results give the users
information and an opportunity for decision making.
Acknowledgments
Part of this research has been supported by the Scientific and Technology Research Council of Turkey
(TUBITAK) the grant number 104M377. The authors thank to TUBITAK for supporting of this project.
557
References
1.
2.
3.
4.
5.
6.
7.
8.
9.
10.
11.
12.
13.
14.
15.
16.
17.
18.
19.
20.
21.
22.
23.
24.
25.
26.
Mnch, L., Zimmermann, J., (2007), Simulation-based Assessment of Machine Criticality Measures
for A Shifting Bottleneck Scheduling Approach in Complex Manufacturing Systems, Computers in
Industry 58 (2007) 644-655.
Pfeiffer, A. et al, (2007), Stability-oriented evaluation of rescheduling strategies, by using simulation,
Computers in Industry 58 (2007) 630643.
Geyik, F., & Cedimoglu, I. H. (1999), A review of the production scheduling approaches based-on
artificial intelligence and the integration of process planning and scheduling. Proceedings on Swiss
Conference of CAD/CAM99, 167-174, Neuchatel University, Switzerland.
Pinedo, M. (1995), Scheduling Theory. Algorithms and Systems. Prentice-Hall, pp. 378, New Jersey.
Baker, K.R. 1974. Introduction to Sequencing & Scheduling. New York: John Wiley.
Al-Turki, U.; Andijani, A.; Arifulsalam, S. (2004). A new dispatching rule for the stochastic singlemachine scheduling problem. Simulation, Vol. 80, Issue 3, 165-170.
Baker, K.R. and Trietsch, D. (2009). Principles of sequencing & scheduling. New York: John Wiley.
Olafsson, S. and Li, X. (2010). Learning effective new single machine dispatching rules from optimal
scheduling data. International Journal of Production Economics128 (2010) 118126.
Li, Y.-C. E., & Shaw, W. H. (1998), Simulation modeling of a dynamic job shop rescheduling with
machine availability constraints. Computers &Industrial Engineering, 35, 117-120.
Li, H., Li, Z., Li, L. X., & Hu, B. (2000), A production rescheduling expert simulation systems.
European Journal of Operational Research, 124, 283-293.
Allaoui, H. & Artiba, A. (2004). Integrating simulation and optimization to schedule a hybrid flow
shop with maintenance constraints. Computers & Industrial Engineering, 47, 431450.
Gharbi, A., & Kenne, J. P. (2005), Maintenance scheduling and production control of
multiplemachine manufacturing systems. Computers &Industrial Engineering, 48, 693707.
Yildirim, M.B., Cakar, T. Doguc, U., & Meza, J.C. (2006), Machine number, priority rule, and due
date determination in flexible manufacturing systems using artificial neural networks. Computers &
Industrial Engineering, 50, 185194.
Yang, S., Wang, D., Chai, T., Kendall, G. (2010). An improved constraint satisfaction adaptive
neural network for job-shop scheduling. Journal of Scheduling (2010) 13: 1738.
Vinod, V., Sridharan, R., (2011), Simulation Modeling and Analysis of Due-Date Assignment
Methods and Scheduling Decision Rules in a Dynamic Job Shop Production System, International
Journal of ProductionEconomics129(2011)127146.
Akkaya, G. and Gken, T. (2006). Job shop scheduling design with artificial neural networks.
Sigma Engineering and Natural Sciences, 2006/4.
Geiger, C., D. et al, (2006), Rapid Modelling and Discovery of Priority Dispatching Rules: An
Autonomous Learning Approach, Journal of Scheduling 9: 734, 2006.
Holthaus, O. and Ziegler, H. (1997). Improving job shop performance by coordinating dispatching
rules., International Journal of Production Research, 35: 2, 539 549.
Moghaddam, R. T. and Mehr, M. D., (2005), A Computer Simulation Model for Job Shop
Scheduling Problems Minimizing Makespan, Computers & Industrial Engineering 48 (2005) 811
823.
Xing, L.,N. et al, (2008), Multi-objective flexible job shop schedule: Design and evaluation by
simulation modeling, Appl. Soft Comput. J. (2008), doi:10.1016/j.asoc.2008.04.013.
Weng, M.X. and Ren, H. (2006). An efficient priority rule for scheduling job shops to minimize
mean tardiness. IIE Transactions, 38: 9, 789 795.
Chen (Gary), S.J. and Lin, L. (1999). Reducing total tardiness cost in manufacturing cell scheduling
by a multi-factor priority rule. International Journal of Production Research, 37: 13, 2939 2956.
Penn, M., Raviv, T., (2008), An Algorithm for The Maximum Revenue Jobshop Problem, European
Journal of Operational Research (2008), doi:10.1016/j.ejor.2007.11.039.
Thiagarajan, S., Rajendran, C., (2005), Scheduling in dynamic assembly job-shops to minimize the
sum of weighted earliness, weighted tardiness and weighted flowtime of jobs, Computers &
Industrial Engineering 49 (2005) 463503.
Weigert, G. and Henlich, T.(2009). Simulation-based scheduling of assembly operations.
International Journal of Computer Integrated Manufacturing, 22: 4, 325 333.
Reeja, M. K. and Rajendran, C. (2000). Dispatching rules for scheduling in assembly jobshops - Part
1. International Journal of Production Research. 38: 9, 2051 2066.
558
27. Yang, W.H. and Liao, C.J. (1999). Survey of scheduling research involving setup times. International
Journal of Systems Science, 30: 2, 143 155.
28. Vinod, V. and Sridharan, R. (2009). Simulation-based metamodels for scheduling a dynamic job
shop with sequence-dependent setup times. International Journal of Production Research, 47: 6, 1425
1447.
29. Balas, E. et al, (2008), Job Shop Scheduling with Setup Times, Deadlines and Precedence
Constraints, Journal of Scheduling (2008) 11: 253262.
30. Cakar, T., Yldrm, M., B., Barut, M., (2005), A Neuro-Genetic Approach to Design and Planning of
a Manufacturing Cell, Journal of Intelligent Manufacturing, 16, 453462, 2005.
31. Sabuncuoglu, I.(1998). A study of scheduling rules of flexible manufacturing systems: A simulation
approach. International Journal of Production Research, 36: 2, 527 546.
32. VDI. (1983), Richtlinie -3633, Dsseldorf, VDIVerlag.
33. Witte, T. (1990), Lexikon der Wirtschaftsinformatik, Hrsg. Mertens P., Berlin, Springer Verlag,
2.Auflage.
34. Law, A. M., & Kelton, W. D. (1991), Simulation Modeling and Analysis. McGraw-Hill, 2nd Edition,
New York.
35. Schmittbetz, M. (1998), Simulation wird zum Triebwerk fr Innovation. VDI-Nachrichten (52), 1824.
36. Zlch, G., Bogus, T., & Fischer, J. (2002), Integrated Simulation and Workforce Assignment for the
Evaluation of Flexible Working Time Models, in: Chen, Z. et al., System Simulation and Scientific
Computing, International Academic Publishers,Beijing, Vol. I, pp. 353-357.
37. Koruca, H. I., Ozdemir, G., Aydemir, E., Cayirli, M. (2010), The Simulation-Based Performance 10
Measurement in An Evaluation Module for Faborg- Sim Simulation Software, Expert System with
Applications, 37 (12), pp. 8211-8220.
38. Frantzen, M., Ng, A., H., C., Moore, P., (2011), A simulation-based Scheduling System for Real-time
Optimization and Decision Making Support, Robotics and Computer-Integrated Manufacturing,
doi:10.1016/j.rcim.2010.12.006.
39. Kapanolu, M., and Alikalfa, M., 2011, Learning IFTHEN Priority Rules for Dynamic Job Shops
Using Genetic Algorithms, Robotics and Computer-Integrated Manufacturing 27 (2011) 4755.
40. Schuh, G., 2008. Produktionsmanagements I, WZL/FIR, Acchen Universitat.
41. Panwalkar, S.S., Iskander, W. (1977). A survey of scheduling rules. Operations Research, Vol. 25,
No. 1 (Jan. - Feb., 1977), pp. 45-61.
559
Abstract. Many models are suggested to solve course timetabling problems, but a comprehensive solution
could not be found because of the variety of constraints. In this study, a new approach was proposed to solve
course timetabling problems that encountered by educational institutions frequently. Graph coloring
algorithm is applied for produce an initial solution to achieve vital constraints by grouping unlike courses.
Then an optimization algorithm is used on initial course groups for obtain better solution which satisfies all
constraints, and finally the optimized course groups are assigned to appropriate classes at appropriate days
and hours by considering all of the constraints.
Keywords: course timetabling, graph coloring, optimization
1 Introduction
Course timetabling that is a classical problem for educational institutions has been interested by researchers
for more than a decade. So, many models have been proposed in the literature. University course timetabling
problem (UCTP) is a large scale optimization problem and it is computationally NP-Hard. Usually, a staff
spends a couple of days to arrange a timetable that meets all of the constraints, and then the timetable may be
inappropriate by addition or modification of a single constraint. Considering that each institution has its own
constraints, a general algorithm approach may be incapable for find a suitable model to all course timetabling
problems. This is the reason of many different types of models in the literature. The university course
timetabling is a scheduling problem that assignment of courses in timeslots with considering some constraints.
These constraints are categorized into two major groups as hard and soft by Burke et al. [1]. Hard
constraints are indispensible constraints for course timetabling problems such as no one attends more than one
course at any time and hard constraint must be fulfilled to illustrate appropriate course timetables. According to
the timetabling problem, soft constraints may exist in different forms and soft constraints are overruled, if
necessary. Soft constraints are integrated into timetabling problems with penalty functions and penalty cost
shows that the fulfillment degree of soft constraints. Gotlieb has published the first paper about timetabling
problems [2]. Then, many methods have been issued for timetabling problems [1, 3, 4, and 5]. These methods
are classified in four groups such as sequential methods, cluster methods, constraint-based methods and metaheuristics [5]. Graph coloring is categorized as a sequential method. Welsh and Powell [6] and Neufeld and
Tartar [7] are published early papers about graph coloring applications at timetabling problems. Also, Wood [8]
and De Werra [9] published the early papers about how to solve course timetabling problems by graph coloring.
When illustrating a timetable by using graph coloring, courses are represented by vertices and constraints are
represented by edges and two vertices are connected by an edge if there is at least one conflict. In recent years,
researchers are developing graph-based heuristic methods with hope that illustrate better timetables. Burke et al.
560
[10], Qu et al. [11] have developed graph-based hyper-heuristic methods to raise better solutions. There are some
obvious things about timetabling. It is an unstable problem that can be formed by the needs of era and the new
approaches will be proposed in the future as in the past.
In this article a novel approach of graph coloring is presented to solve course timetabling problems. As all
graph based models, each course is represented by a vertex and constraints such as assignment of lecturers,
the academic year of courses are represented by edges. The main idea is the adjacent courses cannot be
scheduled at the same time. In addition, timetabling problems have external constraints such as only one
courses at any time, free days of lecturers, capacity of classrooms, types of classrooms and duration of
courses. Some of these constraints cannot be represented by edges. In our work, the courses are colored
/clustered without considering external constraints, but ignored constraints are integrated in the solution by using
an optimization algorithm(penalty function).
Specifications
Number of courses
Duration of courses
Number of departments
Number of lecturers
Number of classrooms
Number of days in a week
Number of timeslots in a day
Total capacity of classrooms
#
74
3
4
23
7
5
8
432
Remarks
(including electives)
4 hours, 3hours, 2hours
MIS, PSPP, Economy, and Business
(every lecturer have a free day in a week)
2 x type1, 1 x type2, 3 x type3, and a lab.
Monday-Friday
08:00-12:00 and 13:00-17:00
2x42, 1x66, 3x84 and 30
If we introduce somewhat more details on the constraints; a student or a faculty member can only have one
course only at one time, and only one course can be assigned to any of the classes at one time. And the courses
of the same academic year should not be same time. Thus, a student of that academic year can follow all of the
courses of his/her period. The courses are held at the weekdays, Monday to Friday, in 10 timeslots. Five days of
a week are divided into two timeslots as morning (8:00 to 12:00 a.m.) and afternoon (1:00 to 5:00 p.m.). Lecturer
of each course is determined in advance, and they can never be changed. And the courses which are requiring
use of computers can only be given in the laboratory. The number of students should be taken into account, and
the classrooms should be determined in accordance with the number of students who took courses. In addition,
there are certain soft constraints. In case study, lecturers can choose 2 of the 10 timeslots as free time and
thecourses should not be assigned to them at these time periods. Also the lesson hours of courses should be held
in same day sequentially.For example, a three creditcoursemust be scheduled from 8:00 to 11:00 a.m. without
giving a break between lesson hours.
561
Type
Hard
Hard
Hard
Hard
Hard
Hard
Hard
Hard
Soft
Soft
Constraints
A student attends only one course at one time.
A lecturerpresents only one course at one time.
There may be only one course in a classroom at one time.
The courses of the same class/department cannot be heldat the same time.
The courses can be given between 08:00 a.m. to 5:00 p.m. (except 12:00
a.m. to 1:00 p.m.)
The courses can be held between Monday and Friday.
Lecturers of each course is fixed, cannot be changed.
The courses, that computer use is necessary, are given in the laboratory.
An instructors course cannot be assigned itsown free times/days.
Hours of courses cannot be separated
Representation of Criteria
Lecturers
Credits of courses
Departments
Semester / Year
Size of classroom
Free time of lecturer
#
1 to 23
2 to 4
1 to 4
1 to 2
1 to 4
1 to 10
Remarks
Every lecturer represented by a number
2 credit, 3 credit or 4 credit
1=Economy, 2=Business, 3=PSPP, 4=MIS
1=first year, 2=second year
1=small, 2=medium, 3=large, 4=laboratory
01=Monday(08.00-12:00),
02=Monday(13:00-17:00),
..
09=Friday(08:00-12:00)
10=Friday(13:00-17:00)
The developed encoding system includes hard constraints, such as course credits, classroom size, course
instructor, and department of course, and semester/year. Also, the picked free times of lecturers take place in it.
The constraints are separated by "-" mark and anencoding system with 4 parts had been obtained.
562
The Coloring algorithm searches a relationship by comparing 1st and 3rd parts of each course codes. Because,
the 1st part identifies course's department and year/semester, and the 3rd part identifies the instructor of the
course. Assignment of a student or a lecturer to more than one course at a time is prevented by taking these two
constraints into account.
Pseudo code of coloring (grouping) algorithm.
#program Coloring
int
GColor = 1;
array CODE1[74], CODE2[74], ColorLesson[74];
BEGIN
FOR from first lesson to the last lesson //i=1 to 74
SET the first part of lessons code CODE1[i]
SET the third part of lessons code CODE2[i]
END FOR
SET K=1
FOR from first lesson to the last lesson //k=1 to 74
SET M=K+1
FOR from first lesson to the last lesson //m=k+1 to 74
IF CODE1[k]!= CODE1[m]||CODE2[k]!= CODE2[m]
STORE GColor in ColorLesson[m]
INCREMENT GColor
END IF
END FOR
END FOR
3.3
The course groups, which can be assigned at the same time slot, is determined by the application of coloring
/groupingalgorithm. However, the number of course groups we have obtained should not excess the number of
available timeslots. If the number obtained course groups is greater than the available timeslots, the courses in
excess groups are added to the appropriate course groups, and the number of groups are synchronized with the
number of timeslots in a week. Consequently, the course groups are created, as shown in Table 4. However, the
assignment of courses on cannot be performed because of the -free day- constraints of lecturers. In this case,
each group can include all free days as constraint. So, courses of lecturers, who demand same days as free times,
563
must be grouped together as much as possible. Although the free time demands of lecturers are soft constraints in
current problem, they should be fulfilled. Because, the time constraints of lecturerscan be transform to hard
constraints. A lecturer may need to spend time for doctoral courses and advisory duties, which are not taken into
account in the case study. As a result, the time constraints taken into account by the demands of lecturers are
considered as hard constraints, at the optimization process of course groups. Hence, the goal of the optimization
algorithm is the gathering the courses, that have similar free day constraints, in the same group by substituting of
courses and considering a penalty cost.
Group1
21-33-11-0102
11-33-14-0304
42-23-02-0708
22-33-18-0308
12-33-01-0102
32-33-05-0607
41-23-22-0506
31-23-23-0708
Group6
21-23-21-0102
11-23-06-0304
31-23-05-0607
41-64-03-0910
32-31-07-0708
22-33-17-0304
42-31-02-0708
3.4
Group2
41-33-11-0102
31-33-14-0304
11-33-15-0400
22-33-01-0102
21-33-04-0203
42-33-09-0102
12-33-06-0304
32-33-10-0304
Group7
41-23-21-0102
31-33-13-0506
21-64-02-0708
11-33-06-0304
12-33-07-0708
42-23-17-0304
32-33-19-0506
22-31-04-0203
Group3
21-33-15-0400
32-33-01-0102
11-33-04-0203
31-33-05-0607
41-33-12-0304
22-33-09-0102
12-33-10-0304
42-33-02-0708
Group8
21-33-08-0910
11-64-20-0102
31-23-22-0506
41-23-23-0708
42-31-03-0910
22-31-11-0102
12-31-06-0304
32-31-05-0607
Group4
42-33-01-0102
41-33-04-0203
11-33-12-0304
21-33-19-0506
31-23-21-0102
22-33-11-0102
12-23-07-0708
32-33-13-0506
Group9
31-33-08-0910
11-23-22-0506
22-32-07-0708
21-23-23-0708
32-31-10-0304
41-33-15-0400
Group5
31-33-12-0304
11-23-21-0102
21-23-04-0203
41-23-02-0708
12-31-07-0708
42-33-03-0910
22-23-10-0304
32-33-13-0506
Group10
41-33-08-0910
21-23-22-0506
11-23-23-0708
12-31-07-0708
31-64-20-0102
Each course group is regrouped, according to their similarities by the optimization process. The next step is
placement of course groups to the appropriate hours, by taking the other constraints into account, which are not
included in the solution, during the encoding. Firstly, each course group must be assigned to the available time
slots of the week, and then the courses are placed to appropriate classrooms by considering the 3rd, 5th, 6th, 8th
and 10th constraints given in Table 4. During the assignment process, the following operations are performed in
a sequence.
STEP 1. Select the classroom, according to number of students of thecourse or according to the requirements of
the course as laboratory.
a. Assign the courses with 3 credits to 8:00-11:00 a.m. / 2:00-5:00 p.m. time slots.
b. Assign the courses with 2 credits to 8:00-10:00 a.m., 10:00-12:00 a.m. / 1:00-3:00 p.m., 3:00-5:00
p.m. time slots.
c. Assign the 2 credit courses, which were not assigned in the course groups, to 11:00-12:00 a.m. and
1:00-2:00 p.m. in appropriate classrooms where the 3 credit courses has been assigned.
d. Collect the courses that cannot be assigned, into a temporary group.
STEP 2. Repeat the 1st step until all of the courses are assigned.
STEP 3.Search for the appropriate time slots to assign the courses in the temporary group, and relocate if
necessary by neglecting the free time constraint of faculty members.
564
565
References
1.
Burke, E.K., Jackson, K., Kingston, J. and Weare, R. (1997). Automated university timetabling: the state of the art.
The Computer Journal, 40(9): 565-571.
2. Gotlieb, C.C. (1963). The construction of class-teacher time-tables. In Proceedings of IFIP Congress. Munich,
Germany, Amsterdam: North Holland Pub. Co., 73-77.
3. Schaerf, A. (1999). A Survey of Automated Timetabling. Artificial Intelligence Review 13: 87127.
4. Burke, E.K., & Petrovic, S. (2002). Recent research directions in automated timetabling. European Journal of
Operational Research, 140: 266280.
5. Carter, M. W. and Laporte, G. (1998). Recent Developments in Practical Course Timetabling. In Practice and
Theory of Automated Timetabling II, Lecture Notes in Computer Science: Burke, E., Carter, M. (eds.), SpringerVerlag, 319.
6. Welsh, D. J. A. and Powell, M. B. (1967). An upper bound for the chromatic number of graph and its application to
timetabling problems. The Computer Journal, 10(1): 360-364
7. Neufeld, G. A. & Tartar, J. (1974). Graph Coloring Conditions for the Existence of Solutions to the Timetable
Problem. Communications of the ACM, 17(8): 450453.
8. Wood, D. C. (1969). A technique for coloring a graph applicable to large scale timetabling problems. The
Computer Journal, 12(4): 317-319
9. De Werra, D. (1995). Some combinatorial models for course scheduling, In Practice and Theory of Automated
Timetabling I, Lecture Notes in Computer Science :Burke, E. K. and Ross, P. (eds.), 296-308
10. Burke, E. K., McCollum, B., Meisels, A., Petrovic, S. and Qu, R. (2007). A graph-based hyper-heuristic for
educational timetabling problems. European Journal of Operational Research, 176: 177-192
11. Qu, R. and Burke, E.K. (2009). Hybridizations within a graph based hyper-heuristic framework for university
timetabling problems, Journal of the Operational Research Society 60:12731285.
Biographies
Zafer BOZYER was born in Kayseri on July 18, 1983. He is currently working as research assistant of Management
Information Systems Department at the Bartn University in Turkey. He received his B.Sc. in Industrial Engineering from the
University of Gaziantep, in 2010.His M.Sc. is in progress in the Department of Industrial Engineering at the Kocaeli
University.
His research interests are: Information Systems Design, Artificial Intelligence and Expert Systems.
M. Sinan BAAR was born in Erzurum on October 03, 1970. He is presently an Assistant Professor of Management
Information Systems Department at Bartn University, Turkey. He received his B.Sc.in Economy from the stanbul
University, in 1996. He received his M.Sc. and Ph.D. in Business from the Atatrk University, in 1999 and 2004.
His research interests are: Hashing Algorithms and Information Systems. He has published technical papers in academic
journals and academic meetings.
Assist. Prof. Dr. Baar is presently Vice-Dean of the Faculty of Economics and Administrative Sciences at Bartn
University.
Alper AYTEKN was born in Nide on June 24, 1973. He is presently an Associated Professor of Management Information
Systems Department at Bartn University, Turkey. He received his B.Sc.in Forest Industry from the Karadeniz Teknik
University, in 1994. He received his M.Sc. in Natural and Applied Science from the Karadeniz Teknik University, in 1994
and Ph.D. in Natural and Applied Science from the Zonguldak Karaelmas University, in 2002.
His research interests are: Computer Programming, CAD, ANN. He has published technical papers in academic journals
and academic meetings.
Assoc. Prof. Dr. Aytekin is presently Vice-Dean of the Faculty of Economics and Administrative Sciences at Bartn
University.
566
Abstract. Parasuraman, Zeithaml and Malhotra [1] developed E-S-QUAL and E-RecS-QUAL scales for
capturing electronic service quality. Akinci, Inan and Aksoy [2] also reassessed Parasuraman et al. [1]s study
for electronic service quality offered by 13 banks in Turkey. In this study, we used the results of Akinci et al.
[2]s study to evaluate the criteria of electronic service quality through decision-making trial and evaluation
laboratory (DEMATEL) method. A questionnaire based on E-S-QUAL and E-RecS-QUAL scales was
developed. Firstly, we phoned the head office of banks to inform them about our study. Then, we sent
developed questionnaire to 17 different banks departments via e-mail to obtain questionnaires completed by
experts. After approximately one month (9 February 2011 to 10 March 2011), we obtained 39 completed
questionnaires from 11 different banks. Finally, casual relationships among criteria for both E-S-QUAL and
E-RecS-QUAL scales are constructed. In addition, the importance of the criteria for both E-S-QUAL and ERecS-QUAL scales are prioritized.
Keywords: DEMATEL, Electronic Service Quality, Internet Banking, Survey
1 Introduction
Measuring electronic service quality is quite important to improve customer satisfaction. Especially, internet
banking service quality is gaining more attraction by costumers day by day. Therefore, measuring internet
banking service quality correctly is quite significant to understand the quality level of provided services and to
improve it by feedback. Parasuraman, Zeithaml and Malhotra [1] developed E-S-QUAL and E-RecS-QUAL
scales for capturing electronic service quality. Researchers use E-S-QUAL scale to measure the perceived
electronic service quality. On the other hand, they use E-RecS-QUAL to measure perceived electronic service
quality by customers who experienced problems with assistance from banks Web sites. E-S-QUAL includes
four dimensions and 22 items, while E-RecS-QUAL includes three dimensions and 11 items. Akinci, Inan and
Aksoy [2] reassessed Parasuraman et al. [1]s study for electronic service quality offered by 13 banks in Turkey.
The statements were reduced 8 items from original 22 item E-S-QUAL scale and 7 items from original 11 item
E-RecS-QUAL scale by Akinci et al [2]. This study is based on the results of Akinci et al. [2]study. The aim of
this study is to use decision-making trial and evaluation laboratory (DEMATEL) method to find relationships
among dimensions and to prioritize them for both E-S-QUAL and E-RecS-QUAL scales.
567
2 Literature Review
Akinci et al. [2] stated that studies related to electronic service quality in literature fall into three main categories:
technical quality of web sites, the dimensions that affect customer satisfaction, and Web site service quality. In
this study, we examined the dimensions that affect customer satisfaction for Web site service quality in internet
banking by using E-S-QUAL and E-RecS-QUAL scales. Ancarani [3] and Buckley [4] examined e-services in
the public sector. There are some studies developed a scale to measure the quality of e-services (Ho and Lin, [5];
Swaid and Wigand, [6]; Cristobal, Flavian and Guinalu, [7] ). Details on developed scales can be found in
Ladhari [8] and Barrutia and Gilsanz [9]. Hu and Liao [10] found critical criteria of evaluating electronic service
quality of Internet banking using fuzzy multiple-criteria decision making (FMCDM). Hu [11] also found critical
criteria for assessing service quality of travel websites using fuzzy multiple-criteria decision making (FMCDM).
Sheng and Liu [12] conducted a amprical study to measure e-service quality by using E-S-QUAL scale. Calisir
and Gumussoy [13] examined internet banking versus other banking channels. Jun and Cai [14] researched the
key determinants of Internet banking service quality by using content analysis.
The basic steps of used DEMATEL method to evaluate the dimensions electronic service quality is summarized
as follows based on Liou et.al. [15] and Wu [16].
Step 1: Compute average initial direct-relation (matrix A). Firstly, a survey study is conducted by asking experts
who are expert about interested topic. The survey includes comparison scale to find influence and direction
among criteria with respect to expert opinions. Comparison scale includes four levels, namely (0) no influence,
(1) low influence, (2) medium influence, and (3) high influence. The basic notation to compute matrix A is given
as follows.
Xij =influence degree of factor i to factor j.
H = total responds (experts), n = number of factors.
Xk = nxn non- negative matrix for kth respond.
Xijk = influence degree of factor i to factor j with respect to k th respond.
Matrix A = [aij] can be computed as follows.
H
aij = (1/H) x Xijk
(1)
k=1
Step 2: Compute normalized initial direct-relation (matrix D).The formulation to compute matrix D is as follows.
n
S= 1 / ( max aij )
(2)
1<i<n j=1
D= AxS
(3)
Step 3: Compute factor total-influence (matrix T). The formulation to compute matrix D is given as follows.
I= identity matrix
T=D(I-D)-1
(4)
Step 4: Compute C, R, C+R, and R-C values and also set threshold value to obtain digraph of showing causal
relations among criteria. The basic notations to conduct this step are given as follows.
C= sum of column of the matrix T, cj represents direct and indirect effects by factor j from the other factors.
R= sum of row of the matrix T, ri represents direct and indirect effects given by factor i to the other factor.
ri+cj= the importance of factor i
ri-cj= the net effect of factor i
568
3 Case Study
In this study, we used the results of Akinci et al. [2]s study to evaluate the dimen-sions of electronic service
quality through decision-making trial and evaluation la-boratory (DEMATEL) method. A questionnaire based
on E-S-QUAL and E-RecS-QUAL scales was developed. Firstly, we phoned the head office of banks to inform
them about our study. Then, we sent developed questionnaire to 17 different banks departments, namely
department of Internet and mobile banking, department of Internet and e-banking, department of telephone and
internet banking, department of system planning and analysis, department of system development, etc. via e-mail
to obtain completed questionnaires by experts. After approximately one month (9 February 2011 to 10 March
2011), we obtained 39 completed questionnaires from 11 different banks. 6 banks did not answer our request
although we phoned them more than once. Studies which used DEMATEL method did not apply their
questionnaires to large samples in literature, while survey studies were applied to enough number of samples to
interpret accurate comment about interested system. For example, Shieh, Wu and Huang [17] applied their
DEMATEL questionnaire to 19 managerial personnel to evaluate factors of hospital service quality. Wu et al.
[18] also applied their DEMATEL questionnaire to 13 experts to construct casual relationships among criteria
for outreach personnel in Employment service outreach program. Therefore, obtained 39 completed
questionnaires from 11 different banks are enough to conduct this study.
There are eight criteria for E-S-QUAL scale. These are: (A1) good organization of the information in the web
site, (A2) easy usage of the web site, (B1) availability of the web site, (B2) launcing and running duration of the
web site, (C1) accuracy of the information in the web site , (C2) keeping the promises given in the web site, (D1)
protecting information related to customers web shopping behavior, and (D2) not sharing the customers
personal information with other web sites.
The computation results using DEMATEL method are given in Table 1- 4. Table 1 shows the average initial
direct-relation matrix A based on equation (1). Table 2 shows the normalized initial direct-relation matrix D
based on equation (3). The fac-tor total-influence matrix T depicted in Table 3 based on equation (4). Table 2
summarizes the sum of influences given and received among the criteria. According to Table 4, the importance
of the criteria can be prioritized as A1>A2>B2>B1>D1>C2>C1>D2 base on (C+R) values. On the other hand,
A1, A2, B1 and B2 are net causes, while C1, C2, D1, and D2 are net receivers based on (R-C) values. Finally,
the threshold value is determined to draw Figure 1, which depicts the digraph of showing causal relations among
eight criteria. The average of the elements in matrix T (Table 3) is 0.381. Figure 1 was drew by considering
influences values of which are above 0.381.
Table 1.The average initial direct-relation matrix A for E-S-QUAL scale
A1
A2
B1
B2
C1
C2
D1
D2
A1
0.000
2.826
1.056
1.778
0.467
0.667
1.389
0.294
A2
2.063
0.000
1.000
1.500
1.000
1.571
1.250
0.500
B1
1.619
1.286
0.000
1.400
1.400
1.250
2.000
0.750
B2
1.650
1.387
1.441
0.000
1.500
0.667
1.200
1.250
C1
0.917
0.806
0.412
0.306
0.000
1.875
1.500
0.923
C2
0.792
0.469
0.486
0.278
0.767
0.000
1.750
1.333
D1
0.810
0.355
0.371
0.294
0.667
1.069
0.000
1.533
D2
0.864
0.545
0.514
0.543
0.815
1.258
1.167
0.000
Table 2. The normalized initial direct-relation matrix D for E-S-QUAL scale
A1
A2
B1
B2
C1
C2
D1
D2
A1
0.000
0.291
0.109
0.183
0.048
0.069
0.143
0.030
A2
0.213
0.000
0.103
0.155
0.103
0.162
0.129
0.052
B1
0.167
0.132
0.000
0.144
0.144
0.129
0.206
0.077
B2
0.170
0.143
0.149
0.000
0.155
0.069
0.124
0.129
C1
0.094
0.083
0.042
0.031
0.000
0.193
0.155
0.095
C2
0.082
0.048
0.050
0.029
0.079
0.000
0.180
0.137
D1
0.083
0.037
0.038
0.030
0.069
0.110
0.000
0.158
D2
0.089
0.056
0.053
0.056
0.084
0.130
0.120
0.000
569
A1
A2
B1
B2
C1
C2
D1
D2
D2
0.362
0.382
0.427
0.449
0.328
0.333
0.325
0.211
Table 4.The sum of influences given and received among the criteria of E-S-QUAL scale
C
R
C+R
R-C
A1
3.494
3.723
7.217
0.229
A2
3.160
3.719
6.879
0.559
B1
2.217
3.931
6.148
1.713
B2
2.540
3.795
6.336
1.255
C1
2.675
2.616
5.291
-0.059
C2
3.407
2.285
5.691
-1.122
D1
4.101
2.034
6.135
-2.068
D2
2.816
2.309
5.125
-0.507
Figure 1. The digraph of showing causal relations among these eight criteria.
In addition to the analysis for E-S-QUAL scale, we also conduct same method for E-RecS-QUAL scale. There
are seven criteria for E-RecS-QUAL scale. These are: (E1) responding speed to customers request and
questioning which customers made by e-mail or other channels, (E2) telling customers what to do if their
transaction is not processed, (E3) taking care of problems related to web site promptly, (F1) compensating
customers for problems caused by the web site, (F2) compensating customers when their transactions are not
completed on time, (G1) providing online customer service representatives, and (G2) offering the customers to
be able to speak to a liveperson in case of a problem.
The computation results obtained by DEMATEL method for E-RecS-QUAL scale are given in Table 5- 8. Table
5, 6, and 7 are obtained by using equation (1), equation (3) and equation (4), respectively. According to Table 8,
the importance of the criteria can be prioritized as G1>E3>G2>F2>E1>F1>E2 base on (C+R) values. On the
other hand, E1, E2, E3 and G2 are net causes, while F1, F2, and G1are net receivers based on (R-C) values. The
570
threshold value is calculated by averaging the elements in ma-trix T (Table 7). The result is 1.312. Finally,
figure 2 was drawn by considering influ-ences values of which are above 1.312.
Table 5.The average initial direct-relation matrix A for E-RecS-QUAL scale
E1
E2
E3
F1
F2
G1
G2
E1
0.000
1.435
1.913
1.316
1.316
2.091
2.306
E2
2.500
0.000
1.808
1.536
1.440
1.471
1.760
E3
2.750
2.231
0.000
1.533
1.688
2.176
2.061
F1
1.700
2.000
2.462
0.000
1.765
2.081
2.219
F2
2.400
2.154
2.542
2.500
0.000
2.114
2.000
G1
2.000
1.600
2.000
3.000
2.333
0.000
2.059
G2
2.000
1.909
2.286
1.333
2.333
2.318
0.000
E1
E2
E3
F1
F2
G1
G2
Table 8. The sum of influences given and received among the criteria of E-RecS-QUAL scale
C
R
C+R
R-C
E1
8.071
10.056
18.128
1.985
E2
8.062
8.674
16.736
0.612
E3
9.359
9.786
19.145
0.428
F1
9.322
8.570
17.892
-0.753
F2
10.266
8.396
18.662
-1.871
G1
9.877
9.349
19.226
-0.528
G2
9.317
9.444
18.761
0.127
571
Figure 2. The digraph of showing causal relations among these seven criteria.
4 Conclusions
Casual relationships among criteria for both E-S-QUAL and E-RecS-QUAL scales are constructed by using 39
completed questionnaires by means of DEMATEL meth-od. The diagraph of showing casual relations among
criteria is also constructed. In addition, the importance of the criteria for both E-S-QUAL and E-RecS-QUAL
scales are prioritized. The results of this study can be very useful for enterprises to improve perceived service
quality by customers and to gain customer loyalty. Our research results show that there are interdependent
among criteria for both E-S-Qual and E-RecS-Qual scales. As can be seen from computed total-influence
matrixes (see Table 3 and Table 7), all of the criteria are both affected and affect each other. However, the
digraphs of showing causal relations among criteria (see figure 1 and figure 2) was drawn by considering
threshold value to improve intelligibility of relation among criteria and to construct most relation criteria.
For future study, casual relationships and the importance of the criteria for both E-S-QUAL and E-RecS-QUAL
can be constructed for other e-service system.
References
[1] Parasuraman, A.; Zeithaml, V.A.; and Malhotra, A. (2005). E-S-QUAL : A mul-tiple-item scale for assessing electronic
service quality. Journal of Service Research 7(3): 213-233.
[2] Akinci, S.; Inan, E.A.; and Safak Aksoy. (2010). Re-assessment of E-S-Qual and E-RecS-Qual in a pure service setting.
Journal of Business Research 63: 232240.
[3] Ancarani, A. (2005). Towards quality e-service in the public sector: The evolution of web sites in local public service
sector. Managing Service Quality 15(1): 6-23.
[4] Buckley, J. (2003). E-service quality and the public sector.Managing Service Quality 13(6): 453-462.
[5] Ho, C.T.B.; and Lin,W.C. (2010). Measuring the service quality of internet bank-ing: scale development and validation.
European Business Review 22(1): 5-24.
[6] Swaid,S.I.; and Wigand, R.T. (2009). Measuring the quality of e-service: scale development and initial validation. Journal
of Electronic Commerce Research 10(1): 13-28.
[7] Cristobal, E.; Flavian,C.; and Guinalu, M. (2007). Perceived e-service quality (PeSQ) Measurement validation and
effects on consumer satisfaction and web site loyalty.Managing Service Quality 17(3): 317-340.
[8] Ladhari, R. (2010). Developing e-service quality scales: A literature review. Journal of Retailing and Consumer Services
17:464477.
[9] Barrutia,J.M.; and Gilsanz, A. (2009). e-Service quality: overview and research agenda. International Journal of Quality
and Service Sciences 1(1): 29 50.
[10] Hu, Y.C.; and Liao, P.C. (2010). Finding critical criteria of evaluating electronic service quality of Internet banking
using fuzzy multiple-criteria decision making. Ap-plied Soft Computing Journal, doi:10.1016/j.asoc.2011.02.008.
572
Biographies
Serkan Altuntas was born in Malatya in 1984, Turkey. He got the B.Sc degree in industrial engineering, from Eskisehir
Osmangazi University, Eskisehir, Turkey. He got M.Sc degree in industrial engineering, from Dokuz Eylul University in
Izmir, Turkey. His Ph.D. is in progress
in the department of Industrial Engineering at the University of Gaziantep. He has been working as a research assistant at the
Industrial Engineering Department of the Bayburt University since 2009.
His research interests include assembly line, facility layout, simulation, multi-criteria decision making and technology
evaluation.
Trkay Dereli is a professor of Industrial Engineering, Department at the University of Gaziantep, Turkey. He has published
numerous technical papers in professional academic journals and conferences and has several textbooks on CAD/CAM, ICT
and quality management. He is also an active referee for many professional journals and edited several conference
proceedings.
His current research interests include innovation and technology management, CAD/CAM, process planning, feature
technology, TQM, agile and responsive manufacturing and management, soft computing, informatics and applications of
artificial intelligence.
Mustafa Kemal Ylmazwas born in Trabzon in 1975. He got B.Sc degree in Faculty of Economics and Administrative
Sciences, from Karadeniz Technical University in Trabzon, Turkey. He got M.Sc and Ph.D. degrees in Department of
Business, from Ataturk University, in 2005 and 2010, respectively. Since 2001, he had been a lecturer at Ataturk University,
Erzurum. Recently, he accepted a position of Assistant Professor at the Bayburt University, Bayburt, Turkey.
His major field of study is e-services, industrial market, and consumer behavior.
573
1 Introduction
Iterative feedback tuning (IFT), is a method for tuning parameters of a controller in feedback control systems
without needing an explicit model of the system that initially proposed by Hjalmarsson[1]. It has proved to be
very effective in practice and is now widely used in process control [2-5]. and it has been applied to many
experiments such as DC-servo with backlash[4], inverted pendulum[6], and two-mass spring with friction[7]. In
particular, the IFT method is appealing to process control engineers, because under this scheme the controller
parameters can be successively improved without ever opening the loop. The key feature of IFT is that the
closed-loop experimental data are used to compute the estimated gradient of a performance criterion that usually
is a quadratic criterion. In this performance criterion there is a user defined scalar parameter ( ) that expresses
the relative importance of the penalty on the control effort. This parameter used to be adjusted by trial and error
that usually is time consuming and doesn't have enough accuracy. This paper presents a method based on
analysis of variance (ANOVA) and nonlinear regression to find an equation or expression for tuning parameter
in terms of model parameters. Recently, it has been proved that ANOVA is a useful method in tuning
parameters in control[8, 9]. This strategy is derived for First Order plus Dead Time (FOPDT) models of an
industrial plant that can model many of industrial processes.
The paper is organized as follows: An overview on the IFT algorithms used in tuning the linear 2-DOF
controllers is presented in Section 2. The third part of paper introduces the main ideas of ANOVA. Section 4
elaborates on tuning procedure which is followed by a case study simulation that illustrates the effectiveness of
574
the method. In this section IFT performance using the proposed strategy is compare to Relay feedback method.
The paper is concluded in section 5 with a summary of key points and results.
(1)
Where G is a linear time invariant operator, {vt } is an immeasurable (process) disturbance. A block diagram of
the closed loop system is represented in Fig. 1 Where Cr ( ) and C y ( ) are linear time invariant transfer
functions parameterized by some parameter vector ( ) and {rt } is an external deterministic reference signal
v
u
r
y
Cr
Cy
Fig. 1. Block diagram of the closed loop
For a controller of some fixed structure parameterized by , it is natural to formulate the control design
objective as a minimization of some norm of ~
yt ( ) over the controller parameter vector . For this purpose the
following quadratic criterion is adopted:
N
N
1
J
E
(Ly ~
yt ( )) 2
( Lu ut ( )) 2
(2)
2 N t 1
t 1
* arg min J ( )
(3)
To obtain the minimum of J ( ) we would like to find a solution for to the equation:
N
N
~
y
u
J
1
~
0
E
yt ( ) t ( ) ut ( ) t
(4)
N t 1
t 1
In this equation to facilitate the notation we assume that L y Lu 1 . If the gradient J / could be
computed, then the solution of (4) would be obtained by following iterative algorithm:
i 1 i i Ri1
J
( )
(5)
Here Ri is some appropriate positive definite matrix, typically a Gauss-Newton approximation of the Hessian
of J, while is a real positive scalar that defines the step size of algorithm. It is clear that ~
y ( ) is obtained by
i
taking the difference between the achieved response from the system operating with the controller C ( i ) and the
desired response, while the quantities ~y / and u / depend on the unknown system and are thus not
computable. To achieve these quantities IFT estimates them using some special closed loop experiments. For
more details refer to[4]. After estimating the needed signals, an experimentally based estimate of the gradient of
can be formed as follow:
J 1
est
N
~y ( )est u( )est
i
(6)
t 1
575
2 Analysis of Variance
In statistics, analysis of variance (ANOVA) is a collection of statistical models, and their associated procedures,
in which the observed variance in a particular variable is partitioned into components attributable to different
source of variation[10]. In this paper multi-way or n-way ANOVA has been used to reach our purpose. In multiway ANOVA it is determined whether means in a set of date differ when grouped by multiple factors. If so, it
can be verified which factors or combination of factors are associated with the difference. In other words, it
measures the effects of multiple factors on the mean of data[11]. Finally, the result of Analysis of Variance can
be observed in a table named ANOVA table.
3 Tuning procedure
As mentioned before, lambda tuning is carried out using Analysis of Variance and nonlinear regression. Here the
objective is to find an equation or expression for control effort weighting parameter ( ) in terms of
FOPDTmodel parameters. An FOPDT model can be shown as following:
K
G( s)
e td s
(7)
Ts s 1
Where K is the static gain of system, Ts is time constant and t d denotes time delay of system. To utilize
ANOVA for our goal, some experiments have to be performed that involve FOPDT model parameters and
optimal tuning factor [8], [9]. The optimal lambdas are defined using a new cost function that will be
introduced in next section. These experiments lead to a bank of data that needed for using in ANOVA.
3.1 Experiment setup
The experiment setup for FOPDT model is shown in table 1. Considering 5 levels for each parameter, 5 3=125
FOPDT models are generated according to the table and 125 simulations were executed. It is clear that a smaller
value gap between levels would result in more accurate model prediction. In every step, one of the 125 models is
simulated with IFT.
Table 1. Experiment setup for FPODT models
FOPDT
Parameters
K
Ts
td
Level
1
1
1
1
Level
2
3
2.5
2.5
Level
3
5
4
4
Level
4
7
5.5
5.5
Level
5
9
7
7
In every simulation a tuning parameter is obtained that minimizes the following cost function[12]:
* 2
CFj ( M m ( ) M m
) ( M c ( ) M c* ) 2
(8)
j 1...125
Where CFj is the cost function for the j th model. In above equation M m ( ) is modulus margin, which is
defined as the inverse of the infinity-norm of the sensitivity function. This value is simply the shortest distance
from the Nyquist curve of the loop transfer function to the critical point -1. M m* is specified value of the modulus
margin that it is chosen equal to 0.65.Mc() is complementary modulus margin that is equal to the inverse of the
infinity- norm of the complementary sensitivity function. Its value is related to the response peak of the transfer
function from setpoint to process output and thus constitutes an important indicator of the response to setpoint
changes and M c* is the desired complementary modulus margin that is chosen equal to 0.72. So, a tuning
parameter that minimizes the cost function (8) is defined as the optimal .
576
9
9
3.2
Ts
1
1
7
7
td
1
2.5
5.5
7
CFj
CF1
CF2
CF124
CF125
opt
opt1
opt2
opt124
opt125
Analysis of Variance
After constructing the bank of data in previous section, an analysis of variance is performed on the optimal
tuning parameter as a response vector or dependent
variables and FOPDT model parameters as independent
variables. Therefore, model parameters or combination of them that have a significant influence on the optimal
tuning parameter set would be fined using analysis of variance. Furthermore, the level of contributed influence
and its validity is also determined by ANOVA. The result for this analysis is depicted in table 3. In this table up
to 2 way interactions between the model parameters are considered.
It can be observed that there are F-values and P-values associated with each parameter and interaction. These
values indicate the effect of model parameters on optimal lambda. In ANOVA table a parameter or interaction
larger F value has larger influence on optimal tuning parameter opt . Typically there is a cut-off value of 0.05 for
P index. Any parameter or interaction having a value under this cut-off value is considered to be significant. The
remaining parameters and interactions are neglected and their effects are considered as the error. It can be
deduced from the date that 5 terms have P-values less than the cut-off rate and hence should be considered
significant. Also, the corresponding F value shows that most influential parameter is K and after it td andTs have
a considerable F value. The effect of combination of parameters has been disregarded. As mentioned above,
opt will depend on FOPDT model parameters and is a function of it and can be expressed in terms of this
model parameters.
Table. 3: ANOVA results
Source
k
Ts
td
k Ts
Ts t d
k td
Error
Total
Sun of
Squares
104.565
7.18
D.O.F
4
4
Mean
suares
26.141
1.795
88.713
6.092
0.000
0.000
12.838
3.21
10.8921
0.000
5.573
16
0.348
1.182
0.306
24.131
16
1.508
5.118
0.000
10.083
16
.63
2.139
0.017
18.859
318.576
64
125
0.295
To find this function or expression, nonlinear regression is performed on the opt and model parameters using
many possible combinations of the significant factors, until a good correlation coefficient was obtained. The
tuning expression that best fit the optimal values of the optimal parameter opt is:
(9)
Where fit parameters with 95% confidence bounds and R 0.98 are:
2
577
(10)
[1 2 3 4 5 6 7 ] [1.998 0.0517 1.998 0.044 1.991 - 2.22 1.609]
To study the accuracy of this equation and prediction ability of it, in Fig. 2.a, the estimated lambda set is
compared to optimal lambda set for FOPDT models. It can be observed that estimated values for lambda are
close to optimal values. For a better deduction, a new set of model parameters that didnt have any role in
extracting equation (9) is used to probe the accuracy of the method. In Fig. 2.b the estimated lambda set is
compared to optimal lambda set for the test FOPDT model parameters. It is
observed that for some models
the estimation error seems to be high. But this error doesn't make any problem in system response.
Illustrativeexample. In this section, we attempt to test our proposed tuning expression on a sample case study.
The case study system is a gas fire boiler where the gas valve position is used to control the temperature of the
water[13]. The transfer function e 5s \ (1 3s)(1 s) is used to represent the relationship between the variations in
gas valve position and temperature of the water when the flow of water is constant. The step response of this
system has a non-oscillatory behavior and a small overshoot. Therefore, it can be estimated with a first order
model. The FOPDT model estimated by frequency response method for this system is as follow:
Pest ( s)
e 5.837s
3.162s 1
(11)
Estimated lambda
Optimal lambda
Estimated lambda
y = 01900121900r1lx +
01900451900r1l
R = 01900271900r1l
Optimal lambda
Using model parameters and equation (9) the optimal lambda for this model is equal to 0.053 . Fig.3
depicts the step response and corresponding input signal of simulated system controlled by a PI controller that
has been tuned by IFT, in presence of disturbance. Also, it can be seen in Fig.3 that the case study system has
been simulated using Relay feedback method and modified(Hang) Relay feedback method for dominant time
delay process[14]. Performance of IFT method tuned by the proposed strategy can be compare to these methods.
578
0
-0.5
50
100
150
Time
control signal
1.4
IFT
R-F
Modified R-F
1.2
1
0.8
0.6
0.4
50
100
150
Time
Conclusion
In this paper a new systematic method for tuning , that is an important parameter in IFT algorithm, has been
proposed. This parameter used to be tuned by try and error method that was a time consuming method. ANOVA
strategy has been derived for FOPDT models. In this method, analysis of variance is utilized for finding the
parameters that have significant effect on tuning , which performed using 125 IFT simulations. After
producing this information, nonlinear regression was performed, using many possible combinations of the
significant factors, until a good correlation coefficient was obtained. Therefore it can be possible to derive an
expression for tuning . The simulation results show the effectiveness of the proposed tuning method for
FOPDT models.
References
1.
2.
3.
4.
5.
6.
7.
Hjalmarsson, H., S. Gunnarsson, S.et al. (1994). A convergent iterative restricted complexity control
design scheme. Proc. 33rd IEEE CDC, FL, USA.
Hjalmarsson, H. (1999). Efficient tuning of linear multivariable controllers using iterative feedback
tuning. International journal of adaptive control and signal processing 13(7): 553-572.
Hjalmarsson, H. (2002). Iterative feedback tuning-an overview. International journal of adaptive control
and signal processing 16(5): 373-395.
Hjalmarsson, H., M. Gevers, et al. (1998). Iterative feedback tuning: Theory and applications. Control
Systems Magazine, IEEE 18(4): 26-41.
Lequin, O., M. Gevers, et al. (2003). Iterative feedback tuning of PID parameters: comparison with
classical tuning rules. Control Engineering Prac.11(9): 1023-1033.
Codrons, B., F. De Bruyne, et al. (1998). Iterative feedback tuning of a nonlinear controller for an
inverted pendulum with a flexible transmission. Proc. of the IEEE International Conference on Control
application, Trieste, Italy, IEEE.
Hamamoto, K., T. Fukuda, et al.(2003). Iterative feedback tuning of controllers for a two-mass-spring
system with friction. Control Engineering Prac.11(9):1061-1068.
579
8.
9.
10.
11.
12.
13.
14.
Iglesias, E. J., M. E. Sanjuan, et al. (2006). Tuning equation for dynamic matrix control in SISO loops.
Ingenieri y Desarrollo (19): 88-100.
Neshasteriz, A., A. Khaki-Sedigh, et al. (2010). An analysis of Variance Approach to tuning of
Generalized Predictive Controllers for Second Order plus Dead Time models. 8th IEEE International
Conference on Control and Automation, Xiamen.
Hogg, R. and J. Ledolter (1987). Engineering statistics MacMillan, New Jersey.
Montgomery, D. C. and G. C. Runger (2007). Applied statistics and probability for engineers, Wiley
New York.
Garcia, D., A. Karimi, et al. (2005). PID controller design with specifications on the infinity-norm of
sensitivity functions. 16th IFAC World Congress, Prague, Czech Republic.
Normey-Rico, J. E. and E. F. Camacho (2007). Control of dead-time processes, Springer Verlag.
O'Dwyer, A. (2009). Handbook of PI and PID controller tuning rules, Imperial College Pr.
580
Abstract. The nozzle is the part of the air-jet and Jetring yarn production systems. In these spinning
systems, the pressurized air is fed into the nozzle and the yarn is twisted by the airflow arising in the nozzle.
The nozzle design and air pressure level affects the yarn quality. Present work analysed the airflow variables
and character in the nozzle having different injector angles computationally. Nozzle injector angle is one of
important nozzle parameter affecting the yarn properties. In the analysis, ANSYS 12.1 package program and
Fluid Flow (CFX) analysis method was used. The airflow variables, particularly air pressure, air velocity and
air density, and the airflow character of the nozzles having different injector angles were obtained.
Keywords: Nozzle, CFD, airflow, Jetring, yarn spinning
1 Introduction
In any spun yarn production, there are mainly two requirements. One of them is to lay the fibres as parallel
form while the latter is to hold the fibres together by twist. Twist is defined as the measure of spiral rotations of
the fibres. Fibre strands are twisted to induce the lateral forces. Lateral forces act to bring the fibres closer
together and hence friction is created among the fibres. These forces control the fibre slippage and so it will be
difficult to break the twisted fibre strand. Many different types of equipment have been using to insert the twist
into the fibre strand. One of them is the pneumatic unit and nozzles are widely used as a pneumatic twister.
Nozzles are the main component of air-jet and Jetring yarn production systems. In these spinning systems,
pressurized air having certain value is injected into the nozzle to create an air vortex. The fibre strand is passed
through the nozzle and twisting and so yarn spinning is realized.
The nozzle component of air-jet and Jetring spinning systems consists of nozzle head and nozzle body (Fig.
1). Nozzle head transfers the pressurized air coming from the compressor into the nozzle body by means of the
injectors. Nozzle body has a cylindrical cross section comprising of main hole, injectors, twisting chamber and
nozzle outlet. Main hole lies starting from the nozzle inlet to nozzle outlet and the nozzle has a constant
diameter. Injectors are positioned at certain angles with respect to the nozzle axis and so they lie tangentially.
Basically, there are two expected phenomena in the nozzle. One of them is to strengthen the sucking of the
drafted fibre strand from the main hole into the nozzle by the pressurized air. Second is to produce the twist
resulted from the swirling air flow depending on the injector design and the air pressure level.
581
(a)
(b)
(c)
Fig. 1. Nozzle head (a), nozzle body (b) and the nozzle (c) (1: Main hole, 2: Injectors, 3: Twisting chamber, 4:
Nozzle outlet)
Sucking and twisting cases of fibre strand depend on the nozzle structural parameters. The most important
nozzle parameters are the diameter of main hole and injectors, injector angle, nozzle length and injector position.
Many researchers studied these parameters and the main subjects of these studies are the effect of air pressure
level and main hole diameter on yarn quality. The findings concerning with the air pressure and main hole
diameters are agreed with each other [1-5]. Regarding with the injector angle, many researchers widely centre
upon the angle range between 30 and 60 [3, 6-9]. Therefore, we believe that it should be analysed wider
injector angles to attain general results about the effect of injector angle. Base on above reasons, we aimed to
analyse the effect of injector angle on airflow character and fluid variables in the nozzle. In the analysis, ANSYS
12.1 package program and Fluid Flow (CFX) analysis method was used.
2 Method
2.1 Nozzle Geometry
In the study, each nozzle configuration is firstly created in ANSYS Workbench as a fluid. As it was
mentioned in previous part, the nozzle used in present work has a cylindrical cross section and consists of main
hole (1), injectors (2), twisting chamber (3) and nozzle outlet (4) (Fig. 1).
To determine the effect of injector angle, three different injector angles were analysed. In addition to 30 and
45 angles, 15 was selected which is different from the literature (Fig. 2). When the angle was changed, the
other nozzle structural parameters were kept constant.
In literature, it is assumed that the yarn occupies a small area in the nozzle hence, the flow inside the nozzle
affects the yarn, but not vice versa [8-9]. Therefore, the yarn was not modelled in the numerical analyses
concerning with Jetring or air-jet spinning systems. However, the yarn is the one of main component of these
spinning systems. Present work includes the area from where the yarn is passing, and studies the flow field
around the yarn (Fig. 3). Additionally, it was given a speed which is equal to production speed of spinning
machine to the yarn (approximately 0.2 m/s).
(a)
(b)
(c)
582
583
The direction of the airflow at the nozzle inlet varies depending on the nozzle configuration and so it is not
possible to specify accurately at the beginning of the analysis. Therefore, nozzle inlet was defined as opening
and the cases of air coming in or leaving was being the optional depending on the main flow conditions in the
nozzle. At the nozzle inlet, the air pressure was the atmospheric pressure conditions. Atmospheric pressure was
assumed at the nozzle outlet. Non-slip boundary condition was applied to the nozzle walls and also to the yarn.
In the work, iteration number was taken five as a minimum value and 300 as a maximum. Depending on the
converging condition of the solution, iteration was performed over 300.
Sensitivity of the solution was set at 10-5.
Injectors
(kg/m3)
1.99
1.927
1.949
Nozzle inlet
(kg/m3)
1.23
1.21
1.2
Nozzle outlet
(kg/m3)
1.207
1.208
1.208
584
585
(a)
(b)
Fig. 6. The changes in the air velocity of the nozzles having 15 (a) and 45 (b) injector angles
When the airflow on the yarn is analysed, it was found that the air goes forward without swirling from nozzle
inlet to the injectors (Fig. 7). Suction effect at the nozzle inlet may lead to this airflow case. And then the airflow
starts to swirl beyond the injectors and moves to the nozzle outlet. Particularly, in 15, the airflow covers the
yarn completely and regularly. As the injector angle increases, the airflow character gets irregular and intensity
of the airflow on the yarn decreases.
(a)
(b)
(c)
Fig. 7. The airflow on the yarn for 15 (a), 30 (b) and 45 (c) injector angles
4 CONCLUSION
In this study, the airflow in the nozzle of Jetring and air-jet spinning systems was simulated by ANSYS Fluid
Flow (CFX) analysis method. There are also studies based on the numerical analysis of the nozzle. On the
contrary of the literature, present work takes into consideration the effect of the yarn on to the airflow and
therefore the yarn was included to the analysis. Shear Stress Transport (SST) turbulence model encompassing
both the k- and the k- models was selected as a solution algorithm. Hence the airflows at the near-wall region,
and at the outer wake region and in free shear layers were simulated. Mass flow values of the pressurized air
were measured for all analysed injector angles and the air pressure values were used as a boundary condition. At
the end of the analysis, measured and calculated mass flow values were compared to validate the analysis.
In the study, it was found that airflow variables such as air pressure, velocity and density change along the
nozzle length for all analysed injector angles. Air pressure and velocity are the lowest at the nozzle centre while
they are the highest close to the walls. At the nozzle inlet and also nozzle outlet, a suction effect was observed
depending on the injector angle. When the airflow character in the nozzles is analysed, three types of airflow
structure was observed. These airflows depend on the injector angles. These mentioned fluid behaviours lead to
the air displaying different flow attitudes at the nozzle injectors, main hole and nozzle outlet associated with
different airflow properties.
586
References
1.
2.
3.
4.
5.
6.
7.
8.
9.
10.
Oxenham, W., Basu, A. (1993). Effect Of Jet Design On The Properties Of Air-Jet Spun Yarns, Textile
Research Journal, 63 (11): 674-678.
Zeng, Y.C., Yu, C.W. (2003). Numerical Simulation Of Air Flow In The Nozzle Of An Air-Jet
Spinning Machine, Textile Research Journal 73 (4): 350-356.
Rengasamy, R.S., Kothari, V.K., Patnaik, A., Ghosh, A., Punekar, H. (2005). Reducing Yarn Hairiness
in Winding By Means Of Jets: Optimisation Of Jet Parameters, Yarn Linear Density And Winding
Speed, AUTEX Research Journal 5 (3): 127-132.
Guo, H.F., Chen, Z.Y., Yu, C.W. (2007). Numerical Study Of An Air-Jet Spinning Nozzle With A
Slotting-Tube, Journal Of Physics: Conference Series 96: 1-5.
Patnaik, A., Rengasamy, R.S., Kothari, V.K., Bhatia, S.K. (2008). Some Studies On Hairiness
Reduction Of Polyester Ring Spun Yarns By Using Air-Nozzles During Winding, TJTI 99 (1): 17-27.
Jeon, B.S. (2000). Effect Of An Air-Suction Nozzle On Yarn Hairiness And Quality, Textile Research
Journal 70 (11): 1019-1024.
Zeng, Y.C., Yu, C.W. (2004). Numerical and Experimental Study on Reducing Yarn Hairiness With
The JetRing and JetWind, Textile Research Journal 74 (3): 222-226.
Patnaik, A., Rengasamy, R.S., Kothari, V.K., Punekar, H. (2006). Airflow Simulation In Nozzle For
Hairiness Reduction Of Ring Spun Yarns. Part II: Influence Of Nozzle Parameters, Journal Of Textile
Institute 97 (1): 97-101.
Rengasamy, R.S., Patnaik, A., Anandjiwala, R.D. (2008). Simulation Of Airflow In Nozzle-Ring
Spinning Using Computational Fluid Dynamics: Study On Reduction In Yarn Hairiness And The Role
Of Air Drag Forces And Angle Of Impact Of Air Current, Textile Research Journal, 78 (5): 412-420.
Garg, V.K., Ameri, A.A. (2001). Two-Equation Turbulence Models For Prediction Of Heat Transfer On
A Transonic Turbine Blade, International Journal of Heat and Mass Transfer, 22: 593-602.
587
Introduction
Space agencies throughout the world operate optical imaging satellites that capture high-resolution images of
the earth for improving our knowledge and for better utilization of the earths resources. These satellites gather
massive amounts of geographical information on a regular basis and there is a need to automate the process of
extracting structures appearing in these images. Our focus is on automatic extraction of major roads. Currently
existing approaches for this problem can be categorized as local and global methods. Some of the local methods
proposed so far are generic edge detectors [1,2], crest detectors [3], and morphological operators [4,5]. Neural
networks and statistical classifiers have also been proposed [6]. A popular global method is the one introduced in
[7] where the problem is formulated over a graph in the underlying pixel lattice. Another graph-based iterative
global method that our classifier is inspired from is the one proposed in [8] that is later referred to in [9]. In this
method, a relatively small number of paths emanating from the current root in the underlying decision tree
representation are examined, after which some are kept and the others are discarded. As discussed in [9], the
fundamental difficulty in automatic road extraction is that pixels representing the road are not systematically
brighter or darker than those representing the surrounding background in practice.
588
Problem Definition
The scenario is a statistical model for road tracking. A root node associated with a ternary tree of fixed depth
D is given. Starting with the root node, we assume that each tree segment goes either straight, or turns left or
right at each junction relative to the direction of the previous segment. Note that under this assumption, there
are exponentially many paths, 3D to be exact, and out of these paths, only one of them is the true path. Our
goal is to find which one it is based on the statistical characteristics of the segments of the ternary tree.
2.1
Statistical Model
The space Y of classes is the set of paths emanating from the root node to a leaf node in the ternary tree of
depth D. Thus, there areY = 3D possible classes. Segments of the ternary tree will be denoted by s for s S.
For example, for D = 3, there are 33 = 27 paths, and 39 segments. In general, two distinct paths (or classes) can
share up to D-1 segments in common. We associate a direction with the root of the tree and then assign angles to
the two turns and lengths to the segments, so each class can be interpreted as an actual piecewise linear, planar
path. With this setup, the set Y can be considered as a representation of a model for road tracking.
We let a 3x3 transition matrix Q specify the transitioning probabilities from each one of the three directions to
the direction in the next road segment. It will be assumed that Q is uniform, i.e., all of its entries are 1/3.
Motivation for this specific assumption on the transition matrix Q will be discussed in the section describing the
classification algorithm. Note that the random variables Y1, Y2, , YD form a first-order Markov chain for the
state of the segments. These random variables are independently and identically distributed and each of the 3D
paths has an equal likelihood of p(y) = 3-D for y Y. Here, Yl {Straight, Left, Right} for 1 l D
denoting the direction of segment l relative to the previous segment in the path.
2.2
Data Model
For each segment s S, the observation Xs will denote a measurement obtained by a local filter. We suppose
that Xs {1, , M}, 2 M, with high values indicating path and low values indicating background. We
assume two conditions in our model: space invariance and conditional independence. The first assumption
implies that given Y = y, every Xs such that s y (i.e., s is a segment on the path y) has density f1(x), x = 1, ...,
M; and every Xs such that s y has density fo(x), x = 1, ..., M. The second assumption states that the random
variables X = {Xs, s S} are conditionally independent given Y. These two assumptions yield the following
likelihood model:
p(x y) = f1( xs ) f 0 ( xs )
sy
(1)
sy
In this study, the observations Xs will denote the color of the segment. We assume that Xs {G, D, B} where
G stands for gray, D stands for dark gray, and B stands for black (i.e., M = 3). That is, for both the true
path and the background, the output of the local filter is one of the three possible colors; gray, dark gray, and
black. Darker colors will indicate "the true path" and lighter colors will indicate "background".
In the subsequent sections, computational experiments with different distributions for the true path and the
background will be presented. However, a typical problem instance has probability mass functions as in Fig. 1.
589
Fig. 1. f1(x): mass function for the true path, f0(x): mass function for the background.
For simplicity, we will assume that the probability masses for the true path and the background are symmetric,
as in above example. In other words, we assume that
f1(B) = f0(G), f1(D) = f0(D), and f1(G) = f0(B).
This assumption is solely for the purpose of easily recognizing various likelihoods in the computational
experiments section. The symmetry assumption has no implications on the classification algorithm we present.
Clearly, the quality of solutions obtained by any classifier within the context of the tracking problem is heavily
dependent on the degree of overlap between the supports of the true path and the background. In particular, if
they have disjoint supports then tracking is trivial and if they are very similar, it is exceedingly hard. In the
computational experiments section, we present the results of our classification algorithm for various supports for
the true path and the background.
In this section, we present our classification algorithm, which is based on successively approximating the
Maximum Likelihood Estimator (MLE). We call our algorithm MLE-Successive Approximator (MLE-SA). The
MLE-SA algorithm is essentially based on the algorithm discussed in [9]. We first note that the optimal solution
for the road tracking problem can be obtained theoretically by computing the maximum a posteriori (MAP)
estimator:
X MAP arg max P X x Y1 ,..., YD
xX
(2)
xX
As discussed earlier, we assume that our priors are uniform (recall that the random variables Y1, Y2, , YD are
assumed to be independently and identically distributed). Accounting for the uniform prior, we notice that the
MAPestimator is equivalent to the MLE:
X ML arg max P Y1 ,..., YD X x
x X
arg max
x X
arg max
x X
arg max
x X
f1 (Ys ) f 0 (Ys )
f1 (Ys )
f0 (Ys )
f 0 (Ys ) sS
f1 (Ys )
f 0 (Ys )
s:xTs
s:xTs
s:xTs
(3)
s:xTs
Therefore, we conclude that under the uniform prior assumption, the optimal strategy is to maximize the
product of the likelihood ratios (note that Ts above denotes the set of paths that contain the segment s, s S). If
the depth of the ternary tree D is 50, for instance, then there are a total of 3 50 paths. It is clearly impossible to
ML
compute X
over this many possible paths. Therefore, MLE is computationally infeasible for practical values
of D. In order to handle large values of D, we successively approximate the MLE: we approximate X
by
successively tracking over relatively smaller pieces of length d, where d is a parameter such that d 7 (due to
computer-memory limitations).
ML
590
For instance, in the first step, we compute the following MLE approximation:
X d arg max
xX , x d
f1 (Ys )
f 0 (Ys )
s:xTs
(4)
However, the segments near the bottom of X are not reliable enough to re-initiate tracking. For this reason,
we keep only the first k-many segments of each maximizing approximation path, and discard the remaining d k
segments. Here, k is a parameter between 1 and d. Thus, our algorithm has two parameters: d and k. In the
computational experiments section, we show results obtained by MLE-SAfor different values these parameters.
d
During each iteration of the algorithm, we assume that the color of the segments is given up to the next d
levels in the ternary tree. Thus, our algorithm processes only 3 d levels at any given time during the iteration and
does not require that the complete tree is available to start execution. This property of the algorithm enables us to
evaluate new data as it comes along. Prior to the execution of the actual algorithm, we randomly generate the
true path, which is of depth D. Our classifier is computed from the data model described above, not from
samples. That is, for each segment on the true path, its color is sampled from the probability mass function f1(x),
and colors of background segments are sampled from f0(x). Recall that for either type of segments, the color will
be one the colors gray, dark gray, and black. In our methodology, rather than a 0-1 loss function, we define the
distance between the true path (TP) and the approximate path (AP) obtained by the MLE-SA algorithm as:
Dist(TP, AP) = D APACS
(5)
where APACS denotes the number of the segments that AP classifies accurately. Of course, if APACS = D, then
Dist(TP, AP) = 0, and the algorithm has correctly found the true path. For each problem instance, we run the
algorithm 100 times, each time with a different true path and different color readings. After 100 runs, the
following two values are computed:
AVGDIST (Average Distance): average distance between the true path and the overall approximation path
over these 100 runs.
ACCRT (Accuracy Rate): the percentage of times MLE-SA was able to accurately identify the complete true
path over the 100 runs.
Computational Experiments
This section presents performance of MLE-SA with different parameters on different data sets. Our purpose is
to gain insight about the quality of solutions obtained by our algorithm, as well its sensitivity to model and
algorithm parameters.
Recall that for simplification, we assume that f1(x) and f0(x) are symmetric with three possible values: gray
(G), dark gray (D), and black (B). These two masses will be denoted as a single cumulative distribution function
with two parameters, CDF(.,.) such that summation of the two parameters of CDF() gives the percentage of
591
overlap. For example, the masses in Figure 1 above will be denoted as CDF(0.1, 0.4).Also, CDF(0.1, 0.4)
implies that the percent of overlap between f1(x) and f0(x) is 50%.
Various masses for the true path and the background will be considered, with overlap percentages varying
from .01% to 50%. Two different values will be used for D; 7 and 50. The reason we are interested in D = 7 is
that in this case MLE-SA with d = 7 reduces to MLE and thus we obtain the actual optimal solution. As for the
algorithm, it is clear that d = 7 and k =1 will yield the best results. However, in order to evaluate the sensitivity
of the algorithm to its parameters, we will run the algorithm for both d = 5 and d = 7 as well as for k = 1 and k =
2for D = 50. These options will provide us valuable insight about the quality of solutions obtained by MLE-SA
as well its sensitivity to different model and algorithm parameters. As mentioned earlier, throughout our
computational experiments, we run the algorithm for each problem instance for 100 times, each time with a
different true path and different color readings.
k
1
2
1
2
AVGDIST, ACCRT
2.5, 81%
43.62, 0%
0.9, 87%
46.12, 0%
As can be observed from the above table, the algorithm is extremely sensitive to k. In particular, the accuracy
rate is 0% for k = 2 for both d = 5 and d = 7. We experimented with various other likelihoods, and each time we
observed a similar behavior. Therefore, we conclude that k should always be chosen as 1 for better solution
quality. However, it should also be noted that this doubles the execution time compared to
k = 2.
We observe from the above table (as well as from several other likelihoods that we experimented with) that
the algorithm is sensitive to d as well, even though not as sensitive as to k. In this specific case, accuracy rate
goes up 6% when d is increased from 5 to 7. In the case of this parameter, however, execution-time trade-off is
much more significant compared to that of k; execution time of the above problem instance was 4 seconds for d
= 5 whereas it was 46 seconds for d = 7 (approximately 3(7-5) = 32 = 9 times slower, as expected).
592
Overlap %
CDF
Parameters
0.15
1.5
7
15
16
20
25
31
35
41
45
50
0005, .001
.005, .01
.02, .05
.05, .1
.01, .15
.05, .15
.1, .15
.01, .3
.05, .3
.01, .4
.05, .4
.1, .4
AVGDIST,
ACCRT
forD = 50
0, 100%
0, 99%
0.01, 96%
0.9, 87%
0.01, 95%
2.8, 69%
14, 37%
1.3, 86%
15.4, 39%
4.2, 65%
29.7, 10%
42.03, 1%
AVGDIST,
ACCRT
for D = 7
0, 100%
0, 100%
0, 100%
0.02, 99%
0, 100%
0.03, 99%
0.16, 94%
0.07, 98%
0.3, 93%
0.16, 97%
1.09, 73%
2.52, 39%
Our overall results from extensive Monte Carlo simulations with different MLE-SA parameter values and
various overlaps in road and background supports are as follows: (1) The algorithm is extremely sensitive to k,
so k should always be chosen as one despite the fact that this doubles the execution time compared to k = 2. (2)
The algorithm is somewhat sensitive to d as well, yet the execution time is tripled for each unit increase in d.
Values of d greater than 7, on the other hand, are computationally infeasible. (3) The algorithm performs
remarkably well for road-background overlaps up to 20%. Quality of the solutions obtained by the algorithm
gradually deteriorates as the overlap between the supports increases.
We therefore conclude that MLE-SA is an efficient algorithm for supports not overlapping for more than 20%
(of course, for the case where M = 3). For supports overlapping for more than 20%, the algorithm is still
satisfactory when the first overlap parameter is small. If this is not the case, then algorithm loses track rather
quickly and gives low-quality results.
593
References
1.
2.
3.
4.
5.
6.
7.
8.
9.
Neviata, R. (1982). Locating structures in aerial images. IEEE Transactions on Pattern Analysis and
Machine Intelligence, 476-484.
Wang, E.; and Newkirk, R. (1988). A knowledge-based system for highway network extraction. IEEE
Transactions on Geoscience and Remote Sensing, 26(5): 525-531.
Blanc, H.V. (1993). Application du groupement perceptuel la reconnaissance de routes sur une image
satellite SPOT. Technical Report, University de Montpellier.
Destival, I and Le Men, H. (1986). Detection of linear networks on satellite images. InProceedings of
Conference on Pattern Recognition.
Garnesson, P. (1991). MESSIE: un systeme danalyse de scenes. Technical Report, University de NiceSophia Antipolis.
Boggess, J.E. (1993). Identification of roads in satellite imagery using artificial neural networks: a
contextual approach. Technical Report, Mississippi State University.
Fischler, M.; Tenenbaum, J. and Wolf, H. (1981). Detection of roads and linear structures in lowresolution aerial imagery using a multisource knowledge integration technique. Computer Graphics and
lmage Processing, 15: 201-223.
Serendero, M.A. (1989). Extraction dinformations symboliques en imagerie SPOT reseaux de
communication et aglomations, These de Doctorat, University de Nice.
Geman, D. and Jedynak, B. (1996). An Active Testing Model for Tracking Roads in Satellite Images.
IEEE Transactions on Pattern Analysis and Machine Intelligence, 18(1): 1-14.
Biographies
Vural Aksakall received the B.S. degree in mathematics from Middle East Technical University and the M.S.
degree in industrial engineering and operations research from North Carolina State University. He received a
second M.S. degree and a Ph.D. degree in applied mathematics and statistics from Johns Hopkins University in
Baltimore, Maryland. He is currently an assistant professor in the Industrial Engineering Department at Istanbul
ehir University. His research interests are in data mining, stochastic optimization, and electronic commerce.
Ali Fuat Alkaya received the B.S. degree in mathematics from Ko University, the M.S. degree in computer
engineering and Ph.D. degree in engineering management from Marmara University. He is currently an assistant
professor in the Computer Engineering Department and Vice Dean of the Engineering Faculty at Marmara
University. His research interests include scheduling, analysis of algorithms, and combinatorial optimization.
Ahmet Bulut received the B.S. degree in computer engineering from Bilkent University and Ph.D. degree
from University of California Santa Barbara. He is currently an assistant professor in the Computer Science
Department at Istanbul ehir University. His research interests include information technologies for smart cities,
service platforms enabling utility computing, intelligent transportation systems, and applications of wireless
sensor networks.
594
Abstract. Data mining tools has been extensively used for quality improvement in manufacturing environments
due to need of effectively managing huge amounts of data gathered from the manufacturing processes. Especially in
the last decade, a great variety of data mining tools have been used for different purposes in the area of quality
improvement. The main problem in the area is to be able choose the right tool for the specific function aimed to be
realized like classification, prediction, clustering and association. The aim of this study is to make a critical
literature survey and perform a classification about the data mining tools applied for quality improvement and
provide a guideline for enabling the users to select the most appropriate tools for achieving the specific data mining
functions. Finally, conclusions and future directions are presented.
Keywords: quality control, data mining, manufacturing systems
1 Introduction
Todays manufacturing environment is being more and more competitive due to the effects of globalization
and rapid advances in technology. Quality can be identified as a crucial concept for strategic planning to survive
in this competitive environment by ensuring that the final products meets the required standards at minimum
manufacturing cost and time [1].
Traditionally, quality standards are mainly being provided by Statistical Process Control (SPC). However, in
complex manufacturing processes, product quality is influenced by many factors that involve casual relationship
and mutual interaction between each other. Most statistical tools utilized in quality control processes must meet
several assumptions such as the normality and constant variance of the variables. But in fact, it is hard to meet
these assumptions in practice [2], [3]. In addition, the need to analyze large amounts of data brings an extra
difficulty to the quality control process. The need for understanding complex interactional effects and extracting
useful knowledge free of any assumptions from huge data sets led to the development of new techniques of
knowledge discovery in databases (KDD). Data mining methodologies constitute an important step involved in
the application of KDD [4], [5].
Data mining is a technology to transform the huge data gathered from the manufacturing processes into useful
knowledge that can be acquired by realizing its functions like classification, prediction, clustering and
association. The results acquired through these functions may be helpful in solving the problems about many
applications of manufacturing such as fault detection, detailed design, production planning, quality assurance
and scheduling[6],[7]. Moreover, an increasing number of manufacturing data are also available to extract
relevant knowledge for quality improvement which can be analyzed efficiently by the use of data mining tools
[3],[8]. But, the main problem for such an application remains as the difficulty of determining the most efficient
data mining tool among many strong alternatives. In the quality literature, comprehensive surveys about the data
mining tools applied for quality improvement are still required.
595
The aim of this study is to make a critical literature survey and perform a classification about data mining tools
applied for quality improvement and provide a guideline for enabling the users to select the most appropriate
tools for achieving the specific data mining functions.
Briefly, this study aims to answer the following research questions:
1.
2.
Which data mining tools are frequently used in quality improvement? What are the reasons for selecting
these tools?
What kinds of functions are frequently being implemented with these data mining tools?
This study organized as follows: Section 2 describes the review methodology. Most common data mining tools
used in quality improvement are explained in Section 3. The conclusions and research directions related to the
data mining applications in quality improvement are given in the last section.
2 Review Methodology
The method of literature survey in this study depends on a research about the data mining applications
performed in the quality improvement area during the last 15 years. The authors reviewed 27 papers published in
refereed journals and conference proceedings related to the applications of data mining tools for quality
improvement in manufacturing sector. The papers included in this review were classified with respect to the data
mining tool they had applied and the specific function utilized by this tool. The results of the classifications are
given Table 1.This table stands as the main source to answer the research questions in this study.
In this study,among the many available data mining tools, more frequently used tools such as decision trees,
fuzzy sets, neural networks, regression and association rules are considered.
Table 1.Taxonomy of data mining tools with respect to their data mining functions
Data Mining Functions
Prediction
[1]
Decision
Trees
Classification Clustering
Association
Rules
[10],[13],[14],
[23],[24],[25],
[26],[30],[34]
[29]
[32]
Fuzzy Sets [17],[9]
[1],[15],[18],[19],[20], [23],[25],[28], [31],[32]
[21],[22], [33]
[34]
Neural
Networks
Regression [12],[34]
[9],[11],[16],
[27]
Association
Rules
596
Decision Trees
Fuzzy Sets
Neural Networks
Regression
Association rules
In this section, the specific efficiencies of these tools and the specific problems addressed during their execution
are discussed in details. The results regarding the functions implemented by these tools are given in the next
section.
597
The review study shows that data mining tools have been successfully applied to quality improvement
activities[35] due to their efficiency in solving complex quality problems arousing through the manufacturing
activities.
Taxonomy of data mining tools with respect to their data mining functions is given Table 1. The below
conclusions can be drawn from the table:
Neural Network is the most common tool in quality improvement application due to its capability
for efficiently constructing complex relations between the main parameters of manufacturing
processes in a practical way.
Decision tree is the second most commonly used tool due to its simplicity of application and
interpretation.
Fuzzy sets are used for hybrid approaches especially with the support of neural networks. The main
reason for this may be that they are inherently suitable for coping with linguistic domain
knowledge efficiently and producing more interpretable solutions at the end.
Classification is the most common data mining function realized by the utilization of these tools.
Prediction function comes in the second rank when the frequency of usage is evaluated. Decision
tree is mostly preferred for classification, whereas neural network approach is mostly used for
prediction. Generally, for both functions, neural network approach gives more accurate results than
the decision trees.
Association rule has come forward to be an effective tool for identifying associations within the
quality parameters in manufacturing area.
The research in this paper has concentrated on identifying the most frequently used data mining tools, and
discussing their efficiency in the implementation of several data mining functions. The results of this study can
both be used in developing new hybrid models for dealing with specific quality problems and act as a guideline
leading the users to select the most efficient data mining tool for the specific function they require.
References
1.
2.
3.
4.
5.
6.
7.
Tsang, K.F.; Lau, H.C.W.; and Kwok, S.K. (2007) Development of a data mining system for continual
process quality improvement. In Proceedings of the Institution of Mechanical Engineers; Feb 2007;
221, B2; ProQuest Central pg. 179
Huang, C.C.;Fan, Y.N.;Tseng, B.; Lee, C.H.; and Chuang, H.F. (2008)A Hybrid Data Mining
Approach to Quality Assurance of Manufacturing Process. In Proceedings of the IEEE International
Conference on Fuzzy Systems . 818-825
Wang,K.;Tong, S.;Eynard, B.;Roucoules, L.; and Matta, N. (2007) Review on Application of Data
Mining in Product Design and Manufacturing. In Fourth Proceedings of the International Conference
on Fuzzy Systems and Knowledge Discovery.Haikou, Hainan, China, 24-27 August.
Fayyad, U.; Shapiro, G.P.; and Smyth, P. (1996). The KDD process for extracting useful knowledge
from volume of data. Communications of the ACM, 39, 2734.
Kantardzic, M. (2003), Data Mining: Concepts Models, and Algorithms, IEEE Press and John Wiley,
New York.
Kusiak, A. (2006) Data Mining in Design of Products and Production Systems.In Proceedings of
INCOM2006: 12th IFAC/IFIP/IFORS/IEEE Symposium on Control Problems in Manufacturing,
Saint-Etienne, France, Vol. 1, 49-53.
Harding , J.A; ,Shahbaz, M.;Srinivas; and Kusiak, A,(2006).Data Mining in Manufacturing: A review .
ASME Transactions: Journal of Manufacturing Science and Engineering,128(4),969-976.
598
8.
9.
10.
11.
12.
13.
14.
15.
16.
17.
18.
19.
20.
21.
22.
23.
24.
25.
26.
27.
28.
29.
30.
Porziol, G.C.; and Ragozini, G. (2003) Visually Mining Off-line Data for Quality Improvement, Quality
and Reliability Engineering International, Wiley InterScience, New Jersey, USA, 19(4): 273-283.
Chen,W.C.;Tseng, S.S.; and Wang, C.Y. (2005).A novel manufacturing defect detection method using
association rule mining techniques. Expert Systems with Applications, 29: 807815
Chien, C.;Li, H. ; and Jeang, A. (2006). Data Mining for Improving the Solder Bumping Process in the
Semiconductor Packaging Industry . Intell. Sys. Acc. Fin. Mgmt. 14: 4357
Shahbaz, M.; Srinivas; Harding, J.A. ; and Turner, M. (2006) Product design and manufacturing
process improvement using association rules. In Proceedings of the Institution of Mechanical
Engineers. Feb 2006; 220, B2; ProQuest Central. pg. 243
Charaniya, S.; Lea, H.;Rangwalab, H.; Millsc, K.; Johnsonc, K.; Karypisd, G.; and Hua , W. (2010).
Mining manufacturing data for discovery of high productivity process characteristics. Journal of
Biotechnology 147 :186197
Chien, C. F.; Wang, W. C.; and Chang, J. C. (2007). Data mining for yield enhancement in
semiconductor manufacturing and an empirical study. Expert System With Applications, 33(1), 192198
Ciflikli, C.; and Ozyirmidokuz, E. K (2010) Implementing a data mining solution for enhancing
carpet manufacturing productivity. Knowledge-Based Systems. . 23 : 783788
Shi, X.;Schillings, P.; and Boyd,D. (2004) Applying artificial neural networks and virtual
experimental design to quality improvement of two industrial processes, Int.J. Prod. Res, 42, (1):101
118
Chen,W. C.;Tseng, S.S.; and Wang, C.Y. (2005).A novel manufacturing defect detection method using
association rule mining techniques. Expert Systems with Applications, 29: 807815
Huang, C.C.;Fan, Y.N.;Tseng, B.;Lee, C. H.; and Chuang, H.F. (2008)A Hybrid Data Mining
Approach to Quality Assurance of Manufactur ing Process. In Proceedings of the IEEE International
Conference on Fuzzy Systems . 818-825
Abburi, N.R.; and Dixit, U.S. (2006) A knowledge-based system for the prediction of surface roughness
in turning process. Robotics and Computer-Integrated Manufacturing. 22 : 363372
Chen , W.H.; Tai, P.H.; Wang , M.W.; Deng , W.J.; and Chen, C.T. (2008) A neural network-based
approach for dynamic quality prediction in a plastic injection molding process. Expert Systems with
Applications. 35 :843849
Sterjovski, Z.;Nolan, D.; Carpenter, K.R.;Dunne, D.P.; and Norrish, J. (2005)Artificial neural networks
for modelling the mechanical properties of steels in various applications Journal of Materials
Processing Technology. 170 :536544
Correa, M.;Bielza , C.; and Pamies-Teixeira, J.(2009) Comparison of Bayesian networks and
artificial neural networks for quality detection in a machining process Expert Systems with Applications
.36 :72707279
Erzurumlu, T.; and Oktem, H. (2007). Comparison of response surface model with neural network in
determining the surface quality of moulded parts. Materials and Design 28 : 459465
Braha , D. and Shmilovici, A. (2002 ) Data Mining for Improving a Cleaning Process in the
Semiconductor Industry IEEE Transactions on semiconductor manufacturing. 15(1): 91-101
Rokach, L.; and Maimon, O. (2006). Data mining for improving the quality of manufacturing: A feature
set decomposition approach. Journal of Intelligent Manufacturing, 17, 285299
Li, T. S.;Huang, C. L.; and Wu, Z. Y. (2006). Data mining using genetic programming for construction
of a semiconductor manufacturing
yield rate prediction system. Journal of Intelligent
Manufacturing,17, 355361
Kusiak, A.; and Shah, S. (2006). Data-Mining based system for prediction of water chemistry faults.
IEEE Transactions on Industrial Electronics, 15(2), 593603
Al-Salim, B.; and Abdoli, M. (2005) Data Mining for Decision Support of the Quality Improvement . In
Proceedings of the Eleventh Americas Conference on Information Systems, Omaha, NE, USA August
11th-14th.
Kwack, S.G.;Chung, G.B.; Choi, J.; and Choi, G. (2008)Development of Artificial-Intelligent Power
Quality Diagnosis Equipment for Single-Phase Power System . In Proceedings of the 2nd IEEE
International Conference on Power and Energy (PECon 08), December 1-3, Johor Baharu, Malaysia
Sebzalli, Y.M.; and Wang, X. Z. (2001). Knowledge discovery from process operational data using
PCA and fuzzy clustering. Engineering Applications of Artificial Intelligence, 14, 607616
Chen, R.S.; Yeh, K.C.; Chang, C.; and Chien, H.H. (2006) Using Data Mining Technology to improve
Manufacturing Quality A Case Study of LCD Driver IC Packaging Industry . In the Proceedings of
the Seventh ACIS International Conference on Software Engineering,Artificial Intelligence,
Networking, and Parallel/Distributed Computing
599
31. Hsu,, S.C.; and Chien, C.F. (2006) Hybrid data mining approach for pattern extraction from wafer bin
map to improve yield in semiconductor manufacturing Int. J. Production Economics 107 :88103
32. Hsieh, K.C.(2010) Incorporating ANNs and statistical techniques into achieving process analysis in
TFT-LCD manufacturing industry. Robotics and Computer-Integrated Manufacturing 26 :9299
33. Perzyk, M.; Biernacki, R.; and Kochanski, A.(2005) Modeling of manufacturing processes by learning
systems:The naive Bayesian classifier versus artificial neural networks Journal of Materials Processing
Technology 164165 :14301435
34. Bakr, B.;Batmaz, .;Gntrkn, F.A.;peki, .A.;Kksal, G.; and zdemirel, N.E.(2006) Defect
Cause Modeling with Decision Tree and
Regression Analysis .In Proceedings of World Academy
Of Science
Engineering and Technology. December (17): 1307-6884
35. Koksal, G.; Batmaz , I; and Testik, M.C. (2011) A review of data mining applications for quality
improvement
in
manufacturing
industry
Expert
Systems
with
Applications
doi:10.1016/j.eswa.2011.04.063
Biographies
Burcu Devrim tenba- was born in zmir, 1980. She received the B.Sc.and M.Sc the degree in statistics from
Dokuz Eyll University, zmir, Turkey, in 2003 and 2006 respectively. She is currently doctorate student at
Atlm University Modeling and Design of Engineering Systems PhD Program.
She is interested in data mining, quality management and quality function deployment.
Instructor Devrim tenba still works for Atlm University, Department of Industrial Engineering.
Cenk Gray-was born in Ankara, 1973. He received his B.Sc., M.Sc. and Ph.D. degrees in mining engineering
from Middle East Technical University, Ankara, Turkey, in 1995, 1998 and 2003 respectively.
He is interested in Operations Research, Statistics and Economy applications in Industrial Engineering.
Asst. Prof.Dr. Guray still works for Atilim University, Department of Industrial Engineering.
600
Abstract. It is hard to see traditional grocery stores at the present time. Hypermarkets, supermarkets etc.
have taken the place of traditional grocery stores. These progresses have naturally affected the relations
between buyer and seller. For example bargaining relation between buyer and seller has worn off. We think
that bargaining relation is the main distinction between a customer and a consumer. Correspondingly, these
progresses have squeezed out customer to evolve being consumer. Recent developments show that the next
step of these progresses will be internet retailing, i.e., e-commerce. For that reason, the researchers are
focused on the automated negotiation because automated negotiation is an important function for e-commerce
to be efficient. These efforts, especially, on negotiation for bargaining, can also be seen as an opportunity for
the re-evolution since the answer of how we can build up buyer-seller negotiation for bargaining is also the
answer of how we can accomplish re-evolution of consumers into customers. In this study, we aim to help
taking one step ahead at the re-evolution by modifying Even-Swaps method for bargaining in multi-issue
negotiation. We selected this method since it allows not only considering one offer to be better or worse than
another, but also it allows finding the difference between them. We presented a modified Even-Swaps method
which it utilizes the difference for bargaining.
Keywords: Even-Swaps, modified Even-Swaps, automated negotiation, bargaining, customer, consumer
1 Introduction
In a recent study, it is reported that buyers prefer shopping and buying online over traditional stores to: avoid
social interactions, buy unobserved, and experience immediate positive feelings [1]. We think bargaining
relation is the main distinction between a customer and a consumer. Avoiding social interaction prevents
bargaining since doing social interaction is necessary for it.
Bargaining is one of the oldest negotiation mechanisms in human history. Negotiation is one of the most active
research areas in multi agent system (MAS). Automated negotiation has been becoming an increasingly popular
method with the developments of information and communication technologies (ICT), e.g., the evolution of
electronic commerce (e-commerce) on the web and the rise of interest in intelligence of software agents. Before
development of e-commerce, the role of negotiation in MAS was to solve conflicts of interest among agents
during task allocation in distributed problem solving and resource allocation among agents [2]. With the
development of e-commerce, to study of negotiation became broader, especially the study of open electronic
marketplaces where humans can delegate their agents to negotiate with other agents [2].
Agents provide the abilities of autonomous operations, learning, adaptation, intelligent and adaptive decisionmaking, and cooperation and coordination in highly dynamic environments [3]. Buyer-seller negotiation may be
601
carried out by using agents. MAS provide negotiation between buyers and sellers without face to face social
interaction which nowadays buyers dont like it. The interaction of a set of agents to reach an agreement can be
preferred instead of face to face social interaction. So the developments on the agent technology can also help reevolution of consumers into customers.
Selecting and integrating the most suitable negotiation protocol, decision making algorithm, and data model are
the requirements for developing an automated negotiation mechanism [4]. Making trade-offs is one of the
important steps of decision making processes. Making rational trade-offs are the heart of developing acceptable
management strategies and plans [5]. Even-Swaps is developed as a rational and simple method for making
trade-offs that help the decision maker in considering the value of one objective in terms another [see 6-9]. In
this study, Even-Swap method is modified for automated negotiation for bargaining. Making practical trade-offs
during negotiation can be possible with the proposed Modified Even-Swap method. And, taking one step ahead at
the re-evolution process may be possible with it.
The rest of the study is organized as follows: Some brief description of automated multi-issue negotiation and
state-of-the-art on it are presented in section 2. How Even-Swaps is modified for automated negotiation are
given in section 3. Conclusions and future study are given in the last section.
602
function is used in order to evaluate the validity of the multi-issue negotiation in the model. Teuteberg [19]
presented a multilateral negotiation model on an agent-based job-marketplace. This model is based on fuzzy
utility scoring method and simultaneous multi-issue negotiation with multi negotiation partners.
Chen and Huang [20] developed an automated bilateral multi-issue negotiation scheme for facilitating agent
technology in an outsourcing partner selection case. They use trade-off to find out buyers partner and determine
which issues need to change. They generated trade-offs rules according to the some strategies which are
determined by buyers attitude toward risk.
Luo et al. [21] developed a fuzzy constraint based model for bilateral multi-issue negotiation in trading
environments. Their model uses prioritized fuzzy constraints to represent trade-offs between the different
possible values of the negotiation issues and to indicate how concessions should be made when they are
necessary.
It can be seen from the literature that buyer preferences in multi-issue negotiation can be expressed in a variety
of ways. However, multi-issue negotiation require the management of expressive agreement preferences
regarding multiple issues so that they capture the relationships between issues and, hence, enable making tradeoffs during negotiation [4]. In this study, we propose a novel way for multi-issue negotiation by modifying a
well-known trade-off methodology which is clear, rational and easy-to-use, namely Even-Swaps.
Step 5:
Step 6:
Step 7:
Problem initialization
Eliminate dominated alternatives
Eliminate irrelevant issues
More than one remaining alternative?
If yes, then go to Step 5
Else, go to Step 7
Determine alternatives and issue to perform even swap on them
Determine the required change and perform swap.
Then, go to Step 2
The most preferred alternative is found
Nowadays, MCDM methods based on the utility or desirability scores are commonly used in the multi-issue
negotiation [13]. Especially, they are used to provide support to negotiators while evaluating and comparing the
offers in multi-issue negotiation. Moreover, in a recent study, the Even-Swaps method has been chosen since it
allows negotiators to avoid comparing issues and resolution levels by means of some abstract values and requires
only the analysis in terms of the objectives the parties decided to negotiate [23]. In the study, it is also claimed
that the substantial advantage of Even-Swaps is it allows not only considering one offer to be better or worse
than another, but also it allows finding the difference between them [23]. However, this advantage has not been
used in the study because of the structure of the method. The objective of the method is to determine the most
preferred alternative. Actually, the negotiator does not only want to know what the best option is but also the
differences among alternatives [24]. With the method, the negotiator selects one of the offers, but he/she does
not use the differences among the offers for bargaining. What is more, in the method, trade-offs are performed
with only one-sided by invoking the decision maker. Although bargaining can be possible on some issues,
owners of the offers are not informed while performing issue trade-offs. Some modifications on the method are
603
needed to overcome these obstacles. In this study, we propose a modified Even-Swaps method for bargaining in
multi-issue negotiation. Main steps of modified Even-Swaps method:
Step 1:
Step 2:
Step 3:
Step 4:
Step 5:
Step 6:
Step 7:
Step 8:
Step 9:
Step 10:
Problem initialization
Determine bargainable issue
Determine a range for bargainable issue for bargaining
Eliminate dominated alternatives
Eliminate irrelevant issues
Is there any relevant unbargainable issue?
If yes, then go to Step 7
Else, go to Step 9
Determine alternatives and issue to perform even swap on them
Determine the required change and perform swap.
Then, go to Step 4
The most preferred alternatives are found
Start bargaining
An example is given for better understanding in the following: Issue 4 is determined as bargainable issue.
Therefore, all even swaps are performed on this issue. In the next step, it is assumed that the range for bargaining
on the issue is 1000. Afterward, it is observed that Alternative 1 is dominated by Alternative 2. The
column of Alternative 1 is eliminated from the decision matrix [see Figure 1a]. An even swap is performed by
adjusting Issue 1 Alternative 2 value to Issue 1 - Alternative 4 value, and by changing Issue 4
Alternative 2 value in accordance with the decision makers assessment [see Figure 1b]. Since there is no
dominated alternative on the revised decision matrix, an even swap is needed [see Figure 1c]. An even swap is
performed by adjusting Issue 1 Alternative 3 value to Issue 1 Alternative 4 value, and by changing Issue
4 Alternative 3 value. After the change, it is observed that Issue 1 values for any alternatives are equal. The
row of Issue 1 is eliminated from decision matrix since the issue is irrelevant [see Figure 1d]. An even swap is
performed by adjusting Issue 2 Alternative 3 value to Issue 2 Alternative 4 value, and by changing Issue
4 Alternative 3 value [see Figure 1e]. The row of Issue 2 is eliminated from decision matrix since it is
irrelevant. An even swap is performed by adjusting Issue 3 Alternative 3 value to Issue 3 Alternative 4
value, and by changing Issue 4 Alternative 3 value [see Figure 1e]. After the change, it is observed that
Alternative 3 is dominated by Alternative 4. The column of Alternative 4 is eliminated from the decision
matrix [see Figure 1f]. After that, it is observed that Issue 3 is irrelevant. The row of Issue 3 is eliminated
from decision matrix [see Figure 1f]. Only bargainable issue remains in the decision matrix after the elimination
[see Figure 1f]. The alternatives which remain in the decision matrix reflect the decision makers subjective
preference. The differences between the bargainable issue values of the alternatives will be used to specify
unaccepted area [see Figure 1g]. In our example, the difference between Alternative 2 and Alternative 4 is
500. For that reason, Alternative 2 is acceptable if the Issue 4 value of it is less than 19500 (20000
500) [see Figure 1g]. With the modified Even-Swaps method, the negotiator is able to assess the alternatives
with considering bargaining and select the most preferred ones. The proposed modified Even-Swaps method also
provides identification of the acceptable and unacceptable areas of the issue values of alternatives which will be
bargained for. Moreover, the method helps to select the best alternative, once the negotiator has reached an
agreement for each alternative [see figure 1g].
604
d
Alt. 1
Alt. 2
Alt. 3
Alt. 4
Alt. 2
Issue 1 (max.)
2007
2007
2008
2009
Issue 1 (max.)
2009
Issue 2 (max.)
Issue 2 (max.)
Issue 3 (max.)
35
45
35
45
Issue 3 (max.)
45
35
45
Issue 4 (min.)
20000
20000
21000
21500
Issue 4 (min.)
22000
22000
21500
Alt. 2
Alt. 3
Alt. 4
2007
2008
2009
Alt. 3
2009
Alt. 4
2009
b
Issue 1 (max.)
Alt. 2
Alt. 3
Alt. 4
Issue 2 (max.)
Issue 2 (max.)
Issue 3 (max.)
45
35
45
Issue 3 (max.)
45
35
45
Issue 4 (min.)
20000
21000
21500
Issue 4 (min.)
22000
21800
21500
Alt. 2
Alt. 3
Alt. 4
Alt. 2
Alt. 3
Alt. 4
Issue 1 (max.)
2009
2008
2009
Issue 2 (max.)
Issue 3 (max.)
45
35
45
Issue 4 (min.)
22000
21000
21500
Issue 3 (max.)
45
45
45
Issue 4 (min.)
22000
23000
21500
0
00
20
19
50
0
18
50
Alternative 2
0
50
21
50
20
19
19
Acceptable area
Unacceptable area
Alternative 4
21
40
0
Buyers offer
605
alternatives for bargaining on some issues. One restriction of the proposed method is that it considers bargaining
over a single item/product. In the future study, how to extend this method for multi-items can be addressed.
References
1.
2.
3.
4.
5.
6.
7.
8.
9.
10.
11.
12.
13.
14.
15.
16.
17.
18.
19.
20.
21.
22.
23.
24.
Kukar-Kinney, M., Ridgway, N.M., Monroe, K.B. (2009). The Relationship Between Consumers
Tendencies to Buy Compulsively and Their Motivations to Shop and Buy on the Internet. Journal of
Retailing, 85 (3) 298-307.
Winoto, P., McCalla, G., Vassileva, J. (2002). An extended alternating-offers bargaining protocol for
automated negotiation in multi-agent systems. Lecture Notes in Computer Science, 2519 179-194.
Zhang, J., Luo, J. (2008). Agent based automated negotiation for grid. Proc. of 12 th Int. Conf. on CSCWD,
330-336.
Resinas, M., Fernandez, P., Corchuelo, R. (2011). A bargaining-specific architecture for supporting
automated service agreement negotiation systems. Science of Computer Programming,
doi:10.1016/j.scico.2010.09.008.
Gregory, R., Wellman, K. (2001). Bringing stakeholder values into environmental policy choices: a
community-based estuary case study. Ecological Economics, 39 37-52.
Hammond, J.S., Keeney, R.L., Raiffa, H. (1998). EVEN SWAPS: a rational method for making trade-offs.
Harvard Business Review, 76 (2) 137-150.
Hammond, J.S., Keeney, R.L., Raiffa, H. (1999). Smart Choices, a practical guide to making better decisions.
Harvard Business School Press, Boston, MA.
Kajanus, M., Ahola, J., Kurtilla, M., Pesonen, M. (2001). Application of even swaps for strategy selection in
a rural enterprise. Management Decision, 39 (5) 394-402.
Mustajoki, J., Hamalainen, R.P. (2005). A preference programming approach to make the even swaps
method even easier. Decision Analyses, 2 (2) 110-123.
Cheng, C.B., Chan, C.C.H., Lin, K.C. (2006). Intelligent agents for e-marketplace: negotiation with issue
trade-offs by fuzzy inference system. Decision Support Systems, 42 626-638.
Lopez-Carmona, M.A., Marsa-Maestre, I., Ibanez, G., Carral, J.A., Velasco, J.R. (2010). Improving tradeoffs in automated bilateral negotiations for expressive and inexpressive scenarios. Journal of Intelligent and
Fuzzy Systems, 21 (3) 165-174.
Carbonneau, R., Kersten, G.E., Vahidov, R.. (2008). Predicting opponents moves in electronic negotiations
using neural networks. Expert Systems with Applications, 34 1266-1273.
Lopes, F., Wooldridge, M., Novais, A.Q. (2008). Negotiation among autonomous computational agents:
principles, analysis and challenges. Artif Intell Rev, 29 1-44.
Zhang, C.J., Ji,S.J., Zhu, H.M., Zhao, M.Q. (2007). The state of arts and the challenges of automated
negotiation strategy. Proc. of the second International Conference on Game Theory and Applications, 273278.
Kersten, G.E., Lo, G. (2003). Aspire: an integrated negotiation support system and software agents for ebusiness negotiations. International Journal of Internet and Enterprise Management, 1 (3) 293-315.
Faratin, P., Sierra, C., Jennings, N.R. (2002). Using similarity criteria to make issue trade-offs in automated
negotiations. Artificial Intelligence, 142 205-237.
Chen, Y.M., Wang, S.C. (2007). Framework of agent-based intelligence system with two-stage decisionmaking process for distributed dynamic scheduling. Applied Soft Computing, 7 229-245.
Zhang, H., Qui, Y. (2005). Intelligent automated negotiation mechanism in e-commerce. Proceedings of the
2005 IEEE/ASME International Conference on Advanced Intelligent Mechatronics, 1395-1400.
Teuteberg, F. (2001). Experimental evaluation of a model for multilateral negotiation with fuzzy preferences
on an agent-based marketplace. Electronic Markets, 13 (1) 21-32.
Chen, Y.M., Huang, P.N. (2009). Agent-based bilateral multi-issue negotiation scheme for e-market
transactions. Applied Soft Computing, 9 1057-1067.
Luo, X., Jennings, N.R., Shadbolt, N., Leung, H., Lee, J.H. (2003). A fuzzy constraint based model for
bilateral, multi-issue negotiations in semi-competitive environments. Artificial Intelligence, 148 53-102.
Luo, C.M., Cheng, B.W. (2006). Applying Even-Swap method to structurally enhance the process of
intuition decision-making. Systemic Practice and Action Research, 19 (1) 45-59.
Wachowicz, T. (2007). Even swaps method for developing assessment capabilities of e-negotiation system.
Proceedings of the International Multiconference on Computer Science and Information Technology, 597606.
Li, H.L., Ma, L.C. (2008). Visualizing decision process on spheres based on the even swap concept. Decision
Support Systems, 45 354-367.
606
Abstract. In underground mining, one of the major concerns is effective use of vehicles to produce and transport
the mine ore out of the mine to the ground surface. In metal mining in most of the cases, Load-Haul-Dump (LHD)
vehicles are employed to draw the mine ore from the drawing points and carry it to transfer points to load the ore
onto trucks. From that point, trucks are in charge to transport the ore out of the mine. In this paper, a simulation
model is developed to analyze the movements of LHD vehicles and trucks in underground mines. Certain rules and
strategies are employed to assign the LHD vehicles to draw points and trucks to transfer points. Once the current
operation of the mine is modeled, different scenarios are tested with different number of LHD vehicles and trucks.
The results are analyzed in terms of makespan which is the termination time of the system.
Keywords: Load-Haul-Dump vehicle scheduling, underground mining, simulation, scenario analysis
1 Introduction
Load-Haul-Dump (LHD) vehicles are ore transporting machines adapted for underground conditions. These
vehicles are started to be used first in 1950s and are widely spread by 1960s. Today, they are employed in most
of the metal mines all around the world. A typical LHD vehicle can be seen in Figure 1.
607
The ore reserve located in the mine is known ahead. Therefore, production amount is already known. There is
nothing to be determined for production maximization. What we should determine is how to make this
production in minimum time. In other words, all the ore reserve should be taken out of the mine in the minimum
time. By this way, LHD and truck utilization would be maximized.
In this study, simulation modeling of a metal mine operating in this manner is developed. After verification and
validation of the model, scenario analysis is made on the number of LHD vehicles and trucks.
2 Literature review
A similar scheduling problem in mining industry occurs between trucks and shovels in open pit mines which is
addressed by many researchers [1, 2, 3]. However, LHD scheduling problem is not addressed by many
researchers. There are very few studies in this area. Scheding et al [4] describes the theoretical development and
experimental evaluation of an autonomous LHD based on the results obtained by field trials. Saayman et. al. [5]
examined the LHD and truck scheduling using simulation. In this study the tool employed to study this problem
is also simulation. However, after validating the simulation model of the current system, scenario analysis is
carried out in order to find the effective number of trucks and LHD vehicles.
The rest of the paper is organized as follows. In Section 3, problem overview is given. In Section 4, the proposed
approach is explained in detail. Computational results are given in Section 5. Finally Section 6 is the conclusion.
3 Problem Overview
In this study, a simulation software called ARENA is used to simulate the system and to determine the effective
number of LHDs and trucks used in the system. Proposed approach is to simulate the real system using
predefined distances between the draw points and transfer points, number of LHDs and trucks, then change the
number of LHDs and trucks to obtain an effective makespan, which refers to simulation terminating time. In
addition, the utilization of LHDs are examined. Simulation terminating time or makespan is calculated when all
of the mine ores are transported to the trucks at the transfer points.
Operations chart for LHD and truck in the underground mine can be seen in Figure 2. The ore reserve located in
the mine is known ahead. Also, LHD and truck capacity are known. What we should determine is how to make
this transportation in minimum time. By this way, LHD and truck utilization would be maximized.
608
4 Proposed Approach
The basic operation of the simulated mine is as follows: Ore collects at the drawpoints because of blasting
and caving. The ore is loaded by LHDs and transported to the transfer points. At the transfer points an LHD
dumps its load into a truck. After the truck has received enough loads to fill its bucket, it transports the ore to the
crusher. The load is dumped into the crusher, from where the ore is crushed and processed further. The
simulation is only concerned with everything from the point where the ore is available at the drawpoint until it is
dumped into the crusher. The main objective is to test different dispatching strategies to obtain an effective
completion time of the ore collection.
The simulated environment is based on the layout of the mine. The specific details of the type of the mining
technique are not important for the scope of this research.The principles used in the dispatching of vehicles will
be very similar for many different types of mines and mining techniques. The results discussed in this paper are
therefore not limited to this specific mining scenario. It is known that the movement of the vehicles is confined
to tunnels. The possible routes between different points are therefore fixed. The simulation environment has 8
tunnels that are in production with 8 drawpoints These numbers may not closely reflect real-life mining
environments.
To simulate this mining enviroment, simulation software called ARENA is employed. Using transportation
modules, distances between drawpoints and transfer points are constructed. When simulation runs, LHD goes to
the drawpoints, takes ore from the drawpoints as much as its capacity and takes the ore to the trucks until they
fills the truck capacity.When all of the ore is taken from the drawpoints, simulation terminates. The termination
time of the several simulation runs refer to the makespan.
Simulation is used, since the loading and unloading times, and truck travelling time to the crusher are
random.
5 Computational Results
In the solution of the model, initially it is assumed that there are three LHD vehicles and three trucks in the
mine. Once the model is run with these parameters, different number of trucks and LHDs are tested in order to
fineffective numbe rof LHDs and trucks fort he mine operation. As number of LHDs increase, makespan would
surely decrease. On the other hand as the number of trucks increase, the truck tour time decreases.
The experiments carried out with different number of LHDs and trcuks and the corresponding makespans are
calculated. The change in makespan values according to different number of LHDs can be seen in Figure 3
below. The tradeoff between the benefits obtained by the decrease in makespan and the cost of an
additionalLHD vehicle should be analyzed carefully and is beyond the scope of this paper. However, at first
glace, as number of LHD increases form 3 up to 7 , the decrease in makespan is quite sharp. But , as number of
LHD increases from 7 to 10 the decrease in the makespan is not very considerable. There for it may be benefical
to operate 7 LHDs the mine
609
No.of LHD
80000
60000
40000
20000
0
3
10
Makespan
Makespan
200000
150000
100000
50000
0
3
Number of Trucks
References
[1] Nasuf E. & Krmanl, C. Computer Aided Truck Dispatching Systems in Open Pit Mines and World Applications, Mining
Journal, 1993, 32(2), 5-11 (In Turkish).
[2] Ta, C.H., Kresta, J.V., Forbes, J.F., Marquez, H.J. A Stochastic Optimization Approach to Mine Truck Allocation.
International Journal of Mining, Reclamation & Environment, 2005, 19(3), 162-175.
[3] Aksoy, C.O., Kose, H., Malli, T., Ozfirat, K., Gonen, A.,Gurgen, S. Truck-Shovel Optimization Accounting For Cost,
Safety, and Operational Practice, CIM Bulletin, 2005, 98, (1091).
[4] Scheding S., Dissanayake G., Nebot E. M., and Durrant-Whyte H. An Experiment in Autonomous Navigation of an
Underground Mining Vehicle, IEEE Transactions on Robotics & Automation, 15(1), 1999.
[5] Saayman, P., Craig, I.K., and Camisani-Calzolari, F.R. Optimization of an Autonomous Vehicle Dispatch System in an
Underground Mine, The South African Institute of Mining & Metallurgy, 2006, 106, 77-86.
610
Abstract. A Web Content Management System (WCMS) facilitates the publication of content to a website
and removes the need for the user to possess technical knowledge of editing HTML files. Such systems allow
content creators to directly add or update content on a website without involving an IT department or external
company. Many of these systems are built as monolithic, stand-alone applications that lack the dynamicity
support and are difficult to scale and interoperate with other enterprise systems. In order to tackle these
issues, a service-oriented approach is introduced for a web content management system that is composed
from a number of web services. The proposed model consists of browser, web service layer and content
server. Browser includes backend sites and frontend sites. Web service layer includes authentication and
authorization service, repository service, management service and publishing service. Content server includes
content database and content files. In this paper, we first present the overall system architecture and detail of
each service.
Keywords: web content management system (WCMS), service-oriented architecture (SOA)
1 Introduction
Since the dot-com boom of the late 1990s, corporate websites have become commonplace for almost any type of
company, large or small, across the globe. Almost every enterprise these days needs a website to communicate
with customers, partners, shareholders, and so on, providing up-to-date information on the enterprise, its
products and services. Increasingly, commercial activities and order transactions are conducted on enterprise
websites.
Building and setting up a website is not a one-time project. Different departments in the enterprise will have
areas of content they need to add to and update. Plus, websites have to be maintained and updated on a regular
basis due to the dynamic nature of modern business.
In the early days of website maintenance, the task of uploading and updating site content usually fell to the IT
department. Traditionally, technical staff would have to assist a content editor who needs to update a site by
translating the content into a suitable web page format (i.e. HTML) and uploading it to the web server on their
behalf. Managing the website updating process is another problem with older approach. The Content
Management Systems (CMS) that have appeared more recently are designed to tackle these problems, and make
it easier to collaboratively update a website. [1]
While Web content management systems may be effective in creating, editing, managing and publishing content,
they can also be expensive, complex to implement and configure, difficult to maintain, and complicated to use
611
and adopt within an organization. [2] Additionally, most of these management systems are designed as
monolithic, stand-alone systems. The need for a more flexible and extensible approach becomes increasingly
critical. [3]
Recently, service-oriented architecture (SOA) and web services are increasingly recognized as a suitable
architectural principle for developing modern enterprise applications. [11] SOA concept is to share and exchange
the runtime results of executing software as services rather than the code itself that produces these results. The
sharing and exchange is effectively achieved using standard XML format and HTTP communication protocol.
By composing many low-level services, one can easily handle higher level business services or processes. The
simplicity, reusability, interoperability and flexibility are just a few named benefits provided by this SOA
approach.
However, there are some technical challenges when adopting SOA to realize content management system. First,
how to identify the basic services that support content management is very critical. Second, for different types of
documents, the requirements on content management may be varied. In order to tackle these issues, we introduce
a web service-oriented content management system. This service-oriented implementation approach shows
promise as a new way to realize application systems with better interoperability and flexibility while minimizing
development and maintenance costs. [3]
612
A web service is a technology that allows applications to communicate with each other in a platform and
programming language independent manner. A web service is a software interface that describes a collection of
operations that can be accessed over the network through standardized XML messaging. It uses protocols based
on the XML language to describe an operation to execute or data to exchange with another web service. A group
of web services interacting together in this manner defines a particular web service application in a ServiceOriented Architecture (SOA). [7]
Web service architecture involves many layered and interrelated technologies. There are many ways to visualize
these technologies, just as there are many ways to build and use Web services. Fig. 1 below provides one
illustration of some of these technology families.
613
614
the use of a registry such as a Universal Description, Discovery, and Integration (UDDI) registry to manage,
share, and track service information artifacts.
Though this is an improvement over the less formal Web site approach, it has a major limitation: a registry can
only store links or pointers to service information artifacts. The actual artifacts must reside outside the registry,
using informal and inconsistent means such as Web sites. This makes the actual artifacts ungovernable by the
registry. What is missing is a repository that stores the artifacts. Using the library analogy, imagine a library that
has only a card catalog but no books, and all the books were kept in peoples homes.
A repository provides an integrated solution able to store metadata such as links or pointers to artifacts, as well
as the actual artifacts.
An SOA repository should provide governance capabilities that enable organizations to define and enforce
organizational policies governing the content and usage of the artifacts throughout their life cycles. Since
organizational policies vary, an SOA repository should enable organizations to enforce custom policies for the
governance of any type of service information artifact throughout its life cycle. In particular, it should enforce
conformance to such policies when a service information artifact is published or updated.
All manners of service acquisition need to be supported by strong cross-organization authentication and
authorization, so that legitimate users can be correctly identified, and only legitimate users can access the
services in an authorized manner according to the agreed upon policies between cooperating organizations.
3.2.4 Management Service
Management service is the application logic and business services provided by a CMS that aid in the
management of content and other CMS resources. Specifically, the management services let you create, update,
and delete the rules and scripts governing workflow, personalization, user rights and access, repository
maintenance and archiving and connectivity as well as the database and structures within the repository. [10]
Once a page has been created, it is saved into a central repository in the CMS. This stores all the content of the
site, along with the other supporting details. Most importantly, the CMS provides a range of workflow
capabilities.
When the page is created by an author, it is automatically sent to their manager for approval, and then to the
central web team for their editorial review. It is finally sent to the legal team for their sign-off, before being
automatically published to the site.
At each step, the CMS manages the status of the page, notifying the people involved, and escalating jobs where
required.
In this way, the workflow capabilities allow more authors to be involved in the management of the site, while
maintaining strict control over the quality, accuracy and consistency of the information. Workflow rules bring
order to the chaos of manual processes.
3.2.5 Publishing Service
Publishing service is the application logic and business services provided by a CMS that aid in the creation of
publications from the content and metadata in the repository. The publication services process templates, files
and database records with the goal of marshalling the correct resources to create a targeted publication. They
also execute personalization, conversion and user access routines that are called by particular templates. In the
case of a dynamic publication, the publishing services are invoked via a request from a browser and produce a
single page. In static publications (static Web sites as well as other publications) a staff member triggers the
publication services that then produce a complete publication. [10]
Once the final content is in the repository, it can then be published out to either the website or intranet. Content
management systems boast powerful publishing engines which allow the appearance and page layout of the site
to be applied automatically during publishing. It may also allow the same content to be published to multiple
sites.
615
CMS lets the graphic designers and web developers specify the appearance that is applied by the system. These
publishing capabilities ensure that the pages are consistent across the entire site, and enable a very high standard
of appearance. This also allows the authors to concentrate on writing the content, by leaving the look of the site
entirely to the CMS.
3.2.6 Content Database
The databases within the repository that hold the systems content. Content databases can be standard relational
databases where content components are stored in tables. Generally, each kind of meta information will have a
field in the table and there will be a single field for the internal information of a component. Content databases
can also be object databases where components are fully represented as XML and are stored in a large hierarchy.
4 Concluding Remarks
In this paper, we introduced a service-oriented web content management system. The system is developed as a
web application on a common service platform that is supported by a number of services. The proposed model
includes browser, web service layer and content server. Browser includes backend sites and frontend sites. Web
service layer includes authentication and authorization service, repository service, management service and
publishing service. Content server includes content database and content files.
This for a CMS is a refreshingly flexible model as it offers solutions to issues that have been bothersome to most
organizations. It does everything that a traditional CMS vendor cant; lowers cost, allows faster deployment,
enables transparent upgrades and patches, and achieves rapid ROI and adoption through quick proof of concept
applications. Simultaneously, it also reduces risk by giving organizations the power to unplug a technology or
feature almost effortlessly.
References
1.
2.
3.
4.
5.
6.
7.
The Government of the Hong Kong Special Administrative Region. (2008). Web Content Management
System, 3-11.
CrownPeak. (2011). Software as a Service: The Alternative for Content Management System,
Whitepaper, 3-5.
Chieu T.C. and Zeng L. (2008). Service-Oriented Approach for Implementing an Extensible Content
Management System. In IEEE Congress on Services Part II. 23-26 September 2008. IEEE Computer
Society.
Scanlan, L. (2008). Why use a Web Content Management System, SiteLeads Whitepaper, 2-4.
Browning P. and Lowndes M. (2001). Content Management Systems. JISC TechWatch Report, 2-4.
W3C Working Group. (2004). Web Services Architecture. W3C. Retrieved February 24, 2011 from the
World Wide Web: http://www.w3.org/TR/ws-arch/.
IBM. (2004). New to SOA and web services. Retrieved March 11, 2011 from the World Wide Web:
http://www.ibm.com/developerworks/webservices/newto/service.html.
616
8.
Kim, A. (2007). An Architecture for Web Services Authentication and Authorization in a Maritime
Environment. In ITNG '07. Fourth International Conference. 2-4 April 2007. Information Technology.
9. Sun Microsystems. (2005). Effective SOA Deployment Using an SOA Registry-Repository, Sun
Microsystems, 3-18.
10. Matsumoto, A. (2008). The Content Management Possibilities Poster. Retrieved April 10, 2011 from
the World Wide Web: http://www.metatorial.com/pagea.asp?id=poster.
11. Cherbakov L.; Galambos G.; Arishankar R.; Kalyana S. and G. Rackham. (2005). Impact of Service
Orientation at the Business Level, IBM System Journal 44(4): 653-667.
Biographies
Mehmet Ta 1985, Adapazar, MSc, Research Assistant, Department of Industrial Engineering, Sakarya
University.
Mustafa Torun 1976, Denizli, Electronic Engineer, Founder of Bulu Atlyesi.
617
Abstract. The Capacitated Lot Sizing Problem with setup carryover and backordering (CLSPCB) deals with multiple
products produced on a single machine. A setup of a product is assumed to be carried over from one period to the next and
the partial sequencing of the first and last product is incorporated. When the demand of a period can not be satisfied on time,
there is an option of backordering, meaning that the unmet part of the demand can be satisfied in the following periods in the
planning horizon. In this study, a number of hybrid approaches is proposed to solve the CLSPCB. The proposed hybrid
approaches combine Genetic Algorithms (GA) with a Mixed Integer Programming (MIP) based heuristic, namely Fix-andOptimize heuristic. It is well known that GAs are good at finding the promising regions for global optima, however they are
not good at locating the minimum of these optima in a large solution space. To overcome this difficulty, we propose to
hybridize the GA with the Fix-and-Optimize heuristic. The values of GA parameters are set experimentally through pilot
experiments and the performance of the proposed approaches are compared to pure GAs by generating small, medium and
large size problem instances.
Keywords: Lot sizing, genetic algorithms, setup carryover, backordering, Fix-and-Optimize heuristic.
1. Introduction
The Capacitated Lot Sizing Problem (CLSP) consists of determining the production quantity and timing for
several items on a single facility over a finite number of periods so that the demand and capacity constraints can
be satisfied at a minimum cost. Besides lot sizing decisions, in recent years this problem has been extended to
include semi-sequencing (i.e. the first and last product produced in a period) of the products produced in a period
on a single machine. This problem which is known as Capacitated Lot Sizing Problem with Setup Carryover
(CLSPC) involves producing more than one product per period but with at most one setup carryover activity per
period. Another extension to this problem is to permit backordering by allowing the unsatisfied part of the
demand to be satisfied in the following periods in the planning horizon. The resulting problem is called the
Capacitated Lot Sizing Problem with setup carryover and backordering (CLSPCB). Unlike the CLSP which is
extensively studied in the relevant research, the number of studies addressing the solution of the CLSPCB is very
limited. Considering the perceived research gap in this area, in this study a number of Genetic Algorithm (GA)based heuristic approaches is proposed for solving the CLSPCB. Although GAs have been applied to many
different lot sizing problems [5], to the best of our knowledge there is not any study employing GA-based
approaches to solve this problem.
618
2. Problem Definition
The CLSPCBdetermines the production quantities and production timing of the products along with the
semi-sequencing (i.e. first and last product produced in a period) in a period. If the demand can not be satisfied
in a period, it is backordered meaning that it is satisfied by production in the following periods during the
planning horizon. In solving the CLSPCB, we modify the model proposed by Suerie and Stadtler [16] and add
backordering costs and constraints.
Throughout the paper, the indices for a product, and period are denoted by j=1,2,3,,K and t= 1,2,3,.P,
respectively. We define Xjt as the quantity of product j produced in period t. The variables
I jt and I jt denote
inventory and backorder levels for product j at the end of period t, respectively. The binary variable Yjt is equal to
1 if a setup for product j is performed in period t. The setup carryover variable, which is a binary variable, is
denoted by Wjt. This variable is set to 1 if a setup state for product j is carried from period (t-1) to (t). The binary
variable Qtis the single product variable which equals one if the resource is occupied solely by product i in
period t. Moreover, scj, hjt , Ct , aj , stj, djt ,bj denote the setup cost for product j, the unit holding cost for product j
in period t, the amount of resource available in period t, the time to process one unit of product j, the setup time
of product j, the demand for product j in period t, and the backordering cost of product j, respectively. Finally,
M is a sufficiently large number.
K
Min ( sc jY jt h j I jt b j I jt )
(1)
j 1 t 1
s.t.
I j ,t 1 X jt I jt I j ,t 1 I jt d jt
(a X
jK
W
jK
jt
jt
j K ; t P
(2)
s jY jt ) Ct t P
(3)
1 t {2,...., P}
(4)
W jt Yjt 1 W jt 1 j K , t {2,...., P}
(5)
W jt 1 W jt 1 Qt
(6)
t {1,...., P 1}
Yjt Qt 1 t {1,...., P 1}
(7)
X jt M (Y jt W jt ) j K , t {1,...., P}
(8)
Qt 0 t {1,...., P 1}
(9)
W jt 0,1 (W j1 0) , Y jt {0,1} j K , t P
X jt , I jt 0, I jt 0 j K ; t P
(10)
(11)
I jt 0, j K ; t P
(12)
The objective function (1) minimizes the total inventory, backorder and setup costs. Inventory balance
equations (2) describe the evolution of the inventory and backorder over time and require that the demand is
satisfied either by production, inventory or backorder. There is a limited capacity available in each period, and
we take into account both production and setup times (Constraint 3). At most, one setup state can be carried over
from one period to the next on the resource (Constraint 4). Constraints (5) impose that a setup can be carried
over to period t only if either the item is setup in the previous period or if the setup state is already carried from
period t-2 to t-1. A setup state for a specific item can only be preserved over two bucket boundaries if that
product is the only one produced in the middle period (Constraint 6), which is only possible if there is no setup
in the middle period (Constraint 7). Constraints (8) enforce either a setup or a carryover if we have strictly
positive production for a specific item. Further, we have the non-negativity and binary conditions (Constraint 9,
10 and 11). There are no setup carryovers in the first period (Constraint 10) and all demand should be satisfied
during the planning horizon (Constraint 12).
619
3. Literature Review
The CLSP with different modeling features such as set-up times, set-up carryover, backordering, etc. has
attracted the attention of many researchers in lot sizing area.
Sox and Gao [15] reformulate the mathematical programming model for the CLSPC by using a shortest-route
representation and propose a Lagrangean decomposition heuristic. Gopalakrishnan et al. [4] propose a Tabu
Search (TS) heuristic which consists of five basic move types- three for the sequencing and two for the lot-sizing
decisions. Suerie and Stadtler [16] propose an extended formulation and valid inequalities and time
decomposition heuristic to solve this problem both for single and multi level case. Gupta and Magnusson [7]
consider sequence-dependent setup costs and times in solving the CLSPC. The sequence dependent setup times
and costs are also added to the CLSPC by Almada-Lobo et al. [2] for a real world problem in the glass container
industry. A GRASP meta-heuristic is proposed in Nascimento and Toledo [11] to solve the CLSPC in a multi
plant setting. Gren et al. [6] propose a hybrid approach for solving the CLSPC.
As mentioned earlier, the CLSP with backordering has not attracted the attention of the researchers in this
field. Pochet and Wolsey [12] present two formulations for the uncapacitated and capacitated lot sizing problem
with backordering. Millar and Yang [10] develop Lagrangean heuristics for solving the problem. Cheng et al. [3]
deal with a capacitated lot sizing problem with backorder consideration under the various types of production
capacities such as regular time, overtime and subcontracting. Absi and Sidhoum [1] deal with a special type of
the CLSP in which demand can not be backordered, but will be totally or partially lost. During the survey of
current relevant literature, we noted only a few studies focusing on both setup carryover and backordering,
abbreviated to CLSPCB [8, 9, 13]. Karimi and Ghomi [8] propose a greedy heuristic consisting of four stages to
deal with this problem. In another study [9], Karimi et al. propose a Tabu Search (TS) based approach and use
this greedy heuristic to find a feasible initial solution. The problem is extended to the parallel machine
environment by a real life example in Quadt and Kuhn [13].
In this study, a number of hybrid approaches is proposed to solve the CLSPCB. The proposed hybrid
approaches combine GA with a MIP based heuristic, namely Fix-and-Optimize heuristic. It is well known that
GAs are good at finding the promising regions for global optima, however they are not good at locating the
minimum of these optima in a large solution space. To overcome this difficulty, we propose to hybridize the GA
with the Fix-and-Optimize heuristic. This work extends the work presented in Gren et al. [6] in two ways; 1.
This study considers the backorder version of the lot sizing problem presented in Gren et al. [6]. 2. Unlike
Gren et al. [6] which employs only the time decomposition in the Fix-and-Optimize heuristic, this study uses
the Fix-and-Optimize heuristic in several ways.
620
an individual is randomly selected from the new population and it is used as an initial solution of the Fix-andOptimize heuristic.
In this study, we use two types of decomposition schemes in the Fix-and-Optimize heuristic. The first one is
the time decomposition where the whole planning horizon is divided into time windows. The other one is the
product decomposition where the decomposition is done according to the specific information of products
instead of time periods.
Time decomposition: The basic idea in time decomposition is to divide the whole planning horizon into
time windows. Each time window corresponds to a sub-problem and by solving the sub-problems in an iterative
manner, the overall solution is obtained. As in Gren et al. [6], each sub-problem is defined for a time window of
five consecutive periods with respect to the setup and setup carryover variables. For example, in one iteration ( l
) of the Fix-and-Optimize heuristic, three sub-problems are solved for an instance with 15 periods.
Product decomposition: In contrast to time periods, this decomposition scheme focuses on product specific
information to form and order the sub-problems. As a result of a pilot experiment, it has been decided to form
and order the sub-problems based on set up costs of the products (i.e. the higher the set up cost, the higher the
priority of the product)
Start GAs
Random Solution from the new
population = Initial Solution
Initial Population
The Fix-and-Optimize
Genetic Operators
heuristic
No
Repair Operators
Terminate?
No
Yes
New Population
Terminate?
Yes
Report the OverallBest
Solution
621
By using two decomposition schemes mentioned above, four hybrid approaches are proposed in this study to
solve the CLSPCB:
Hybrid Approach 1 (H1): GA embedded hybrid approach where the Fix-and-Optimize heuristic employs time
decomposition only.
Hybrid Approach 2 (H2): GA embedded hybrid approach where the Fix-and-Optimize heuristic employs
product decomposition only.
Hybrid Approach 3 (H3): GA embedded hybrid approach where the Fix-and-Optimize heuristic employs two
decomposition schemes in the sequence of product and time decomposition.
Hybrid Approach 4 (H4): GA embedded hybrid approach where the Fix-and-Optimize heuristic employs
product decomposition in one generation and time decomposition in other generation.
5. Computational Experiments
5.1 Benchmark Problems
Since there are no available data for testing the efficiency of the proposed approaches, the problem instances
given in Trigeiro et al. [17] were modified by adding the backorder costs. Specifically, to introduce
backordering, the demands were modified and multiplied by 1.1. The backorder cost was defined as a linear
function of the holding cost (b=fh), where f=2 as in Millar and Yang [10]. Table 1 presents the features of the
data instances studied.
Number
products
Number
periods
Number
instances
6
24
of
15
15
15
30
30
30
of
As a result of some preliminary tests the computational time to solve each problem using Fix-and-Optimize
heuristic was limited to 2 seconds. All computations were carried out on a PC with Dual Core, 2 GHz
microprocessor and 2 GB RAM. The pure and hybrid GAs were coded in Visual C++ 2008 Express Edition and
all sub-problems were solved using Concert Technology of Cplex 11.2.
The parameter settings of the proposed approaches were done through some pilot experiments on a medium
size problem with 10 products and 20 periods. Based on these experiments, a population size of 100, a crossover
rate of 0.95 and a mutation rate of 0.005 were used for 100 generations.
5.2 Comparative Experimental Results
In this section, the experimental studies comparing the performances of the proposed GA based hybrid
approaches to that of Pure GAs (PGA) are given. The average gaps over lower bounds obtained from the SPL
formulation are given in terms of both the average of 10 runs and the best result among these runs. The overall
results obtained for all problem classes are given in Table 2. To obtain compatible results the performances of
the proposed hybrid approaches and PGA were tested under the same run time limitation. For instance, the time
limitation to solve problem instances in Class 1 was set to 252.37 s. (see Table 2). As seen in Table 2, the large
gaps from the lower bounds indicate that the performance of PGA is poor for solving CLSPCB.
As seen in Table 2, the proposed embedded hybrid approaches outperform PGA in all problem classes. This
could be attributed to the success of using the Fix-and-Optimize heuristic in the loop of GAs. As expected, the
Fix-and-Optimize heuristic acts like a diversification strategy and helps GAs direct the search to promising areas
in the search space.
622
Class
1
Class
2
Class
3
Class
4
Class
5
Class
6
302.75
357.05
5.06 %
5.26 %
4.58 %
4.80 %
6.91 %
6.30 %
6.19 %
6.51 %
Among the proposed embedded hybrid approaches, while H2 employing product decomposition in the loop
of GAs has the worst performance, H3 which uses both decomposition schemes in the loop of GAs slightly
outperforms H1 and H4. Overall, it can be stated that compared to using a single decomposition scheme in the
loop of GAs, using both decomposition schemes sequentially has more potential to improve the solution quality.
Furthermore, it is observed that the solution quality of the proposed hybrid approaches is affected by the type of
decomposition scheme and the proposed hybrid approaches employing time decomposition scheme have better
performance.
6. Conclusion
Lot sizing is one of the most well-known optimization problem in production planning. Most of the studies in
this area are devoted for solving the CLSP which is known to be NP-Hard. The CLSP becomes more complex
when setup carryover and backordering are considered. Hence exact algorithms are not capable of producing
good quality solutions in reasonable computational time. In this study, a number of GA based hybrid approaches
are proposed for solving the CLSPCB which is the least studied lot sizing problem in the literature. An extensive
computational study is made to compare the performance of these hybrid approaches to the pure GAs. The
general conclusion which can be drawn from this comparative experimental study is that hybridizing the PGA
with the proposed MIP-based heuristic significantly improves its performance in solving the CLSPCB. In a
future research, the robustness of the proposed hybrid approaches can be investigated under different
backordering costs.
References
1. Absi N., Kedad-Sidhoum S. (2007).MIP-based heuristics for multi-item capacitated lot-sizing problem
with setup times and shortage costs. RAIRO Operations Research41 (2): 171-192.
2. Almada-Lobo B., Klabjan D., Carravilla M.A., Oliveira J.F. (2007). Single machine multi-product
capacitated lot sizing with sequence-dependent setups. International Journal of Production Research
45 (20): 4873-4894.
3. Cheng C.H., Gupta Y.M., So S. (2001). Solving the capacitated lot-sizing problem with backorder
consideration. Journal of Operational Research Society 52: 952-959.
4. Gopalakrishnan M., Ding K., Bourjolly J-M., Mohan S. (2001). A Tabu Search Heuristic for the
Capacitated Lot-Sizing Problem with Set-up Carryover. Management Science 47 (6): 851-863.
5. Gren Gner H. Tunal S., Jans R. (2010). A Review of applications of genetic algorithms in lot sizing.
Journal of Intelligent Manufacturing 21 (4): 575-590.
623
6. Gren Gner H. Tunal S., Jans R. A Hybrid Approach for the Capacitated Lot Sizing Problem with Setup
Carryover. International Journal of Production Research (Accepted).
7. Gupta, D. and Magnusson, T. (2005). The capacitated lot sizing and scheduling problem with sequencedependent set-up costs and set-up times. Computers and Operations Research 32: 727-747.
8. Karimi B., Ghomi Fatemi. (2002). A new heuristic for the CLSP problem with backlogging and set-up
carryover. International Journal of Advanced Manufacturing Systems 5 (2): 66-77.
9. Karimi B.,Ghomi Fatemi SMT, Wilson JM. (2006). A tabu search heuristic for the CLSP with
backlogging and set-up carryover. Journal of the Operational Research Society 57: 140-157.
10. Millar H.H., Yang M. (1994). Lagrangean heuristics for the capacitated multi-item lot sizing problem
with backordering. International Journal of Production Economics 34: 1-15.
11. Nascimento M.C., Toledo F.M.B. (2008). A Hybrid Heuristic for the Multi-Plant Capacitated Lot Sizing
Problem with Setup Carry-Over. Journal of the Brazilian Computers Society 14 (4): 7-15.
12. Pochet Y., Wolsey L.A. (1988). Lot-size models with backlogging: Strong reformulations and cutting
planes. Mathematical Programming Series A 40: 317-335.
13. Quadt D., Kuhn H. (2009). Capacitated lot-sizing and scheduling with parallel machines, back-orders,
and setup carry-over. Naval Research Logistics 56 (4): 366-384.
14. Sahling F., Buschkhl L., Tempelmeier H., Helber S. (2009).Solving a multi-level capacitated lot sizing
problem with multi-period setup carry-over via a fix-and-optimize heuristic.Computers and Operations
Research 36 (9): 2546-2553.
15. Sox C.R. and Gao Y. (1999). The Capacitated Lot-Sizing Problem with Set-up Carryover. IIE
Transactions 31: 173-181.
16. Suerie C., Stadtler H. (2003). The Capacitated Lot-sizing problem with Linked Lot sizes. Management
Science 49 (8): 1039-1054.
17. Trigeiro W., Thomas L.J., McClain J.O. (1989). Capacitated Lot Sizing with setup times. Management
Science. 35: 353-366.
BIOGRAPHIES
Hacer Gner Gren Hacer Gner Gren is a Ph.D. student at the Department of Industrial
Engineering at Dokuz Eyll University. She received her B.Sc. from Dokuz Eyll University in 2002
and her M.Sc. from Pamukkale University in 2005.
Her research interests are mainly production planning problems, specifically lot sizing problems, metaheuristics and hybrid approaches. She presented papers in domestic and international conferences and
published in her areas of research.
Semra Tunal Semra Tunal is a full time professor at Izmir University of Economics at Department
of Business Administration. After graduating from Ege University, Industrial Engineering in 1978, she
received M.Sc. Degree from Ege University and her MBA degree from University of Missouri, St.
Louis, Missouri, USA. Then she received Ph.D. from Ege University.
Prof. Tunal, who is an expert in the fields of Using Simulation and Meta Heuristics for Manufacturing
Optimization, Simulation-Based Scheduling, Simulation Optimization Using Response Surface
Methodology, Balancing and Sequencing of Mixed-Model Assembly Lines, Decision Support Systems
for Production and Service Systems became an Associate Professor at Dokuz Eyll University, in 1998
and received her Professor title in 2004. Prof. Dr. Tunal has published in her areas of research in many
cited journals, presented papers in conferences.
Raf Jans Raf Jans is an associate professor at Department of Logistics and Operations Management at
HEC Montreal. He received his B.Sc. and M.Sc. in Business Engineering from Katholieke Universiteit
Leuven, Belgium between 1992 and 1997 and his Ph.D. in 2002 from the same university.
His research interests are production and inventory planning, product variety, logistics, combinatorial
optimisation, decision and risk analysis. Dr. Jans has published in his areas of research in many
journals.
624
Abstract. In this paper, a typical dynamic problem which is known as vehicle deployment problem is
studied. In classical vehicle deployment problems, vehicle development process that parks a finished
product into the plant yard before its shipment to dealer is considered. However, in real world, the location of
a vehicle is not static and it changes continuously from production release to shipment. Therefore, this
problem is indeed a dynamic optimization problem and should be solved by using real time information to
develop more promising solutions. Unfortunately, conventional (classical) vehicle deployment strategies are
based on current location of vehicles already parked and do not consist online information. In this respect,
this paper focuses on possible strategies to solve this problem effectively by using real time information.
Since agent based modeling and simulation (ABMS) approach has been a favorable approach to deal with
dynamic problems, the paper discusses the availability of applying ABMS for creating more realistic results.
In this content, as being a multi-methods simulation language, AnyLogicTM, is specifically considered
through the study. This study also discusses the possible improvements that can be obtained with the change
of auction mechanisms with simulated annealing approach.
Keywords: agent based simulation, AnyLogic, vehicle deployment
1 Introduction
In this age of global competition, managers of various industries have faced with lots of complex problems with
the increasing uncertainty and under operational dynamics, timeliness, lateness, and ineffectiveness in sharing
information. In the past, when uncertainty and unpredictability were relatively less, companies used classical
solution methods like linear programming, integer programming in order to solve decision making problems.
However, in todays competitive environment, obtaining solutions to the real-world problems with classical
techniques is nearly impossible. Since with the advance of technological development, most of the real-world
problems have become distributed, dynamic and unpredictable which makes these problems more complex.
The characteristics of real-world problems, such as vehicle deployment problem, make the problem domain
particularly challenging and suitable for exploring different solution approaches. Nowadays, both researchers
and commercial organization have used agent based approaches in order to solve complex problems. Agents that
act in an autonomous fashion to perform delegated tasks have aroused much attention over the last decade [1]. In
this regard, agent based computing has often been suggested as a promising technique for problem domains that
are distributed complex and heterogeneous [2-4].
625
In this paper, vehicle deployment problem which was previously investigated by Kim et al. in 2010 [5] is
studied. In this problem vehicle development process that parks a finished product into the plant yard before its
shipment to dealer is considered. As stressed by [5], conventional vehicle deployment strategies are based on
vehicles current location already parked. However, in real world, the location of a vehicle is not static and it
changes continuously from production release to shipment. Therefore, this problem is indeed a dynamic
optimization problem and should be solved by using real time information to develop more promising solutions.
Unfortunately, conventional (classical) vehicle deployment strategies are based on current location of vehicles
already parked and do not consist online information. In this respect, this paper focuses on possible strategies to
solve this problem effectively by using real time information. Since agent based modeling and simulation
(ABMS) approach has been a favorable approach to deal with dynamic problems, the paper discusses the
availability of applying ABMS for creating more realistic results. In this content, as being a multi-methods
simulation language, AnyLogicTM, is specifically considered through the study.
The rest of the paper is organized as follows. Section 2 describes literature on Dynamism. Section 3 presents
current vehicle deployment practices and the solution approach provided by [5]. Section 4 discusses the
application of the same model using AnyLogicTM with a novel agent decision strategy. Finally, the conclusions
and the future work are drawn in Section 5.
626
Consequently, the vehicle moves to a general truck buffer or a general train buffer according to its predetermined delivery transportation mode. The general buffers store a number of vehicles until they move to
loading buffers where they are loaded onto trucks or trains for delivery. Every finished vehicle has its delivery
destination, with a corresponding transportation mode, assigned by a delivery schedule. Before actual loading
onto trucks, the vehicles first move from the general buffer to a truck loading buffer consisting of many loading
locations.
In their study [5], they validate that using the RFID tracking system together with the intelligent deployment
algorithms improve the performance. They benefit from the advantage of RFID technology which is capable of
detecting location changes of a vehicle. They employ a market-based control algorithm which is specifically
developed for capturing the dynamics of the deployment process. They also developed a multi-agent
computational architecture for processing and coordinating a large amount of real-time information.
4 Model
In order to succeed, agent modeling, agent technology must not only be philosophically attractive but it
should also be computationally affordable, easy to develop, and have an intuitive way of modeling knowledge
for solving a problem [11]. Hence, tools to design, develop, implement, run, and debug agent systems are as
critical as the description of the properties that the agents should exhibit [11]. In this respect AnyLogic TM has
been agent-modeling software with numerous advantages. Different simulation approaches such as system
627
dynamics (SDs), discrete events (DEs) and agent based simulation (ABS) can be used in the same model.
AnyLogics visual development environment reduces development cost and time. On the other hand, it is
possible to use JAVA code at any place of the program which gives the flexibility to user to adapt their models
to their needs. Beside this, it is possible to create a JAVA applet of a working model and run it anywhere.
In this paper, considering their case [5] (some conditions and data are modified) we used AnyLogicTM. The
model described in the previous section has been adapted with AnyLogic. In the AnyLogic model developed for
this study, there are interactions between agents and the system dynamics components, and interactions between
agents. The Main interface is given in Fig. 2. As it can be seen from Fig. 2, produced vehicles are stocked in
temporary buffer using the stock variable TemporaryBuffer, and slots in general buffer modeled using stock
variable Slots, and LoadingBuffer is the stock variable as well and they are all the parts of system dynamics.
Fig. 3. Statecharts of the vehicle agent and yard manager agent in AnyLogic model
628
Kim et al. in 2010 [5] considered minimization of consolidated operational time in their market-based control
algorithms as decision objective. Therefore as they mentioned for gathering more efficient deployment process
other decision objectives can be considered. Here same objective is used in the model.
Computational experiments show that the computational time increases with the size of the set of the candidate
relocation sites. On the other hand, for large size problems it is hard to obtain the optimal solution due to the
large size of the problem files. In order to reduce the computational time, heuristic approaches can be used for
communication in the modeling instead of auction mechanisms, since in auction mechanisms all the solutions are
evaluated. Simulated annealing is appropriate heuristic for yard manager agent determining the winner agent.
One of the appropriate heuristic techniques for yard manager agent determining the winner agent is Simulated
Annealing (SA). As it is known, the ideas that form the basis of simulated annealing dates back to the research
performed by [12] in 1953 in an algorithm to simulate the cooling of material in a heat bath. It is similar to the
conventional search method but the most important distinctive feature of SA is accepting worse solution within
certain a probability. As a result, the chance of escaping from local optimum increases in order to find global
optimum [13]. In this respect, the simulated annealing approach proposed here includes the comparison of
existing assignments of vehicles (which is representing the initial solution) with the utility of randomly assigned
offers received from the agents. For instance, if there are five agents proposing themselves to be assigned to
certain slot, then an agent is randomly selected out of these five agents and assigned to that certain slot. If this
random move yields better results than it is accepted, else it is can be accepted with a certain probability. In the
crowded systems with multi-factorial fitness consideration, the expected computational time will certainly be
less than the existing auction mechanism. In this regard, the experimental study to check the validity of this
hypothesis is still in its infancy and will be finalized sooner.
References
1.
2.
3.
4.
5.
6.
7.
8.
Xu, K., & Liu, H. (2006). A Multi-Agent Particle Swarm Optimization Framework with Applications.
In 3 (pp. 1-6). Presented at the First International Symposium on Pervasive Computing and
Applications.
Weiss, G. (1999). Multiagent Systems: A Modern Approach to Distributed Artificial Intelligence. The
MIT Press.
Wooldridge, M. (2002). An Introduction to MultiAgent Systems (1st ed.). John Wiley & Sons.
Persson, J. A., Davidsson, P., Johansson, S. J., Wernstedt, F., Center, S., & Ronneby, S. (2005).
Combining agent-based approaches and classical optimization techniques. In Third European Workshop
on Multi-Agent Systems.
Kim, J., Ok, C., Kumara, S., & Yee, S. (2010). A market-based approach for dynamic vehicle
deployment planning using radio frequency identification (RFID) information. International Journal of
Production Economics, 128(1), 235-247.
Younes, A. (2006). Adapting Evolutionary Approaches for Optimization in Dynamic Environments.
PhD Thesis, University of Waterloo.
Psaraftis, H. (1995). Dynamic vehicle routing: Status and prospects. Annals Operations Research 61,
143164
Bianchi, L. (1990). Notes on dynamic vehicle routing - the state of the art. Tech. Rep.idsia 05-01, Italy.
629
Biographies
Adil Baykasolureceived his B.Sc., M.Sc. and PhD degrees in Mechanical Engineering and Industrial
Engineering areas from Gaziantep and Nottingham, UK. From 1993 to 2011, he was with the Department of
Mechanical Engineering and Industrial Engineering at the University of Gaziantep. He is presently a full
Professor at the Industrial Engineering Department of Dokuz Eylul University, Izmir. He has published around
300 academic papers, 3 books and edited several conference books on operational research, computational
intelligence, metaheuristics, quality, and manufacturing systems design. He is an active referee for many
scientific journals and serving on the board of several academic journals. He is one of the editors of Turkish
Journal of Fuzzy Systems. He is also an active member of many academic and professional institutions like
International Society of Agile Manufacturing, Turkish Chamber of Mechanical Engineers, Turkish Fuzzy
Systems Association, Turkish Operational Research Association etc. He is also awarded by TUBA,
TUBITAK and METU for his scientific contributions.
Zeynep D. U. Durmuolu She was born in Malatya, Turkey. She received her BSc. in Industrial Engineering
from University of Gaziantep. She has also received her M.Sc. in Industrial Engineering from the Universityof
Gaziantep in 2009. Her Ph.D. is in progress in the departmentof Industrial Engineering at the University of
Gaziantep. Her research interestsare: Simulation optimization, heuristic optimization, negotiation, agent based
simulation and modeling. Mrs. U. Durmuolu, is currently working as research assistant of Industrial
Engineering department at University of Gaziantep, Turkey.
Latife Grkemli She was born in Kayseri, Turkey. She received her BSc. in Industrial Engineering
from University of Erciyes in 2007. She has also received her M.Sc. in Industrial Engineering from the
Universityof Erciyes in 2009. Her Ph.D. is in progress in the departmentof Industrial Engineering at the
University of Gaziantep. Her research interestsare: Operations research models and techniques, simulation and
system dynamics, agent based simulation. Ms. Grkemli, is currently working as research assistant of Industrial
Engineering department at University of Erciyes, Turkey.
630
Abstract.A mobile phone is a portable electronic device that uses a network of cell sites and supports SMS
services for text messaging and MMS for photos and videos. This paper focuses on mobile's sales points to
predict and economically schedule numbers of sales executives for highest customer satisfaction. Primary
goal is to determine the best number of opened workstations to decrease customers waiting time. Data were
collected from Qmatic system at a sales point in Jeddah for 2009. Simulation was done using Arena. Required
data were average number of tickets taken, average service time, total number of customers served, total
number of customers not show and number of resources. The best scenario attained a net profit of SR
1,509,506 with total system time of 453 seconds which was decreased by 72%. Number of unsatisfied
customers was decreased by 77%, number of satisfied customers increased by 374% and net profit increased
by 6.2% despite increasing number of employees.
Keywords: mobile phones, simulation, customer satisfaction
1 Introduction
1.1 Importance of the Study
Customer satisfaction is a measure of how products and services supplied by a company meet or surpass
customer expectation. It is seen as a key performance indicator within business. Satisfied customers usually lead
to more sales and profit, making it a strong indicator of company performance in a competitive marketplace
where businesses compete for customers. Customer satisfaction is seen as a key differentiator and has
increasingly become a key element of business strategy. Organizations are increasingly interested in retaining
existing customers while targeting non-customers. Measuring customer satisfaction provides an indication of
how successful the organization is at providing products and/or services to the marketplace.
In Mobily, the study and analysis of the sales points is very important to determine the state of the customer
satisfaction. If a customer waits for a long time in the queue to be served, there is a possibility that he/she quits
or leaves and that will build bad experience in his/her mind and this might lead to a higher probability of loosing
others customers. Research about the impact of satisfied and dissatisfied customers on sales experiences reveals
the following facts [1]: Satisfied customers will tell an average of 5 persons. Dissatisfied customers will tell
about 9 persons. Dissatisfying experience will be remembered 10 times. 13% of the unsatisfied customers will
share their unhappy experience with at least 20 persons. 10% of the unsatisfied people never complain; they just
switch to a competitor.
631
2 Literature Review
The impact of social identity on consumer satisfaction was investigated by Al Otaibi [2]. It revealed that
customers' relationship is a topic that has attracted much attention in the service literature. This study sought to
further our understanding as how social identity theories could be applied in the context of service interactions.
This study further sought to investigate the extent to which culture has an impact on customers satisfaction with
hotel service attributes. Regression results revealed that there is a relationship between social identity and
satisfaction with the service attributes.
In another research by Al-Nasser, results indicated a significant difference between Europeans and Saudi
Arabians satisfaction with service attributes [3]. This study was a first step towards a better understanding of the
impact of social relations on the evaluation of service outcomes. It has concentrated mainly on customers
satisfaction. There is no doubt that many companies have lost their competitive advantage because they
neglected to meet the specific needs or requirements of the customer.
Another research examined the relationship between time orientation and time allocation [4]. A questionnaire
was first developed to assess the main aspects of time orientation. Time allocation was measured using a 24 hour
time diary over one week. Because time allocation can be affected by a number of exogenous and endogenous
variables, a narrow sampling method was adopted where age, gender, employment and education were
controlled.
632
Mobily launched its services on the 25th of May 2005 and in less than a year, Mobily's customers have
reached 3.8 millions and more than 6 millions by 2006, beating the record in the Middle East and North Africa.
Mobily's workforce exceeds 2700 employees with 82% of them from Saudi Arabia, and was the first to introduce
female sales outlets in the telecom sector.
Mobily Company currently provides many services such as missed call notification, voice mail, conference
call, call forwarding, call waiting, call hold, call baring, video call, streaming, 3G, SMS (Short Message System),
MMS (Multimedia Message system), Friend finder, My location [10].
Once the customer visits the sales point, he has to take a ticket from the Qmatic (automated ticket taken
machine) then move to get full information about products, services and promotions. Then the customer goes to
the waiting area till his ticket number is shown in the display. After that, the customer is served in the
workstation by the sales executive. Eventually, the customer leaves the sales point [11].
Payment
Update
Customers
No
No
No
Customer
Customer
Served
Show
Show
Show
Served
Served
18845
18664
181
331
8
274
17
19007
18688
319
3194
21
4332
221
18972
1879
93
2051
13
453
16
19500
19390
110
1963
15
104
6
19127
19035
92
2274
6
54
12
17426
17361
65
4093
16
101
12
17224
17146
78
3719
10
92
17
11958
11877
81
2730
13
65
7
8976
8931
45
1997
12
57
14
16781. 7 16663.4 118.2 2483.6
12.7
614.7
35.8
Tickets
Taken
633
Jan
Jul
Aug
08:30-09:30 AM
09:30-10:30 AM
10:30-11:30 AM
11:30-12:30 PM
12:30-01:30 PM
01:30-02:30 PM
02:30-03:30 PM
03:30-04:30 PM
04:30-05:30 PM
05:30-06:30 PM
06:30-07:30 PM
07:30-08:30 PM
08:30-09:30 PM
09:30-10:30 PM
10:30-11:30 PM
53
56
45
39
16
46
29
41
63
59
71
57
39
60
56
53
56
45
39
16
46
29
41
63
59
71
57
39
60
56
53
55
48
51
14
47
31
35
68
51
69
52
47
68
57
52
51
50
40
23
37
29
42
61
62
70
56
43
64
58
58
56
63
42
14
39
36
37
64
57
66
55
47
64
63
58
60
59
35
24
43
23
38
58
56
77
51
43
72
59
49
54
56
37
20
41
19
40
55
54
71
49
40
69
64
60
59
54
46
18
37
25
34
58
62
65
52
34
56
59
Sep AVG
42
53
47
33
19
33
19
31
59
53
58
47
30
61
49
53.1
55.6
51.9
40.2
18.2
41.0
26.7
37.7
61.0
57.0
68.7
52.9
40.2
63.8
57.9
Arena software combines the ease of use found in high-level simulators with the flexibility of simulation
languages and even on the way down to general-purpose procedural languages like the Microsoft Visual
Basic programming system or C.
For case of display and organization, modules are typically grouped into panels to compose a template. By
switching panels, one gains access to a whole different set of simulation modeling constructs and capabilities
[14]. In addition, Arena Professional Edition, Rockwell Software offers full suite of products to provide
enterprise wide simulation, optimization, and 3D model animation. A recent addition to the Arena family of
products is OptQuest for Arena, an optimization software package [13].
634
635
6 Simulation Results
6.1 Base Case Results
The obtained simulation results show the Base case result (5 sales executives in the AM shift and 6 in the PM
shift). The results are the number of customers that enter the system, the number that leaves the system, the
number of customers that wait more than 10 min, number of customers that wait less than 10 min, number of
customers service no show, number of customer service served, number of payment machine no show, number
of payment machine served, number of update service no show, and number of update service served. The
results also indicate the branch profit per month, unsatisfied cost, salaries for additional employees, money saved
from increasing of sales executives, and the net profit.
6.2 Different Scenarios
We initially divided our scenarios into two types which are:
Type one: Dependent on an employee training course which helps reduce customer service time and
leads to a decrease in the number of unsatisfied customers. From scenario 1 to scenario 4 a customer is
considered unsatisfied if he waits more than 10 minutes in customers service queue.
Type two: Dependent on an additional number of sales executives per shift in order to reach optimal
customer satisfaction from scenario 5 to scenario 9.
The scenarios are shown in table 3.
Table 3. The different initial scenarios
Scenarios
Service Time
Reduced (Sec)
AM Shift
PM Shift
Base Case
Scenario 1
Scenario 2
Scenario 3
Scenario 4
Scenario 5
Scenario 6
Scenario 7
Scenario 8
Scenario 9
0
-15
-30
-45
-60
0
0
0
0
0
0
0
0
0
0
1+
2+
3+
4+
5+
0
0
0
0
0
1+
2+
3+
4+
5+
For each scenario, we obtained results for the number of customers who waited more than 10 min, number of
customers that waited less than 10 min, number of customers service no show, number of customer service
served, number of payment machine no show, number of payment machine served, number of update service no
show and number of update service served, the total number seized, and the utilization of each resource.
636
AM Shift
2+
3+
3+
4+
PM Shift
3+
2+
4+
3+
Figure 4 shows average waiting time in the customer service queue, number of customers waiting more than
10 minutes, and the net profit for all scenarios applied.
637
116
0
118776
434 56000 83885
+374
-30
Loss (SR)
392
92
-77
Net Profit
(SR)
Additional
Salary (SR)
PM shift
0 0 1612
+3 +4 453
- 72
Satisfied
Cust./day
0
0
Unsatisfied
Cust./d
Base Case
Scenario 12
Improvement %
Output
Total time (sec)
Scenario
AM shift
Service Time
Reduced (sec.)
Input
1420979
1509506
+6.23
Payment Machine Service time was 183 + 226 * BETA (0.69, 0.761).
Update Information Service time was 306 + 286 * BETA (0.767, 0.617).
2- A simulation model was constructed; the model consisted of the following sub-models: customer arrivals,
customer service, payment machine, and update service.
3- Thirteen different scenarios were considered by decreasing the service time due to training, and by adding
additional employees.
4- The best scenario attained a net profit of SR1,509,506 with an acceptable total system time of 453 seconds.
5- Comparing the best scenario with the base case, the following improvements were achieved:
* Total system time was decreased by 72%,
* Number of unsatisfied customers was decreased by 77%,
* Number of satisfied customers was increased by 374%,
* Net profit was increased by 6.2% despite increasing the number of employees which incurred extra costs.
7.2 Points for Future Researches
1.
2.
3.
4.
Make necessary coordination of the work schedule and increase the number of staff members at peak times.
Recommend a perfect way to serve VIP customers without affecting workflow.
Apply the same procedure for other branches in Mobily Company in Jeddah and other cities in Saudi Arabia.
Prepare a user-friendly decision support system to help executives to use such a complicated interface for
decision making.
References
1.
2.
3.
4.
5.
Glanz, B. A. (1994). Building Customer Loyalty- How You Can Help Keep Customers Returning, 1st
ed., McGraw-Hill, USA.
Al Otaibi, E. (2005). An investigation of the impact of social identity on consumer satisfaction, MBA
thesis, Business administration, University of surrey.
Al-Nasser, S. (1994). Customer satisfaction and competitive advantage: The case of SABIC (Saudi
Basic Industries Corporation), Master thesis, Business and information technology, Sheffield Hallam
University.
Omer, O. M. A. (1998). An examination of consumer time allocation,Ph.D. thesis,Business
Administration, University of Manchester.
ITU website: http://www.itu.int/en/pages/default.aspx, accessed on 27/01/2010.
638
6.
7.
8.
9.
10.
11.
12.
13.
14.
15.
Biographies
Said Ali Hassan El-Quliti received his BS in Mechanical Engineering from Cairo, Egypt, in May 1971,
M.Sc. in Operations Research Applications, Cairo, Egypt, 1974, M.Sc. in Management Development, Euro-Arab
Management School, Spain, 2000, and Ph.D. in Informatics, National Institute Polytechnic, Toulouse, France,
1981. He is currently working as a Professor, Industrial Engineering Dept., Faculty of Engineering, King
Abdulaziz University, Kingdom of Saudi Arabia,and as a Professor, Decision Support Dept., Faculty of
Computers and Information, Cairo University, Cairo, Egypt. His major research interest includes optimization,
simulation, decision making, feasibility studies, and strategic planning.
He has published over 65 papers in refereed national/international journals and conferences. In addition, he
has reviewed manuscripts for a number of journals and conferences and supervised more than 50 Master and
Ph.D. thesis in Egypt and Saudi Arabia.
Prof. El-Quliti is a member in many scientific societies: The Egyptian Society for Operations Research
Applications, Cairo, Egypt, The Egyptian Society for Operations Research and Decision Support, Cairo, Egypt,
The Scientific Society for Electronic Computers, Alexandria, Egypt, Management Engineering Society, Cairo,
Egypt, International Association of Science and Technology (IASTED), Canada. Association for Advancement
of Modeling and Simulation Techniques in Enterprises (AMSE), France, The International Society on Multiple
Criteria Decision Making, Georgia, USA, World Scientific and Engineering Academy and Society (WSEAS),
USA, Europe, Australia, Mexico, Asia, and Association for Human Resources Management (ASHRM), Saudi
Arabia
Ibrahim Abdulaziz Al-Darrab, PhD, received his BS in Systems Engineering from King Fahad University
of Petroleum and Minerals, Saudi Arabia, MS in Operations Research from Stanford University, California,
USA, MS in Statistics from Stanford University, California, USA and PhD in Engineering Economic Systems
from Stanford University, California, USA. Presently, he is working as the Chairman of the Industrial
Engineering Department, King Abdulaziz University, Jeddah, Saudi Arabia. His major research interests include
optimization of design and process variables using design of experiments, quality control, information systems,
decision analysis, network analysis, reliability and fuzzy modeling.
He has published over 25 papers in refereed national/international journals and conferences. In addition, he
has reviewed manuscripts for a number of national journal and conferences and he is currently preparing
manuscripts of two new books.
Dr. Al-Darrab is a member of the Saudi Society for Industrial and Systems Engineering (SSISE).
639
640
Chapter 4
MECHANICAL
ENGINEERING
641
Abstract. The aim of the present study is to present a methodology using the Waibull and Rayleigh probability
density functions for estimating the wind power potential. In this regard wind power density in the Mediteranean
region of Turkey was investigated. The wind characteristics were statistically analyzed using the wind speed data
collected of three meteorological stations, namely Mersin and Anamur in Turkey during the period 2010 2011.
Mean wind speed, and wind power potential variations of all the stations were estimated by the Weibull and
Rayleigh models.
1 Introduction
The interest in renewable energy increases steadily all over the world, since they dont cause green house effect in
contrary to the fossil fuels. Because of their sustainability, the renewable are also considered as a source suitable to
meet energy needs of the whole world [1]. One of the main renewable energy resources all over the world is wind
which has played a long and important role in the history of human civilization.The wind is an inexhaustible
resource that can provide significant quantities of energy to support a countrys needs. An accurate wind resource
assessment is an important and critical factor to be well understood for harnessing the power of the wind [14].
Turkish energy policy is concentrated mainly on ensuring a supply of reliable, sufficient, economic and clean
energy. Because more than half of the energy demand is met through imports to Turkey, it is necessary to consider
and review the use of potential wind energy resources. However, the potential sites of wind generation of the
country have not been completely investigated in detail yet.Inthe wind energy literature, many relevant works have
been developedin this aim. The wind characterization in terms of speed, direction and windpower is the first step to
obtain the initial feasibility of generatingelectricity from wind power through a wind farm, in a given region [17].
Although several studies investigating the wind potential of the Aegean region of Turkey are available, these
studies do not cover all sites where a substantial rate of wind power potential are available. In this paper the wind
energy potential of two locations in the Mediterranean sea region of Turkey wasstudied based on the Weibull and
the Rayleigh models. The obtained results were compared with measured data.
642
2 Literature
Many researchers have shown interest in the estimation of the wind energy potential of several sites over the world.
Gokcek et al. [1] investigated the wind characteristics and wind potential of Krklareli province in the
MarmaraRegion, Turkey. It was found that annual mean power densitybased on observed data and Weibull
function. Kose et al. [2] examined the wind characteristics and energy potential of Kutahya in Turkey. Ahmed et
al. [3] carried out a similar work in Pakistan to check the possible use of wind energy for irrigation deep well
pumping and small scale power generation, Zhou et al. [4] in the Pearl River Delta region in China and Elamouria
and Ben Amar [5] in Tunisia. Changet al. [6] presented an assessment of wind characteristics and windturbine
characteristics in Taiwan and Fawzi and Jowder [7] realized a similar study in the Kingdom of Bahrain using the
graphical method and Weibull distribution for the wind statistical study. Ucarand Balo [8] investigated the wind
characteristics and energypotential in Uludag-Bursa, Turkey. Shata and Hanitsch [9]developed a work regarding
the electricity generation and windpotential assessment by using wind data from 10 coastal meteorological stations
in Hurghada, Egypt. Migoya et al. [10] estimated the wind energy resource in Madrid region in Spain using the data
collected by the Spanish National Meteorological Institute.Incecik and Erdogmus [11] carried out the wind power
potential on the western coast of Anatolia at eight stations. Karsli and Gecit [12] determined the wind power
potential of the Nurdagi/Gaziantep district located in the south of Turkey using Weibull parameters of the wind
speed distribution. Celik [13] compared the monthly probability density distribution obtained from theWeibull and
Rayleigh models in order to study their suitability. Sahin et al. [14] investigated the wind energy potential of the
east Mediterranean region of Turkey and identified locations with the best wind source. Lima and Filho [15]
presented a wind energy assessment and a wind farm simulation in the northeast region of Brazil. Dahmouni et al.
[16] evaluated the most important characteristics of wind energy in the central coast of the Gulf of Tunis.
643
establish an adequate statistical model for obtaining the wind speed frequency distribution which may later be used
for predicting the wind energy potential. Two of the commonly used functions for fitting a measured wind speed
probability distribution in a given location over a certain period of time are the Weibull and Rayleigh distributions.
The Weibull distribution function that is a special case of generalized gamma distribution for wind speed is
expressed with Eq. (1).
k v
f ( ) =
c c
k 1
v k
exp ,
c
(1)
where v is the wind speed, c is a Weibull scale k is a dimensionless Weibull shape parameter f (v) is the probability
of observing wind speed. The Rayleigh function is a special and simplified case of the Weibull function. It
isobtained when the shape factor c of the Weibull function is assumed to be equal to 2.The Rayleigh probability
density function is given by equation 2. In this study, least mean squares method was used for the parameter
estimation.
v 2
f (v ) = 2 exp .
4 v
2v
(2)
p (v ) =
1
Av 3 .
2
(3)
The wind power per unit area in any windy site is of importance in assessing of the wind power projections for the
power plants. The wind power density of the considered site per unit area based on any probability density function
can be expressed as:
P (v ) =
where
1 3
v f (v )dv ,
2 0
(4)
is the standart air density, 1,225 kg/m3 and v is wind speed, m/s.
In the current study, the monthly probability density distributions obtained from the Weibull and Rayleigh models
were compared to the measured distributions to study their suitability. The mean absolute percentage error (MAPE)
was calculated to see the convergence between the measured data and the results determined by the Weibull and
Rayleigh models.
644
Variations of the wind speed frequency distributions for all stations are presented in fig. 2. It is seen that in this
figure the mean wind speed varies between 1,7 m/s and 3.2 m/s. The diurnal wind speeds of Anamur station are
higher than Mersin station. In addition there is a significant difference in the mean wind speeds among stations and
months.
Anamur
Mersin
months
Then, wind speed frequency distributions for all stations were determined by using the Weibull and Rayleigh
probability density functions. The monthly variations of the mean power densities, which are calculated using
measured data. Weibull and Rayleigh models were presented in figures 34.
645
The power densities calculated from the measured probability density distributions and those obtained from the
models were shown in figures 3-4. The minimum power densities accour 6 W/m2, respectively. The highest power
density values accour 10 W/m2.
Stations / Models
Mersin
2,62
5,07
Anamur
4,14
8,63
5 Conclusions
In this study, assesment of the monthly wind characteristics and wind power potential during the period of 2010
2011 in the two stations of Turkey were determined. The probability density distribution have been derived and the
distributional parameters were indentified. The wind energy potential of the location has been studied based on
Weibull and Rayleigh models. It is seen that the Weibull model provide beter power density estimations than the
Rayleigh model for two stations. The maximum value of monthly mean wind speed determined as 3,2 m/s at
Anamur in January and a minimum value of 1,7 m/s is occurred at Anamur in Agust. And maximum power density
is occurred at Anamur with 27 W/m2. A maximum value of annualmean wind speed is obtained at Anamur as 32,8
m/s.
646
References
1.
2.
3.
4.
5.
6.
7.
8.
9.
10.
11.
12.
13.
14.
15.
16.
17.
Gkek, M. Baylken, A. and Bekdemir, S. (2007). [Investigation of wind characteristics and wind energy potential
in Kirklareli Turkey]. Renewable Energy, 32, 17391752.
Kose, R. Ozgur A. M. Erbas, O. Tugcu, A. (2004). [The analysis of wind data and wind energy potential in
Kutahya, Turkey]. Renewable and Sustainable Energy Reviews, 8, 277288.
Ahmed, A. M. Ahmad, F. Akhtar, W. M. (2006). [Assessment of wind power potential for coastal areas of
Pakistan]. Turkish Journal of Physics, 30, 127.
Zhoua, W. Yanga, H. Fangb, Z. (2006). [Wind power potential and characteristic analysis of the Pearl River Delta
region, China]. Renewable Energy, 31, 739753.
Elamouria, M. Amar, F. B. (2008). [Wind energy potential in Tunisia]. Renewable Energy, 33, 758768.
Chang, T. Wu, Y. Hsu, H. Chu, C. Liao, C. (2003). [Assessment of wind characteristics and wind turbine
characteristics in Taiwan]. Renewable Energy, 28, 851871.
Jowder, A. Fawzi, L. (2009). [Wind power analysis and site matching of wind turbine generators in Kingdom of
Bahrain]. Applied Energy, 86(4), 538545.
Ucar, A. Balo, F. (2009). [Investigation of wind characteristics and assessment of windgeneration potentiality in
Uludag - Bursa, Turkey]. Applied Energy, 86, 333339.
Shata, A. S. Hanitsch, R. (2008). [Electricity generation and wind potential assessment at Hurghada, Egypt].
Renewable Energy, 33, 141148.
Migoya, E. Crespo, A. Jimnez, A. Garca, J. and Manuel, F. (2007). [Wind energy resource assessment in Madrid
region]. Renewable Energy, 32, 1467-1483.
ncecik, S. Erdomu, F. (1994).[An Investigation ofWind Power Potentialin Western Coast of Anatolia]. Renewable
Energy, 6, 863-865.
Karsli, M. V. Gecit, C. (2003). [An investigation on wind power potential ofNurdagi-Gaziantep, Turkey]. Renewable
Energy, 28 (5), 823830.
Celik, A. N. (2003). [A statistical analysis of wind power density based on the Weibull and Rayleigh models at the
southern region of Turkey]. Renewable Energy, 29, 593604.
Sahin, B. Bilgili, M. and Akilli, H. (2005). [The wind power potential of the eastern Mediterranean region of
Turkey]. Journal of Wind Engineering and Industrial Aerodynamics, 93 (2), 171-183.
Lima, L. A. and Filho, C. R. B. (2010). [Wind energy assessment and wind farm simulation in Triunfo
Pernambuco, Brazil]. Renewable Energy, 35 (12), 2705-2713.
Dahmouni, A.W. Ben Salah, M. Askri, F. Kerkeni, C. and Ben Nasrallah, S. (2010). [Wind energy in the Gulf of
Tunis, Tunisia]. Renewable and Sustainable Energy Reviews, 14(4), 1303-1311.
Ulgen, K. Genc, A. Hepbasli, A. Oturanc, G. (2004). [Assessment of Wind Characteristics for Energy Generation].
Energy Sources Part A, 26(13), 1227 1237.
Biographies
Aysun SABA Dr. Sagbas is an Associate Professor in Industrial Engineering Department at Namk Kemal University,
Turkey. She received his Ph.D.(2003) in Industrial Engineering from Cukurova University, Turkey. Dr. Sagbas's major research
areas are optimization and modeling, statistical quality improvement, experimental design, artificial intellegent.
Tunca KARAMANLIOLU He was born in Sultanhisar Aydn / TURKEY at 1985. He finished University of Mersin,
Department of Mechanical Education with 83 points. Masters education at the University of Mersin began in 2009 continues.
647
Abstract. Our country is located in active fault line geography. Many losses have happened due to the earthquakes
in recent years. This situation indicates how important it is to design buildings and determine seismic performances
of existing structures and strengthen them if necessary. Linear and non-linear evaluation methods have been
introduced according to different codes. Consistency between the results of the methods used for seismic evaluation
of existing buildings is of prime importance. For this reason, a 3 storey structure which has A2 irregularity in plan is
evaluated according to Turkish Earthquake Code-2007 (TEC-2007) and FEMA-356 in this study. The structure is
analysed by applying both linear and non-linear evaluation procedures. After the damage states are determined
according to codes, the results are compared by each other and suggestions are proposed. Furthermore non-linear
procedure results are compared between TEC-2007 and FEMA-356 Displacement Coefficient Method (DCM).
648
1 Introduction
Performance based design and evaluation terms have become important in recent years. Performance evaluation is
used to determine the seismic safety of existing buildings. Generally, structures are designed to survive without any
damage after minor earthquakes. They are aimed to provide life safety after medium intense earthquakes and to stand
without fully collapsing after major earthquakes.
Big scaled earthquakes have occurred in many countries in the past 30 years. After a number of major earthquakes,
investigation of earthquake codes and scientific research projects have been started. As a result, displacement based
procedures have been developed instead of forced based ones. In addition, ATC-40 by Applied Technology Council,
FEMA-273, FEMA-356 and FEMA-440 by Federal Emergency Management Agency procedures are introduced
[1,2,3]. In parallel with these procedures, Turkish Earthquake Code came into effect in 2007 [4]. Calculation steps and
strengthening design rules for buildings under the effect of earthquakes in seismic zones are described in 7th Chapter
of Turkish Earthquake Code-2007.
Linear and non-linear evaluation methods are widely used to determine the seismic performances of structures [5-14].
While linear methods are force based and damage regions are decided according to r = demand/capacity values, nonlinear methods are deformation based and more parameters are needed to evaluate the structural system.
In the scope of this paper, performance evaluation of an existing concrete structure which has A2 irregularity in plane
according to Turkish Earthquake Code-2007 and FEMA-356 Displacement Coefficient Method is studied. SAP2000
and SEMAp analysis programs are used for solutions [15,16].
After modeling the structure in SAP2000 analysis program, linear and non-linear analysis procedures are
implemented for the building. Linear elastic analyses are combined with the capacity analysis so as to determine the
column and bream section capacities of the building under seismic effects. Afterwards, member damage regions are
649
decided by comparing the demand/capacity ratios according to TEC-2007. For non-linear analysis procedures, SEMAp
analysis program is used to obtain moment-curvature diagrams instead of default values for hinge properties to
determine the exact damage states. By using the real moment-curvature results, performance level of the building is
estimated by comparing strain values corresponding to plastic rotations with the limit values according to TEC-2007
and FEMA-356.
Periods of vibration modes and related effective mass ratios according to each direction are determined after
analysing the structure. Third, sixth and ninth modes appear to be torsional ones as given in Table 1.
Table 1. Periods of vibration modes and related effective mass ratios
X Direction
Y Direction
Period
ux
(sec)
(%)
0.48
89
0.16
0.09
Mode
Z Direction
Period
uy
(sec)
(%)
0.47
90
0.42
10
0.16
0.14
0.09
0.08
Mode
Mode
Period
(sec)
3 Evaluation Procedures
There are several procedures available for performance assessment in the literature. Generally evaluation procedures
are divided into two main parts such as linear and non-linear ones.
3.1 Linear Evaluation Procedures
There are two linear analysis methods given in TEC-2007. These methods are Equivalent Seismic Load Method and
Mode Superposition Method. These analyses are employed for performance assessment. Performance limits of sections
are obtained by determining the demand/capacity ratios (member r factors) of each of them. Demand/capacity ratios
are used in determining the member damage levels of ductile member using linear analysis methods.
Equivalent seismic load method analysis is limited to 8 story buildings with total height not exceeding 25m without
basement, and not possessing torsion irregularity. Mode superposition method can be applied to all buildings without
any restrictions. The signs of internal member forces and capacities under an earthquake excitation direction are taken
as the signs consistent with the dominant mode shape at this direction.
Structural members are classified as ductile and brittle with respect to their mode of failure in determining the
damage limits. Member damage levels for ductile structural members are divided into three limits. They are 3
Minimum Damage Limit (MN), Safety Limit (SF) and Collapse Limit (CL) as indicated in Fig. 2. The calculated
member r factors for beams and columns are compared with the r limits given in TEC-2007 to determine the member
damage regions.
650
t = C0 C1C 2 C3Sa
Te2
42
(1)
Non-linear flexural behavior in frame members is confined to plastic hinges, where the plastic hinge length Lp is
assumed as half of the section depth. Pre-yield linear behavior of concrete sections is represented by cracked sections.
Strain hardening in the plastic range may be ignored, provided that the plastic deformation vector remains normal to
the yield surface.
The objective is to carry out non-linear static analysis under incrementally increasing seismic forces distributed in
accordance with the dominant mode shape in the earthquake excitation direction. Seismic forces are increased until the
earthquake displacement demand is reached. Internal member forces and plastic deformations are calculated at the
demand level. Concrete compressive strains and steel tensile strain demands at the plastic regions are calculated from
the moment-curvature diagrams. Moment-curvature diagrams of the critical sections are obtained by using appropriate
stress-strain rules for concrete and steel. Finally, plastic rotation values and calculated strain demands are compared
with the damage limit values according to related codes.
4 Findings
The building is evaluated according to linear and non-linear evaluation procedures in accordance with TEC-2007 and
FEMA-356 (DCM). After the damage states of the members are calculated, damage limits of the structure are
determined.
651
Performance Levels
Moment
Performance Levels
Plastic Rotation (P )
Displacement ()
Member damages are calculated at each direction. There are total results of beams and columns for linear and nonlinear evaluation procedures in the following figures. The results are the superposition of the x and y directions
according to TEC-2007 and FEMA-356 (DCM).
652
determined for each one. Its seen that linear methods according to TEC-2007 give more conservative results as
expected. TEC-2007 also gives conservative results when non-linear damage ratios are compared.While damage limits
are decided according to concrete and steel strain values according to TEC-2007, these damage limits are determined
by using plastic rotation values according to FEMA-356. For further studies, solution steps between codes and
guidelines shall be studied. In addition to this, low, middle and high structures which have other irregularities shall be
studied too. So that, performance levels of concrete structures can be generalized.
References
1.
2.
3.
4.
5.
6.
7.
8.
9.
10.
11.
12.
13.
14.
15.
16.
ATC-40. 1996. Seismic evaluation and retrofit of concrete buildings vols. 12. California: Applied Technology Council.
FEMA-356. 2000. Prestandard and Commentary for Seismic Rehabilitation of Buildings. Washington (DC): Federal
Emergency Management Agency.
FEMA-440. 2004. Improvement of Nonlinear Static Seismic Analysis Procedures. Washington (DC): Federal Emergency
Management Agency.
Turkish Earthquake Code-2007. 2007. Specifications for Buildings to be Built in Seismic Areas. Ankara: Ministry of
Public Works and Settlement.
Kasil, H.B. 2008. A comparative study of seismic safety of reinforced concrete structural systems, Master Thesis, Istanbul
Technical University, Department of Civil Engineering. Istanbul.
Korkmaz, K.A.; and Dzgn, M. (2007). Dorusal olmayan yapsal analiz yntemlerinin deerlendirilmesi. stanbul
Teknik niversitesi Dergisi (Mhendislik-D) 6(3): 3-10.
Sezer, F.; Genolu, M.; and Celep, Z. (2007). A comparative study of seismic safety evaluation of concrete buildings
according to Turkish Seismic Code (2007). In Proceedings of the Sixth National Conference on Earthquake Engineering.
Istanbul, Turkey, 16-20 October.
Kappos, A.J.; and Panagopoulos, G. (2004). Performance-based seismic design of 3D RC buildings using inelastic static
and dynamic analysis procedures. ISET Journal of Earthquake Technology 41(1): 141-158.
Tuncer, O.; Celep, Z.; and Ylmaz, M.B. (2009). A comparative evaluation of the methods given in The Turkish Seismic
Code (2007). In Proceedings of the WCCE ECCE TCCE Joint Conference : EARTHQUAKE& TSUNAMI. Istanbul,
Turkey. 22-24 June.
Uygun, G.; and Celep, Z. (2007). A comparative study of seismic safety evaluation of a concrete building according to the
linear and the non-linear methods of Turkish Seismic Code (2007). In Proceedings of the Sixth National Conference on
Earthquake Engineering. Istanbul, Turkey. 16-20 October.
Fardis, M.N. (2003). Seismic assessment and retrofitting of existing buildings according to Eurocode 8. In Proceedings of
the Fifth National Conference on Earthquake Engineering. Istanbul, Turkey, 26-30 May.
Fajfar, P.; and Fischinger, M. (1988). A method for non-linear seismic analysis of regular RC buildings. In Proceedings of
the Proc. 9thWorld Conference in Earthquake Engineering. Tokyo-Kyoto.
Kim, S.; and DAmore, E. (1999). Pushover analysis procedures in earthquake engineering. Earthquake Spectra 15(3):
417-434.
nel, M.; and zmen, H.B. (2006). Effect of plastic hinge properties in non-linear analysis of reinforced concrete
buildings. Engineering Structures 28(11): 1494-502.
SAP2000. 2000. Three Dimensional Static and Dynamic Finite Element Analysis and Design of Structure. Berkeley:
Computers and Structures Inc.
nel, M.; zmen, H.B.; and Bilgin, H. 2008. Sarg Etkisi Modelleme Analiz Program (SEMAp) Kullanm klavuzu,
Denizli.
653
Biographies
Muhiddin Bac Muhiddin Bac was born in Ankara. He graduated from Dokuz Eyll University in 1991. He completed his
graduated thesis in 1996 and his doctorate thesis in 2001. He started working as a research assistant since 1993. He has been
working as an assistant professor since 2006 in Celal Bayar University. He is studying concrete structures, performance evaluation
and finite elements method. He knows English as a foreign language.
He has seven pressed scientific articles, eighteen national articles. His five papers have been accepted and presented in national
conferences and six in international conference. He has two graduate and two doctorate students this year.
Assistant Professor Dr. Muhiddin Bac is a member of Chamber of Civil Engineers.
R. Turul Erdem R. Turul Erdem was born in 1985 in zmir. He was graduated from Celal Bayar University in 2008. He is a
doctoral student in civil engineering department in the same university. He has started working as a research assistant in Celal Bayar
University since 2010. He is studying concrete structures, performance evaluation and artificial neural network analysis. He knows
two foreign languages such as English and German.
He has two scientific, two international and two national articles. He has four papers in international conferences.
Research Assistant R. Turul Erdem is a member of Turkish Chamber of Civil Engineers since 2008.
Ali Demir Ali Demir was born in Karabk in 1983. He was graduated from Celal Bayar University in 2005. He completed his
graduate thesis in 2008. He is still a doctoral student in Celal Bayar University. He has been a research assistant in Celal Bayar
University for 6 years. He is studying concrete structures and performance evaluation of buildings. Ali Demir knows English as a
foreign language.
He has seven published articles. He also has accepted and presented papers in national and international conferences. He has
completed a scientific research project in Celal Bayar University.
Research Assistant Ali Demir is a member of Turkish Chamber of Civil Engineers since 2005.
654
Abstract. Many methods have been developed about analysis of ground vibration effects which arise from blasting
operations. One of the most popular of these methodologies is the scaled distance approach, which involves the
prediction of the peak particle velocity (PPV) as a function of the explosive charge and distance from the explosion
source. The use of scaled distance approach in the field of civil engineering has been limited so far. Acceleration
records have been preferred to particle velocities in estimating the behavior of non-buried structures under dynamic
loads. In this study, a relationship between the amount of explosives, the distance, and the resulting horizontal peak
acceleration was obtained. Then, the dynamic responses of four and six storey reinforced concrete structures that
consist of frame and shear wall type structural systems to blast-induced ground accelerations were investigated using
finite element analysis.
1 Introduction
Blasting is frequently used in civil engineering applications during the construction of dams and tunnels in addition
to mining and quarrying disciplines. These operations, which involve surface and underground applications of blasting,
are possible sources of threat for existing structures nearby. It is well known that the blasting-induced effects on a site
within a given distance depend on a number of factors that work hand in hand. Among the most important of these
factors are the amount and the type of the explosive employed, the distance between the blasting source and the site,
the topography, and the physical and mechanical properties of the material that rests between the source and the site.
The uses of explosives are not limited to civil construction procedures only, they are also employed for malicious
purposes. Unfortunately, a number of structures were targeted by terrorist attacks through the use of significant amount
of explosives during the past few years. Therefore, blast-resistant design of important structures such as high-rise
buildings and bridges, in addition to military complexes, has recently become a major research field [1-6].
A literature survey indicates that the PPV is the principal concept that has been used to investigate the blast-induced
dynamic effects especially on buried structures [7-12]. However, PPV is not a frequently employed parameter in civil
engineering applications. The authors are not aware of any studies that involve the measurement of the blast-induced
accelerations and the consideration of these accelerations as a dynamic load on the structures. Therefore, within the
scope of this study, the use of acceleration records instead of PPV has been considered.
In this paper, the maximum horizontal accelerations that result from 19 TNT explosions were correlated with the
amount of explosive, the distance from the explosion, and the depth of the explosion. The obtained correlations were
used to estimate the blast induced accelerations that will act on the structures, especially for similar geotechnical
655
conditions These correlations were then employed to estimate the acceleration profiles that can be utilized as the
dynamic loading within the finite element analyses of 4 and 6 storey reinforced concrete structures that consist of
frame and shear wall structural systems respectively.
2 Effects of Blasting
A number of experiments that are based on the results of trial blasting have indicated that the PPV of any component
of ground vibrations is a function of the amount of explosives, and the distance to the blasting point. The scaled
distance (SD) is a combined representation of the effects of the amount of explosives at a single detonation (W) and
the distance from the blasting point to the measurement station (R) are given in Eq.1. below.
SD =
(1)
Investigations on blasting related damage of surficial structures from a mining engineering point of view concluded
that one of the better damage parameters is the PPV, and utilized a power scaling law between PPV and SD. On the
other hand, the current study uses the definition given by Equation 1 to obtain a relationship between SD and the
measured values of maximum horizontal ground acceleration. As shown in Fig. 1, when plotted on logarithmic scale
axes, this relationship results in the estimation of two constants that are specific for the geotechnical conditions
considered. These constants can then be used to calculate the expected horizontal accelerations for other blasting
operations that are performed using different amounts of explosives and at different distances for structures located at
similar geotechnical conditions.
Maximum and minimum ground acceleration measurements obtained from TNT explosions are given in Fig. 2.
656
Selected column
Selected column
for comparison
for comparison
The story heights are 3.0 m. Three dimensional views of the four storey buildings are shown in Fig. 4. For all
buildings, the dimensions of corner columns are 400400 mm. The rest of the columns are 300500 mm and 500300
mm. Shear walls are 2002000 mm and 2000200 mm. All beams are rectangular in shape, and their dimensions are
250 mm500 mm. The slab thickness is 150 mm in all structures.
657
The vertical loads consist of dead and live loads of slabs, wall loads on beams, as well as the dead loads of columns
and beams. In addition to the dead and live loads, the acceleration time history that was measured during the
experimental blasting operations was acted on the structures as a dynamic loading. For this purpose, the accelerationtime history with the highest peak horizontal acceleration value among the nineteen experiments was chosen.
4 Evaluation of Results
Within the context of the current study, the modeled reinforced concrete structures were loaded with the acceleration
time history having the highest peak acceleration value among the measured ones.
The interstory drift ratio values are used to investigate and compare the deformation profiles of the structures due to
blast-induced loading. The interstory drift ratios for four and six storey buildings belonging to both structural types are
given in Table 1 and 2.
Table 1.Drift ratios of 4 storey structure
Acceleration 1
Story
Direction
Acceleration 2
Structure with RC
Walls
Frames
Interstory
Total
Interstory
Total
Interstory
Total
Interstory
Total
Drifts
Drifts
Drifts
Drifts
Drifts
Drifts
Drifts
Drifts
0.00201
0.00201
0.00093
0.00093
0.00289
0.00289
0.00134
0.00134
0.00222
0.00222
0.00082
0.00082
0.00319
0.00319
0.00117
0.00117
1
x
0.00194
0.00395
0.00145
0.00239
0.00279
0.00568
0.00209
0.00343
0.00199
0.00421
0.00137
0.00219
0.00287
0.00605
0.00197
0.00314
2
x
0.00124
0.00519
0.00135
0.00373
0.00179
0.00746
0.00194
0.00537
0.00115
0.00536
0.00135
0.00354
0.00166
0.00771
0.00194
0.00509
3
X
0.00046
0.00565
0.00102
0.00475
0.00067
0.00813
0.00147
0.00683
0.00029
0.00565
0.00110
0.00464
0.00042
0.00813
0.00158
0.00667
Story
Direction
Acceleration 2
Structure with RC
Walls
Frames
Interstory
Total
Interstory
Total
Interstory
Total
Interstory
Total
Drifts
Drifts
Drifts
Drifts
Drifts
Drifts
Drifts
Drifts
0.00185
0.00185
0.00093
0.00093
0.00267
0.00267
0.00134
0.00134
0.00206
0.00206
0.00081
0.00081
0.00296
0.00296
0.00116
0.00116
1
X
0.00195
0.00380
0.00142
0.00235
0.00280
0.00546
0.00204
0.00338
0.00200
0.00406
0.00130
0.00211
0.00288
0.00583
0.00188
0.00304
2
X
0.00106
0.00486
0.00124
0.00359
0.00152
0.00698
0.00178
0.00516
0.00099
0.00505
0.00121
0.00332
0.00143
0.00726
0.00174
0.00478
3
X
0.00050
0.00536
0.00082
0.00441
0.00072
0.00771
0.00118
0.00634
0.00406
0.00501
0.00088
0.00421
0.00066
0.00792
0.00127
0.00605
4
X
0.00025
0.00561
0.00043
0.00483
0.00036
0.00807
0.00061
0.00695
0.00015
0.00566
0.00058
0.00478
0.00021
0.00813
0.00083
0.00688
5
X
0.00014
0.00575
0.00020
0.00504
0.00021
0.00827
0.00029
0.00724
0.00008
0.00573
0.00038
0.00517
0.00011
0.00825
0.00055
0.00743
The analyses results indicate that the buildings with the frame structural system only, experienced significantly
higher interstory drift ratios than those observed in the building with shear walls. The difference in the interstory drift
658
ratios decreases with increasing height from ground level. It is clear that the inclusion of a reinforced concrete shear
wall system with an area of 1.0% of the total floor area has been quite beneficial in decreasing the deformations due to
blast induced dynamic loading.
To further investigate the effect of blast induced dynamic effects on the structures, a vertical load carrying member
was also selected in all four buildings, to compare and observe the variation of cross sectional moment and shear force
values over the height of the structures. The comparisons of shear force and moment values that were observed in the
selected column for different structural types are presented in Tables 3 and 4, for four and six storey buildings
respectively.
Table 3.Shear force and moment values under the effect of blasting for 4 storey structures
Acceleration 1
Story
No
Direction
Structure with
Structure with
Structure with
Structure with
RC frames
shear walls
RC frames
shear walls
Shear
Force
(kN)
1
Acceleration 2
Moment
(kNm)
Shear
Force
(kN)
Moment
(kNm)
Shear
Force
(kN)
Moment
(kNm)
Shear
Force
(kN)
Moment
(kNm)
49.02
65.573
46.25
61.428
70.49
94.288
66.51
88.328
54.38
72.990
49.27
67.257
78.19
104.953
70.84
96.710
31.75
47.984
28.17
46.841
45.65
68.996
40.50
67.353
32.28
53.284
30.19
50.779
46.42
76.617
43.41
73.016
19.05
34.677
18.91
33.600
27.39
49.863
27.19
48.314
20.98
36.763
21.27
36.111
30.16
52.862
30.58
51.924
2.88
4.428
1.77
3.081
4.14
6.367
2.55
4.431
4.48
5.571
3.12
4.916
6.45
8.010
4.49
7.069
Table 4.Shear force and moment values under the effect of blasting for 6 storey structures
Acceleration 1
Story
No
Direction
Structure with
Structure with
Structure with
Structure with
RC frames
shear walls
RC frames
shear walls
Shear
Force
(kN)
1
Acceleration 2
Moment
(kNm)
Shear
Force
(kN)
Moment
(kNm)
Shear
Force
(kN)
Moment
(kNm)
Shear
Force
(kN)
Moment
(kNm)
50.70
65.730
45.56
62.638
72.90
94.514
65.52
90.068
56.81
73.956
50.47
71.104
81.69
106.341
72.58
102.241
33.89
53.917
31.04
51.119
48.73
77.528
44.64
73.504
36.92
57.676
34.30
55.051
53.09
82.933
49.32
79.158
18.97
32.950
17.47
31.969
27.28
47.379
25.12
45.969
20.67
35.929
19.52
34.774
29.72
51.663
28.07
50.002
5.36
12.439
4.39
11.022
7.71
17.886
6.31
15.849
7.45
16.374
6.09
13.935
10.71
23.544
8.76
20.038
2.46
1.675
2.17
1.480
3.54
2.409
3.12
2.127
3.87
3.778
3.09
3.159
5.56
5.432
4.44
4.543
0.23
0.858
0.18
0.786
0.33
1.234
0.26
1.131
0.95
1.386
0.87
1.218
1.37
1.993
1.25
1.751
659
These tabulated values are the column upper-end cross sectional moments and shear forces at the corresponding
floor slab level. Both the shear force and moment values are observed to be higher in the selected column for the
structures without the shear walls.
References
1.
2.
3.
4.
5.
6.
7.
8.
9.
10.
11.
12.
13.
DOE/TIC-11268. 1992. A Manual for the Prediction of Blast and Fragment Loadings on Structures, Washington DC:
U.S. Department of Energy.
Longinow, A.; and Mniszewski, K.R. (1996). Protecting buildings against vehicle bomb attacks. Practice Periodical on
Structural Design and Construction, ASCE 51-54.
Mendis, P.A.; and Ngo, T. (2002). Assessment of tall buildings under blast loading and aircraft impact. International
Association of Bridge and Structural Engineering 375-376.
Mills, C.A. (1987). The design of concrete structure to resist explosions and weapon effects. In Proceedings of the 1st Int.
Conference on Concrete for Hazard Protections. Edinburgh, UK.
Newmark, N.M.; and Hansen, R.J. 1961. Design of Blast Resistant Structures. New York: McGraw-Hill.
Ngo, T.; Mendis, P.; and Itoh, A. (2004). Modelling reinforced concrete structures subjected to impulsive loading using
the modified continuum lattice model. In Proceedings of 18th Australasian Conference on the Mechanics of Structures
and Materials. Australia.
Das, B.M. 1993. Principles of Soil Dynamics. CA, USA: Brooks/Cole Thomson Learning.
Duvall, W. I.; Devine, J. F. 1972. Surface Mining. New York: E. Pfleider ed.
Dowding, C.H. (1985). Blast vibration monitoring and control. Prentice Hall, Eaglewood Cliffs 40-110.
Aung, K.K.; Rahardjo, H., Leong, E.C.; and Toll, D.G. (2001). Relationship between porosimetry measurement and soilwater characteristic curve for an unsaturated residual soil. Geotechnical and Geological Engineering 19: 401-416.
Fredlund, D.G.; and Rahardjo, H. 1993. Soil Mechanics for Unsaturated Soils. New York: John Wiley and Sons Inc.
Lee, M.C.; and Penzien, J. 1980. Stochastic seismic analysis of nuclear power plant structures and piping systems
subjected to multiple support excitations (Report No. UCB/EERC-80/19). California, Berkeley, Earthquake Engineering
Research Center.
Turkish Earthquake Code-2007. 2007. Specifications for Buildings to be Built in Seismic Areas. Ankara: Ministry of
Public Works and Settlement.
660
Biographies
Erkan Kantar Erkan Kantar was born in 1972 in Kahramanmara. He was graduated from Uluda University in 1993. He
completed his doctorate thesis in 2009 in Gazi University. He started working as a research assistant in 1994 in Celal Bayar
University. He is studying structure of materials, concrete and concrete buildings. He knows English as a foreign language.
He has two pressed scientific and two national articles. He has two papers in international conferences
Assistant Professor Dr. Erkan Kantar is a member of Turkish Chamber of Civil Engineers.
R. Turul Erdem R. Turul Erdem was born in 1985 in zmir. He was graduated from Celal Bayar University in 2008. He is a
doctoral student and started working as a research assistant since 2010 in Celal Bayar University. He is studying concrete structures,
performance evaluation and artificial neural network analysis. He knows two foreign languages as English and German.
He has two scientific, two international and two national articles. He has four papers in international conferences.
Research Assistant R. Turul Erdem is a member of Turkish Chamber of Civil Engineers.
zgr Anl zgr Anl was born in 1975 in Ankara. He took his BSc. degree at 1996 from Gazi Universtiy. He continued his
MSc. And PhD. study at the same university and completed them at 2004. He started working as a research assistant at Gazi
University starting from 1997. His expertise subjects are strengthening of reinforced concrete structures, experimental studies and
finite element modeling of structures. He knows English.
He has thirty-nine scientific and conference paper totally. He has eight completed scientific research projects in Gazi University.
S. Ouzhan Akba Sami Ouzhan Akba was born in 1976 in Ankara. Dr. Akba, after receiving his B.S. degree in Civil
Engineering from Middle East Technical University in 1997, an M.Eng. in 1999, a Ph.D. in Civil Engineering in 2007 at Cornell
University, and working as a post-doctoral researcher in University of Milan for one year, joined Gazi University Civil Engineering
Department in the same year. His research interests include foundation behavior, soil-structure interaction and reliability-based
design. He knows English.
He has more than twenty refereed publications totally. He has fifteen papers in national and international conferences.
Assistant Professor Dr. Akba is a member of ASCE, Geo-Institute, ISSMGE and ZMTM.
661
murat.sarigul@bayar.edu.tr, 2hakan.boyaci@bayar.edu.tr
1,2
Abstract. In this study, transverse vibrations of beam carrying a concentrated mass were handled. The beam
was simply supported at both ends. The Hamilton principle was employed to obtain the governing equation,
which was a nonlinear partial-differential equation due to the geometric nonlinearity caused by finite stretch
of the beam. Damping and harmonic excitation terms were added to the equations of motion, and then
approximate solutions of the mathematically formulated problem were sought. In case of primary resonance,
analytical solutions were derived by means of Method of Multiple Scales(a perturbation method). New
Symplectic Method has been used to solve coupled differential equations. Vibrational characteristics of the
system were investigated in the region of steady-state vibrations via the amplitude-phase modulation
equations of the beam. For different mass ratios, mass locations, frequencyamplitude graphs were plotted.
1 . Introduction
The symplectic elasticity approach was first used by Feng [1,2] in computational solid mechanics.
Subsequently, the symplectic approach was widely applied by many investigators to the static analyses of twodimensional problems after the pioneering works of Zhong [3,4] in analytical solid mechanics using the
symplectic approach. Xu et al. [5] studied the SaintVenant problem in circular cylinders and analyzed the zero
eigenvalue solutions and their Jordan normal forms of the corresponding Hamiltonian operator matrix. The
similar problems in anisotropic composite laminated plates in cylindrical bending were also investigated by
Zhong and Yao [6], and Yao and Yang [7]. Most recently, Lim et al. [10] proposed a new symplectic approach
for the bending analysis of thin plates with two opposite edges simply supported. In their analysis, a series of
bending moment functions were introduced to construct the ProHellingerReissner variational principle, which
is an analogy to plane elasticity. This new symplectic approach was subsequently extended to the bending
analysis of corner supported. Kirchhoff plates due to uniformly distributed loading [8]. The exact explicit
solution of the deflection was derived for the first time. Furthermore, the zero natural support conditions at the
free edges are satisfied and the twisting moment condition at the support corners are predicted with high
accuracy, which could not be exactly satisfied in many papers for more than half a century in both analytical and
numerical analysis including the finite element method. As for vibration analysis of plates, Zou [9] reported an
exact symplectic geometry solution for the static and dynamic analysis of Reissner plates, but it was not exactly
the same as the symplectic elasticity approach described above because trial mode shape functions for the simply
supported opposite edges were still adopted in his analysis.
Our model is beam-mass systems which was studied by Pakdemirli[12], zkaya et al[13] and
zkaya[14]. Approximate analyses have been carried out for nonlinear vibrations of a beam-mass system under
662
simple end conditions. In case of immovable end conditions nonlinearities arise due to the axial stretching of the
beam during the vibrations. The other source of the nonlinearities includes stretching.
In this study, transverse vibrations of beam carrying a concentrated mass were handled. The beam was
simply supported at both ends. The Hamilton principle was employed to obtain the governing equation, which
was a nonlinear partial-differential equation due to the geometric nonlinearity caused by finite stretching of the
beam. Damping and harmonic excitation terms were added to the equations of motion, and then approximate
solutions of the mathematically formulated problem were sought. In case of primary resonance, analytical
solutions were derived by means of Method of Multiple Scales(a perturbation method). New Symplectic Method
was used to solve coupled differential equations. Vibrational characteristics of the system were investigated in
the region of steady-state vibrations via the amplitude-phase modulation equations of the beam. For different
mass ratios, mass locations, frequencyamplitude graphs were plotted.
2 . Equations of Motion
w1 (x, t )
w1 (0)= 0
w1(0 )= 0
w2 (x, t )
w2 (1)= 0
w2 (1)= 0
zkaya et al.[12] used Energy approach and invoked this energy via Hamilton principle. After
nondimensionalization, they obtained the following equations of motion and conditions;
iv
1
1
2
2
&&i = . (w1 ) .dx + (w2 ) .dx . wi 2. . wi + . Fi .cos(.t )
wi + w
2 0
2
w1 (0) = w1(0) = 0,
w1 ( )= w2 ( ),
w1 ( )= w2 ( ),
w1( )= w2 ( ),
(1)
w2 (1)= w2 (1)= 0
&&1 ( )= 0
w1( ) w2( ) .w
(2)
where was defined as ratio between concentrated mass(M) and beam mass.
3 . Analytical Solutions
3.1 Multiple Scales Method
We apply the method of multiple scales (MMS), a perturbation technique (Nayfeh-[15]), directly to the
partial-differential equations and conditions. Impending i=1,2 in this method, we assume expansions as follows;
3
(3)
663
where is a small book-keeping parameter artificially inserted into the equations. This parameter can be taken as
1 at the end upon keeping in mind, however, that deflections are small. Therefore, we investigated a weak
nonlinear system. T0=t is the fast time scale, T1=.t and, T2=2.t are the slow time scales in MMS.
Now consider only the primary resonance case and hence, the forcing and damping terms are ordered so
that they counter the effect of the nonlinear terms
r
2
3
i = . i ,
Fi = .Fi
(4)
d dt = D0 + 2 .D2 ,
(5)
Substituting Eqs.(9)-(11) into Eq.(7) and separating each order of , one obtains the followings;
order :
(6)
(0 ) = 0 w 21 (1) = w 21
(1) = 0
w11 (0 ) = w11
( ) = w21
( ), w11
( ) = w21
( ), w11
( ) w21
( ) .D0 w11 ( ) = 0
w11 ( ) = w21 ( ), w11
2
(7)
order 3:
1
2 .dx + w21
2 .dx . wi1 + Fi .cos(.t ) (8)
D02 wi 3 + wiv3 = 2.D0 D2 wi1 2. . D0 wi1 + w11
(0 ) = 0 , w 23 (1) = w 23
(1) = 0
w13 (0 ) = w13
( ) = w 23
( ), w13
( ) = w 23
( )
w13 ( ) = w 23 ( ), w13
( ) w23
( ) .D0 w13 ( ) 2. .D0 .D2 w11 ( ) = 0 (9)
w13
2
i. .T0
(10)
Yi 4 2 .Yi = 0
(11)
(12)
It is noted that the conjugates of the mode shapes are same for both modes.
664
order 2:
3 2
1
2
.i = 2.i..( A + . A).Yi + . A . A. (Y1) .dx + (Y2 )2 .dx .Yi+ .Fi .e i. .T2
2
2
v
i
(13)
Yi
i
i
= i ,
= i , , i = i ,
= 2 .Yi
x
x
x
x
(15)
Y1 (0 ) = 1 (0 ) = 0 , Y2 (1) = 2 (1) = 0
Y1 ( ) = Y2 ( ), 1 ( ) = 2 ( ), 1 ( ) = 2 ( ), 1 ( ) 2 ( )+ . 2 .Y1 ( ) = 0 (16)
Eq.(15) can be expressed in the matrix form as
X
= H .X
x
(17)
X = e .x
(18)
gives us
X
= .X
x
(19)
where is eigenvalue and X (x ) is the corresponding eigenvector(see details in ref. [Li 2009]). Eq.(19) gives
following matrix form:
0
2
0
0
0
1
0 Yi 0
0 i 0
=
1 i 0
i 0
(20)
From eigenvalue problem at Eq.(20), there are four eigenvalues. Thus, solution of linear problem can be written
as follows;
Yi (x ) = ci1.e 1. x + ci 2 .e 2 . x + ci 3 .e 3 . x + ci 4 .e 4 . x
(21.a)
i (x ) = d i1.e 1. x + d i 2 .e 2 . x + d i 3 .e 3 . x + d i 4 .e 4 . x
(21.b)
i ( x) = f i1.e 1.x + f i 2 .e 2 . x + f i 3 .e 3 .x + f i 4 .e 4 .x
(21.c)
i (x ) = g i1.e 1. x + g i 2 .e 2 .x + g i 3 .e 3 .x + g i 4 .e 4 .x
(21.d)
665
Two of the eigenvalues are complex, thus eigenmode can be written in cosin form;
(22)
~
i ~ ~i ~ i ~
= i ,
= i ,
= i ,
x
x
x
1
~i
~
3
1
2
2
= 2 .i 2.i..(A + . A).Yi + . A 2 . A (Y1) .dx + (Y2 ) .dx .Yi+ .Fi .e i. .T2 (23)
x
2
2
~
~
~
~
1 (0)= 1 (0)= 0 , 2 (1) = 2 (1) = 0 , 1 ( )= 2 ( ) , ~1 ( )= ~2 ( ) , 1 ( )= 2 ( ) ,
(24)
The solvability procedures (Nayfeh and Mook[16]) for these equations are as follows;
2 ~
1 ~
x 2
x 1
~
~
2 ~
1 ~
1
2
1
21
22
23
24 ~
2 ~
1 ~
0
2
1
x
x
~
~
2 2 .~
1 2 .~
2
1
x
x
0
0
1
0
0
= {u11 u12 u13 u14 }. .dx + {u 21 u 22 u 23 u 24 }. .dx
0
0
0
0
0
(25)
From Eq.(25), following trial functions can be obtained in the form of differential equation forms and their
conditions:
u 2 .u i 4 = 0
u i1 + 2 .u i 4 = 0 , u i1 = u i 2 , u i 2 = u i 3 , u i 3 = u i 4 ,
4 i4
x
x
x
x
x
u11 ( ) u 21 ( ) . 2 .u14 ( ) = 0, u12 ( ) u 22 ( ) = 0,
u13 ( ) u 23 ( ) = 0, u14 ( ) u 24 ( ) = 0,
(26)
666
u14 = Y1 ,
u 24 = Y2 .
(27)
For mode shapes and other processes, these functions are sufficient. Then, trial functions can be used for
nonhomogeneous problem. After necessary calculations, one obtains the following equations;
1
1
2
2
2
2
2
0
2
1
= . Y1.F1.dx + Y2 .F2 .dx .e i. .T2 (28)
2 0
3
1
2.i..k . A + 2.i. . . A + . A 2 . A .b 2 = . f .e i. .T2
2
2
(29)
where
b=
(Y1)
.dx + (Y2 ) .dx , f = Y1.F1.dx + Y2 .F2 .dx , 1= Y12 .dx + Y22 .dx ,
1
A(T2 ) = . a(T2 ).ei. (T2 )
2
(31)
Substituting Eq.(31) into Eq.(29), and separating real and imaginary parts, following amplitude-phase
modulation equations can be finally obtained;
1
2
.k .a = . . a + . f .sin , .k . .a = .k . .a
3 2 3 1
.b .a + . f .cos
16
2
(32)
where
= .T2
(33)
667
nl = + = +
3 b2 2
.
.a 0
16 .k
(35)
Damped-forced vibration
Considering damping and external excitation case under Eq.(46), one obtains following equations;
1
2
. . a0 = . f .sin , . k . .a0 +
3 2 3 1
.b . a0 = . f . cos
16
2
(36)
= 1 ).
Fig. 3. Nonlinear frequency-amplitude curves for different mass ratios (first mode, = 0 . 5 ).
After some manipulation for steady state case, we obtained following detuning parameter expression;
668
3 b2 2
.
. a0
16 . k
1 f 2
.
2
2
2
4 .a
0
(37)
f = , =
k
k
(38)
= 1 ).
669
G1
a
G
2
a
G1
G2
a==a0
(39)
where it is assumed that a = G1 and = G 2 , and terms with 0 indices define fixed points of the steady state. If
eigenvalues of the Jacobian matrix have negative real parts, these fixed points are stable.
In Figs.(23), the non-linear frequencyamplitude curves are drawn for different mass locations and
ratios. As seen in these figures, an increase in concentrated mass ratio(=0.5=5.6796, =1=6.9660)
decreases the linear and non-linear frequencies. While mass gets closer to the midpoint (increasing mass
location(=0.2=7.4541, =0.5=5.6796)) the linear and non-linear frequencies decreases. Using Eq.(37),
the frequencyresponse graphs were drawn at Figs.(4-5). f=1 and =0.2 were taken at those graphs.
Frequencyresponse
graphs
via
different
concentrated
mass
locations(=0.2=7.4541,
=0.5=5.6796) were given in Fig. 4 for =1. While mass gets closer to on midpoint (increasing mass
location()), the system has hardening type behavior as seen from graph. In Fig. 5 for =0.5, maximum
amplitude value increases while concentrated mass increases (=0.5=5.6796, =1=6.9660) jump region
gets wider.
In this paper the new symplectic method was applied for nonlinear vibrations of beams. Vibration
problems of the beams carrying concentrated masses, which had been modelled in a previous study(zkaya[14]),
were handled. Unlike the previous studies, this study is a first in using the New Symplectic Approach via
Perturbation method. The present analysis has provided a significant extension of the symplectic approach in
future such as our study based on the Timoshenko beams carrying concentrated masses.
References
1. Feng,K.(1985). On difference scheme and symplectic geometry, in Proceedings of the Beijing Symposium Differential
Geometry and Differential Equations, Beijing: 4258.
2. Feng,K.(1986). Difference schemes for Hamiltonian formalism and symplectic geometry, J. Comput. Math. 4: 279
289.
3. Zhong,W.X.(1991). Plane elasticity problem in strip domain and Hamiltonian system, J. Dalian Univ. Tech. 31: 373
384 (in Chinese).
4. Zhong,W.X.(1992). On the reciprocal theorem and adjoint symplectic orthogonal relation, Acta Mech. Sin. 24: 432
437 (in Chinese).
5. Xu,X.S., Zhong, W.X., Zhang, H.W.(1997). The SaintVenant problem and principle in elasticity, Int. J. Solids Struct.
34: 28152827.
670
671
1,2,3,4
1
Abstract. Investigation of curvature ductility of reinforced concrete members and the factors which affect the
ductility have been an important research field in civil engineering. Therefore, confined and unconfined concrete
models have been taken into consideration and analyses have been completed for concrete sections. In this study,
ductility values of 51 different sections which are under the sway of different parameters are calculated by using
moment-curvature diagrams by SEMAp analysis program. Furthermore, effects of different parameters on ductility
are given with details. Artificial neural network structure is generated to predict the ductility values which are
obtained by SEMAp. In this artificial neural network, 40 data for training and 11 data for testing have been chosen
to evaluate the results according to artificial neural networks analysis. Calculated and predicted ductility values have
been compared by each other and its seen that a strong relationship is established between them.
Keywords: artificial neural networks, ductility, moment-curvature diagram, reinforced concrete section
1 Introduction
Since, concrete is durable against pressure and can be formed easily, it is a widely used concrete material.
Unconfined concrete behavior shall be known to understand confined concrete behavior much better. Eventual unit
contraction value of unconfined concrete is so low that concrete materials can not reach enough ductility level.
Improved Hognestad Model and Mander Model are used frequently in the literature [1,2].
It is proved by studies that strength and ductility of confined concrete are considerably better than unconfined
concrete [3,4]. Confining is providing the lateral pressure on concrete core to increase the strength and ductility. With
this lateral pressure, more ductile behavior can be obtained by improving the cracks in the concrete core. Researches
about confined concrete models have started in 1903. Considere published a research about the effect of confining on
concrete columns with spiral reinforcement in 1903 [5]. Richart et al made many tests with concrete cylindrical
specimen which do not have any reinforcement under cylinder hydraulic compression in 1928 [6].Confined concrete
models are; Improved Kent&Park Model, Confined Mander Concrete Model, Saatiolu&Razvi Model and
Reinforcement Model [7,8].
Moment-curvature diagram for a cross-section envelope gives information about the changes in force capacity with
deformation in non-linear analysis. The relationship between moment-curvature indicates strength, ductility, energy
dissipation capacity and rigidity of the section.
Artificial neural networks (ANN) are computing systems that simulate the biological neural systems of the human
672
brain [9]. The prevalent type of artificial neural network consists of three or more layers, including an input layer, an
output layer and a number of hidden layers in which neurons are connected to each other with modifiable weighted
interconnections [10].
Neural Networks types are widely used for engineering problems [11,12]. The most common neural network
learning algorithm is the Back Propagation. This is due to its relatively simplicity and together with its universal
approximation capacity [13].
In this paper, after determining the ductility values for reinforced concrete sections, the values are compared with
several data points that are used in training a multi-layer, feed-forward and back propagation artificial neural network
(ANN) algorithm. Finally, the relationship between results is shown extensively in figures.
2 Methods
Sizes of the beam model which is taken into consideration to evaluate the design variables that can be seen in Fig. 1.
Effect of concrete compression strength, yield strength of tension reinforcement, tension and compression
reinforcement ratio on ductility is investigated. These variables are analysed by using an analysis program named
SEMAp [14].
673
Member
fc
2
(kg/cm )
fy
(kg/cm2)
C.R.
Stirrup
(uy)
(umax)
Ductility
A1
300
0.0026
4200
212
8/20
0.002193
0.056147
25.60
A2
200
0.0026
4200
212
8/20
0.002243
0.055943
24.94
A3
160
0.0026
4200
212
8/20
0.002361
0.056140
23.78
A4
300
0.0052
4200
212
8/20
0.002410
0.042642
17.69
A5
200
0.0052
4200
212
8/20
0.002120
0.034893
16.46
A6
160
0.0052
4200
212
8/20
0.002133
0.032751
15.35
A7
300
0.0065
4200
212
8/20
0.002333
0.032520
13.94
A8
200
0.0065
4200
212
8/20
0.002286
0.026405
11.55
A9
160
0.0065
4200
212
8/20
0.002192
0.024766
11.30
A10
300
0.0090
4200
212
8/20
0.002305
0.022064
9.57
A11
200
0.0090
4200
212
8/20
0.002336
0.017894
7.66
A12
160
0.0090
4200
212
8/20
0.002334
0.016707
7.16
Data of A1 member are defined by using SEMAp program and moment-curvature diagram that is shown in Fig. 2 is
obtained.
Yield values of the tension reinforcement are given in point B and the values at breaking time can be seen in point C.
Coordinates in brackets give rotation-moment values which are used in hinge definitions.
Ductility capacity of concrete sections is defined by curvature ductility coefficient as given in Eq. 1.
u
(1)
= max
uy
In the equation; , umax and uy represent coefficient of curvature ductility, curvature at breaking time and curvature at
yielding time respectively.
674
fc
(kg/cm2)
(kg/cm2)
B1
4200
B2
B3
B4
Member
C.R.
Stirrup
(uy)
(umax)
Ductility
200
212
0.0052
8/20
0.002120
0.034893
16.46
3600
200
212
0.0052
8/20
0.001825
0.032211
17.65
3000
200
212
0.0052
8/20
0.001522
0.027784
18.25
2200
200
212
0.0052
8/20
0.001549
0.028717
18.54
fc
C.R.
Stirrup
(uy)
(umax)
Ductility
4200
212
8/20
0.002243
0.055943
24.94
200
4200
212
8/20
0.002392
0.051264
21.43
200
4200
212
8/20
0.002120
0.034893
16.46
0.0065
200
4200
212
8/20
0.002286
0.026405
11.55
0.0200
200
4200
212
8/20
0.002708
0.006970
2.57
Member
C1
(kg/cm )
(kg/cm2)
0.0026
200
C2
0.0039
C3
0.0052
C4
C5
675
fc
Stirrup
(uy)
(umax)
Ductility
4200
8/20
0.002373
0.039153
16.50
200
4200
8/20
0.002301
0.043470
18.89
160
4200
8/20
0.002114
0.048596
22.99
314
300
4200
8/20
0.002170
0.056061
25.83
200
4200
8/20
0.002061
0.019850
9.63
0.0065
212
160
4200
8/20
0.002206
0.025073
11.37
D7
0.0065
216
300
4200
8/20
0.002041
0.043218
21.17
D8
0.0065
318
200
4200
8/20
0.002133
0.057015
26.73
Member
C.R.
D1
0.0039
D2
D3
(kg/cm )
(kg/cm2)
300
0.0039
210
0.0039
212
D4
0.0039
D5
0.0065
D6
D9
0.0086
160
4200
8/20
0.002375
0.014304
6.02
D10
0.0086
214
300
4200
8/20
0.002137
0.025806
12.08
D11
0.0086
218
200
4200
8/20
0.002204
0.026857
12.19
D12
0.0086
418
160
4200
8/20
0.002080
0.057154
27.48
fy
(kg/cm2)
(kg/cm2)
0.0026
300
10/20
0.0026
12/20
0.0026
E4
8/20
E5
10/20
E6
Member
Stirrup
C.R.
(uy)
(umax)
Ductility
E1
8/20
4200
212
0.002573
0.056922
22.12
E2
E3
200
4200
212
0.002445
0.049852
20.39
160
4200
212
0.002584
0.047593
18.42
0.0052
300
4200
212
0.002333
0.032520
13.94
0.0052
200
4200
212
0.002182
0.026042
11.93
12/20
0.0052
160
4200
212
0.002404
0.024441
10.17
E7
8/20
0.0065
300
4200
212
0.002083
0.023496
11.28
E8
10/20
0.0065
200
4200
212
0.002261
0.018738
8.29
E9
12/20
0.0065
160
4200
212
0.002546
0.017763
6.98
676
fc
C.R.
(uy)
(umax)
Ductility
4200
212
0.002266
0.055226
24.37
200
4200
212
0.002392
0.051264
21.43
160
4200
212
0.002413
0.057019
23.63
0.0065
300
4200
212
0.002841
0.054813
19.29
0.0065
200
4200
212
0.002285
0.026405
11.55
8/15
0.0065
160
4200
212
0.002286
0.029048
12.71
F7
8/10
0.0086
300
4200
212
0.002175
0.040036
18.41
F8
8/20
0.0086
200
4200
212
0.002459
0.021080
8.57
F9
8/15
0.0086
160
4200
212
0.002206
0.019051
8.64
Member
Stirrup
F1
8/10
F2
F3
(kg/cm )
(kg/cm2)
0.0039
300
8/20
0.0039
8/15
0.0039
F4
8/10
F5
8/20
F6
After calculating the ductility values of 51 sections according to moment-curvature diagrams, the values are also
evaluated by using artificial neural network analysis and as explained below.
3 Numerical Analysis
In this study, a multilayer feed forward neural network is used. The study is concerned with the prediction of the
ductility using ANN. The database includes 40 data for training and 11 data for testing the network. The most
convenient architecture of network was determined as 5-7-1 after trials. The network consists of input layer, single
hidden layer and one output layer. Four neurons in input layer, seven neurons in hidden layer and one neuron in output
layer take place in architecture of network. Input parameters of the network are; concrete compression strength, yield
strength of tension reinforcement, tension reinforcement ratio, compression reinforcement ratio and volumetric ratio of
stirrup. On the other hand, ductility is the output parameter of the network. Inputs and outputs are normalized in the (1:1) range by using simple normalization methods. Scaled Conjugate Gradient (SCG) is used as the learning algorithm.
2000 iterations are performed to find out the optimum result.
The performance of training and testing sets of the ANN are given in Fig. 3 and 4. As its seen in both of the figures,
calculated and predicted values are close each other.
677
References
1.
Hognestad, E. (1951). A study of combined bending and axial load in reinforced concrete members (Tehnical Report
Bulletin No. 399). University of Illinios Engineering Experiment Station.
2. Mander, J.B.; Priestley, M.J.N.; and Park, R. (1988). Theoretical stress-strain model for confined concrete. Journal of
Structural Engineering, ASCE 114(8): 1804-1825.
3. Aktan, A.E.; Karlsson, B.A.; and Sozen, M.A. (1973). Stress-strain relationship of reinforced concrete bars subjected to
large strain reversals. Civil Engineering Studies, Structural Research Series 397.
4. Ersoy, U; and Ozcebe, G. (1997). Moment-curvature relationship of confined concrete sections. In Proceedings of the
First Japan-Turkey Workshop on Earthquake Engineering. Turkey, March. stanbul: T Publication.
5. Considere, A. 1903. Experimental Researches on Reinforced Concrete. New York: McGraw Publishing Company.
6. Richart, F.E.; Brandtzaeg, A.; and Brown, A.L. (1928). A study of failure of concrete under combined compression
stresses (Technical Report Bulletin No. 185). University of Illinois Engineering Experimental Station
7. Kent, D.C.; and Park, R. (1973). Cyclic load behaviour of reinforcing steel. Strain 9(3): 98-103.
8. Saatcioglu, M.; and Razvi, S.R. (1992). Strength and ductility of confined concrete, Journal of Structural Engineering,
ASCE 118(6): 1590-1607.
9. Bishop, C.M. (1994). Neural networks and their applications. Review Scientific Instruments, American Institute of Physics
65(6): 1803-1832.
10. Lee, S.H.; Hong, H.K.; and Kim, M.W. (1992). Design rules of multi-layer perception. Science of Artificial Neutral Nets
in Structural Mechanics, Structural Optimization 1710: 329-339.
11. Karlk, B.; Ozkaya, E.; Aydn, S.; and Pakdemirli, M. (1998). Vibration of beammass systems using artificial neutral
networks. Computers & Structures 69(3): 339-347.
678
Biographies
R. Turul Erdem R. Turul Erdem was born in 1985 in zmir. He was graduated from Celal Bayar University in 2008. He
completed his graduated thesis in 2010. He has started working as a research assistant since 2010 in Celal Bayar University. He is
studying concrete structures, performance evaluation and artificial neural network analysis.
His has two scientific, two international and two national articles. He has four papers in international conferences.
Research Assistant R. Turul Erdem is a member of Turkish Chamber of Civil Engineers.
Soner eker Soner eker was born in Uak in 1983. He was graduated from Celal Bayar University in 2005. He is still a doctoral
student in Celal Bayar University. He has been a research assistant in Celal Bayar University since 2005. He is studying steel
structures and artificial neural network analysis.
He has one scientific and three published national articles. He also has 3 accepted papers in international conferences.
Research Assistant Soner eker is a member of Turkish Chamber of Civil Engineers.
Engin Gcyen Engin Gcyen was born in Manisa in 1983. He was graduated from Celal Bayar University in 2005. He is still a
doctoral student in Celal Bayar University. He has been a research assistant in Celal Bayar University for 5 years. He is studying
offshore structures and artificial neural network analysis
He has one international and one national published articles. He also has seven papers in national conferences.
Research Assistant Engin Gcyen is a member of Turkish Chamber of Civil Engineers.
Muhiddin Bac Muhiddin Bac was born in Ankara. He was graduated from Dokuz Eyll University in 1991. He completed his
graduated thesis in 1996 and his doctorate thesis in 1996. He is studying concrete structures, performance evaluation and finite
elements method.
He has seven pressed scientific articles, eighteen national articles. His five papers have been accepted and presented in national
conferences and six in international conference. He has two graduate and two doctorate students this year.
Assistant Professor Dr. Muhiddin Bac is a member of Chamber of Civil Engineers.
679
TOBB ETU, Mech. Eng. Dept., Stz Cad. No: 43 (06560) Ankara, Turkey
otopcu@etu.edu.tr, 2 ersanaslan@kku.edu.tr
Abstract.This study focuses on the integration of a virtual manufacturing (VM) lathe system with the Internet and
proposes a web-based and also a platform-independent VM system. A prototype web-based virtual lathe has been
developed using java and java3D application programming interface (API). The user can easily observe the
machining process on the internet with any kind of operating system by setting up machining conditions and
operating the virtual lathe machine in a web browser. Using the web-based lathe system, lecturers and students can
simulate machining process and can easily analyze the outcomes of the given conditions in this dynamic
environment.
1 Introduction
Utilization of the virtual reality (VR) technology improves learning by letting the user to experience the situations
before experiencing it in real life. In addition practicing under real conditions is expensive and is also dangerous in
some situations. It is always preferable to practice with a VR model instead of exhausting a real one. The aim of the
project is a virtual laboratory where lecturers and students can gain experience in a classroom or even at home before
studying with real machines. In order to accomplish a virtual world, which demands less computational power, Java
programming language, which is an object-oriented programming language and supports web-based application
development, is selected.
This work is produced to achieve more accessible manufacturing course materials and to let the user to experience
machining process by integrating three-dimensional (3D) computer graphics and real world machining process models
with VR methods. The prototype of the virtual lathe presented in this study provides an easy to develop online
platform and cost free way for students to experience machining process.
Virtual reality is an emerging reality, which has the potential to improve and change the methods in which the
students are educated and lecturers deliver the information, as discussed by Pantelidis [1]. VR technologies especially
help students and instructors by allowing them to study the process and to realize the actual consequences by the
immediate feedback of the system [2, 3]. Also VR technology have a great potential in aiding manufacturing planning,
system testing, process control and factory layout [4]. Manufacturing industries prefer well trained workforces but in
the same time they do not want to waste precious systems resources as well [5]. Virtual manufacturing (VM) is a field
of VR which exclusively deals with integration, computer-based representation, physical and logical demonstration of
680
a real manufacturing system [6]. After construction of a VM system, it is possible to use it for virtually operating a
CNC machine, verifying G-codes, avoiding collisions in machining, training operators and with appropriate code and
algorithms predicting tool life, etc. It is also possible with VM systems to monitor dangerous or remote environments
by associating it to a real system. A VM system can be presented in a Web-based CAM system to visualize the
machined workpiece by using any of the existing 3D modeling interfaces. Some of these 3D modeling interfaces are
virtual reality modeling language (VRML) structure, X3D and Java 3D. Web-based CAM systems are capable to
animatedly simulate machining processes in solid graphics from through the internet [7]. The main advantages of the
VM systems are the estimation and the optimization of manufacturing processes in advance without using any real
manufacturing facilities. In addition, high speed data transmitting among computers via networks or World Wide Web
and recent developments in the personal computer arena, has made it possible for common users to use these services
and it also accelerated the development of Web-based software tools [8, 9].
n=
Vc
.D
(1)
where Vc is the cutting speed and D is the diameter of the workpiece. In the start screen user enters the values within
the limits and the revolution speed is calculated afterwards.
The html code used to attach the virtual lathe application inside a web page is given below. The dimensions
represent the width and the height of the applet.
<P><APPLET codebase=../.. code="graduationThesis/helen/helena.class" width=652 height=524></APPLET></P>
681
Fig. 1. (a) The tool holder, the cutting tool and accessories and (b) the tool holder and the cutting tool as a single object
682
Fig. 2. The Virtual Universe object representation of the cutting tool and the tool holder
(a)
(b)
Fig. 3. (a) The 2D representation of the workpiece and the cutting tool during the machining process and afterwards and (b) related
point data of the procedure
683
The algorithm keeps track of the intersection points and then it replaces the workpieces points between these
intersection points with the cutting tools points when requested. After obtaining a set of boundary points for the new
workpiece, the algorithm updates 2D geometry of the workpiece. The process repeats itself until the application is
terminated. Fig. 4(a) shows the BG node of the raw workpiece which is a cylindrical geometry and the machined
workpiece with its five faces as geometric objects is shown in Fig. 4(b).
(a)
(b)
Fig. 4. (a) the BG node of the raw workpiece and (b) wireframe visualization of the machined workpiece
Fig. 5 (a) shows the start screen of the web-based lathe system, in which the user sets the process parameters. These
parameters with their upper and lower limits are cutting speed (Vc) (7-31 mm/s), feed (f) (0.05-2 mm), workpieces
length (l) (100-200 mm) and diameter of the workpiece (D) (10-120 mm). The view of the virtual lathe inside a
browser window is shown in Fig. 5(b). Fig. 5(c) shows the wireframe visualization of the workpiece, which is a feature
of the applet that enables the noticeability of the small depth of cuts. At the end of the machining process workpiece is
visualized as solid, as shown in Fig. 5(d).
Camera controls of the virtual lathe can be activated by clicking on the screen of the virtual lathe application with the
left mouse button. The keys, which are assigned to manipulate the viewing point of the camera, are w a s d or W A
S D. Also the user can alter the cameras position by pressing i j k l or I J K L keys. Java 3D provides mouse
navigation in tutorials but this method of control preferred to experience the keyboard input capabilities for future
improvements.
(a)
(b)
684
(c)
(d)
Fig. 5. Desktop screen captures of the VM system (a), the virtual lathe (b), during the machining process (c) and the end of the
machining process (d)
In this study, the positioning of the cutting tool is intentionally kept simple. This simplicity comes from the working
principles of step motors which are mainly used to control movements of a CNC machine. Also this simplicity allows
modifications of the code and additions for new features. The user inputs the replacement value of the cutting tool only
in one of the text boxes and then presses the activation button, which is KU button. The cutting tool only moves in XAxis and Y-Axis in both directions. The user must enter negative values to retreat the cutting tool away from the
workpiece.
5 Conclusion
This article addressed the importance of using web-based VR technology in manufacturing education where real
machine experience possibility is doubtful. The presented Web-based lathe system is capable of displaying workpiece
during the machining process and afterwards. Since the system is still at its early stages of development, the future
improvements can be addition of multiple materials, various cutting tools with tool holders, machining theories and
process knowledge and chip formations both in 3D environment and as data. The aim of the project is a virtual
laboratory but due to technical obstacles such as computational and programming demands of the virtual systems and
mainly lack of development time, the only presented application, which is a virtual lathe machine, facilitates only one
cutting tool at the current state.
References
1.
V. S. Pantelidis, (1997), Virtual reality and engineering education, Computer Applications in Engineering Education, 5(1), 3
12.
2. C. Rasi, E. Imperio, M. Sacco, (2000), The representation of virtual reality in education, Education and Iformation
Technologies, 5(4), 317-327.
3. A. Antonietti, E. Imperio, C. Rasi, M. Sacco, (2001), Virtual reality and hypermedia in learning to use a turning lathe, Journal
of Computer Assisted Learning, 17, 142-155.
4. Z. Xu, Z. Zhao, R. W. Baines, (2000), Constructing virtual environments for manufacturing simulation, 38(17), 4171-4191.
5. G. Chryssolouris, D. Mavrikios, D. Fragos, V. Karabatsou, K. Pistiolis, (2002), A novel virtual experimentation approach to
planning and training for manufacturing processes the virtual machine shop, Int. J. Computer Integrated Manufacturing, 15
(3), 214-221.
6. S. K. Ong, L. Jiang and Y. C. Nee, (2002), An internet-based virtual cnc milling system, Int. J. Adv. Manuf. Technol., 20, 2030.
7. Y. Seo, D.-Y. Kim, S.-H. Suh, (2006), Development of web-based cam system, Int. J. Adv. Manuf. Technol., 28, 101-108.
8. S-H. Suh, Y. Seo, S-M. Lee, T-H Choi, G-S. Jeong, D-Y. Kim, (2003), Modeling and implementation of internet-based virtual
machine, Advanced Manufacturing Technology, 21, 516-522.
9. K.I. Lee, S.D. Noh, (1997), Virtual manufacturing system--a test-bed of engineering activities, Annals of the CIRP, 46(1), 347350.
10. H. Sowizral, K. Rushforth, M. Deering, (1998), The java 3d api specification, Corporate & Professional Publishing Group,
Addision Wesley Longman, Inc. ISBN 0-201-32576-4.
685
Biographies
Okan Topu was born in Ankara, Turkey in 1980. He completed his primary and high school education in Ankara and
received his B.Sc. degree from the Department of Mechanical Engineering at Krkkale University in Krkkale, Turkey 2005. He is
currently working on a doctoral thesis which focuses on additive manufacturing processes in mechanical engineering department at
TOBB ETU, Ankara, Turkey.
He is the author of 3 scientific papers. His research focuses on micro devices, micro fabrication, precision engineering, optimal
design and control, advanced mechanisms, robotics, computer aided engineering, scientific computing, artificial intelligence,
powder technology and discrete element method.
Mr. Topu is a graduate student member of ASME.
Ersan Aslan was born in Bursa, Turkey in 1962. He received a BSc degree from Gazi University, Ankara, Turkey in 1983. He
carried out technical and pedagogy education under the patronage of World Bank and YOK, abroad. He also both received his MSc
and PhD degree in mechanical engineering from Gazi University, Ankara, Turkey. He has been working in the field of High Speed
Machining.
He is currently involved in many research projects related to CAD, CAM, High Speed Machining and Manufacturing Processes.
He is the author of more than 34 scientific papers.
Prof. Dr. Aslan is a referee of the Journals published by Gazi University, Pamukkale University, Dokuz Eylul University,
Karabuk University and Selcuk University.
686
Abstract. The determination procedures of cancellous bone properties are to compare with theoretical
outcomes and experimental outcomes. So, there are numerous alternatives for handling of image data and
setting up the finite element (FE) model that variation of computed outputs is unavoidable. Uncertainty of
results can cause to errors while mechanical properties or behavior of tissue are settled. In this paper, the
influence of Micro-CT slice thickness in connection with the reconstruction process on the modal behavior
and bone tissue properties were studied. Results have indicated that investigated variables have significant
influences on not only numerical values of the free vibration frequencies but also behaviors. Further, it has
been concluded that natural frequency values and mode shapes must be deliberated together for appropriately
understanding the vibration behaviour of bone tissue.
Keywords: image to FEM, slice thickness, modal analysis, voxel based mesh
1 Introduction
The assessment methods with vibrational implementations are applied today more than before. It is important
that the utilization of assessment methods including vibration procedures but also determination of the
vibrational behavior of bones for bone related problems.
Papers printed by Cornelissen et al. [1, 2] and Van der Perre and Lowet [3, 4] principally concerned the
assessment of free vibration frequencies and corresponding mode shapes of human bones, especially the tibia
and the radius. The outcomes are indicated that the free vibration frequencies of bone decline with enlarging
level of osteoporosis. Assessment of bone tissue properties has a crucial role in examination and healing of bone
disease. The function of free vibration frequency of bone for detection of bone fracture treatment was deliberated
by Singh et al. [5]. Mc Donnell et al. [6] studied with rapid prototype replicas manufactured based on Micro-CT
images for showing alterations in the mechanical properties due to considerable bone loss.
Throughout technological advancements, high-resolution images can be acquired by Micro-CT or Micro-MRI
systems in the order of a few microns per pixel. Micro-architectural components considered in the studies, and
detailed models can be attained by high-resolution image data. Ulrich et al. [7] used a Micro-CT scanner to
examination the relationship between micro-architectural items Partt [8] and mechanical properties of
cancellous bone. They investigated micro-architectural items as well as them into Micro Finite Element (FE)
models. The FE models were implemented to calculate orthotropic mechanical stiffness; implementation of
FE models were to guarantee that stiffness coefficients calculated was only due to the Micro-Architecture, and
not to trabecular bone properties and inherited experimental artifacts. Chevalier et al. [9] investigated that,
687
voxel-based FE models of trabecular bone are combined with physical properties of the volume fractions, MicroCT reconstructions, uniaxial mechanical tests and sample specic nanoindentation experiments for suitable
validation method.
FE estimations of apparent modulus were previously declared to be strongly affected by the threshold procedure
applied for segmentation of CT data to acquire the FE mesh Hara et al. [10], and are considerably sensitive to
errors due to the strong interaction between volume fraction and mechanical properties.
Literature review indicates that there is no any study addressed the influence of slice thickness on free vibration
properties of bone tissue. This study constituted the aim of assessing the importance of slice thickness in natural
vibration properties and behaviors of trabecular bone tissue. Additionally, it is examined that the influence of
slice thickness on model geometry and FE mesh.
Since the pixel width and pixel length are known, each pixel can be assumed to shape a basis of a brick with the
length equal to the slice thickness. Micro- images (37 m isotropic voxel size) were digitized according to the
midpoints of histogram peaks of bone intensities. Anisotropic voxels with length 74, 111, and 148 m were
calculated while keeping the base area 37x37 m2. FE models were acquired from the image dataset obtained
by Micro-CT by transforming the voxels (slice to slice) into a brick formed elements. Every voxel was divided
into five tetrahedra. To guarantee that neighboring cells match each other, the orientations of the tetrahedra
rotated Cutler [11]. The voxels acquired by threshold routine do not ensure interconnected bone tissue. 3D flood
fill script was coded in Matlab for assuring interconnected tissue model. Moreover, Flood fill algorithm takes
off islands from the reconstructed model. Islands can arise from noisy pictures or threshold mistakes and cause
examination downfalls. Only one variable slice thickness is used for taking off the impact of others.
Segmentation and reconstruction routines are applied in its simplest form for concentrating only influence of
slice thickness furthermore secondary algorithms like surface smoothing and others are not used in
investigations. In some stages of this study, software developed by the NIH/NCRR Center for Integrative
Biomedical Computing was used.
688
Modal analysis was carried out by the commercial FEA tool ABAQUS. Bone tissue model with the material
properties (Young modulus 17 GPa, Possion Ratio 0.32, Density 1800x106) and free boundary conditions was
analyzed. The model and the simulation outcomes were done by ABAQUS CAE that can display color plots of
the displacement contours.
19
5
(a)
2310
2280
2250
2220
83
68
6
2190
22
05
.
2160
2070
12
1
.1
6
148
84
81
09
80
.
70
63
81
.0
(b)
.0
81
111
Slice Thickness (m)
51
81.15
81.1
81.05
81
80.95
80.9
80.85
80.8
80.75
80.7
80.65
80.6
80.55
80.5
80.45
74
80
.
Volume (mm3)
37
21
39
2100
20
91
.1
7
2130
19
9
2340
23
75
.
2370
81
37
74
111
148
Fig. 2.Effect of slice step size on geometric properties: (a) area, (b) volume
Principal geometric properties of FE models are displayed in Fig. 2 for different slice thickness. Surface area of
the FE models decline while slice thickness increases and the volume of the models increased for declining of
slice thickness.
A variation for surface area between FE models with 37 m and 148 m slice thickness reaches % 11.96 while
the difference in volume is % 0.7 for the same slice thickness values. Obviously, slice thickness is more efficient
on surface area than volume of FE models. This issue can be crucial in problems related with fluid flow and
interaction problems where surface area more dominant than volume. Studied geometrical properties of
reconstructed models indicated that the effect of slice thickness on surface area, and volume is unpredictable and
irregular.
The volume and area of bone tissue do not alter linearly for constant incrementation of slice thickness contrary to
prismatic shapes with a constant cross section. To acquire bone tissue from Micro-CT images threshold
procedure was used. Flood fill algorithm used for taking off islands and ensure interconnections of bone tissue.
689
Voxel Size
Voxel Size
37x37x37
37x37x74
37x37x111
37x37x148
20464.3
20949.9
17483.2
18884.5
24303.3
24536.3
20989.8
20406.4
26773.4
27230.4
21859.6
20519.1
30208.4
31117.2
23194.1
22409.2
31101.1
32448
23497.3
23275.1
32592.6
33573.9
26695.4
24950.7
32994.1
34825.5
27518.4
25763.6
33933
35194.9
30988.8
26461.9
34613.2
36314
31719.9
27701.8
10
35875.8
38346.1
32361.6
28576.2
Table 1 displays free vibration frequencies of reconstructed models for various slice thickness. For enlarging
slice thickness, constant number of modes took place in a reducing range of frequency. Briefly there is no
obvious association between free vibration frequencies of models with different meshes, hence, the more
comprehensive table below created to see modal behavior of modes by observing mode shapes and numbers
together.
Table 2.Mode shapes and maximum displacement regions of models for different voxel size
37x37x37
37x37x148
Mode 1-20464.3 Hz
Mode 1-18884.5
690
Mode 2-24303.3
Mode 2-20406.4
Mode 3-26773.4
Mode 3-20519.1
In Table 2, natural frequencies and associated mod shapes with displacement contours are displayed together.
Maximum displacement field of every mod shape of the bone tissue marked with ring.
The maximum displacement field of every mode shape is a moderately small field for showed modes in Table 2.
Above the maximum displacement field, displacements are an nearly constant, and behaves like a support. This
issue can be an important for the case of osteoclasis and micro implementations with vibration.
It can be observed from the Table 2, there is no obvious connection between mode shapes and mode orders for
every model. Each model has its individual mode order, and number of frequencies is different also for same
mode types. One of the significant outcomes acquired from this work is that the free vibration frequency values
and mod shapes must be considered together for free vibration behavior of models that obtained from sets with
different slice thickness is astonishingly different from each other. These differences contain not only free
vibration frequencies but also mode types and sequences.
Additionally, influence of modal behavior irregularity is vibration behaviors of models imponderable by using
results of other models obtained from datasets with different slice thickness.
Primary effect of slice thickness variation is the element size alteration but the principal changes took place in
model geometry. Influence of segmentation techniques on geometric components is crucial when the amount of
lost details is increased due to utilization of large slice thickness. This matter is particularly important in the case
of fatal bone loss when weak connections come out. Influence of e size on FE techniques is well known in
established literature but this interesting outcome cannot be clarified by only the variation of element size.
The variations of element size and model geometry simultaneously united effect cause irregular modal behavior,
contrary to the situation of element size variation in customary FE techniques. In the application of FE method,
the variation of element size influences the convergence level, but it does not change trends generally. The
variation of slice thickness accompanied by segmentation techniques can cause not only the alteration of element
size but also ignorance of weakly connected parts or construction of parts, which is not existent in real. The
united effects of variation of slice thickness and segmentation method led to irregular trends and convergence
691
levels. Finally, the outcomes revealed that 3D solid FE models must be reconstructed by using Micro-CT images
with using smallest slice thickness.
4 Conclusions
This paper was clarified the central importance of slice thickness in free vibration attributes of cancellous bone.
The most obvious finding rising from this work is the natural vibration behavior of bone tissue models with
varied slice thickness is surprisingly different from each other in terms of not only the values of natural
frequencies but also modal behavior.
One of the more consequential results of this study is that alterations of element size and model geometry
simultaneously causes inconsistent modal behavior. The results of this study have a number of significant
implications for in the process of reconstruction and analysis in future practice.
The present study was particularly formed to assess the influence of slice thickness on natural vibration attributes
connected to voxel size. Thus, the results might not be transferable to virtual models obtained by utilizing
different solid model evaluation routines. This study has thrown up numerous questions in need of further
investigation. Taken together, these outcomes imply that 3D model reconstruction must be done shortest slice
thickness and model behavior consideration must be incorporated both free vibration frequency values and mode
shapes.
References
1.
Cornelissen, M.; Cornelissen P.; Van der Perre, G.; Christensen, A.B.; Ammitzboll, F.; and Dyrbye, C. (1986).
Assessment of tibial stiffness by vibration testing in situ III. identication of mode shapes in different supporting
conditions. Journal of Biomechanics 20: 333342.
2. Cornelissen, M.; Cornelissen P.; Van der Perre, G.; Christensen, A.B.; Ammitzboll, F.; and Dyrbye, C. (1986).
Assessment of tibial stiffness by vibration testing in situ II. inuence of soft tissues, joints and fibula, Journal of
Biomechanics 19: 551561.
3. Van der Perre, G.; and Lowet G. (1994). Vibration, sonic and ultrasonic wave propagation analysis for the
detection of osteoporosis. Clinical Rheumatology 13: 4553.
4. Van der Perre, G.; and Lowet G. (1996). In-vivo assessment of bone mechanical properties by vibration and
ultrasonic wave propagation analysis. Bone 18: 29 35.
5. Singh, V.R.; Yadav, S.; and Adya, V.P. (1989). Role of natural frequency of bone as a guide for detection
of bone fracture healing, Journal of Biomechanical Engineering 11: 457-461.
6. Mc Donnell, P.; Liebschner, M.A.K.; Tawackoli, W.; and Mc Hugh, P.E. (2009). vibrational testing of
trabecular bone architectures using rapid prototype models. Medical Engineering & Physics 31: 108-115.
7. Ulrich, D.; Van Rietbergen, B.; Laib, A.; and Ruegsegger, P. (1999). The ability of three-dimensional structural
indices to reect mechanical aspects of trabecular bone. Bone 25: 5560.
8. Partt, A.M. (1988). Bone histomorphometry: Standardization of nomenclature, symbols and units. Bone 9: 6769.
9. Chevalier, Y.; Pahr, D.; Allmer, H.; Charlebois, M.; and Zysset, P. (2007). Validation of a voxel-based FE Method
for prediction of the uniaxial apparent modulus of human trabecular bone using macroscopic mechanical tests and
nanoindentation. Journal of Biomechanics 40: 33333340.
10. Hara, T.; Tanck, E.; Homminga, J.; and Huiskes, R. (2002). The inuence of microcomputed tomography
threshold variations on the assessment of structural and mechanical trabecular bone properties. Bone 31: 107109.
11. Cutler, B.; Dorsey, J.; Mcmillan, L.; Mller, M.; and Jagnow R. (2002). A procedural approach to authoring solid
models. ACM Transactions on Graphics 3: 302311.
Biographies
Gkhan Altnta Gkhan Altnta was born in 1971 in zmir. He was graduated from Dokuz Eyll University in 1992. He
completed his master thesis in 1996 and his doctorate thesis in 2002. He started working as a research assistant in 1993 in
Celal Bayar University. He has been working as an associate professor since 2010. His expertise subjects are finite difference
method, finite elements method, vibration and mechanics of composite materials, image processing, three dimensional image
to solid reconstruction.
He has eleven international and four national articles. He has one published papers in international and two papers in
national conferences. He also has one completed and two continuing scientific research projects.
Associate Professor Dr. Gkhan ALTINTA is a member of Chamber of Civil Engineers.
692
selim.solmaz@gediz.edu.tr, 2scaglarb@hacettepe.edu.tr
Abstract. In this paper we present a method for decentralized control design for systems with multiple
dynamical modes, which guarantees robustness with respect to structural uncertainty. We consider the
implementation of this method to the decentralized control designs for the automotive lateral and roll
reference dynamics tracking. The respective control rules that we utilize are based on simplified, 2-state roll
and lateral dynamics models of the vehicle. We utilize a method for checking the overall stability of the
integrated controllers based on a frequency domain criterion. We also give a numerical example for the
integrated automotive tracking control designs based on PI feedback, which utilize active suspension and
active steering actuators. Finally, we show how this result can preserve robustness with respect to sensor
failure in such applications.
Keywords: Decentralized Control, Stability Theory, Diagonal Lyapunov Functions, Vehicle Emulation, Integrated
Chassis Control, Active Vehicle Safety
1 Introduction
It is an irrefutable fact that decentralized control is a feature of the control engineering practice. A basic problem
in the design of control systems is the lack of simple methods for designing decentralized controllers that are
robust with respect to certain types of structural uncertainties. Most complex engineering systems can be
described by a number of interacting dynamical modes and sub-processes. When control of such systems are
considered, the engineering practice is to utilize a number of control systems, each of which are designed to
control a particular sub-process or a dynamical mode. Such designs arise for a number of reasons. Firstly, it is
human nature to divide a complicated problem into a set of simpler ones, each of which can be solved
independently of each other. Second, even when a centralized controller is possible, in practice, a low orderdecentralized controller is often preferred due to their simplicity and ease of implementation. Third, when
networked control systems are considered, it is often the case that interconnected subsystems must be designed
to be individually stable, even when interconnected together, so that they are robust to the effects of unreliable
communication between each of the subsystems. Finally, from the perspective of the industry, it is often the case
that different sub-contractors are assigned separate parts of a complex control task; the automotive industry
provides a very good example of this latter point. In terms of vehicle dynamics control applications, it is
common in the automotive industry that some parts suppliers design and manufacture the roll over prevention
systems, and some other manufacturers supply the lateral dynamics control systems [1] for the same vehicle.
Even though each of these control systems affect one another [2], and even sometimes these utilize common
693
sensor and actuator units, they are often designed independently of each other. Also, it is a common practice in
the automotive industry to utilize sensor measurements to artificially decouple dynamical interactions of the
vehicle, and the control task at hand then becomes the development of methods for the integration of these units
so that the overall performance objectives, and also a certain degree of robustness are met with respect to
unreliable sensor measurements.
Motivated by these facts, our objective in this paper is to present one such method and to explore how some of
the methods developed in this paper can be integrated with other automotive control systems in a structured
manner. To aid the exposition, we utilize decentralized controllers based on the two-state single track model and
the two-state roll plane model to simultaneously track reference vehicle states. Therefore, here we discuss about
how to integrate a controller for the roll degree of freedom, with that for the lateral degree of freedom in a
vehicle equipped with both active suspension and active front wheel steering actuators, while at the same time
achieving a decentralized design that is not restricted by the structural constraints and uncertainties imposed on
the problem.
Let be a Hurwitz stable matrix, which represents the overall closed loop system matrix of a dynamical
process with two interacting constituent subsystems, which have the following structure
(1)
(2)
(3)
where
and
. In other words we want to stabilize each of the constituent
subsystems, and the entire interconnected system, simultaneously. If such a block diagonal, positive matrix
P=PT>0exists, then not only is the system stable, but also it is stable with respect to structural uncertainties in
A12 and A21, which in effect, can be considered as robustness with respect to system failure. The choice of a
diagonal P also simplifies the analysis. Despite having a long history in control theory, this problem, namely the
problem of finding vector Lyapunov functions, remains open to this day. Fortunately, in our situation, one may
find feedback strategies to ensure a block diagonal Lyapunov function as described in the sequel.
Let be the matrix of all zeros except for the last m rows and m columns which are set to be the
elements of mxm identity matrix, and which we denote by
, such that B matrix is expressed as
June 1 -4, Kusadasi, Aydin, Turkey http://iscse2011.gediz.edu.tr
694
follows
(4)
Further we denote , as the scalar variable in the frequency domain. Suppose now that one can find a
such that
symmetric matrix
(5)
(6)
is SPR.
695
control gains are obtained by LQR (Linear Quadratic Regulator) design techniques.
The well known two-state single track model is the simplest model that represents the lateral dynamics of a car
in the horizontal plane, where the effects of heave, roll, and pitch motions are all ignored. It is also assumed that
only the front tire is used for steering the vehicle, and that the steering angle is small. In this model, we represent
the horizontal dynamics in terms of the state variables and , that is the sideslip angle and the yaw rate,
respectively. Both of these states are assumed to be small for linearization. The corresponding state space
representation of the model with active steering input is given as follows
(8)
where , is the driver steering command, and u1 is the steering command from the active steering actuator. Also,
(9)
For further description of the parameters of the model, refer to [2] and [4].
The purpose of the control design we consider here is to follow a reference yaw rate trajectory (It is possible to
consider the tracking of both the yaw rate , and the sideslip angle given that there is more control authority
such as the differential braking and/or active front and rear wheel steering. However, we assume neither of these
actuators in this discussion). This choice is motivated by the fact that the yaw rate , along with the lateral
acceleration ay are responsible for most of the lateral handling feel, (i.e., the lateral response) of an automotive
vehicle, and therefore are the natural targets for emulation. We can achieve this by introducing a new integrator
state xI as a function of the yaw rate tracking error.
(10)
where !"# denotes the reference yaw rate trajectory, which is a linear function of the driver steering input $% .
Thus, the scalar is the steady-state gain from the driver steering input $\delta_d$ to the reference yaw rate !"#
for the vehicle that we want to emulate. Further, we can define an augmented state x1 given as below.
(11)
(12)
Now in order to track the reference yaw rate trajectory, it is possible to design a PI linear state feedback rule
696
such that the closed loop system is stable, and also that the quadratic cost function
(14)
where * + +, is the state, ayis the lateral acceleration input, and u2 is the roll torque input generated by the active
suspension actuator. Also, Jxeq denotes the equivalent roll moment of inertia defined as Jxeq = Jxx+mh2
The purpose of the control design we consider here is to follow a reference roll angle trajectory. In order to achieve this we
propose the following lead-lag control structure
(16)
where Kr1, Kr2 and Kr3 are the lead-lag gains, e=-ref is the roll angle tracking error, and ref is the reference roll trajectory.
Now we can define an augmented state x2 given as below
(17)
which results in the following augmented feedback system description
697
(18)
(19)
The characteristic polynomial corresponding to the closed loop system matrix A22 is
(20)
Denoting the closed loop eigenvalues (i.e., the target poles) as 1, 2 and 3, the target characteristic polynomial of the closed
loop system can be expressed as
(21)
Comparing (21) with (22) we obtain the following fixed control gains in terms of the target poles and the elements of the
system matrix
(22)
Now taking the Laplace transform of the controller (16) we can express the resulting lead-lag controller in frequency domain
as follows
where u2(s) and e(s) are the Laplace transforms of the control input, and the tracking error, respectively. As a
final remark we note that the design parameters for this tracking controller are the target poles 1, 2 and 3; in
order for the closed loop system to be stable these are set be negative real. In what follows, we implement the
integrated control methodology based on the decentralized control designs introduced in the current and the
preceding subsections.
3.3 Robust Integration of Controllers
In this subsection we give an implementation of the design methodology described in Proposition (2.1) for
simultaneous, and structurally robust emulation of the reference vehicle states related to the lateral and the roll
dynamics. In doing so, we utilize the simple decentralized control structures introduced in the preceding two
subsections and show that the resulting controller integration approach is robust with respect to certain types of
structural uncertainties.
For the controller integration we consider the four-state single track model with the roll degree of freedom to
698
represent the real vehicle. This choice is motivated by the fact that the four-state single track model is the
simplest model that considers the interactions of both the roll and the lateral dynamics, and thus it is an ideal
choice for exposing the control integration idea. We utilize the version of the model with the states -
.// , and the model assumes control inputs from both the active front wheel steering actuator and the
active suspension actuator. Then, the linearized equations of motion corresponding to this model with two inputs
can be expressed follows
(23)
(24)
(25)
(26)
where u1 represents the mechatronic-angle-superposition type active steer input, and u2 represents the active
suspension roll torque input. See [4] for a detailed derivation of this model.
The task of the integrated controller considered here is to follow reference roll angle ref and reference yaw rate
ref trajectories corresponding to a different vehicle. In our numerical studies we simulate this scenario by
utilizing two four-state single track models, each with a different parametrization; we shall refer to the vehicle
that we want to emulate as the reference vehicle, which generates the reference trajectories, and the other vehicle
is referred to as the controlled vehicle. In Table 1 we give the numerical values for the model parameters
corresponding to each vehicle. Figure 1 below shows a gradual step steering input d of the the driver that is
applied to both vehicles at a constant speed of vx = 20 m/s (where a constant steering ratio of 1/20 assumed for
both); this results in the dynamical responses shown on Figure 2, which clearly indicates that both vehicles have
distinct dynamical characteristics.
Now in order to implement the controller integration procedure described in Proposition 2.1 we need to express
the full state feedback system as a function of the control gains. To do so, we first introduce a new state x
representing the full controlled vehicle states as defined below
699
(27)
Figure 2. Comparison of the lateral and the roll dynamics responses of the reference and the uncontrolled vehicles
Next we substitute u1 from (13) and u2 from (16) in the four-state vehicle model given in (23)-(26), which results in the
following closed loop state space system describing the vehicle dynamics
(28)
where
(29)
(30)
700
(31)
(32)
In Figure 3 below the integrated control structure and the corresponding closed loop feedback system is shown schematically.
Now we are in a position to numerically implement the control design as outlined in Proposition 2.1. First, we
start with designing the roll tracking controller based on Section 3.2. Recall that the target poles (eigenvalues)
are the only design parameters for the roll controller. Without loss of generality we set 1=-40, 2=-50 and 3=60 as the target poles; then from (22) the lead-lag control gains are calculated as
For this numerical solution, without loss of generality, we set Q22 = 10 I3, where I3 denotes the identity matrix. This solution
results in
(33)
Next we design the control gains for the lateral PI tracking controller described in Section 3.1, according to the
item (b) of Proposition 2.1. In order to do so, we first utilized the Matlab control system toolbox to compute the
LQR controller gains described in 13), which minimizes the quadratic cost function (14). Again, without loss of
generality, we assumed a diagonal structure for the weighting matrices Q and R in conjunction with the LQR
solution (note here that, as the lateral tracking controller is based on a single input, then R is a scalar). We used
the following values for these;
(34)
Then one can numerically show that condition (7) satisfied. We note that the resulting matrix pair (12,B) is controllable and
the pair (12,BP) is observable. In Figure 4, we show the variation of the eigenvalues of 03 4, 56 7 89: 7 12 4 &
701
4= demonstrating SPR condition in a section of the frequency domain for varying 6 , where
In order to test the integrated controller, we considered an obstacle avoidance maneuver scenario conducted by
the driver. The vehicle speed was assumed to be fixed at vx= 20 m/s during the course of the maneuver and the
driver steering input is shown in 5. The resulting dynamical responses of the reference and the controlled
vehicles is shown in Figure 6, where we observe a good agreement in the reference and the controlled states.
Note here that the lateral acceleration in simulations was computed using the following relationship
Figure 6 Comparison of the lateral and the roll dynamics of the reference and the controlledvehicles
702
(35)
was a
Based on the simulation results, we observe that the decision of following a yaw rate reference >?@
reasonable one, as this also resulted in good tracking results for the lateral acceleration of the reference vehicle.
Considering the fact that the lateral acceleration is one of the most important variables for the lateral dynamics
response, the effectiveness of the controller is evident.
References
1.
2.
3.
4.
Van Zanten, A., (2002) Evolution of electronic control systems for improving the vehicle dynamic behavior, In
Proceedings International Symposium on Advanced Vehicle Control (AVEC02), Hiroshima, Japan, 2002.
Solmaz, S., (2007) Topics in Automotive Rollover Prevention: Robust and Adaptive Switching Strategies for
Estimation and Control, PhD Thesis, NUI-Maynooth, Co. Kildare, Ireland,
Khalil, H. K (2001), Nonlinear Systems, 3rd Ed., Prentice Hall.
Kiencke, U. and Nielsen L, (2005) ,Automotive Control Systems: for Engine, Driveline and Vehicle, 2nd Ed.,
Springer.
Biographies
Dr. Selim Solmaz Dr. Solmaz was born in Izmir, Turkey in 1978. He received his BSc (High Honours) degree in
aeronautical engineering from Middle East Technical University, Ankara, Turkey in 2001 and MSc from Purdue University
at West Lafayette, School of Aeronautics and Astronautics in 2003. Dr. Solmaz received his PhD in 2007 from Hamilton
Institute, National University of Ireland-Maynooth. As part of his PhD project he spent a year as a visiting researcher at
Daimler Chrysler Research AG in Esslingen, Germany in 2003-2004. Upon his return to Ireland he conducted research work
on theoretical and applied problems arising from automotive vehicle dynamics estimation and control.
Dr. Solmaz was employed as a Postdoctoral Research Fellow at the Hamilton Institute, NUIM for during 2007-2010. During
this time he secured and completed two independent EU research grants (co-funded by Enterprise Ireland) with him as the
Principal Investigator (PI).
Dr. Solmaz is currently with Gediz University, Mechnanical Engineering Department, where he is employed as an Assistant
Professor. His current research interest includes vehicle dynamics control, vehicle state and parameter estimation, stability
theory, switched systems, mechatronics, robotics, renewable energy generation, and human-machine interfaces. Dr. Solmaz
has authored and co-authored about 30 peer reviewed journal and conference articles, 2 research project reports and 2 patents.
He is a past member of IEEE and AIAA.
Dr S. alar Balaml Dr. Balaml was born in Izmir, Turkey in 1977. He received his BSc (High Honours) degree in
mechanical engineering from Middle East Technical University, Ankara, Turkey in 1999 and MSc from the same department
in 2001. Dr. Balaml received his PhD in 2007 from the department of mechanical engineering at Boazii University,
stanbul Turkey. His PhD research consisted of the solution of vehicle dynamics control problems, especially in the gain
scheduled LPV control framework.
Dr. Balaml is currently with Hacettepe University, Department of Mechanical Engineering, where he is employed as an
Assistant Professor. His current research interest includes vehicle dynamics control, vehicle state and parameter estimation,
stability theory, switched systems, and human-machine interfaces. Dr. Balaml has authored and co-authored about 30 peer
reviewed journal and conference articles. He is involved in a number of industrial consultancy activities related to the design
to heavy construction vehicles and electrical vehicles.
703
selim.solmaz@gediz.edu.tr, 2t_coskun@ug.bilkent.edu.tr
Abstract. In this paper we describe the use of a cheap RC (remote control) model car as a vehicle Dynamics
analysis platform. We then implement two safety control applications based on this platform; namely an ABS
(Antilock Brake System) system using open-loop pulse brake control and a roll over prevention system
utilizing lateral acceleration feedback. Both applications are presented with sensor measurements.
Keywords: vehicle dynamics, active safety, feedback control, rollover prevention, brake control, scale
models.
1 Introduction
Analysis of vehicle dynamics is an important aspect of road vehicle design. It helps in understanding the vehicle behavior
under changing road and vehicle operating conditions. Once a safety critical condition is identified, it is either remedied by a
design modification or by employing an active feedback safety control system. Design modifications usually end up with a
tradeoff between vehicle safety and comfort; therefore, use of active safety systems is usually preferred. Active safety
systems typically sense the adverse dynamical condition with on-board inertial sensors and then utilize actuators such as
active steering, active suspension and differential braking to mitigate the problem. See [1] for a review and [2] and [3] for
examples of such vehicle dynamics control applications.
In modern road vehicles it is a common practice to have many active safety control systems as part of the standard vehicle
configuration. Some examples of these are ABS (Anti-lock Braking System), ESC (Electronic Stability Control) [4] and ROP
(Roll Over Prevention) [2],[3],[5] systems. Development and tuning of such active safety control systems are usually done on
accurate dynamical simulators and real pre-production vehicles, both of which are a great source of expense. In this paper we
describe the development of a cheap RC (remote control) model car as a HIL (Hardware-In-Loop) vehicle dynamics analysis
platform. Such a platform can be used as a low-cost rapid prototyping system for advanced active vehicle safety system
development. We demonstrate implementation of two safety control systems on such a platform, namely an ABS system
using an open-loop pulse brake control system and a roll over prevention system utilizing lateral acceleration feedback.
Throughout the paper we describe all the development stages of the vehicle platform including the design of interface
between the RC car and computer, integration of an inertial wireless motion sensor SHAKE (Sensing Hardware Accessory
for Kinesthetic Expression) SK6 (see http://code.google.com/p/shake-drivers/) on the vehicle and coding the feedback control
system using Matlab Simulink. We demonstrate the efficacy of our experimental setup with real measurement data.
704
Fig.2. Labjack U3-LV I/O board (on the left) and Shake SK6 sensor (on the right).
705
706
707
708
Fig. 6. Throttle signal and longitudinal acceleration for pulse and full brakes.
The pulse brake system is observed to have a better performance than the full brake both in terms of the duration and the
distance it takes to stop the car. The full brake takes about 1.0sec to bring the car to a full stop while the pulse brake takes
about 0.8sec to the same (see Fig. 6). Moreover it was observed that the car takes up to 1.5m less distance to stop with the
pulse brake switched on. It is also noted from Fig. 6 that the peak accelerations measured was lower during pulse brake test
than the full brake one.
4.2
Rollover Prevention System based on Active Steering Compensation and Utilizing Lateral
Acceleration Feedback
The second application we implemented on the car was a rollover prevention control system based on active steering
control inputs and utilizing lateral acceleration thresholds, which aims to prevent possible vehicle turnovers. The system
monitors the lateral acceleration (ay) measurements in real time and decides whether to limit the steering inputs or not,
depending on the thresholds we specify. If the lateral acceleration value is greater than |0.3*g| then the steering input is
limited at either 1V or 2V depending on the turning direction.
To test the system, we designed a J-turn maneuver scenario. During the test, the car goes straight for 2 seconds, then
makes a right turn for 0.3 seconds and finally makes a full left turn. The aim is to give a steering input signal which will be
able to roll the car over. We implemented two experiments to show the efficiency of the applied rollover prevention control
system and the lateral and vertical acceleration data and the steering input data were recorded during both experiments.
709
Fig. 7. The steering signal and the acc data for controlled and uncontrolled systems.
710
References
1.
2.
3.
4.
5.
6.
Gustafsson F. (2009). Automative safety systems: replacing costly sensors with software algorithms. IEEE Signal
Processing Magazine 26(4): 32-47.
Solmaz S., Corless M. and Shorten R., (2006). A methodology for the design of robust rollover prevention
controllers for automotive vehicles: Part 1-Differential Braking. In Proc. 45th IEEE Conference on Decision and
Control. San Diego, 13-15 Dec.
Solmaz S., Corless M. and Shorten R. (2007). A methodology for the design of robust rollover prevention
controllers for automotive vehicles: Part 2-Active Steering. In Proc. American Control Conference. New York, 1113 July.
Anton Van Zanten (2002). Evolution of electronic control systems for improving the vehicle dynamic behavior. In
International Symposium on Advanced Vehicle Control (AVEC02). Hiroshima, Japan.
Palkovics L., Semsey A. and Gerum E. (1999).Rollover Prevention System for Commercial Vehicles - Additional
Sensorless Function of the Electronic Brake System. Vehicle System Dynamics, 32(4-5): 285-297.
Ackermann J. and Odenthal D. (1998). Robust steering control for active rollover avoidance of vehicles with
elevated center of gravity. In Proceedings of International Conference on Advances in Vehicle Control and Safety.
Amiens, France, July.
Biographies
Dr. Selim Solmaz Dr. Solmaz was born in Izmir, Turkey in 1978. He received his BSc (High Honours) degree in
aeronautical engineering from Middle East Technical University, Ankara, Turkey in 2001 and MSc from Purdue University
at West Lafayette, School of Aeronautics and Astronautics in 2003. Dr. Solmaz received his PhD in 2007 from Hamilton
Institute, National University of Ireland-Maynooth. As part of his PhD project he spent a year as a visiting researcher at
Daimler Chrysler Research AG in Esslingen, Germany in 2003-2004. Upon his return to Ireland he conducted research work
on theoretical and applied problems arising from automotive vehicle dynamics estimation and control.
Dr. Solmaz was employed as a Postdoctoral Research Fellow at the Hamilton Institute, NUIM for during 2007-2010. During
this time he secured and completed two independent EU research grants (co-funded by Enterprise Ireland) with him as the
Principal Investigator (PI).
Dr. Solmaz is currently with Gediz University, Mechnanical Engineering Department, where he is employed as an Assistant
Professor. His current research interest includes vehicle dynamics control, vehicle state and parameter estimation, stability
theory, switched systems, mechatronics, robotics, renewable energy generation, and human-machine interfaces. Dr. Solmaz
has authored and co-authored about 30 peer reviewed journal and conference articles, 2 research project reports and 2 patents.
He is a past member of IEEE and AIAA.
Trker Cokun Mr. Cokun was born in Kardzhali, Bulgaria in 1988. He received his BSc degree (High Honours) in
electrical and electronics engineering in Bilkent University, Ankara, Turkey in 2011. He received full scholarship from
Bilkent University between 2007 and 2011 for his undergraduate studies. He attended the SIAR research internship program
in 2010 and he worked at the Hamilton Institute, NUIM as a researcher under the supervision of Dr. Selim Solmaz. He
received a scholarship from SFI (Science Foundation Ireland) under the SIAR grant. Mr. Cokun is a member of IEEE.
711
Abstract.In the present study, performance evaluation of three different metaheuristic techniques is carried
out in finding the solution of optimization problems taken from the literature. Three metaheuristic techniques
selected are the classical harmony search algorithm, improved harmony search algorithm and particle swarm
optimizer. Metaheuristic search techniques of optimization are non-deterministic methods and they rely on
heuristics in finding the better solutions in the search space. They use random or probabilistic parameters
while they search for the optimum solution rather than deterministic quantities. The source of random
variables may be several depending on the nature and the type of problem. The heuristics behind these
innovative techniques is borrowed from the nature or physics. Himmelblaus nonlinear optimization problem
and the welded beam design problem taken from the literature are solved by using the metaheuristic
techniques mentioned and their performance is evaluated and compared.
Keywords: Metaheuristic search techniques, harmony search algorithm, particle swarm method, optimization
problems
1 Introduction
A vast amount of techniques are now available in the optimization literature to deal with various types of
problems encountered in computational structural mechanics [1]. The long-standing research on computational
efficiencies of these techniques has clearly evinced that performance of any technique is affected by a high
number of factors, such as dimensionality, multi-modality, differentiability, continuity, etc. and that each
technique has some weaknesses apart from its strengths in searching for the optimum solution. For example, the
well-known mathematical programming techniques, which have long been used in engineering optimization
applications, are characterized by their rapid convergence rates. Nevertheless, poor convergence reliability to the
global optimum associated with a gradient based search is the major drawbacks of these techniques [2].
Recently, a group of novel techniques that make use of nature as a source of inspiration to develop numerical
search algorithms have emerged as robust and versatile methods for discrete and global optimization [3]. These
metaheuristic search algorithms come up with an increased promise of finding the global optimum due to their
lack of dependence on gradient information during the search. Heuristic methods are quite suitable and powerful
for obtaining the solution of engineering optimization problems. These methods do not require the derivatives of
the objective function and constraints [4]. Many engineering design problems can be formulated as constrained
optimization problems. The presence of constraints may significantly affect the performances of any
optimization algorithms that may work well for unconstrained problems. With the increase of the research and
712
applications based on stochastic search techniques [5,6], constraint handling used in metaheuristic computation
techniques has been a hot topic in both academic and engineering fields [7-11].
is
subjected
to
gi(x)0,
i=1,2,,p
Here, f(x) is the objective function, x denotes the decision solution vector, Nd is the number of decision
variables, Lxk and Uxk, are the lower and the upper bound of each decision variable, respectively. p is the
number of inequality constraints and m is the number of equality constraints.
Step 1: Initialization. Initialize an array of particles with random positions and their associated
velocities.
Step 2: Local best updating. Evaluate the fitness function of the particles and update local best
position Pikaccording to the best current value of the fitness function.
Step 3: Global best updating. Determine the current global minimum fitness value among the current
positions and update Pgk, the global best position.
Step 4: Solution construction. Change the velocities and move each particle to the new position
considering the related velocity.
Step 5: Terminating criterion controlling. Repeat Steps 24 until a terminating criterion is satisfied.
713
Step 1. Initialization. Initialize the optimization operators of HS algorithm includes the harmony
memory (hm), the harmony memory size (hms), the harmony memory considering rate (hmcr), and the
pitch adjusting rate (par).
Step 2.Harmony memory matrixis initialized. Each row of harmony memory matrix contains the
values of design variables which are randomly selected feasible solutions from the design pool for that
particular design variable.
Step 3.New harmony memory matrix is improvised.Generate a new harmony vector from the hm,
based on memory considerations, pitch adjustments, and randomization.
(14)
where, par(i) is the pitch adjusting rate for generation i, parmin is the minimum adjusting rate, parmax is the
maximum adjusting rate, and i is the generation number. The degree is updated according to the following
expression:
degree ( i ) =
( HCost ( i ) HCost
max
( HCost ( i )
max
mean
(15)
HCost min ( i ) )
where, HCostmax(i) and HCostmin(i) are the maximum and minimum function objective values in generation
i, respectively; HCostmean is the mean of objective function value of the harmony memory.The improvisation of
hmcr is carried out using the following expression;
(16)
where, hmcr(i) is the harmony memory considering rate for generation i, hmcrmax is the maximum
considering rate, hmcrmin is the minimum considering rate, and i is the generation number.
Step 4.Harmony Memory matrix is updated. If a new harmony vector is better than the worst harmony
in the hm, judging in terms of the objective function value, the new harmony is included in the hm and
the existing worst harmony is excluded from the hm.
714
5 Design Examples
and
which is subjected to
side
constraints
are
where;
g1 ( x) = 85.334407 + 0.0056858 x2 x5 + 0.0006262 x1 x4 0.0022053 x3 x5
g 2 ( x) = 80.51249 + 0.0071317 x2 x5 + 0.0029955 x1 x2 + 0.0021813 x32
g3 ( x ) = 9.300961 + 0.0047026 x3 x5 + 0.0012547 x1 x3 + 0.0019085 x3 x4
Himmelblau [24] first solved this problem by using the generalized reduced gradient (GRG) method. Then it is
studied by Gen and Cheng [25] using genetic algorithms (GAs). Runarsson and Yao [26] proposed an
evolutionary strategies (ES) algorithm with stochastic ranking for the solution of this problem.
Table 2.Optimal results for Himmelblaus nonlinear optimization problem.
Optimum solutions obtained by different metaheuristic methods
Design
variables
IHSO
PSO
HSO
ES
GRG
GAs
x1
78.0000
78.0000
78.0000
78.0000
78.6200
81.4900
x2
33.2777
33.0003
33.0000
33.0000
33.4400
34.0900
x3
27.3060
29.9962
29.9950
29.9953
31.0700
31.2400
x4
44.9988
44.9999
45.0000
45.0000
44.1800
42.2000
x5
44.2466
36.7734
36.7760
36.7758
35.2200
34.3700
g1(x)
91.9544
92.0000
91.7147
92.0000
91.7927
91.7819
g2(x)
100.4151
98.8402
98.8405
98.8405
98.8929
99.3188
g3(x)
20.0000
20.0000
19.9999
20.0000
20.1316
20.0604
-31004.19
-30665.40
-30665.50
-30665.54
-30373.95
-30183.58
f(x)
The optimal solution is obtained at x = (78.0000, 33.2777, 27.3060, 44.9988, 44.2466) with corresponding
function value equal to f(x) = -31004.19 using IHSO as demonstrated in Table 1. The comparisons of the best
solutions of this example obtained using the proposed metaheuristic search techniques with previous best
solutions reported in [24-26].
715
0
g 4 ( x ) = 0 . 10471 x 1 + 0 .04811 x 3 x 4 (14 . 0 + x 2 ) 5 0
5
1
,
and
as side constraints,
Minimize
g 6 ( x) = ( x) max 0
as end deflection of the beam, and g 7 ( x ) = P Pc ( x ) 0 as buckling load on the bar. And
also, design variables of the problems are limited as;
0 .1 x1 2 .0,
0 .1 x 2 1 0, 0 .1 x 3 1 0,
0 .1 x 4 2 .0
where;
' =
P
, '' = M R , M = P ( L + x2 ) , R =
J
2
2 x1 x2
( x) =
4.013E
4 PL
6 PL ,
, ( x) =
3
E x3 x4
x4 x32
x 2 x + x
x2
x +x
+ ( 1 3 ) 2 , J = 2 x1 x2 2 2 + ( 1 3 )2 ,
4
2
2
12
Pc ( x) =
( x32 x46 )
36 (1 x3 E )
2
L
2L 4 G
The same problem was also solved by Ragsdell and Philips [28] using geometric programming (GP). Deb [29]
used a simple genetic algorithm (GAs) with traditional penalty function to solve the same problem. Ray and
Liew solved this problem using a society and civilization algorithm (SCA) [30]. Also same problem has solved
by using proposed IHSO, HSO, and PSO algorithms. The optimum solutions and comparison of results for this
problem are tabulated in Table 2.
716
IHSO
PSO
HSO
SCA
GP
GAs
x1 (h)
0.18461
0.23886
0.25381
0.2444
0.2455
0.2489
x2 (l)
3.95998
2.5296
3.07156
6.2379
6.1960
6.1730
x3 (t)
9.10543
9.1796
7.82838
8.2885
8.2730
8.1789
x4 (b)
0.20626
0.2389
0.27413
0.2445
0.2455
0.2533
f(x)
1.77192
1.90340
1.98118
2.3854
2.3859
2.4331
6 Conclusions
The highly promising outcome of this research suggests that the proposed metaheuristic algorithms, which are
classical harmony search algorithm, improved harmony search algorithm, and particle swarm optimizer are
effective alternative for solving engineering optimization problems. Especially the new optimization approach,
improved harmony search, can also be extended to other real-world optimization problems in manufacturing and
design area. Himmelblaus nonlinear optimization problem and the welded beam design problem taken from the
literature are solved by using the metaheuristic techniques mentioned and their performance is evaluated and
compared. The design examples considered to demonstrate the effectiveness and robustness of the improvements
suggested in the optimization techniques in this study.
References
[1]
[2]
[3]
[4]
[5]
[6]
[7]
[8]
[9]
[10]
[11]
[12]
[13]
Arora, J.S. (2002). Methods for Discrete Variable Structural Optimization. Recent Advances in Optimum Structural
Design, ASCE, Ed. S. A. Burns, 1-40, USA.
Sun, W and Y-X, Yuan (2006). Optimization Theory and Methods; Nonlinear Programming, Springer-Verlag.
Saka, M.P. (2007). Optimum Design of Steel Frames using Stochastic Search Techniques Based on Natural
Phenomena: A Review. Civil Engineering Computations: Tools and Techniques, Ed. B.H.V. Topping, Saxe-Coburgh
Publications, pp: 105-147.
Kaveh, A. and Talatahari, S. (2009). Particle swarm optimizer, ant colony strategy and harmony search scheme
hybridized for optimization of truss structures. Computers andStructures, 5-6, 87, 267-83.
Michalewicz, Z. (1995). A survey of constraint handling techniques in evolutionary computation methods. In:
McDonnell, J.R. et al. (Eds.), Proceedings of the fourth Annual Conference on Evolutionary Programming, MIT
Press, Cambridge, MA, pp. 135155.
Coello, C.A.C. (2002). Theoretical and numerical constraint handling techniques used with evolutionary algorithms:
a survey of the state of the art. Computer Methods in Applied Mechanics and Engineering, 191 (11/12), 12451287.
Saka, M.P. and Erdal, F. (2009).Harmony Search Based Algorithm for The Optimum Design of Grillage Systems To
LRFD-AISC. Structural and Multidisciplinary Optimization, 38, 1, 25-41.
Saka, M.P. (2007). Optimum Geometry Design of Geodesic Domes Using Harmony search Algorithm. Advances in
Structural Engineering, 10, 6, 595-606.
arba, S. and Saka, M.P. (2009). Optimum Design of Single Layer Network Domes Using Harmony Search Method.
Asian Journal of Civil Engineering, 10, 1, 97-112.
Hasanebi, O., arba, S., Doan, E., Erdal, F., and Saka, M.P. (2009). Performance evaluation of metaheuristic
search techniques in the optimum design of real size pin jointed structures. Computers and Structures, 87, 5-6, 284302.
Hasanebi, O., arba, S., Doan, E., Erdal, F., and Saka, M.P. (2010). Comparison of non-deterministic search
techniques in the optimum design of real size steel frames.Computers and Structures, 88, 17-18, 1033-1048.
Eberhart R.C. and Kennedy, J. (1995). A new optimizer using particle swarm theory. In proceedings of the sixth
international symposium on micro machine and human science, Nagoya,Japan.
Kennedy, J., Eberhart, R.C., and Shi, Y. (2001), Swarm intelligence. San Francisco (CA), Morgan Kaufman
Publishers.
717
Kaveh, A. and Talatahari, S. (2009). Engineering Optimization with Hybrid Particle Swarm and Ant Colony
Optimization, Asian Journal of Civil Engineering, 10, 6, 611-628.
Lee, K.S. and Geem, Z.W. (2004). A New Structural Optimization Method Based on theHarmony Search Algorithm.
Computers and Structures, 87, 5-6, 284-302.
Mahdavi, M., Fesanghary, M., and Damangir, E. (2007). An improved harmony search algorithm for solving
optimization problems. Applied Mathematics and Computation 188(2):15671579.
Taherinejad, N. (2009). Highly reliable harmony search algorithm. European Conference on Circuit Theory and
Design (ECCTD09), Antalya, Turkey, August 23-27, 818-822.
Omran, M.G.H. and Mahdavi, M. (2008). Global-best harmony search. Applied Mathematics and Computation,
198(2):643656.
Geem, Z.W. (2009). Particle-Swarm Harmony Search for Water Network Design. Engineering Optimization,
41(4):297-311.
Geem, Z.W. and Roper, W.E. (2010). Various continuous harmony search algorithms for web-based hydrologic
parameter optimization. Mathematical Modelling and Numerical Optimisation, 1(3):231-226.
Hasanebi, O., Erdal, F., and Saka, M.P. (2010). An Adaptive Harmony Search Method for Structural Optimization.
Journal of Structural Engineering, ASCE, 136, 4,419-431.
arba, S. and Saka, M.P. (2011). Optimum Topology Design of Various Geometrically Nonlinear Latticed Domes
Using Improved Harmony Search Method. Structural and Multidisciplinary Optimization, (under review).
Coelho, L.S. and Bernert, D.L.A. (2009). An improved harmony search algorithm for synchronization of discrete-time
chaotic systems Chaos. Solitons and Fractals, 41(5):2526-2532.
Himmelblau, D.M. (1972). Applied Nonlinear Programming. Mc-Graw Hill, Daryaganj, New Delhi, India.
Gen, M., and Cheng, R. (1997). Genetic Algorithms and Engineering Design. John Wiley and Sons, NY, USA.
Runarsson, T.P. and Yao, X. (2000). Stochastic ranking for constrained evolutionary optimization. Institute of
Electrical and Electronics Engineers International Transactions on Evolutionary Computation, 4, 284-294.
Doan, E. (2010). Optimum Design of Rigid and Semi-Rigid Steel Sway Frames Including Soil-Structure Interation.
PhD Dissertation, Middle East Technical University, Ankara, Turkey.
Ragsdell, K.M. and Phillips, D.T. (1976). Optimal design of a class of welded structures using geometric
programming. American Society of Mechanical Engineering Journal of Engineering for Industries,98, 1021-1025.
Deb, K. (1991). Optimal design of a welded beam via genetic algorithms. American Institute of Aeronautics and
Astronautics Journal, 29, 2013-2015.
Ray, T. and Liew, K.M. (2003). Society and civilization: An optimization algorithm based on the simulation of social
behavior. Institute of Electrical and Electronics Engineers International Transactions on Evolutionary Computation,
7, 386-396.
Biographies
Serdar arba- Mr. arba was born in Erzurum, Turkey. He received his BSc degree in civil engineering department from
Atatrk University, in 2002, Erzurum, Turkey. His MSc degree is from Engineering Sciences Department of Middle East
Technical University in 2008, Ankara, Turkey. Also, he is currently continuing his PhD studies in same department.
He has carried out researches on shape, size and topology optimization of latticed domes such as lamella, network, and
geodesic domes. He has several proceedings of international and national conferences on optimum shape and topology design
of latticed domes as well as optimization of steel structures. Besides, he has five international journal publication related with
structural optimization using recent metaheuristic search techniques. His research interests focuses on the optimum design of
steel structures, structural optimization of geometrically nonlinear structures such as steel domes, size, shape and topology
optimization of pin and rigidly connected steel structure systems, optimization algorithms, optimum design of cold-formed
steel sections. His present studies concern the use of meta-heuristic optimizations techniques in structural optimization
problems.
Mr. arba is a member of Chamber of Civil Engineers of The Union of Turkish Engineers and Architects.
Erkan Doan- Dr. Doan was born in Erzurum, Turkey. He received his BSc degree in civil engineering department from
Atatrk University, in 2002, Erzurum, Turkey. He is awarded his doctoral degree from Engineering Sciences Department of
Middle East Technical University in 2010, Ankara, Turkey.
718
Ferhat Erdal- Dr. Erdal was born in Gaziantep, Turkey. He graduated from Akdeniz University, in 2003, Antalya, Turkey.
He undertook his both MSc and PhD degrees from Engineering Sciences Department of Middle East Technical University
(METU), in Ankara, Turkey.
His early studies involve optimum design of grillage systems and effect of beam spacing on these systems. And then, he
concentrated on cellular beams. Furthermore, his researches also involve with the optimum design of real-size steel trusses
and frames. He has eight international, two national journal paper and six studies printed in national and international
proceedings. He professionally deals with stochastic search techniques and its applications on structural systems. Lately, he
has carried out an experimental research on cellular beams. He has developed an optimum design algorithm for cellular
beams.
Dr. Erdal got involved in some international courses and seminars. He was in charge of mechanics laboratory of Engineering
Sciences Department in METU.
Mehmet Polat Saka- Prof. Saka was born in Trabzon, Turkey. He graduated from stanbul Technical University in 1969. He
had his PhD degree from Civil Engineering Department of University of Aston in Birmingham, United Kingdom. He is a
professor of structural engineering in the department of civil engineering at The University of Bahrain, in Bahrain. Prior to
his current post, he worked as a professor in the Department of Engineering Sciences at the Middle East Technical
University, in Ankara, Turkey.
He has been carrying out research on structural optimization since he has attained his PhD degree. Since then he has
published 6 chapters in books printed by international publishers, 52 articles in the international journals and 63 papers in the
proceedings of international conferences on structural optimization. His work covers size, shape and topology optimization of
linear and nonlinear skeletal structures as well as evolutionary structural optimization of 3-D continuous structures. He is one
of the early researchers who applied structural optimization to geometrically nonlinear structures. Furthermore, some of his
early works covers shape and topology optimization of pin and rigidly connected steel structures. In his recent publications
he has carried out shape and topology optimization of various braced domes such as geodesic, lamella and network domes
where the geometrical nonlinear behaviors of these domes are taken into consideration. His current research covers the use of
metaheuristic optimizations techniques in structural optimization.
He has membership of so many societies related with his professions and he is in editorial boards of 6 international journals.
719
Abstract. Perforated armor plates are in use for many years to improve ballistic protection level of armored
vehicles. Theoretically, it induces bending stress mainly to break incoming projectile core or at least to divert
it from its incident trajectory and thus substantially reduce its residual penetration capability through the basic
armor. This could have an application in design of light weight armors since a broken or diverted projectile is
considerably easier to defeat. Although, there has been experimental studies related with ballistic
performance of perforated plates, in open literature there is not any numerical simulation based study. In this
paper, the ballistic resistance of perforated plates made of high hardness armor steel was investigated against
7,62x54 armor piercing steel core using ballistic tests and simulations. A 3D numerical model of the
projectile and target was developed using LS-DYNA to investigate transient affects, such as influence of
perforated holes on projectile bending stress. Numerical results correlated with tests.
Keywords: Ballistic simulation, high hardness armor, perforated plates, Johnson-Cook material model
Introduction
In civil and military applications, ballistic penetration has become prime importance in development of armour
solutions and continues to be a challenging research field for engineers. High velocity impact and penetration
problems include large deformation, erosion and high strain rate dependent nonlinear material behaviour. Due to
complexity of the problem three approaches are become quite popular in investigation of ballistic penetration.
These are test based empirical formulations, analytical and numerical methods.
Experimental investigation of ballistic penetration in development of armour and ammunition also requires quite
complicated test and data analysis procedures, therefore new methods in experimental studies are still in
progress. With the tests conducted in laboratory environment, numerous empirical formulations are derived and
used in solution of problems [1]. Although, its difficulty in derivation, analytical models are useful from the view
of their direct applicability. Since the qualitative laws determined by the use of these models can be considered
as basis for the theoretical and experimental investigations, there are many recent studies on this field [2].
Multiple physical phenomena are involved during penetration such as fracture, failure, residual stresses and
friction heating. In reality since empirical and analytic approaches cannot capture all of these phenomena,
720
numerical simulation has become a necessary tool for the study of ballistic penetration. Numerical methods and
corresponding computing technologies have been recently evolved to the level where mentioned complex
deformation and penetration pattern during ballistic impacts can be accurately predicted.
Perforated armour plates are in use for many years in armour systems where they installed at some stand-off
distance from vehicles basic armour. The use of those plates as add-on armour has always been very attractive to
improve ballistic protection level of armoured vehicles. Perforated plates induce bending stress mainly to break
incoming projectile core or at least to divert it from its incident trajectory and thus substantially reduce its
residual penetration capability through the basic armour. If the bending stresses are sufficiently large, the
projectile will fracture. This could have an application in design of light weight armours since a broken projectile
is considerably easier to defeat. The total weight of such an armour system is considerably less than weight of
basic armour with same antiballistic capability. Usually the holes or slits are designed to interact with the
diameter of the threat and performance effectiveness of such armour under impact is dominated by plate material,
thickness and orientation of holes. Although, there has been some patents [3,4,5] and experimental studies related
with ballistic performance of perforated plates, in open literature there is not any simulation based systematic
study to develop a solution to a specific threat. Balos and his colleagues have investigated geometry and
mechanical properties of different hardness perforated steel plates for ballistic application with tests [6]. A
numerical study has been carried out by Chochron and his co-workers [7] to investigate bending affect when
threat impacts an edge. Borvik and Buchar have numerically studied ballistic penetration for high strength steels
and showed that simulations can predict behaviour. [8,9,10]
Numerical Modeling
A review of the literature on impact simulations shows that the most research in this field has focused on the
development and application of continuum hydrocodes using either Eulerian, Lagrangian or Arbitrary
Lagrangian Eulerian (ALE) formulations.
In Euler solver, the numerical mesh is fixed in space and physical material flows through the mesh [14]. This
formulation is generally used to represent large deformations and flow problems. Free surfaces of material
interfaces can move with fixed mesh, which also needs to model void mesh. In order to describe solid behaviour,
solid stress tensor and history of the material must be transported from cell to cell, therefore has less applicability
in ballistic penetration In the Lagrange solver, the numerical mesh moves and distorts with the physical material.
This formulation is widely used because of its advantages, such as low level of computational cost [16], being
able to track accurately and efficiently material interfaces and incorporate complex material models. . In the ALE
solver, solver allows for a type of automatic rezoning which can be quite useful when Lagrangian mesh structure
starts to distort [15].With the recent developments, explicit codes became a reliable tool providing accurate and
stable results for large deformation problems and successfully implemented in many commercial hydrocodes.
2.1
Geometric Model
The geometric model consist of 6 mm perforated plate in front of 9 mm basic armor and Stanag 4569 level 3
7.62x54R AP ammunition. In numerical simulations, for the sake of simplicity only hardened steel core is used.
Simulation performed on full bullet model and simplified hardened core model shows that the penetration on
target is almost identical [11,12]. A similar outcome has been developed by Borvik and his co-workers both with
experimental and numerical studies.[20] In his study, the brass jacket and the lead cap, has removed and the hard
core steel projectile was inserted in a sabot which has ignorable weight. By removing these parts the mass of the
bullet was reduced by 50%, which means that the initial kinetic energy was reduced accordingly. It was found
that when only the hard core was used as projectile the ballistic limit dropped by 35%. This observation
721
indicates that it is conservative to only model the hard core in analytical and FE simulations of 7.62 mm AP
bullet impacting steel targets.
Fig. 1. Geometric cross-section of AP ammunition.For the AP bullet, the ogival nose hardened steel core inserted in a brass
sabot, before the jacket is clamped onto it. In front of the core, a cap of lead-antimony is placed to stabilize the projectile
during flight and in the initial stage of penetration.
Fig. 2. Geometric view of simulation model. The perforated plate consists of 6 mm holes and distance between holes is 10
mm. The distance between perforated and basic plate is 30 mm.
The total mass of AP bullet is about 10.04 g. The muzzle velocity of the ammunition is 854 m/s as specified in
Stanag 4569. The simulation will use the Lagrange method for hardened steel core projectile Lagrange
discretization for steel armor target. High hardness hardened steels are commonly used in ballistics as well as
commercial purposes. In this study, such a material, Secure 500 was used both as basic armour and perforated
plate. Ballistic limit of S500 is given by supplier to be approximately 16 mm against Nato Stanag 4569 level 3
threats.
2.2
The Johnson-Cook model is used for target and projectile as it is well suited to the model metals that are subject
to high strain rate loading. The yield stress at non zero strain rate depends on the strain hardening, strain rate
hardening and temperature softening such that [13]
(1)
& *
Where ~p is the equivalent plastic strain, p is the dimensionless plastic strain rate, TH is homologous
temperature { TH=(T-Troom)/(Tmelting-Troom) }. The five material constants are A, B, C, n and m. The
& *
expression in the first set of brackets gives the stress as a function of p =1 and TH =0. The expression in the
second and third sets of brackets represent the affects of strain rate and thermal softening respectively. Fracture
in the JC material model is based on a cumulative damage law:
722
D=
in which
(2)
(3)
where ~* is the ratio of hydrostatic stress per effective stress. The constants D are experimental constants.
* =
( x + y + z ) / 3
H
=
2
2
2
eq
x + y + z x y y z z x + 3( xy2 + yz2 + zx2 )
(4)
Failure of the elements assumed to occur when D=1. The failure strain and thus the accumulation of damage is a
function of mean stress, strain rate and temperature. The failed elements are removed from the FE model. The JC
material model was used in conjuction with the Mie-Gruneisen equation of state model. The armour material
used in ballistic test is alloyed, liquid-quenched and tempered high-strength special steel for civil use. The
measured hardness values and mechanical properties suit the Armox 500 specified by Skoglund and coworkers.
The JC parameters A = 1470 MPa, B = 702 MPa, n = 0.199, C = 0.00549 and finally m = 0.811 are used in
simulations [17]. The failure parameters D1=0,068, D2=5,382, D3=-2,554 are also taken from Skoglund and
coworkers and compared with tensile test results (Figure 3).
Fig.3. Static tensile test data for armour material. Horizontal axis shows true strain and vertical axis corresponding stress
value in Mpa.
In simulations it is assumed that triaxiality has a larger influence on the strain to fracture than the strain rate in
the investigated loading range, therefore D4=0. Thermal softening affects are also ignored D5=0. For
determining J-C material strength model for bullet, Dynamic Test & Modeling Laboratory at Izmir Institute of
Technology (IYTE) performed experiments. SHPB is a mechanical test instrument used to characterize the
dynamic response of materials at high strain rates (Figure 4).
723
Fig.4. YTE Split-Hopkinson Pressure Bar Apparatus. A gas gun launches a striker bar that impacted on the end of the
incident bar. A stress wave is generated that travels down the bar and is re-corded sequentially by the first and second strain
gages mounted longitudinally on the bar.
The quasi-static stress-strain response was measured for the steel core material. Tensile specimens were
machined, instrumented with strain gauges and strained until failure. The test results are plotted in Figure 5. It
was found that the core material fails approximately 4% strain at a stress of 2.1 GPa. However, there is variation
in strain to failure and associated ultimate stress. Material strength model for bullet has compared with Buchar
and co-workers. [14] The material model is close to Buchar results.
5000
3000
-3
Deformayon hz:1x10 s
-1
y = 2300*(1+m1*ln(M0/1e-3))
Value
Error
0.0049495 0.0018687
87879
NA
0.56621
NA
2500
m1
Chisq
R
4000
Gerilme (MPa)
2000
y = 1900 + 1100* M0^ 0.3
Value Error
1500
Chisq
1.4429e+7
NA
0.7996
NA
1000
3000
2000
1000
500
0
0
0.05
0.1
0.15
0.2
0
0.0001
0.001
0.01
0.1
10
100
1000
10
Deformasyon hz (1/s)
c
-1
Fig.5. Static and dynamic test data for bullet material. In (a) for low strain rates, (0,1 s strain rate) the plastic deformation
versus true stress values are shown. Since material hardness is quite high (880 Vickers), in order to achieve plastic strain
notched specimens were used. In (b) plastic strain versus stress values are shown for 1x10-3 s-1. The blue line is calculated,
the red line is test results. In (c) the stress values are shown for various strain rates.
The relation between pressure, volume and internal energy of a material is defined as the equation of state. In
this numerical modeling, Mie-Gruneisen and linear polynomial equations of state are used. [18]
0C 2 (1 + (1 2) ( 2) 2 )
P=
+ ( + )E0
2
1 (S1 1) S2 2 / ( + 1) S3 3 / (1 + )
E0
S1 , S 2 , S3
are the coefficients of the slope of the shock and particle velocity curve,
coefficient.
Where
(5)
is the internal energy per unit volume, C is the intercept of the shock and particle velocity curve,
is
the Gruneisen
EOS data for the target plate is taken from Reference 19 as 4340 material. For the AP bullet, the linear
polinomial EOS is used.
P=K
(6)
724
of more than 1,3 million elements. The contact during the penetration of bullet into target is achieved using
CONTACT_NODES_TO_SURFACE.
For perforated plates, numerical simulations have been carried out and results are compared with the
experimental data. Ballistic tests were performed in the ballistics test laboratory in Otokar (Figure 6). During the
tests, with aluminum witness plate, the fragmentation and spall behaviour of target is investigated. Both
perforated and basic plates were 300x300 in dimension and placed with 30 mm gap. The 8 mm holes on 6 mm
thickness perforated were machined by laser cut. The distance between centers of holes was 10 mm.
Fig.6. Ballistic test laboratory and target set-up.During the test, projectile speed was measured with sensors placed between
gun and target and achieved with +/-5 m/s precision.
Test results are as expected quite successful. With single plate ballistic limit thickness of Secure 500 against
7,62x54 AP projectile is 16 mm. Six shots were fired and the penetration depth on basic armour measured. In all
cases, perforated plate worked well and mass areal density has decreased by 31 kg/m2. The most regular
deformation pattern on perforated plates is shown in Figure 7, only in one case, the projectile has centered the
hole but even for this case, perforation induced a bending therefore the basic plate not passed.
In numerical simulations, (Figure 8-a) due to stochastic nature of the ballistic, 4 different hitting positions were
investigated. Examination of computational results showed that the bending affect of perforated plate on
projectile is successfully modeled.
725
Fig.8. Damaged area on perforated plate and basic armor. (a) In case 1, the bullet hit the geometric middle of perforated
holes, in case 2 the tip of the projectile hits the side of hole, in case 3 projectile centered the hole and in case 4 hits the middle
of two hole centers. (b-c) results for case 2. (c) results for case 3. (d) results for case 1
As shown in figure 8-b&c, after hitting on perforated plate, due to edge affect, the plate induces a bending force
on the projectile to change the incident angle. Meanwhile the tip of the projectile eroded which decreases the
penetration capability. The penetration depth in simulations is over estimated (4 mm penetration in simulation
versus 2,1 mm in tests). On the other hand the damage patterns both in basic armour and perforated plate is quite
reasonable. Although in tests, some of the projectiles were broken into two or more pieces; in numerical
simulation only tip erosion was achieved, which may due to weakness of failure model used. Although in tests
when the projectile and hole is overlapped (case 3), 9 mm basic plate has stopped the threat, in numerical
simulations failed. (Figure 8-d)
In summary, it is found that the LS-DYNA code and developed model is able to predict the behaviour of
perforated plates and good agreement was obtained between numerical simulations and experimental results. For
the future work, the perforation size and distance between holes will be calculated for various thicknesses to
optimize areal density.
References
1.
2.
3.
4.
5.
6.
7.
8.
Bangash M.Y.H.(2009). Shock, Impact and Explosion Structural Analysis and Design Springer.
Ben-Dor G., Dubinsky A., Elperin T.(2005) Ballistic Impact: Recent Advances in Analytical Modelling of Plate
Penetration Dynamics. ASME Applied Mechanics Reviews Nov 2005 Vol.58 pp 355-371.
Patent No: WO 21010/036411 A2 Perforated Armor With Geometry Modified for Lighter Weight 29 May 2008
Patent No: US 2006/0213360 A1 Perforated Armor Plates, 28 Sep 2006
Patent No: 5,014,593 Perforated Plate Armor 14 May 1991
Balos S., Grabulov V., Sidjanin L. Pantic M., Radisavljevic I.(2009). Geometry, Mechanical Properties and
Mounting of Perforated Plates for Ballistic Application. Materials and Design. Dec 2009 Vol 31 pp.:2916-2924.
Chocron S., Anderson C.E., Grosch D.J., Popelar C.H.(2000). Impact of the 7,62 mm APM2 Projectile Against the
Edge of a Metallic Target. Inernational Journal of Impact Engineering Oct 2000Vol 25, pp.:423-437.
Buchar J., Voldrich J., Rolc S., Lisy J.(2002). Ballistic Performance of Dual Hardness Armor. In 20th International
Symposium on Ballistics. Orlando, 23-27 Sep 2002
726
10.
11.
12.
13.
14.
15.
16.
17.
18.
19.
20.
Borvik T., Hopperstad O.S., Berstad T., Langseth M.(2002). Perforation of 12 mm Thick Steel Plates by 20 mm
Diameter Projectiles with Flat, Hemispherical and Conical Noses Part II: Numerical Simulation. International
Journal of Impact Engineering 2002Vol 7, pp37-64.
Borvik T., Dey S.,Hopperstad O.S., Langseth M. (2006). On the Influence of Fracture Criterion in Projectile Impact
of Steel Plates. Computational Material Science 2006 Vol 38, pp176-191
Deniz T., Kilic N., Yldrm R.O.(2009). Ballistic performance of monolithic RHA armor plates against 7,62 mm
AP. In USMOS 2009 Conference. Ankara Turkiye
Kilic N., Erdik A, Ekici B.(2010). Simulation of Impact and Fragmentation with Meshless Methods In ESDA 2010
Conference. 12-14 July 2010, Istanbul, Turkey
Johnson G.R., Cook W.H., (1983). A Constitutive Model and Data for Metals Subjected to Large Strains, High
Strain Rates and High Temperatures. In Proceedings of the 7th International Symposium on Ballistics. Hague
Holland, April 1983
Buyuk M., Kan S.C.D.,Bedevi N.E., Durmus A.,Ulku S.(2002). Comparison Between the FEM and Meshless
Methods for a Ballistic Impact Simulation, In Proceedings of 8 th International LS-DYNA Users Conference.
Donea J., Huerta A., Phontot J., Ferran A.R(2004). Arbitrary Lagrangian Eulerian Methods. Encyclopedia of
Computational Mechanics Vol1, 2004, John Wiley&Sons
Zukas J. A., Nicholas T., Swift H. F., Greszczuk L. B., Curan D. R.; Impact Dynamics, J. Wiley, New York, 1982
Skoglund P., Nilsson M., Tjernberg A.(2006). Fracture Modeling of High Performance Armour Steel. J.of Physic
IV France Vol 134, 2006, pp197-202
Livermore Software Technology Corporation (LSTC), LS-Dyna Keyword Users Manual, May 2007, Version 971
Li J., Li X.J., Zhao Z., Ou Y.X., Jiang D.A.(2007). Simulation on Projectile with High Rotating Speed Penetration
into the Moving Vehicular Door. Theoretical and Applied Fracture Mechanics 2007Vol 47, pp113-119
T. Brvik , S. Deya, A.H. Clausen (2009) Perforation resistance of five different high-strength steel plates
subjected to small-arms projectiles. International Journal of Impact Engineering Vol 36. pp: 948964
Biographies
Namik Kilic He was born on May 5, 1970 in Agri, Turkey. He obtained a Bachelor of Science degree in Mechanical
Engineering from Bogazici University, in 1993. In 1997, he obtained Master of Science degree in Mechanical Engineering
from Marmara University. Form 1994 to July 1997, he worked as a research engineer in Marmara University Engineering
Faculty. He is a PhD candidate from Marmara University.
Namiks professional interests include computer aided simulation, vehicle engineering, blast and ballistics. Now, he holds
CAE manager position in Otokar.
Namik Kilic is a member of ASME and Automotive Technologies Platform OTEP.
Yilmaz Erbil He was born in skenderun in 1978. He earned his Bachelor Science and Master of Science degree from the
Mechanical Engineering Department of ukurova University (Adana, Trkiye), in 2000 and 2003 respectively. He has
started his PhD study at the Mechanical Engineering Department of ukurova University (Adana, Trkiye), in 2010.
He is currently working as a Senior R&D Engineer in Research and Development Department in Otokar.
Atil Erdik He was born in Sakarya in 1979. He graduated from Uluda University- Bursa as Mechanical engineer. He
completed his mater in science study in mechanical engineering from Istanbul technical University in 2006. He has been
working as section manager in CAE department Otokar since 2005.
Atl has many articles represented in international conferences especially on blast and ballistic simulation.
His current research interests are anthromorphic test dummies, energy absorbing materials, high strain rate dynamic testing
methods, dynamic behavior of metals at high strain rates, shock waves and blast. Mr. Erdik is a member of Chamber of
Mechanical Engineers, Turkey.
Bulent Ekici He was born in 1965. He graduated from Bogazici University in 1988. He has obtained his master in since
degree in 1991 and PhD degree in 1998 from Bogazici University. He has been working as assistant professor in Marmara
University since 1993.
Durmu Ali Bircan He was born in Pozant in 1975. He earned his Bachelor Science, Master of Science and Ph.D. degree
from the Mechanical Engineering Department of ukurova University (Adana, Trkiye), in 1998, 2001 and 2008
respectively. He is currently working as a Lecturer at the Mechanical Engineering Department of ukurova University
(Adana, Trkiye).
727
Anadolu Isuzu Automotive Industry & Trading Co., Kocaeli, Turkey / R&D Department
3
1
Abstract. The focus of this study is to decrease the level of the noise radiated from an air intake pipe. The
first step is to measure the noise level radiated from a basic intake pipe by using three microphones. The
dominant fequency of the noise is found out by examining this noise data in the frequecy domain. Three
different intake pipes are designed and analyzed for understanding the effect of the design details such as
expansion chamber on the noise transmission loss characteristic of the intake pipe. As a final step, the
dimensions of a resonator is calculated using analytical methods, and its effect on the noise transmission loss
is calculated using numerical methods.
Keywords: air intake, noise, Helmholtz resonator, expansion chamber, transmission loss
1 Introduction
Air intake pipes are used to transport clean air to the IC engines. They are directly connected to intake manifold
by air filter. Since theyre connected to engine , they transmit the intake noise to the outside. For this reason
beside their capability of serving the air to the engines they should have muffling capability to reduce the noise
level. The intake air pipe conducts the noise sound at multiple frequencies which change in response to , for
example a rotational speed of the engine. In order to reduce the level of the noisy sound, resonators should be
designed for intake pipes. Resonators reduce the level of the sound at a specific frequency band. An air intake
pipe with a Helmholtz resonator is shown in Figure 1.
Helmholtz resonator
Expansion chamber
Snorkel
Intake systems fulfill a number of roles in an overall vehicle design which are: [1]
- channel air to the engine;
728
- filter particulates;
- enhance engine performance through wave action tuning;
- reduce noise
2 Literature Review
An acoustic filter consists of an acoustic element or a set of elements inserted between a source of acoustic
signals and the receiver, like atmosphere. Clearly, an acoustic filter is analogous to an electrical filter as well as
vibration isolator. An acoustic filter or muffler is therefore also called an acoustic transmission line. The
performance of an acoustic filter or muffler is measured in terms of one of the following parameters.
a) Insertian Loss, IL
b) Transmission Loss, TL
c) Level difference , LD, or noise reduction , NR
Insertian loss is defined as the difference between the acoustic power radiated without any filter and with that
filter. Symbolically,
IL = LW1 LW2 (dB)
IL = 10 log (W1/W2) (dB)
(1)
where LW denotes acoustic power level and subscripts 1 and 2 denote systems without filter and with filter,
respectively. [2]
Transmission loss is independent of the source and presumes (or requires) an anechoic termination at the
downstream end. It is defined as the difference between the power incident on the muffler proper and that
transmitted downstream into an anechoic termination. Symbolically,
TL = LWI-LWO (dB)(2)
where subscripts I and O denote inlet and outlet, respectively. [2]
Level difference or noise reduction is the difference in sound pressure levels at two arbitrary selected points in
the intake pipe. Symbolically,
LD = 20 log(pn / p1)(dB)(3)
Unlike the transmission loss, the definition of level difference makes use of standing wave pressures and does
not require an anechoic termination.
A silencer is an important noise control element for reduction of engine exhaust noise, intake noise, and other
noise sources involving flow of a gas. In general, a silencer may be defined as an element in the flow duct that
acts to reduce the sound transmitted along the duct while allowing free flow of the gas through the flow passage.
A silencer may be passive, in which the sound is attenuated by reflection and absorption of the acoustic energy
within the element.An active silencer is one in which the noise is cancelled by electronic feedforward and
feedback techniques. [3]
Passive silencers may be of the reactive or dissipative type. Reactive type passive silencers main working
mechanism for attenuation of sound is reflecting of the acoustic energy back to the source.
729
Reactive types of silencers or resonators are Helmholtz resonator, side-branch resonator, expansion chambers
mufflers, concentric hole cavity resonator, extended tube resonators. [3]
In this study the Helmholtz resonator and the expansion chamber muffler are used to decrease the level of sound
radiates from snorkel of air intake pipe. Since the only explained analytical calculation is for Helmholtz
resonator in this study, the theorical details of other mufflers and the resonators are skipped.
Helmholtz resonator is one of the reactive passive resonator which is usually used in the vehicles intake
system. The resonant frequency calculation is shown in the Figure 3.
In Fig.3, c denotes sound velocity at air (344 m/s @ 21 oC), A denotes cross section of the neck, V denotes the
volume of the cavity and the L indicates corrected neck length. When using corrected length instead of normal
length of the neck, the analytical and FE-based characteristic frequencies are well matching.
L = L + (A)1/2 (m)
(4)
3 Methods
The first step of this study is to collect noise data at different driving conditions. The second step is to make
analytical calculation for a Helmholtz resonator which works in a specific frequency band. The final step is to
make numerical calculations for different air intake designs includes expansion chamber and Helmholtz
resonator.
730
Since the aim of this study is to decrease the noise level at idling speed , only the measurements at idling speed is
examined. The noise characteristics from three microphones are well matching at idling speed. The noise level
graph is shown in the Figure 5. In this study the dominant frequencies are defined as 185 Hz and 270 Hz as
shown in the graph. For defining the dominant frequency, noise filtering operation is also used. For the analytical
and the numerical calculations 270 Hz is selected as the focus frequency.
1 .0 0
5 0 .0 0
4 5 .0 0
4 0 .0 0
3 5 .0 0
3 0 .0 0
2 5 .0 0
Amplitude
Pa
dB(A)
2 0 .0 0
1 5 .0 0
1 0 .0 0
F
5 .0 0
S p e c tr u m
P o in t 1 : S
( A )
0 .0 0
- 5 .0 0
- 1 0 .0 0
- 1 5 .0 0
1 8 5 .8 2
- 2 0 .0 0
0 .0 0
5 0 .0 0
1 0 0 .0 0
1 5 0 .0 0
2 0 0 .0 0
2 7 2 .1 5
2 5 0 .0 0
3 0 0 .0 0
0 .0 0
3 5 0 .0 0
4 0 0 .0 0
5 0 0 .0 0
H z
731
270 =
344 0,000706
V= 4,36x 10-4 m3
2 V 0,0665
Second model has an expansion chamber. The expansion rate of this intake pipe is 2. Transmission loss
frequency graphic acquired from FEA is shown in the Figure 8.
Fig. 8. Transmission loss graph of the second design with expansion chamber
732
Third model has an expansion chamber and Helmholtz resonator which has 270 Hz resonant frequency
calculated by analytical methods. Transmission loss frequency graphic acquired from FEA is shown in the
Figure 9.
Fig. 9. Transmission loss graph of the third design with expansion chamber and resonator
References
1.
2.
3.
4.
5.
733
Basic Frequency Analysis of Sound , Brel & Kjr Company Lecture Notes
Kopmaz O. Acoustic Seminar Notes
Intake-Exhaust Modeling , LMS Company Sofware Notes
Quantification of intake system noise at Nissan Motor Co., LMS Company Sofware Notes
Nakayama T. , Fukumori S. , Hayashi K. ,Katoh N. , Otsubo M. Intake Muffler : United States Patent Application
Publication , Pub. No : US 2007/0163533 A1
11. Goksu, A. Midibus Emis Borusu Test ve Saysal Analiz Raporu, DTA Engineering Company
12. Akba A. Susturucularn Akustik Performansnn ncelenemsi , Msc. Thesis , Istanbul Technical University
Biographies
Bar Koca was born in Kartal, Istanbul, 1981, graduated from Kocaeli University, Mechanical Engineering Department in
2004. He received MSc. degree from Marmara University on fatigue life calculation of a drag link in frequency domain.
He presented a presentation about fatigue life prediction in frequency domain at ASME 2010 10th Biennial Conference on
Engineering Systems Design and Analysis Conference. He is interested in NVH problems, propeller shafts, fuel systems,
fatigue life calculation and ICE.
Mr. Koca has membership of MMO and TEMA.
Sleyman Sayar was born in Mihalgazi, Eskisehir, 1977, graduated from Department of Automotive (BSc.), Faculty of
Technical Education, Gazi University in 2001.
He has a patent about a muffler at air intake systems. His current working topics are noise reduction on intake and exhaust
sysetms, engine cooling systems, emission systems and ICE.
M. Akif Gksu was born in Tuzluca, 1968, graduated from Uludag University Mechanical Engineering Department,
received MSc degree from U.University on internal acoustic analysis of a passenger car with numerical methods, completed
PhD Thesis from Uludag University , Bursa, Turkey on Sound Intensity Method for Sound source Localisation in
Automotive.
He works on CAE and TEST Methods for Noise, Vibration, Durability , FE Model Updating.
734
1,2,3,4
1
Abstract. Wave forces induce dynamic effects on the quay piles. The paper presents computer aided
numerical analyses which are utilized to investigate the dynamic behavior of pile under wave loads. For this
purpose, constant cylindrical cross-sectioned steel quay pile is buried into the sea bed and modeled with the
scour around it. The analysis accounts for a soil-pile interaction in a simplified way. As the lateral ground
pressure change by delving deep into the ground, soil spring stiffnesses also change. These stiffnesses are
calculated separately for different depths and taken into account in both analyses. Lateral displacement of the
employed pile varying with the time is obtained from both analyses and compatibility of them is observed.
Finally, the frequency of the external load and natural frequency of pile are compared to examine whether the
resonance came true or not.
1 Introduction
The quays are composed of reinforced concrete or steel piles (vertical structural members) and decks
(horizontal structural members). The number, size and capacity of quay piles shall bear safely the axial loads
from the deck, the lateral ship impact loads and the environmental loads.
Although the traffic and ship impact loads take significant part in designing, the environmental loads are
also effective in sizing. In present study only the wave loads amaong wind, earthquake and wave ones are
considered in dynamic design of pile.
Wave forces on the marine structures are the major contribution to the total forces experienced by such
structures, particularly in rough weather. The calculation of the wave loads on vertical cylinders is always of
major concern to ocean engineers, especially recently when such studies are motivated by the need to build solid
marine structures in connection with oil and natural gas productions [1]. Marine structures are often located on
erodible bottoms, the scouring action of waves and currents may lead to their damage or failure. In order to avoid
these risks, the knowledge of the processes involved in the phenomenon of scour is an essential factor, both for
designing coastal structures and for considering preventive measures [2].
Critical conditions in solutions are taken into consideration under the effect of environmental and functional
forces in designing. However, dynamic analysis has been done according to hydrodynamic forces in this paper.
735
Pile is modeled as a single-degree of freedom system with nonlinear external force due to the velocity-squared
drag force. Two different numeric analysis methods are considered; numerically integrated equation of motion of
single degree of freedom system by runge-kutta method and computer aided Time History Analysis based on
Newmark method in dynamic analysis. Pile is modeled with elastic springs and buried into the soil. In this
paper, behavior consistencies of solutions according to single degree of freedom systems and time history
analyses are compared and suggestions proposed.
2. Methodology
In this paper, the steel cylindrical pile is modeled under 300 kN axial load (N) where the water depth (d) is
26.00m. Its diameter (D) is 1.00m. Modulus of elasticity (E) is 2.1x108kN/m2. Length (L) is 39.00m. The buried
length (Lb) is 7.50m. Wall thickness (e) is 0.01m and damping coefficient (a) is 0.05. The scour around pile has
1.5m diameters. The bottom end of the pile is fixed to a ground and the top end of it is simply supported by an
apron. Frictional effects due to bending of piles (external moments on each pile from soil) are ignored. The paper
presents numeric models which are utilized to investigate the dynamic behaviour of piles subjected to forced
harmonic lateral vibration. The lateral vibration caused by wave loads. Wave loads are calculated by well-known
Morison Equation. The analysis accounts for a soil-pile interaction in a simplified way and predicts the response
of such foundations.
FH = FD + FM =
2 C
Du (z,t ) u (z,t ) dz +
D 2
u& (z,t ) dz
4
736
(1)
In the equation above; , represents water density, D pile diameter, water surface elevation, d also water
depth. The wave load as shown in (Fig. 1) is calculated by Eq.(1). Mentioned load is assigned to both analyses as
time varying external load at the point (z) 21.00m height from seabed. This is the junction point of external
force.
10
0.00
1.25
2.00
2.50
3.00
3.75
4.75
5.00
6.25
7.50
4322.86
3602.38
3170.10
2881.91
2593.72
2161.43
1585.05
1440.96
720.48
0.00
Spring
force
(kN)
The values of the spring stiffnesses must be represented as a function. The function is given below. The function
k(z) is inserted to Eq. (4).
(2)
k ( z ) = 4322.86 - 576.38 z
737
(3)
implying that the inertia, damping and restoring forces balance the applied force. Specifically, m is the
mass, k is the stiffness, and c is the damping coefficient. Furthermore, Y(t) is the displacement,
&
Y
(t )
the velocity,
&&
Y
(t)
the acceleration and F(t) the externally applied force. Finally, dots denote time derivatives d/dt. Obviously,
Eq. (3) is a second order differential equation that needs to be solved for the displacement Y(t). Modified
equation of motion for the values of mass, stiffness and damping coefficient, is given below.
Lb
L
L 2
L 2
L 2
2
2
m ( z) dz Y&&(t ) + aEI (z) dz Y&(t ) + EI (z) dz + k( z)(z) dz N (z) dz Y(t ) = Fi( t )i( z)
0
o
o
(4)
The shape function (z) can be assumed or can be obtained by assigning the maximum value of the
external force (27.19kN) as seen in (Fig. 1) on the structure as a static load [12] at the junction point. After this
assignment, ten or more points determined from the base to top on the structure and the displacement values of
these points are calculated. These values are normalized by Least Squares Method to find the shape function as
follows.
(5)
In the present study, the bottom and top end of the pile are considered respectively; clamped and simply
supported. Shape function satisfies geometric boundary conditions. Geometric conditions for adopted system are
given below.
(0) = (0) = ( L) = ( L) = 0
(6)
& =0
Y
(0)
(7)
738
(8)
This second order differential equation is solved by [13] via RungeKutta method. The displacement
function is obtained as follows.
(9)
In this paper it is aimed to obtain displacement function of the structure which varies with time (t) and
location (z). For this purpose generalized displacement function is obtained [14], as written below, by
multiplying shape function with displacement function.
v(z,t) = 0.739sin(0.069z + 0.54)+ 0.484sin(0.134z + 4.053)(0.004cos(0.967t)- 0.183sin(0.967t)) (10)
The values of the time varying displacements at a certain point can be obtained by the equation given
above. These values are compared with the obtained values from other numeric analysis in section 3.
The natural frequency of the structure is calculated as 18.72 sec-1 by the equation given below.
n =
k
m
(11)
The value of the natural frequency is recognized with the value of the vibtarion frequency of the force
acting on the structure in section 3.
Computer aided Time History Model (THM) is taken into consideration as another numeric method. Two
basic types of numerical solution method are used in mentioned method: convolution integral and time step.
Computer aided model is adopted time step method. Most numerical solution methods are step by step methods
because the equation is solved at a succession of values of t, t+t, t+2t, etc. The accuracy of the solution
depends on the length of the step interval (t). It must be short enough for the load time history, the response
time history and in many cases the shortest natural periods to be well defined [15]. Computer aided method is
based on the implicit time step method. Various implicit methods are available. The widely used Newmark
method is utilized in method given below. Starting from a Taylor series [15]:
c
m
+ k
Yt+1 = 2 +
t 2t
1
m m
c
k Ytt
Ft+1 + (1 2)Ft + Ft1... 2 k 2 Yt ... 2 +
2 t t 2t
(12)
The method is therefore very good for linear both single and multidegree of freedom structural dynamics
where the time step can be selected in order to model the structural response and not the shortest natural period
in the structure. The analysis is completed for 20sec with the length of the step interval (t) 0.01sec. The time
history function (shown in Fig. 1) is utilized for numerical analysis.
3.Numerical Results
The maximum values of the displacement are obtained from 25m height from fixed support. Therefore,
these values are compared as given in (Fig. 2). Whereas the compatibility of the displacements can be seen in
(Fig. 2), their compatibility is controlled by Chi-Square Test. Thus, the compliance has been proven.
739
Group classification of Chi-Square Test is shown in Table 2. According to this table; group number (g=8),
degrees of freedom (df=8-1=7), predetermined alpha level of significance (0.995), are taken in displacement
dispersal and test is completed (2find=0.6288<20.995,7 =0.989). This result reveals compatibility of analyses.
Table 2. Group classification of Chi-Square Test
Groups
<0.018
>-0.018 , <0.012
>-0.012 , <0.006
>-0.006 ,
<0
>0 ,
<0.006
>0.006 ,
<0.012
>0.012 ,
<0.018
>=0.018
N~
20
519
234
229
222
238
536
R.K.
145
410
225
217
222
224
398
160
The approximate solutions obtained from SDOF can differ from other stiffnesses solutions [14]. In this
study, it is observed that differences are not exceeding 3%. The values would be different if another shape
function is chosen instead of an approximate one.
Table 3. Varying displacements along the pile
Distance
from support
17
21
25
28
30
33
35
39
0.005
0.014
0.017
0.018
0.017
0.015
0.010
0.008
0.0
0.002 0.005
0.014
0.017
0.019
0.017
0.016
0.012
0.008
0.0
(m)
N~
Displacements 0.0
(m)
0.001
R.K.
Displacements 0.0
(m)
The maximum displacements of the points determined on the structure are seen in Table 3. The natural
frequency of the pile is calculated 18.72 sec-1 by Eq. (11). The vibration frequency of the force given in (Fig. 1)
is obtained 0.15sec-1 by FFT tool of [16].
740
Natural frequency value which is obtained from approximate shape function converges to its real value
from top. Natural frequency approaches to its real value as the mode structure approximately approaches to its
real value.Mode structure of the system during the vibration is indicated with shape function. Since real shape
function cant be calculated precisely, approximate function is taken in solutions.
As a conclusion, it is detected that the results obtained from approximate method (SDOF) do not differ
from other method due to correct assumptions of shape and spring stiffness (s. In spite of the compatibility of
results obtained from both analyses, computer aided Time History Model should be preferred due to not
including any assumptions.
References
1.
2.
3.
4.
5.
6.
7.
8.
9.
10.
11.
12.
13.
14.
15.
16.
Eicher J.A.; Guan H.; Jeng D.S. 2003. Stress and deformation of offshore piles under structural and wave loading.
Ocean Eng 30(3):369385.
Carreiras J.; Antunes do Carmo J.; Seabra-Santos, F. (2003). Settlement of vertical piles exposed to waves Coastal
Engineering 47:355365
Morison, J.R.; OBrien, M.P.; Johnson, J.W.; Schaaf, S.A. (1950). The forces exerted by surface waves on piles.
Journal of Petroleum Technology, Petroleum Transactions, AIME 189:149 154.
Chakarabarti, S.K.; Tam A. (1975). Interaction of waves with large vertical cylinder. J Ship Res 3:19-22.
Tang, W.H. (1989). Uncertainties in offshore axial pile capacity. In: Kulhawy FH, editor. Foundation engineering:
current principles and practices. New York: ASCE Press. 833847.
Lu J-F.; Jeng D. S. (2010). Dynamic response of an offshore pile to pseudo-Stoneley waves along the interface
between a poroelastic seabed and seawater. Soil Dynamics and Earthquake Engineering 30:184201.
Ergin A. 2009. Coastal engineering. Ankara: Metu Press.
ISO 19901-1. 2005. Specific requirements for offshore structuresPart 1: Metocean design and operating
considerations. International Organization for Standardization, Geneva, Switzerland: Petroleum and natural gas
industries
Khan, A. K.; Pise, P. J. (1997). Dynamic behavior of curved piles. Computers and Structures 65(6):795-807.
Prakash, S.; Sharma H. D. 1990. Pile foundations in engineering practice. London: John Wiley.
Tahghighi, H.; Konagai, K. (2007). Numerical analysis of nonlinear soilpile group interaction under lateral
loads. Soil Dynamics and Earthquake Engineering 27:463474.
SAP2000 V14 (Structural Analysis Program). Integrated finite element analysis and design of structures basic
analysis reference manual. Berkeley (CA, USA): Computers and Structures Inc.
Maple 11. 2007. Maplesoft, Waterloo, Canada
Clough, R.W.; Penzien, J. 1993. Dynamics of structures (2nd edition). Singapore: Mc GrawHill, Inc.
Barltrop, N. D. P.; Adams, A. J. 1991. Dynamics of fixed marine structures, 3rd edition, UK, Epsom: Atkins Oil &
Gas Engineering Limited.
Matlab V 6.5. 2002. The MathWorks, Inc.
Biographies
mit GKKU mit Gkku was born in Sivas. He was graduated from Akdeniz University in 1983. He completed his
graduated thesis in 1987 and his doctorate thesis in 1992. He is studying fluid mechanics, wave mechanics, offshore
structures and coastal and port engineering. He has been a professor in Celal Bayar University since 2002.
Engin Gcyen Engin Gcyen was born in Manisa in 1983. He was graduated from Celal Bayar University in 2005. He is
still a doctoral student in Celal Bayar University. He has been a research assistant in Celal Bayar University for 5 years. He is
studying offshore structures, coastal and port engineering.
R. Turul Erdem R. Turul Erdem was born in 1985 in zmir. He was graduated from Celal Bayar University in 2008. He
completed his graduated thesis in 2010. He has started working as a research assistant since 2010 in Celal Bayar University.
He is studying dynamics of structures, concrete structures.
Soner eker Soner eker was born in Uak in 1983. He was graduated from Celal Bayar University in 2005. He is still a
doctoral student in Celal Bayar University. He has been a research assistant in Celal Bayar University since 2005. He is
studying dynamics of structures, steel structures.
741
2,3
Abstract.A stent is meshed cylindrical tube that is commonly used in angioplasty to restore the
blood flow in a narrowed or blocked coronary arteries. Despite the high success rate of stent
applications, stents are subject to extensive research and development activities, which are directly
related with their deformation characteristics. One of the major drawbacks of the stent
implantation is the in-stent restenosis. Vascular wall injury caused by the contact stresses between
the struts of a stent and the vein during the balloon angioplasty is hypothesized as a possible cause
of in-stent restenosis as well as the deep wall injury with rupture of the internal elastic lamina. The
major emphasis of the present study was on some mechanical testing of related stent according to
EN-ISO-25539-2:2008-Vascular stents. Second emphasis was to investigate some mechanical
analysis of stent with the help of Finite Element Analysis method.
1 Introduction
A stent is a small tube that is used to widen arteries supplying the heart that have narrowed. Stents are often
used in angioplasty, which is a procedure that improves blood flow through narrowed or blocked arteries. Stents
help prevent arteries from becoming narrowed or blocked again in the months or years after angioplasty. Stents
are also placed in a weakened artery to improve blood flow and to help prevent the artery from bursting. The
meshwork of stents is usually made of metal, but sometimes a fabric is used. Fabric stents, also called stent
grafts, are used in large arteries. Some stents are coated with medicines that are slowly and continuously released
into the artery. These medicines help prevent the artery from becoming blocked again. Within a few weeks of the
time the stent was placed, the inside lining of the artery (the endothelium) grows over the metal surface of the
stent.
742
743
The word "stent" was first used in 1916. Charles Stent was an English dentist who invented a dental
impression compound known as "Stent's material" and was used in dentistry. The coronary heart stent was
invented by Charles Theodore Dotter, a vascular radiologist. It was developed in 1969 and was implanted in a
dog. His stenting of peripheral arteries was experimental and was not readily accepted at the time. He was a
genius, who believed there were better ways to treat disease.
In 1986, Jacques Puel implanted the first coronary stent in a patient in France. Another inventor, Cesare
Gianturco was also working on a device in the United States. About the same time, Julio Palmaz and Richard
Schatz were working on a stent, that could be used in the human heart. They applied for a patent in 1985, which
was granted three years later. The first Palmaz-Schatz stent was approved for use in the United States in 1994.
Over the next several years, more improved stents were designed and manufactured.
In 2001, the first drug eluting stent was implanted in a patient during a clinical trial. The coated stent is a
normal metal stent that is coated with a pharmaceutical drug, to prevent re-closure after cardiac angioplasty. In
2003, the Cypher, manufactured by Johnson & Johnson was approved by the F. D. A. These new coated stents
are much safer for patients, but cost much more that the traditional stents [3,4].
The mechanical function of a stent deployed in a damaged artery is to provide, over the long-term, a smooth,
metallic tubular mesh structure that fits any existing arterial curvature and that efficiently opposes compressive
arterial forces. The performance of any biomedical material is controlled by two sets of characteristics:
biofunctionality and biocompatibility. Biofunctionality defines this ability of the device to perform the required
function. The clinical value of stents has been widely evaluated and two main advantages have been identified:
(1) restenosis rates are reduced due to better initial luminal patency; and (2) atherosclerotic vessel walls are
rendered more mechanically stable (with a reduced risk of occlusion, cracking of plaque, and dissection) [5].
The two most important features of a coronary stent are basic to its use: the radial force with which it supports
the vessel and its longitudinal flexibility, one of the major determinants of its trackability into the target lesion
before deployment (1). Accordingly, the ideal coronary stent would be very flexible and should have high radial
strength to avoid collapse due to the pressure exerted on it by a diseased artery wall, but it should adapt easily to
changes in diameter and tortuosity due to coronary vasomotion and heartbeats [6].
Stress and deformation analysis in stents are important and an active research area. Stents are commonly used
in our country. In our paper, structural analysis of stents have been asked to be examined and developed. The
following were the particular aims have been investigated.
First, the major emphasis of the present study was on some mechanical testing of reletad to the stent
according to the EN ISO 25539-2:2008 Cardiovascular implants- Endovascular devices Part 2: Vascular stents.
Second, the major emphasis of the present study was to investigate some mechanical analysis of stent with the
help of Finite Element Analysis method. The following were the particular aims have been investigated.
1- The stent samples were applied and performed the following performance experiments according to BS
EN ISO 25539-2:2008- Cardiovascular implants- Endovascular devices Part 2: Vascular stents.
Radial Strength
Flexibility
Dislodgement force
Bursting
2- The following were examined and investigated some mechanical analysis of stent with the help of
Finite Element Analysis method.
Examination of hexagonal, circular and square cross-section the behavior of stents under the same
load
Comparison of experimental and analytical studies on the flexibility of stent.
Simulation of radial strength and realization of the stent radial strength account.
Examining the effect of material.
744
745
Many of the engineering considerations in stent design were adopted to improve the global acceptability of
the device, rather than making a stent design for a specific type of coronary lesion. In clinical practice, the
operator must decide which stent is most appropriate for the patient. This article focuses on the features of stent
design that make a specific stent more or less suitable for a particular type of lesion or anatomy.
R R0 R
=
R
R 0 . These two values are graphically plotted on the computer and
Similarly, strain is calculated as 0
recorded. Test samples are expanded to their required diameters by using 2.5, 3.0 and 3.5 mm balloons. Nominal
diameters have been taken as a reference.
Materials and Sampling: Starting values for stents are given in table 1 (samples are 9.5 mm stent length/
17.5 mm). Furthermore, raw data which contain surfaces tension and tension strength values of stent measured
samples are listed in below table 1, 2, 3, 4.
746
CODE NUMBER
9,5 mm
SC 00 01 001 46
12,5 mm
SC 00 01 004 46
15,0 mm
SC 00 01 002 46
17,5 mm
SC 00 01 006 46
2.5 mm
3.0 mm
3.5 mm
9.5 mm
2.70 mm
3.24 mm
3.72 mm
12.5 mm
2.86 mm
3.30 mm
3.93 mm
15.0 mm
2.85 mm
3.40 mm
3.80 mm
17.5 mm
2.92 mm
3.35 mm
3.93 mm
Evaluation of the Test Results: In measurements made with stents of different length and diameters. It has
been observed that average radial strength values vary between 0.65x105 Pa / 1.35x105 Pa depending on vessel
implement diameters.
0.65 x105 Pa (490 mm Hg), the smallest average stress value of stents produced ( 9.5mm) is over 0.4x105 Pa
(300mm Hg) which is accepted surface tension value. Radial strength measurement radial narrowing is
compatible with surface area in Ephesos Stents.
3.3 Flexibility
Purpose: The purpose of this test is to determine the minimum radius at which the deployed stent can be
flexed without kinking or exhibiting a diameter reduction of greater than 50 % and if the stent recovers its
original geometry after testing. If overlapping of stents can be anticipated in clinical use (e.g. superficial femoral
artery), the kink radius of the stent under study in overlapping configurations should also be determined.
Test Equipment: When the stent delivery system is advanced within the cardiac vessel, firstly the distal end of
the stent delivery catheter is bent by deforming the first strut then this deformation process moves the consequent
struts. Therefore, it is concluded that the flexibility is affected by the strut geometry rather than stent length. The
flexibility measurement system was designed in this basis.
The flexibility measurement system consists of a load cell that is placed perpendicular to stent surface
and a cylindrical bend-angle-fixing system that has 3.0 mm diameter. In this system, bend-angle-fixing cylinder
represents the inner diameter of a vessel or guiding catheter at the bending point. The probe (load cell) represents
the outer diameter of a vessel. When the stent is bent, it exerts force towards the vessel surface at the outer
diameter (See Figure 5.a). For this reason the force measurements are taken from the outer diameter position
(See Figure 5.b). Bending angles, between = 0 - 90 were selected to inspect the forces that are exerted by the
stent delivery system during bending on to vessel wall. Each size of the EPHESOS stent was tested with five
747
samples. Each stent was mounted on the system and was bent to the angles. Reaction forces against the bending
were recorded using load cell with 5 mN sensitivity.
Fixture
Vessel
RB
Rf
Stent Delivery
System
Bend-angle-fixing
System
Reactive
Force
Load
Cell
Stent Delivery
System
Do Di
Fig.5. (a) The illustration of stent delivery system passing through the vessel curvature, Di: Inner diameter of
curvature, Do: Outer diameter of curvature (b) The schematic of the stent flexibility test system, RB: Reactive
Force, : Bending angle
Materials and Sampling: d = 2.75 mm balloon catheter on a crimped 9.0 mm, 12.0 mm, 15.0 mm, 18.0 mm,
20.0 mm and 25.0 mm in the lengths stent, and d = 2.5 mm, 3.0 mm, 3.5 mm and 4.0 mm balloon catheter on the
crimped 25.0 mm in the length stent samples used.
Table 3: Used stents for flexibility test
STENT NAME
CODE NUMBER
9.0 mm
SC 00 01 001 46
12.0 mm
SC 00 01 004 46
15,0 mm
SC 00 01 002 46
18.0 mm
SC 00 01 006 46
20,0 mm
SC 00 01 007 46
25,0 mm
SC 00 01 010 46
Evaluation of the Test Results: According to the test results, the flexibility of Ephesos stent depends on stent
geometry rather than stent length and king was not observed the tested stent. Peak which was observed on the
graphic is shown the transition to plastic phase and value of angle depends on the position of bend-angle fixing
system.
Reactive forces of test samples were found as between 218 and 367 mN. Ormiston et al. conducted some
flexibility experiments on various commercial stents [11]. According to their results, reactive forces were found
in the range of 150 to 1100 mN. The reactive force results (stiffness) of EPHESOS stents were within this
range. So it was concluded that the flexibility of EPHESOS stents met the product specification. Bend angle
fixing system.
748
SC EF 27509 05
SC EF 27512 05
SC EF 27515 05
SC EF 27518 05
SC EF 27520 05
SC EF 27525 05
SCEF 02525 05
SC EF 03025 05
SC EF 03525 05
SC EF 04025 05
Evaluation of the Test Results: According to test results; average of the stent dislodgement force is
determined as 2.60 N.
749
Nominal balloon
diameter tolerance at 6 bar
2.0
2.00 2.15
2.5
2.50 2.65
3.0
3.00 3.15
3.5
3.50 3.70
4.0
4.00 4.25
The test results were evaluated with international testing recommendations in mind. These include: the FDA
guidance document titled Non-Clinical Tests and Recommended Labeling for Intravascular Stents and
Associated Delivery Systems, BS EN ISO 25539-2:2008 ISO 25539-2:2008(E) Titled Cardiovascular implants
Endovascular devices and in EN 14299:2004 Titled Non active surgical implants Particular requirements
for cardiac and vascular implants- Specific requirements for arterial stents. In addition to, ISO 10555-1 Titled
Sterile, single-use intravascular catheters, Part 1: General requirements and ISO 10555-4 Titled Sterile,
single-use intravascular catheters, Part 4: Balloon dilation catheters.
Evaluation of the Test Results: Upon burst test results:
For all balloon sizes, the statistically measured nominal diameter tolerances have stayed below the predefined
tolerance limits.
Accordingly, all balloons are at required standards under 6 bar pressure.
Average burst pressures have been at a minimum of 22.928 bar.
Even if negative tolerances are applied with 99.9% confidence, the minimum burst pressure will not be below
17 bar.
Examination of hexagonal, circular and square cross-section the behavior of stents under the same load
Simulation of radial strength and realization of the stent radial strength account.
750
Cross-sectional Effect: Hexagonal, circular and square cross-section the behavior of Ephesos stents were
examined under the same load. The applied load is 1 bar.
Stent 2.5 mm in diameter under 1 bar pressure has a circular cross section.
751
Stent 2.5 mm in diameter under 5,5 mN FORCE has a circular cross section. 0.08 mm diameter vessels in
three-dimensional image.
752
Fig. 10. 0.08 mm diameter vessels in three-dimensional image under 5.5 mN FORCE
Stent 2.5 mm in diameter under 11 mN FORCE has a circular cross section. 0.08mm diameter vessels in threedimensional image.
753
754
Accordingly, all balloons are at required standards under 6 bar pressure. Average burst pressures have been at
a minimum of 22.928 bar. Even if negative tolerances are applied with 99.9% confidence, the minimum burst
pressure will not be below 17 bar.
This value GUARANTEES that ALVMEDCA LTD. Manufactures balloons which do not burst below 16
bar pressure, even at the worst conditions.
Dislodgement Test:According to test results; average of the stent dislodgement force is determined as 2.60 N.
2- Examination of hexagonal, circular and square cross-section the behavior of stents under the same load.
Stress distribution of polygon cross-section was found 239 MPa which is higher than other crosssections.
3- Comparison results of experimental and analytical studies on the flexibility of stent were found
compatible.
4- Simulation of radial strength and realization of the stent radial strength account.The behaviour of stent
under 1.75 bar circumferential pressure- Stent collapsed on the radial force simulation.
REFERENCES
1.
2.
3.
http://www.texasheartinstitute.org/hic/topics/proced/angioplasty.cfm,June, 2010
http://www.heart.org.in/diseases/stents.html, June, 2010
Recommendations on stent manufacture, implantation and utilization R. Balcon, R. Beyar, S. Chierchia,
I. De Scheerder, P. G. Hugenholtz, F. Kiemeneij, B. Meier, J. Meyer, J. P. Monassier and W. Wijns for
the Study Group of the Working Group on Coronary Circulation European Heart Journal (1997) 18,
15361547
4. http://www.apsf.org/resource_center/newsletter/2007/winter/12_protocol.htm
5. Journal of Biomechanics 34 (2001) 10651075 Mechanical Properties Of Coronary Stents Determined
By Using Finite Element Analysis Erique Etavea, G!Erard Finetb,C,*, Maurice Boivina, Jean-Claude
Boyera, Gilles Rioufolb,C, Gilbert TholletdA Laboratory Of Mechanics, INSA, Lyon, France
Bdepartment Of Hemodynamics, Hospices Civils De Lyon, Claude Bernard University, Lyon, France
Ccreats Research Unit, CNRS, UMR 5515, INSERM, Lyon, France Dgemppm, Laboratory Of
Scanning Electron Microscopy, INSA, Lyon, France Accepted 14 February 2001
6. Longitudinal Straightening Effect of Stents Is an Additional Predictor for Major Adverse Cardiac
Events Mariann Gyongyosi, MD, PHD, Paul Yang, MD, Aliasghar Khorsand, MS, Dietmar Glogar,
MD, FESC, on behalf of *the Austrian Wiktor Stent Study Group and European Paragon Stent
Investigators Vienna, Austria Vol. 35, No. 6, Published by Elsevier Science Inc. PII S07351097(00)00570-2
7. F. Cardarelli, Ferrous metals and their alloys, Materials handbook, Springer London Limited, London
(2000), pp. 2021.
8. Gopinath Mania, Marc D. Feldmanb, c, Devang Patelb and C. Mauli Agrawal Department of
Biomedical Engineering, College of Engineering, The University of Texas at San Antonio, One UTSA
Circle, San Antonio, TX 78249 0619, USA, 22 December 2006
9. A survey of stent designs D. Stoeckel1, C. Bonsignore1 and S. Duda2 1Nitinol Devices and
Components, Fremont, CA USA 2Dept of Diagnostic Radiology, University of Tuebingen, Tuebingen,
Germany Min Invas Ther & Allied Technol 2002: 11(4) 137147
10. www.alvimedica.com, June, 2010
11. Flyer Coronary Stent System RX Technical Data Bulletin Dr. John Ormiston, Mercy Angiography Unit,
Auckland, New Zealand, November 2004. (Data on file at Atrium Medical Corporation).
755
Abstract. A moving finite element for transverse vibration of plates due to an accelerating moving mass is
presented. This technique replaces the moving load by an equivalent moving finite element to take into
account the inertial effects of the moving mass. By using shape functions 16-DOF conforming rectangular
thin plate element with C(1) continuity the mass, damping and stiffness matrices of the moving finite element
are determined by the transverse inertia force, coriolis force and centrifugal force of the moving mass,
respectively. Numerical examples solved in Matlab environment are given for a simply supported plate.
1. Introduction
Dynamic response of structures under moving loads is an important problem in engineering and studied by many
researchers. The analysis of the dynamic response of a plate structure to moving masses is of considerable
interest to many designers. The monograph by Fryba [1], being an excellent reference with many analytical
solution methods for simple cases, is one of the earliest and most comprehensive study in the moving load
problem. Some analytical or finite element studies of moving load problem are given by [2-9]. Esen [10] offered
a solution method using moving finite element for dynamic response of a beam due to an accelerating moving
mass. Esen et al. [11-13] studied on dynamic response of overhead crane beams under moving loads by using
analytical and FEM methods. The most valuable monographs for structural and plate dynamics can be found in
references [14-19]. When inertial effects of mass are considered, the solution of the moving load problem
becomes more complex and the studies on this field are limited. This paper presents a solution method by
assuming the moving mass as a moving finite element. The classical finite element method is combined with a
moving finite element to represent the motion of the accelerating moving mass with all effects. The proposed
method, using the classical FEM, is applicable for all kinds of shapes, boundary and loading and damping
conditions when they are properly modelled.
Fig. 1 shows a rectangular plate, with dimensions lx, ly and thickness h, under an accelerating moving mass mp
moving with a constant acceleration amx parallel to the x-axis at a distance yp from the origin O; and mesh
discretion of the plate under moving mass and the kth plate element on which the accelerating moving mass m
applies at time t. Hooks law and other assumptions of the Kirchhoffs plate theory are valid given by [14].
756
x p (t )
a
mp vmmx
x
k -1
yp
ly
k+1
lx
Mx
z
f1 ,u1
1
f13 ,u13
f3 ,u3
f4 ,u4
xm (t)
k th plate element
f2 ,u2
f5 ,u5
mp v(am
t)
ym
2
f7 ,u7
f15 ,u15
f9 ,u9
f14 ,u14
a
z
f11 ,u11
f8 ,u8
h
f16 ,u16
f6 ,u6
f10 ,u10
3
f12 ,u12
Fig. 1. A rectangular plate under a moving lumped mass mp moving at a constant acceleration amx and a
variable speed vmx parallel to the x-axis at the distance yp from origin O. Finite element discretization of the
plate continuum by a rectangular plate element with dimensions a and b and equivalent nodal displacements and
forces of kth plate element on which the moving mass m applies at time t.
If the system can be modelled as an isotropic plate, then the corresponding differential equation of motion
without considering the damping in structure is [1, 14].
4 w( x, y,t )
4 w( x, y,t ) 4 w( x, y,t )
2 w( x, y,t )
+2
+
= p( x, y,t ),
D
+
4
2
2
4
x
x y
y
t 2
(1)
3
2
where w(x, y, t) - vertical deflection of the plate at point with coordinates x, y, and time t, D = Eh /12(1 ) bending rigidity, E- Youngs modulus, h- thickness,
-Poissons ratio, - mass per unit area, p(x, y, t)external load per unit area of the plate.
f z ( x, y , t ) = [ m p g m p
d 2 wz ( x p , y p , t )
amx t 2
,
2
amy t 2
y p = y0 + v y 0 t +
,
2
x p = x0 + vx 0 t +
dt 2
dx p
dt
dy p
dt
(2)
] ( x x p ) ( y y p ),
= vx 0 + amx t ,
= v y 0 + amy t ,
d2 xp
dt 2
d2 y p
dt 2
= amx ,
(3)
= amy ,
where ( x - x p ) and ( y - y p ) represent the Dirac delta functions in x and y directions, respectively, while
757
f z ( x, y , t )
is the applied force by the moving mass at point x, y, and time t. Besides, x0, y0 and v0x, vy0 are,
respectively, the initial positions and initial speeds of the mass at the time is zero; and amx is the constant
acceleration of the moving-mass. In case the inertial effect of the moving mass is considered, the acceleration
d2w(xp, yp, t)/dt2 is computed from the total differential of the second order of the function w(x, y, t) with
respect to time t, with variable contact point xp ,yp. For uniformly accelerated or decelerated motions according
to (3), the acceleration in (4) is in the form:
d 2 w( x p , y p , t )
dt 2
(4)
(5)
mp am x wx ] ( x x p ) ( y ym ),
where ~, , are, respectively, spatial and time derivatives of deflection; and, since the mass moves along the
&& and 2 m p (v x 0 + a mx t ) w& x are, respectively, the
deflected shape of the plate, m p (vx 0 + amx t ) 2 wx + m p am x wx , m p w
centripetal force, the inertia force and the Coriolis force components of the accelerating moving mass. The sub
symbol x defines that the function of deflection is derived by x. The rectangular plate element in Figure (1) is
a 16- DOF conforming plate element with C(1) continuity conditions at element boundaries. It includes constant
twist 2w/xy at corner nodes. Hence each corner nodal displacements are [17]
ui = [ ui1 ui 2
ui 3 ui 4 ] = w x y xy (i =1- 4),
i
T
(6)
wherewi vertical deflection of i th nodal point, and xi and yi , xyi are, respectively, rotation about x and y-axis
and twist of nodal point i. The equivalent nodal forces of the kth plate element are given by [10, 14]
(7)
where Nn (n=1, 2, 3,, 16) are shape functions of the plate element given by [14]
(8a)
p1 = 1 3 (t ) 2 + 2 (t )3 , p2 = a[ (t ) 2 (t ) 2 + (t )3 ], p3 = 3 (t ) 2 2 (t )3 ,
1 3
1
[b 3ay 2 + 2 y 3 ], q2 = 2 [b 2 y 2by 2 + y 3 ],
3
b
b
1
1
q3 = 3 [3by 2 2 y 3 ], q4 = 2 [ y 3 by 2 ], with (t ) = xm (t ) / a, ym = b / 2
b
b
p4 = a[ (t )3 (t ) 2 ], q1 =
(8b)
where a and b are, respectively the length and width of kth plate element and xm(t) is the variable distance
between the moving mass and the left end of the kth plate element, at time t, as shown in Fig. (1). While ymis the
758
constant distance between the moving mass and x-axis and assumed as in the midline. The relation between
shape functions and transverse displacement of the plate element at position x and y, and time t, is [10, 14]
(9)
whereui (i = 116) are the displacements of the nodes of the plate element at which the moving mass m locates.
Substituting Eq. (8) into Eq. (5) and writing the resulting expressions in matrix form yields:
(10)
with
{ f } = [ f1
f2
{u&} = [u&1
u& 2
N12
[ m] = m p .
N16 N1
...
f16 ]
... u&16 ]
16x1
T
16x1
, {u&&} = [u&&1
, {u} = [u1
u&&2
u2
... u&&16 ]
... u16 ]
16x1
T
16x1
N N
N1 N 2 ... N1 N16
1 1
.
...
. , [ c ] = 2m p v (t ) .
N N
N16 N 2 ... N16 2
16 1
N1 N16
.
...
. ,
N16 N 2 ... N16 N16
N1 N 2
...
.
.
...
.
[k ] = mp
,
2
2
2
N16 (v(t ) N1 +a m N1) N16 (v(t ) N 2 +a m N 2 ) ... N16 (v(t ) N16 +a m N16 )
(11)
where [m], [c] and [k] are, respectively, the mass, damping and stiffness matrices of the moving finite element.
Because rectangular plate element has four displacements DOF at each corner nodal point, the dimensions of the
property matrices of the moving finite element will be 16x16.
(12)
(13)
759
, with fki=mgNi(i=1-16)
3. Numerical example
In this work, the Newmark direct integration method [18] has been used for obtaining the solution of Eq. (12),
with variable time steps t = Tson /400. Tson is travelling time of the accelerating mass on the plate. In all
simulations, the mass has zero initial speed at the left end and when t=Tson it reaches to the right end. Consider
an isotropic simply supported beam plate traversed by a concentrated accelerating moving load of mp = 19.625
kg. The dimensional and material specifications of the plate are lx = 1 m; ly = 0.25 m, h = 0.01 m; E = 206.8
GPa, = 7850 kg/m3. The plate is divided by 100 identical plate elements in dimensions of 5cm x 5cm. The first
four natural frequencies and mode shapes of the plate without any additional mass are illustrated in Figure 4.
Fig. 2. The First, second, third and forth vibration modes and frequencies of the plate
Figure 3 shows vertical deflections of midpoint versus position of the accelerating mass mp with a=1(dashed)
and 1.5 m/s2 (solid line). Because of zero initial speed at first the mass travels very low velocities and causes
lower deflections. Due to the positive acceleration of the mass while the mass travelling the velocity is increased.
The more the velocity increased the more the deflection increases.
760
w(lx/2, ly/2,t)
a=1m/s2
a=1.5m/s2
The maximums of the deflections under different accelerations are depicted in Figure 4. As seen from Fig.4,
deflection behavior of the plate is non-linear. Showing several maximums and minimums the average of the
deflections tends to lower with increasing mass acceleration.
4. Concluding remarks
An accelerating moving mass affects dynamic response of a plate structure. The moving mass is modeled as a
moving finite element. Thus, in the finite element modeling of the entire system, the combination of mass,
stiffness and damping matrices of a plate element with the mass, stiffness and damping matrices of the moving
finite element can easily be made by taking into account the inertial effects of the moving mass.
761
References
1.
2.
3.
4.
5.
6.
7.
8.
9.
10.
11.
12.
13.
14.
15.
16.
17.
18.
19.
Fryba, L., (1999). Vibration solids and structures under moving loads, Thomas Telford House, London.
Takabataka H. (1998). Dynamic analysis of rectangular plates with stepped thickness subjected to moving loads
including additional mass, J Sound Vib 213:82942.
E. Ghafoori, M. Asghari, (2010). Dynamic analysis of laminated composite plates traversed by a moving mass
based on a first-order theory, Composite Structures 92 18651876.
Shadnam MR, Mofid M, Akin JE. (2001). On the dynamic response of rectangular plate with moving mass. ThinWall Struct39:797806.
Wu JS, Lee ML, Lai TS. (1996). The dynamic analysis of a flat plate under a moving load by the finite element
method. Int J Numer Meth Eng193:30714.
Yoshida DM, Weaver W. (1971). Finite element analysis of beams and plates with moving loads. Publ Int Assoc
Bridges Struct Eng 31:79195.
Wu, J.J., (2003). Use of equivalent beam models for the dynamic analyses of beamplates under moving forces,
Computers and Structures81 27492766.
Gbadeyan J.A., Oni S.T., (1995). Dynamic behaviour of beams and rectangular plates under moving loads,Journal
of Sound and Vibration, 182, 5, 677-695.
Renard J., Taazount M., (2002). Transient responses of beams and plates subject to travelling load: Miscellaneous
results, European Journal of Mechanics A/Solids 21, 301322.
Esen , Dynamic response of a beam due to an accelerating moving mass using moving finite element
approximation. Mathematical and Computational Applications 16, 1: 171-182, 2011.
Esen ., (2008). Dynamic analysis of overhead crane beams under moving loads. PhD Thesis, Istanbul Technical
University.
Gerdemeli ., zer D., Esen ., (2008). Dynamic analysis of overhead crane beam under moving loads. Proceedings
of the 5th. European Congress on Computational Methods in Applied Sciences and Engineering (ECCOMAS
2008), Venice, Italy.
Gerdemeli , Esen . and zer D., (2011). Dynamic Response of an Overhead Crane Beam Due to a Moving Mass
Using Moving Finite Element Approximation, Key Engineering Materials450: 99-102.
Szilard R. (2004). Theories and Applications of Plate Analysis, Wiley, New Jersey.
H. Bachmann, et al., (1995). Vibration Problems in Structures, Birkhauser Verlag, Berlin.
Bathe KJ. (1982). Finite element procedure in engineering analysis. Englewood Cliffs (NJ): Prentice-Hall.
Clough RW, Penzien J., (2003). .Dynamics of structures 3rd ed, Computers and Structures, Inc., Berkeley.
WilsonE.L., Static and Dynamic Analysis of Structures, (2002).Chapter 20: Dynamic analysis by numerical
integration, Computers and Structures Inc.,Berkeley.
Cifuentes A.O., (1989). Dynamic response of a beam excited by a moving mass. Finite Elements in Analysis and
Design, 5, 237 246.
Biography
smail Esenwas born in Gmhane region of Turkey in 1971.
Educational Background:
BSc: Istanbul Technical University, Mechanical Engineering Department 1991, Sakarya, Turkey.
MSc, Istanbul Technical University, Institute of Natural Sciences: Mechanical Engineering Department, construction and
manufacturing programme, 1994, Istanbul, Turkey.
PhD, Istanbul Technical University, Institute of Natural Sciences: Mechanical Engineering Department, construction
programme, 2009, Istanbul, Turkey.
He worked for private sector for 17 years until 2009. He designed various new industrial products and machines. He was
awarded, (Technology Development Award), by Istanbul Chamber of Commerce for his studies in new-product development
in 1994. He made many new-product standards for TSE (Turkish Standard Institution). He has worked as Assistant Professor
for Karabuk University since 2009. He teaches some lectures such as Mechanics of materials: Dynamics, Dynamics of
mechanical systems, Suspension system design, Strength of materials at the university in graduate and postgraduate
programs. His interests of study are the finite element analysis, Dynamic systems, the moving load problem, high-speed
transportation and production, vibration and vibration control.
762
Abstract.We analyzed dynamics of a beam that is its material behavior modeled with Kelvin-Voigt viscoelastic model. Kelvin-Voigt visco-elastic model is combination of dashpot and spring in parallel form.In most
researches, non linear damping, stiffness and inertia terms are usually omitted in the non-linear vibration
analysis of a visco-elastic beam. In this study, we show the effect of those types of non-linear terms in Kelvin
visco-elastic beam analyze. The direct perturbation method is used to examine the problem at the first time in
the literature. The non-linear mode shapes and natural frequencies of the beam have been found for three
types of boundary conditions. The numerical results that are the deflection of the beam and its nonlinear
frequencies for the all conditions are presented in graphs forms.
1 Introduction
Many materials have been produced with the development of material science. After elastic model has started
to fail to satisfy the expectations, many new viscoelastic models have been found to explain the behavior of the
material such as Kelvin-Voigt, Maxwell, Zener, Burgers Models etc. In addition, the damping properties of
viscoelastic materials are sometimes used to study vibration absorbing behavior.
There are several papers dealing with transverse vibrations of non-linear viscoelastic beams.[1]
The stability of viscoelastic homogeneous bars subjected to harmonically distributed loading is analyzed by
Suire and Cederbaum [2]. Additionally, Argyris et al. [3] established a mathematical model for a nonlinear
viscoelastic polymer beam and examined its stability. In the meantime Beldica and Hilton [4] analyzed the
bending of non-linear viscoelastic beams with small or large deformations in the presence of viscoelastic
piezoelectric devices placed on the upper and lower beam. In numerous papers presented by Nayfeh et al. [58],
the non-linear vibration of elastic beams is studied. The method of multiple scales is used in order to solve the
non-linear equations of motion of beams with different boundary conditions and besides they considered the
quadratic and cubic non-linearities. They examined the internal resonances associated with non-linear natural
frequencies of the beam. The method of multiple scales can be applied in order to analyze the problem which
includes viscoelastic properties and nonlinearities. There are several papers analyzed by multiple time scale
method to simulate nonlinear behavior of viscoelastic beam/string [9-12].
The method of multiple scales can be used in two different ways. In the first way, the method of multiple
scales is directly applied to the partial differential equations and boundary conditions, and it is called direct-
763
perturbation method. In the second way, the partial differential system is discretized first and then the method of
multiple scales is applied to the obtained set of ordinary differential equations. This application is called
discretization perturbation method. There are some papers talking about comparison of the direct approach and
the discretization approach in the analysis for non-linear vibrations of continuous systems in general models [913]. Also, similar studies on comparisons of these two methods are appeared for various mathematical models
[7-12]. The results of these analogies, the direct-perturbation method and the discretization-perturbation method
would be identical for infinite modes but the direct-perturbation method would yield better results compared to
discretization-perturbation method for finite mode truncations. Nayfeh and co-workers [14-16] have shown that
discretization of continuous systems can lead to incorrect results if the discretization is not performed by using a
complete set of shape functions that satisfy the boundary conditions. For some specific problems, in the context
of static buckling problems, discretization techniques can lead to qualitatively erroneous bifurcation diagrams. In
the direct approach, the principal advantage is the treatment of the boundary conditions at higher orders. In the
recent literature, the direct approach has been extensively used to investigate nonlinear vibrations.
In the paper presented by Mahmoodi et al. [1], a general formulation for the non-linear viscoelastic beams
with nonlinearities in terms of inertia, damping and stiffness is given. It is assumed that the beam is inextensible
and the viscoelastic system follows a classical linear viscoelastic model, i.e. KelvinVoigt model. It is
considered that the system will vibrate due to an initial velocity. The non-linear natural frequencies and mode
shapes are obtained by using discretization-perturbation method for finite mode truncations. This study is
devoted to show ability of application of direct multiple time scale method to given model and to avoid
qualitatively erroneous results. The method is used for three different types of support conditions. The non-linear
mode shapes and natural frequencies of the beam have been found by direct application of multiple time scale
method.
4 B = R S (4 B = R S (
The governing equation without external force is obtained by using energy method in Ref. [9]. In the model T
represents displacements and U shows the mass of the beam. Cis the damping coefficient. As is known, V is the
modulus of elasticity and is the moment of inertia. WX shows the external force and the other terms in Eq.(1)
represents the viscoelastic non-linear terms. Dot denotes differentiation with respect to time and the prime
denotes differentiation with respect to the spatial variable s. The boundary conditions are displayed by linear
operator that YZ shows the [th condition of \th support.
In this section the method of multiple time scales which is a perturbation method is chosen to solve the
system.
Firstly, the governing equation transform into non-dimensional form by using following non-dimensional
parameters;
M M = ]^B B = ]^S S = _
`a
Hbc
d e_
`Hbc
(2)
Then, non-dimensional governing equation and boundary condition are obtained that
764
BC & d B
EF
&B
EF
N N
f
f
& hBK i iB E & BKBC KLMLMj & B E B E B EE E E & d B EE & BK EE (
g
g
O
4 BR S (4 BR S (
Due to applying perturbation technique, weak non-linearity has to show as kl mk where m is a small
dimensionless parameter. The system is considered without external force. If the equation is rewritten, one can
obtain,
E
BC & 8B EF & B EF & IBK J JO B E & BKBC KLMLMP & B E B E B EE E E & 8B EE & BK EE ((4)
N
Substituting Eq. (5) into Eq. (1), and separating at each order of 8, one obtains;
8 O qrO BO & BOEF (
N N
f
7grO r BO 7 rO BOEF 7 hBOE i irO BOE & BOE rO BOE LMLMj
g
7
O
uvw ,x
yyyy
P (8)
BO M oO o n 8 sMI1: o t uvw ,x & 1
: o t
However that is the linear solution of the system.z functions are the mode shape functions
Substituting Eq.(8) into Eq.(6) the mode shape functions satisfy the following equation;
s EF M 7 6: sM ( (9)
765
Eq.(9) is solved by analytically for different boundary conditions which are hinged-hinged,
hinged-clamped and clamped-clamped. Shape functions and natural frequencies are obtained
for the first mode as below,
) )
Y Y
6 g
\}M 7 \}M
) )
Y Y
6 ((
\}M 7 \}M
Upon substitution of Eq.(8) to the right hand side of Eq.(7) yields below equation,
uvw ,x
yyyy
8 qrO B & BEF 7grO r Is: 1: t uvw ,x & 1
P 7 rO ;s:EF 1: t uvw ,x &
:t
N N
uvw ,x
uvw ,x
uvw ,x
yyyy
yyyy
yyyy
< 7 s:E 1: t uvw ,x & 1
J JO rO ;s:E 1: t uvw ,x & 1
< &
1
:t
:t
:t
s:E 1: t uvw ,x
uvw ,x
uvw ,x
yyyy
yyyy
&1
rO s:E 1: t uvw ,x & 1
LMLM 7 ;s:E 1: t uvw ,x &
:t
:t
E E
uvw ,x
uvw ,x
yyyy
yyyy
;s:EE 1: t uvw ,x & 1
< < (10)
1
:t
:t
N N
1: IJ JO s:E LMLMP 7
;7g6: s: r 1: 7 6: s:EF 1: & 6: s:E 1: yyyy
E
766
1: ((13)
r 1: & : 1: 7 : 1: yyyy
where
: JO s: s:EF LM(14)
:
uw
Eq.(13) is found. As its seen, : is pure imaginary, in other words real part of : is equal
to zero.
This suggestion is substituted into Eq.(13) and separating real and imaginary parts,
:E & : : (
7
(17)
: & .E (
(18)
CASE 1: The absence of material damping and the presence of non-linear inertia or stiffness
or combination of these two terms
The absence of material damping leads to : (. Then the solutions of : and : are
obtained as;
: O (19)
: 6: o & O (20)
6:
O (21)
767
0.8
0.8
0.6
0.6
a0
a0
0.4
0.4
0.2
0.2
0
4
10
18
19
20
w1
21
22
23
w1
hinged-hinged beam
hinged-clamped beam
It is observed that for all boundary conditions the frequency of the system shows softening behavior because
of the effects of nonlinear terms. Also it is noted that dynamic behaviors of the systems for all support conditions
are the same. Its expected results because support conditions have only effect on natural frequency values.
1
0.8
0.6
a0
0.4
0.2
0
13
13.5
14
14.5
15
15.5
w1
: o O t w , (23)
w ? w
& .O
(24)
768
w x u
f
w ,
w
R
O t
t
:
g
w
x
) w , w x
w
& ee : R & 8
(25)
Time history graphs for the first mode of vibrations are obtained for three different boundary conditions in Figs.
4-6. The middle of the beam span is chosen to show the displacement vibrations. Boundary conditions are caused
the difference of the time history graphs. One can note here, if the amplitude increases due to dependency on
support conditions, the damping velocity will be faster.
0.4
0.6
0.4
0.2
0.2
deflection
deflection
-0.2
-0.2
-0.4
-0.4
-0.6
-0.6
0
10
time
10
time
(gO ( and (g
0.6
0.4
deflection
0.2
-0.2
-0.4
-0.6
0
10
time
O ( and (g
769
beam
3 Conclusion
In this study, the governing equations of motion for nonlinear KelvinVoigt visco-elastic beams developed by
Ref. [9] are analyzed and the non-linear mode shapes and natural frequencies of the beam have been analytically
found. In these analyses, these different support conditions of the beam are taking into account of. The support
conditions are hinged-hinged, hinged-clamped and clamped-clamped. The numerical results that are the
deflection of the beam and its nonlinear frequencies for the all conditions are presented in graphs forms. These
results are given for two cases. In the first case, the nonlinear damping term are omitted. In the other case which
is the presence of non-linear stiffness or inertia or a combination of these two nonlinearities with nonlinear
damping are considered. It is observed that the nonlinear natural frequency is asymptotically approaches to a
constant value which is the linear natural frequency of the beam due to increasing amplitude.
Natural frequencies, amplitude and effects of non-linear frequencies are obtained.
It is found that, the non-linear natural frequencies are dependent on amplitude. Moreover, the mode shapes are
exponentially dependent on time and amplitude.
References
1.
2.
3.
4.
5.
6.
7.
8.
9.
10.
11.
12.
13.
14.
15.
16.
Mahmoodi, S. N.; Jalili, N.; Khadem, S. E. (2008) .An experimental investigation of nonlinear vibration and
frequency response analysis of cantilever viscoelastic beams. Journal of Sound and Vibration 311: 14091419.
Suire G.; Cederbaum G. (1995). Periodic and chaotic behavior of viscoelastic non-linear (elastica) bars under
harmonic excitations. International Journal of Mechanical Sciences 37: 75372.
Argyris J.; Belubekian V.; Ovakimyan N.; Minasyan M. (1996) Chaotic vibrations of a nonlinear viscoelastic
beam. Chaos, Solitons &Fractals 7: 15163.
Beldica CE.; Hilton HH. (2001). Nonlinear viscoelastic beam bending with piezoelectric controlanalytical and
computational simulations. Composite Structures 51: 195203.
Nayfeh AH.; Nayfeh SA. (1994).On non-linear modes of continuous systems. Journal of Vibration and Acoustics
116: 12936.
Nayfeh AH.; Chin C.; Nayfeh AS. (1995). Nonlinear normal modes of a cantilever beam. Journal of Vibration and
Acoustics 117: 47781.
Nayfeh AH.; Nayfeh SA. (1995). Nonlinear normal modes of a continuous system with quadratic nonlinearities.
Journal of Vibration and Acoustics 117: 199207.
Pakdemirli M.; Nayfeh AH. (1994). Nonlinear vibrations of a beam-spring mass system. Journal of Vibration and
Acoustics 116: 4339.
zhan, B. B.; Pakdemirli, M. (2010). A general solution procedure for the forced vibrations of a system with cubic
nonlinearities: Three-to-one internal resonances with external excitation. Journal of Sound and Vibration 329:
26032615.
Pakdemirli, M. (2003). Comparison of higher order versions of the method of multiple scales for an odd nonlinearity problem. Journal of Sound and Vibration 262: 989998.
Fung, R.F.; Huang, J.S.; Chen, Y.C.; Yao, C.M. (1998). Nonlinear dynamic analysis of the viscoelastic string with
a harmonically varying transport speed. Computers & Structures 66(6): 777784.
Fung, R.F.; Huang, J.S.; Chen, Y.C. (1997). The transient amplitude of the viscoelastic traveling string: an integral
constitutive law. Journal of Sound and Vibration 201(1): 153167.
Boyac, H.; Pakdemirli, M. (1997). A comparison of different versions of the method of multiple scales for partial
differential equations. Journal of Sound and Vibration 204(4): 595-607.
Pakdemirli, M.; Nayfeh, S.A.; Nayfeh, A.H. (1995) Analysis of one-to-one autoparametric resonance in cablesDiscretization vs. direct treatment. Nonlinear Dynamics 8: 65-83.
Nayfeh, A.H.; Nayfeh, J.F.; Mook, D.T. (1992) On methods for continuous systems with quadratic and cubic
nonlinearities. Nonlinear Dynamics 3: 145-162.
Nayfeh, A.H.; Nayfeh, S.A.; Pakdemirli, M (1995) On the discretization of weakly nonlinear spatially continuous
systems. NonlinearDynamics and Stochastic Mechanics (N.S. Namachchivaya and W. Kliemann editors) Boca
Raton: CRC Press: 064_199.
770
Biographies
Nurhan Hacolu Nurhan Hacolu was born in zmir in 1988. She graduated from departments of
mathematics Ege University in zmir in 2010. She is respectively still a graduate student and a student in Celal
Bayar University and Anadolu University in Manisa and Eskiehir. She is studying mathematical
methods,mechanical engineering and dynamic analyses of structure. Nurhan Hacolu knows English as a
foreign language.
Two papers were accepted by Gediz University and will be presented in June.
Smeyye Snr Smeyye Snr was born in Fethiye/Mula in 1986. She graduated from departments of
mathematics and physics at Sleyman Demirel University in Isparta in 2008. She is still a graduate student in
Celal Bayar University in Manisa. She is studying fluid structure interaction, mathematical methods, and
dynamic analyses of structure. Smeyye Snr knows English as a foreign language.
She has two articles in national and international conferences. Three papers were accepted by Gediz University
and will be presented in June.
B. Gltekin Snr B. Gltekin Snr was born in Korkuteli/Antalya in 1971. He graduated from Dokuz Eyll
University in zmir in 1994. He respectively completed his graduate and doctoral dissertations in 1996 and 2004
at Celal Bayar University in Manisa and Dokuz Eyll Universty. He has been assistant professor doctor in Celal
Bayar University for two years. He is studying mechanical engineering, fluid structure interaction, dynamic
analyses of structures, mathematical methods, coastal and harbor engineering. B. Gltekin Snr knows English
as a foreign language.
He has one pressed, an international article, eleven in national and international conferences. Three paper was
accepted by Gediz University and will be presented in June. He completed different scientific research projects
in TBTAK.
771
Abstract. This paper summarises crashworthiness in the world. We are going to see how crash test make and
which agencies make crash test. New Car Assessment Programs (NCAP) in Australia, Europe, Japan, and
USA take majority of this mission. Crashworthiness simulations are made by using FEA programs. There are
many analyses programs but in this paper Abaqus and Ls-dyna are compared, their similarity and differences
are analysed. Abaqus has been adopted by major corporations across all engineering disciplines as an integral
part of the design process. Ls-dyna is a nonlinear dynamic structure analysis program.
1. INTRODUCTION
Crashworthiness is the ability of a structure to protect its occupants during an impact. This is commonly
tested when investigating the safety of aircraft and vehicles. Depending on the nature of the impact and the
vehicle involved, different criteria are used to determine the crashworthiness of the structure. Crashworthiness
may be assessed either prospectively, using computer models or experiments, or retrospectively by analyzing
crash outcomes. Several criteria are used to assess crashworthiness prospectively, including the deformation
patterns of the vehicle structure, the acceleration experienced by the vehicle during an impact, and the
probability of injury predicted by human body models. [1] The vehicle structure should be sufficiently stiff in
bending and torsion for proper ride and handling. It should minimize high frequency fore-aft vibrations that give
rise to harshness. In addition, the structure should yield a deceleration pulse that satisfies the following
requirements for a range of occupant sizes, ages, and crash speeds for both genders: Deformable, yet stiff, front
structure with crumple zones to absorb the crash kinetic energy resulting from frontal collisions by plastic
deformation and prevent intrusion into the occupant compartment, especially in case of offset crashes and
collisions with narrow objects such as trees. Short vehicle front ends, driven by styling considerations, present a
challenging task to the crashworthiness engineer. Deformable rear structure to maintain integrity of the rear
passenger compartment and protect the fuel tank . Properly designed side structures and doors to minimize
intrusion in side impact and prevent doors from opening due to crash loads. Strong roof structure for rollover
protection . Properly designed restraint systems that work in harmony with the vehicle structure to provide the
occupant with optimal ride down and protection in different interior spaces and trims . [1]
The goals of crashworthiness are provide occupant protection by maintaining integrity of the passenger
compartment, optimized vehicle structure that can absorb the crash energy by controlled vehicle deformations,
provide important information regarding the safety performance of the vehicle in the traffic environment.[1]
Australian NCAP (ANCAP) commenced in 1992. ANCAP has recognised the global nature of the car
manufacturing industry and has developed the testing program to align with existing consumer crash testing
programs operating overseas and to suit the types of crashes resulting in fatalities and serious injuries in
Australia. The marked improvements in protection for occupants in passenger cars that are involved in frontal
collisions in Australia shows that ANCAP has been very successful. [2]
772
Since 1996, the Euro NCAP programme has been testing the (crash) safety of the most widely sold cars.
The fact that consumers are familiar with the test results encourages the car manufacturers to produce cars that
are often safer than is required by law. Euro NCAP tests the (secondary) crash safety for adult occupants, for
child occupants and for pedestrians as crash opponents. Since 2009, points can also be earned for the presence of
devices for the prevention of crashes (primary safety), such as electronic stability control and speed limiters.[4]
The National Organisation for Automotive Safety and Victims Aid (OSA) was established by the
Japanese Government in 1973. OSA is funded yearly by the Ministry of Transport but is a separate legal entity.
The primary objectives of OSA are to reduce road trauma through research and to offer assistance to victims of
vehicle accidents. An increase in the road accident fatality rate in 1988 resulted in the Japan NCAP being
established in 1991. [3]
NHTSA is part of the US Department of Transportation and has been conducting NCAP testing using
the full frontal test at 56 km/h since 1978. NHTSA employs four contractors across the US to conduct its NCAP
testing. NHTSA uses a star system to provide consumers with vehicle safety performance information. [2]
Abaqus, is founded in 1978, is the worlds leading provider of software and services for advanced finite
element analysis. The Abaqus software suite has an unsurpassed reputation for technology, quality and
reliability. It has been adopted by major corporations across all engineering disciplines as an integral part of the
design process. Abaqus offers a powerful and complete solution for simple to complex linear and nonlinear
engineering problems, using the finite element method. A wide range of structural, thermal, dynamic and
coupled analyses is supported. The software delivers a unified simulation environment without equal, presenting
a compelling alternative to implementations involving multiple products and vendors. In October 2005,
ABAQUS became a wholly owned subsidiary of Dassault Systmes, a world leader in 3D and Product Lifecycle
Management (PLM) solutions. SIMULIA, a Dassault Systmes brand, offers realistic simulation solutions based
on an open multiphysics platform allowing a unified approach. [3]
Ls-dyna is a nonlinear dynamic structure analysis program developed by Livermer Software
Technology Corp. (LSTC) in the USA. It can analyze large deformation behaviour in the structures by the
explicit time integration method it is widely used in diverse fields car crash analysis to dropped cell-phone
impact analysis. [4] For engineers in product design analysis settings, todays competitive business climate
means delivering more complex analysis in less time than ever before. Nowhere is this more critical than in the
area of automotive crash simulation. Over 60 standard simulation cases are required by regulatory agencies, and
many others are now considered essential business requirements. To meet these growing needs, Livermore
Software Technology Corporation (LSTC) delivers leading-edge crash simulation with its LS-DYNA solution.
[10]
Using crash simulation models, the multicore performance of the newly developed hybrid Ls-dyna is
usually investigated, a method whose speed up arises from both shared memory and massage passing
paralellisms. Theoretically, the hybrid method gains performance advantages over the traditional , massage
passing parallel (MPP) ls- dyna for two reasons: first, the addition of shared memory parallelism to the massage
passing parallelism reduce the number of massage and theri sizes dramatically, which in turn redeces latency and
band width requirements on interconnect. Second, the same addition enhances spatial and temporal localities for
both code and data accesses, which in turn allows the size limited cache to work more efficiently. Armed whit
this theory, we caracterize performance of the hybrid method with respect to problem size, core count, core
placement, and interconnect speed; thus provide users guidance on when and how to use the hybrit method
efficiently.
2. LITERATURE REVIEW
2.1 NCAP PROGRAMS
ANCAP has recognised the global nature of the car manufacturing industry and developed a crash
testing program to align with existing overseas programs.
In addition to the full frontal test, an offset test using a deformable barrier was added in 1994. This test
was developed for the European Experimental Vehicle Committee (EEVC) by the Transport Research
773
Laboratory (TRL) in the UK. The test is recognised internationally and the barrier design is specified for both
consumer crash test programs and for compliance with regulatory standards.
Each vehicle model tested in ANCAP undergoes a full. frontal crash test into a solid concrete barrier
and an offset crash test into a barrier with a deformable aluminium face. The full frontal crash test simulates
hitting a solid object or another vehicle exactly head-on and is conducted at a speed of 56 km/h. In this test the
impact is spread evenly across the front of the vehicle. This test mainly evaluates the vehicles restraint system.
The offset crash test simulates hitting another car and is conducted at a speed of 64 km/h. Forty percent
of the width of the car makes contact with the barrier. In this test the crash forces are concentrated on the drivers
side of the vehicle. This test mainly evaluates the vehicle structures resistance to intrusion.
The ANCAP ratings based on data provided by overseas organisations might differ from the ratings
assigned by these organisations. In particular the ANCAP rating includes assessment of the results of full-frontal
crash tests and takes into account the passenger injury measures and restraint performance in these tests. The
ANCAP ratings also tend to place less emphasis on foot well intrusion and lower leg injury than IIHS ratings and
more emphasis on structural performance than the Euro NCAP procedures.
ANCAP consults with the automotive industry about the program through the Federal Chamber of
Automotive Industries (FCAI), the group which represents the vehicle manufacturers and importers in Australia.
Representatives from the vehicle manufacturing companies are invited to attend the test of their products and are
able to review the results before publication. ANCAP meets with the FCAI before each public launch and in
early 1997 conducted a technical briefing for FCAI members on the new IIHS-style rating system. ANCAP is
also working closely with the Federal Government through the Department of Transports Federal Office of
Road Safety (FORS) which administers the design standards for Australian vehicles. [2]
Euro NCAP officially determines which car is tested in which laboratory. Governments as well as car
manufacturers can suggest cars for testing. In the latter case, Euro NCAP will test three of the twenty vehicles
(of the same model) that the manufacturer suggests as options. Euro NCAP groups the cars in various model
classes, such as passenger car (small and large), multiple purpose vehicle (MPV; small and large), SUV
(small and large), sports car and pick-up truck. Within these categories cars are only allowed to be compared
with each other when they differ less than 150 kg in weight.
By law, all new car models have to meet specific safety requirements (ECE regulations and EC directives)
before they are allowed on the road. However, the European c.q. Dutch legislation only sets a number of
minimumrequirements that the secondary (crash) safety of new cars has to meet also known as passive safety or
injury prevention). It is Euro NCAPs aim to encourage car manufacturers to exceed these minimum (crash)
requirements in the interest of both the car occupants and the other traffic users, among whom pedestrians.
Early 2009, the Euro NCAP tests and assessment system were drastically revised. Gives an overview of
the various Euro NCAP subtests, divided into four groups since 2009. Group 1 looks at secondary safety of adult
occupants, group 2 at that of child occupants and group 3 at that of pedestrians as crash opponents of a car, and
group 4 investigates the presence of a number of intelligent safety devices. Besides seat belt reminders, these
also include a number of primary safety devices, that is to say, for the prevention of crashes (earlier known as
active safety). Prior to 2009, these types of safety devices were not part of the Euro NCAP assessment.
The European car models have become much safer during the last few decades. Especially the stronger
cage construction of European car models protects occupants increasingly better during a frontal collision.
Nevertheless, there are limitations. At present, Euro NCAP does not allow for mutual mass differences in frontal
car-car collisions (incompatibility), whereas this in particular is a very determining factor in the further outcome
of a crash. Another phenomenon is that heavier cars have also become more unyielding (less shock absorbing)
and therefore are at an advantage in a crash with a lighter car in terms of protection of the occupants .It is
therefore important to set high requirements to the crash friendliness (energy absorption) of the fronts of cars and
the strength and the design of cage constructions.
Due to both mass and rigidity, safety for the occupants increases with vehicle mass, whereas safety for
the occupants of the crash opponent car decreases with vehicle mass. As long as Euro NCAP does not test this
incompatibility, the number of stars gives good insight into the safety within the same model and size class, but
not between the various classes. The above incompatibility problem is all the more important because of the
774
trend to make cars smaller and more lightweight for reasons of environmental targets. As a result, Euro NCAP
and various other road safety organisations are now discussing how to deal with these differences in mass
between cars. It is more effective for road safety to prevent crashes than to reduce the severity of injuries of car
occupants in crashes .The attention paid by Euro NCAP to primary safety devices, such as an ESC-system, is
therefore positive. It is, however, important to keep secondary safety at at least the same level. The new Euro
NCAP assessment now awards extra points when this kind of safety device is installed. However, until now no
test methods have been available to determine the actual safe functioning, which is a limitation.
OSA has a Car Safety Assessment Committee with members from the Japan Automobile Research
Institute (JARI) and Japan Automobile Manufacturers Association (JAMA). It also has regular discussions with
JAMA members on NCAP related issues.
OSA conducts its full frontal crash test program at JARI. The test is the same as for US NCAP. The
numerical test results are not published as OSA believes they are of little interest to consumers. A four category
rating system (A/B/C/D) based on head injury criterion and chest acceleration is used.
Recently the A category was split into A, AA and AAA levels to further discriminate vehicle safety
performance.
The 1996 program tested the top selling nine vehicles across all classes and results were published early
in 1997. A total of 18 models have now been reported on. OSA has tested 13 more passenger vehicles during
1997. Results were published in April 1998. OSA believes its options to provide more vehicle safety data to
consumers include:
- conduct more tests (subject to budget constraints);
- provide overseas NCAP data on vehicles which are sold in Japan; and/or
- provide vehicle manufacturer test data.
OSA is reproducing US NCAP data on similar vehicles sold in Japan. OSA believes that left hand drive
data can be used for right hand drive vehicles if the vehicle manufacturer agrees the vehicles are similar.
Japan will have a national regulation for side impact this year. This will be based on the European test.
OSA is interested in conducting a side impact test and is reviewing the UK DOT test. OSA is likely to increase
the test speed over the regulatory speed by about 10 percent which is intended to show crash performance
differences between vehicles. OSA expects to take up to three years to develop this test, which will be discussed
through the Car Safety Assessment Committee. OSA has agreed to exchange crash test information with
ANCAP. [2]
775
Companies face a daunting array of product design and manufacturing challenges today. They need to
deliver innovation and quality while also cutting product cost and getting to market faster. This puts pressure on
automotive engineering design analysis teams to study more options and tests in a shorter amount of time,
including a wide range of crash simulations. Growing rapidly in size and sophistication over the past two
decades, crash simulation has become part of the critical design path for the industry an important decisionmaking tool. In addition to those tests mandated by government, insurance companies need simulations to
evaluate risks, including specific car-to-car and car-to-passenger cases. To be of greatest value, all cases must be
simulated in both the design and post-design phases, and repeated anytime a model is modified. Whats more, a
number of automotive manufacturers are using superior passenger safety in collisions as a differentiator
andpowerful marketing element. This is yet another reason for companies to make greater use of crash
simulations in vehicle design. In servers alone, automotive companies have roughly 50 teraflops of installed
computing power, more than 60 percent of which is used for crash simulation. The key to enabling more timely
and extensive simulation is making effective use of these computational resources through fast, accurate and
affordable engineering solutions. [10]
Using crash simulation models, the multicore performance of the newly developed hybrid Ls-dyna is usually
investigated, a method whose speed up arises from both shared memory and massage passing paralellisms.
Theoretically, the hybrid method gains performance advantages over the traditional , massage passing parallel
(MPP) ls- dyna for two reasons: first, the addition of shared memory parallelism to the massage passing
parallelism reduce the number of massage and theri sizes dramatically, which in turn redeces latency and band
width requirements on interconnect. Second, the same addition enhances spatial and temporal localities for both
code and data accesses, which in turn allows the size limited cache to work more efficiently. Armed whit this
theory, we caracterize performance of the hybrid method with respect to problem size, core count, core
placement, and interconnect speed; thus provide users guidance on when and how to use the hybrit method
efficiently.[6]
3. METHODS
To assess a vehicles frontal impact crashworthiness an integrated set of test procedures is required that
assessesboth the cars self and partner (compatibility) protection. It has been recommended by the International
Harmonisation of Research Activities (IHRA) frontal impact group that the set of test procedures should contain
both full overlap and offset tests.
The European Enhanced Vehicle-safety Committee (EEVC) WG15 has helped co-ordinate work in
Europe to understand and develop a set of test procedures to improve a vehicles frontal impact crash
performance. It has found that the main factors influencing a vehicles compatibility are its structural interaction
potential, its frontal force levels and its compartment integrity. In 2007, EEVC WG15 made a number of
proposals for potential sets of test procedures, all of which contain both full width and offset tests . These were:
Set 1
- Full Width Deformable Barrier (FWDB) test to assess a vehicles structural interaction potential and
provide a high deceleration pulse to test the restraint system.
- Offset Deformable Barrier (ODB) test with EEVC barrier to assess a vehicles compartment integrity
and frontal force levels and also provide a softer deceleration pulse to test the restraint system.
Set 2
- Full Width Rigid Barrier (FWRB) test to provide a high deceleration pulse to test the restraint system.
- Progressive Deformable Barrier (PDB) test to assess a vehicles structural interaction, frontal force
levels and compartment integrity and also provide a softer deceleration pulse to test the restraint system.
Set 3
776
777
test.
The MDB consists of an IIHS deformable aluminum barrier and the cart to which it is attached. The
crash cart is similar to the one used in Federal Motor Vehicle Safety Standard (FMVSS) 214 side impact testing
but has several modifications .The wheels on the cart are aligned with the longitudinal axis of the cart (0
degrees) to allow for perpendicular impact. [9]
REFERENCES
Bois P. , Chou C. C. , Fileta B. B. , Khalil T. B. , King A. I. , Mahmood H. F. , Mertz H. J. , Wismans J.
Vehicle Crashworthiness and Occupant Prodection American Iron and Steel Institute 2000 Town Center.
2) Case M., Griffiths M., Haley J., Paine M., NRMA Ltd., RACV INTERNATIONAL NCAP PROGRAMS IN
REVIEW Vehicle Design & Research Australia , Paper Number 98-S 11 -O-O3
3) ABAQUS Software Adopted by Major Automotive OEM for Crash Simulation Four-year technical
collaboration signals a major technology shift in automotive engineering Providence, R.I. USA, October 25,
2005
4) Euro NCAP, a safety tool SWOV Fact sheet
5) Andreas Hellman, Simulation of complete vehicle dynamics using FE code Abaqus
MASTER OF SCIENCE PROGRAMME Mechanical Engineering
Lule University of Technology Department of Applied Physics and Mechanical Engineering Division of
Functional Product Development 2008:084 CIV ISSN: 1402 - 1617 ISRN: LTU - EX - - 08/084 - - SE
6) Lin Y. Y., Wang J. Performance of the Hybrid Ls-dyna on Crash Simulation with the Multicore
Architecture
7) Mervyn Edwards, DEVELOPMENT OF A HIGH DECELERATION FULL WIDTH FRONTAL IMPACT
TEST FOREUROPE TRL, United Kingdom, On behalf of the APROSYS SP1.2 consortium,Paper Number
09-0105
8) Meyerson S.L., Zuby D.S.,Lund A.K. Insurance Institute for Highway Safety United States
9) Insurance Institude for High Way Safety, www.iihs.org
10) White Paper, Livermore Software Technology Corporation: Leading-edge Crash Simulation Solutions
1)
778
Abstract: In this study, the contribution of the stucco reinforced with polypropylene fibers to the strength of
the masonry wall was investigated, and then the contribution of the masonry wall reinforced with fibrous
stucco to the lateral behavior of the steel structures was determined. Mechanical and material properties of the
normal and fibrous stucco were found by experimental study. Result of the experiments indicated that the
shear strength of fibrous stucco is 2.75 times larger than normal stucco. In analytical analysis, masonry walls
were modeled as a truss member in diagonal direction in the frame, in order to carry and transfer lateral loads.
In the study 2, 4 and 6 storey steel structures were analyzed with the incremental static pushover method.
Since shear strength of masonry walls is lower than compressive strength, the limit shear strength was
accepted as the maximum bearing capacity of the masonry walls. Pushover analyses were conducted in
SAP2000 program. As a result of the analyses, the contribution of the retrofitted masonry walls to the lateral
behavior of the steel structures under earthquake loads was tried to be interpreted.
Keywords: Masonry Wall, Stucco, Fiber, Seismic Performance, Steel Structures
1 Introduction
In structural analysis of structures effect of rigidity of masonry partition walls are ignored only the weight of the
partition walls are taken into account. This assumption is accepted by many building codes because the structural
behaviors of infill masonries under lateral loads cannot be explained exactly. Detailed information about
masonry in fills may found in literature 1 to 5.
In this study, the contribution of masonry wall retrofitted with polypropylene fiber stucco in earthquake behavior
of the steel structures was investigated. In the frame of study, first of all, the contribution of the stucco reinforced
with fibers to the strength of the masonry wall was investigated, then the contribution of the masonry wall
reinforced with fibrous stucco to the lateral behavior of the steel structures was investigated. Mechanical and
material properties of the normal and fibrous stucco were determined by experimental studies. In the study 2, 4
and 6 storey steel structures are analyzed with pushover method. Pushover analyses are conducted with using
SAP2000 V14 program.
2 Methods
2.1 Experimental Study
A unique full-scale study was undertaken to consider the behavior of fibrous stucco partition walls in structural
mechanic Laboratory of University of Celal Bayar. Each specimen was tested in a 250 KN capacity rigid frame
press. Results of experimental study were gathered by using a load cell, 5 LVDTs and 8 channel data accusation
system. Using CODA program the data gathered from the data accusation system was transferred to the
779
computer. Wall specimens turned 45 for tensile splitting strength test. The specimens loaded to fail using load
controlled hydraulic piston (Fig. 1). Six specimens were used to determine the shear and compression strength
capacities of infill walls (Table 1).
Table 1. Test Methods and Specimens
Identification Codes
of Prepared Test
Specimens
Identification Codes
of Prepared Test
Specimens
NTS-D1
LNTS-D1
NTS-D2
LNTS-D2
NTS-D3
LNTS-D3
NTS-D4
Compression Test
LNTS-D4
NTS-D5
Compression Test
LNTS-D5
NTS-D6
Compression Test
LNTS-D6
Test Method
Tensile Splitting
Strength Test
Tensile Splitting
Strength Test
Tensile Splitting
Strength Test
Test Method
Tensile Splitting
Strength Test
Tensile Splitting
Strength Test
Tensile Splitting
Strength Test
Compression
Test
Compression
Test
Compression
Test
Hydraulic
Piston
Steel
Apparatus
Load Cell
Specimen
45o
Rigid Base
Profile
50
In the experimental study, fibrous stucco which has 2 % fiber by volume and normal mortar applied to the square
shape masonry partition wall specimens. Then 4 LVDTs connected to masonry wall specimen. Specimens
turned over 45 and axial load applied for diagonal tensile failure. For compression test, axial distributed load
applied to specimens.
A-A
A
During
Loading
vertical
and
LVDT
horizontal
displacements were
1
3
1
gathered. As a result
bearing capacity and
modulus of elasticity
4
2
2
of specimens were
determined (Fig. 2-3).
A
0
50
780
40000
NSW-D1
NSW-D2
NSW-D3
35000
25000
Load (N)
Load (N)
30000
20000
15000
10000
5000
0
0
2
3
4
Displacement (mm)
130000
120000
110000
100000
90000
80000
70000
60000
50000
40000
30000
20000
10000
0
PFSW*D1
PFSW-D2
PFSW-D3
0 2 4 6 8 10 12 14 16 18 20 22 24 26 28 30 32
Displacement (mm)
a) normal stucco
b) fibrous stucco
Fig.4. Load-Displacement diagram of diagonal tensile (Tensile splitting test) tests on masonry wall specimens
Table 2. Maximum Load-Displacement values of diagonal tensile tests
Identification
Code
NSW-D1
NSW-D2
NSW-D3
PFSW-D1
PFSW-D2
PFSW-D3
36095
35328
35567
87710
119669
87522
3.838
3.995
3.768
17.305
19.484
16.110
As seen from Table 2 experimental results show that fibrous stucco has positive effect to the ductility and
bearing capacity of the specimens. While normal stucco specimens has brittle behavior fibrous stucco specimens
has ductile behavior (Fig. 5) and the bearing capacity of the specimens has increased about 275%.
53619
47219
59316
119015
121232
125673
7.703
6.921
7.750
11.20
12.80
12.65
781
diagonal compressive bar element assumption. As the lateral load and displacement increase gab occurs around
masonry wall and column-beam intersection region. And this turns the wall behave like a diagonal compressive
element. The occurred equivalent virtual diagonal compressive element in masonry wall under large
displacement is shown in Fig. 6.
Using equations of FEMA and Mainstone (1974) infill masonry walls transformed to equivalent virtual diagonal
compressive bars (Eq.1-3).
d
t
Wef
Em
Es
R
H`
H
L`
L
Ic
: Diagonal length
: Widt of masonry wall
: Effective wal width
: Modulus of elasticity (Masonry)
: Modulus of elasticity (Frame)
: Bearing capacity
: Length of masonry wall
: Story height
: Net span width
: Span width
: Angle of diagonal compressive bar
: Moment of inertia of columns
Fig.6. Diagonal compression region in masonry infill wall under lateral load and equivalent virtual diagonal
compressive bar element that represents the masonry wall
d = H 2 + L2
w = 0.175(1 H)
0.4
(1)
H +L
E t sin 2
1 = m
4 E s Ic h
(2)
1
4
(3)
z=
4 E s Ic H`
4
2 E m t sin 2
(5)
Withal, in the literature maximum strain of equivalent bar element is assumed as u=0.01d and associated with
the force-displacement diagram given above. In this study values of y and u were gathered from the
experimental results.
782
Experimental test results showed that compressive strength is higher than shear strength and the bearing capacity
of infill walls were overtaken by cracks occurred because of shear stress. Therefore in the analytical study the
values of tensile splitting strength test results were used given in Table 4.
R (N)
36095
119669
y (mm)
3.838
19.484
u (mm)
5.087
34.367
3 Analytical Study
3-spanned in both directions 2, 4 and 6 storey steel buildings (Fig. 8) were investigated to determine the effect of
infill walls to seismic behavior of steel structures. The span lengths of the structures are 700 mm in both
directions. The columns and beams are formed from U 120 and U100 steel bars, respectively. The width of the
stucco in fill walls are chosen 140 mm and the density of in fill walls are assumed to be 1.4 t/m. In the analyses
a commercial program, SAP 2000 is used. Plastic hinges are created and assigned to diagonal compressive bar
elements end points and these hinges bear only axial forces. Force-plastic deformation diagram of plastic hinges
were determined from experimental studies (Table 4). Identification codes of frames are given below
PFSW-F : Frames with polypropylene fibrous stucco masonry walls
NSW-F : Frames with normal stucco masonry walls
WMW-F : Frames without masonry walls
800
700
A-A
dimensions in mm
700
A-A
700
800
700
700
800
700
700
700
700
700
700
700
A-A
Fig.8. Plan of the buildings and A-A sections of 2, 4 and 6 storey buildings
Collapse cases of the structures were investigated using performance-points of the structures by means of nonlinear pushover analysis results and the results are presented in Table 5.
Table 5. Collapse case of the structures at the performance-point
Building
2 Storey
4 Storey
6 Storey
PFSW-F
%100 Collapse
% 50 Collapse
%50 Collapse
NSW-F
%100 Collapse
%100 Collapse
%100 Collapse
PFSW-F
Beam
Column
0.001025
0.022650
0.001587
0.004204
0.003926
0.010300
NSW-F
Beam
Column
0.001670
0.004629
0.002137
0.005713
0.004633
0.011800
WMW-F
Beam
Column
0.004132
0.009800
0.005030
0.011100
0.006098
0.013200
783
3500
3000
2500
2500
3000
2000
1500
1000
2000
1500
1000
500
500
2 stories
0
0.00 0.01 0.02 0.03 0.04 0.05 0.06 0.07
Displacement (m)
3000
4 stories
0
0.00
0.02
0.04
0.06
Displacement (m)
0.08
0.10
2500
2000
1500
1000
500
6 stories
0
0.00 0.02 0.04 0.06 0.08 0.10 0.12 0.14 0.16
Displacement (m)
WMW-F
NSW-F
PFSW-F
5 Conclusions
As a result of the study, experiments indicated that the bearing capacity of fibrous stucco is 2.75 times larger
than normal stucco. Normal stucco infill walls fail as a brittle material, while fibrous stucco infill walls fail as a
ductile material and sudden failure is not observed (Fig. 4, 5 and 6). The effect of infill walls are ignored in most
of the earthquake codes. In this study it is shown that NSW-F type frames have 20 % more bearing capacity than
WMW-F type frames. The bearing capacity of PFSW-F type frames has 35 % more bearing capacity than
WMW-F type frames (Fig. 10). As the storey height of the structure increases, the positive effects of infill
masonry walls increase. Demir et al. [11] has shown that these proportions are 40 % for WMW-F type and %
200 for PFSW-F type in reinforced concrete buildings. The reason of this difference is assumed to be because of
high ductility of steel structures. Effect of stucco type to the rotation angle of column-beam elements is
investigated. Maximum values of rotation angle of beam elements are observed in WMW-F type frames and
minimum values of rotation angle of beam elements are observed in PFSW-F type of frames (Table 8). Infill
walls has positive effects to structural system by increasing the lateral rigidity of the frame
As a conclusion it is seen that fibrous stucco infill walls increase the bearing capacity and the ductility of the
structures. Therefore, the effects of infill walls may not be ignored to make more economical designs in
earthquake codes.
784
References
1. Trk, M., Ersoy, U., and zcebe, G. (2003). Betonarme erevelerin betonarme dolgu duvarlarla depreme kar onarm ve
glendirilmesi. Beinci Ulusal Deprem Mhendislii Konferans, 26-30 Mays, stanbul.
2. Ersoy, U. (1996). Seismic rehabilitationapplication research and current needs. 11th World Conference on Earthquake
Engineering, Acapulco, Mexico (invited paper).
3. Ersoy, U. and Tankut, T. (1992). Hysteretic response of RC infilled frames. ASCE Journal of Structural Eng.1(18):8.
4. Erkaya, A. (1996). Betonarme ereve yaplarda dolgu duvarlarn deprem davranna etkisi. Master Thesis, Istanbul
Technical University, stanbul.
5. Kaltakc, M.Y., Kken, A., Arslan, M.H., and Korkmaz, H.H. (2007). Dolgu duvarl elik erevelerin yatay ykler
altndaki davrannn deneysel ve dorusal olmayan analitik yntemlerle incelenmesi. 2. elik Yaplar Sempozyumu,
Eskiehir, 10-11 Mays (Tam metin olarak kitapkta yaymland).
6. Kaltakc, M.Y. and Arslan, M.H. (2005). Tayc olmayan tula dolgu duvarlarn yap davran katsaysna olan etkisinin
incelenmesi. Deprem Sempozyumu, Kocaeli, Trkiye, 23-25 Mart, 598-605.
7. FEMA-306, (2005). Repair of Earthquake Damaged Concrete and Masonry Buildings-Basic Procedures Manual. Federal
Emergency Management Agency, Washington.
8. FEMA-356, (2005). Presentandard and Commentary for the Seismic Rehabilitation of Buildings. Federal Emergency
Management Agency, Washington.
9. Mainstone, R.J. (1974). Suplementary note on the stiffness and strengths of infilled frames. Current Paper 13/74, Building
Research Station, UK, Feb.
10. SAP2000, integrated finite element analysis and design of structures basic analysis reference manual, Computers and
Structures Inc. Berkeley (CA, USA).
11. Demir, A., Ercan, E., Arsoy, B., Nuholu, A., Kozanolu, C. (2011). Dolgu Duvar Svalarnn Yaplarn Deprem
Davranna Etkisi. 7.Ulusal Deprem Mhendislii Konferans, stanbul, Trkiye, 30 May.
BIOGRAPHIES
Ali Demir Ali Demir was born in Karabk in 1983. He graduated from Celal Bayar University in 2005. He completed his
graduate thesis in 2008. He is still a doctoral student in Celal Bayar University. He has been an research assistant in Celal
Bayar University for 6 years. He is studying concrete structures, performance evaluation of buildings, experimental study. Ali
Demir knows English as a foreign language.
He has three pressed, five national articles, eleven articles in national and international conferences. Four papers were
accepted by Gediz University and will be presented in June. He also has two accepted presentations and will be presented in
7UDMK in June. He completed different scientific research projects in Celal Bayar University, TBTAK and he has started
two new projects last year.
Research Assistant Ali Demir is a member of Turkish Chamber of Civil Engineers since 2005.
Emre Ercan Emre Ercan was born in zmir in 1977. He graduated from stanbul Techinical University in 2001. He
completed his master thesis in 2004 and doctorate thesis in 2010 in Ege University. He has been an research assistant in Ege
University since 2004. He is studying Impact-Echo, Operational Modal Analysis, FEM and steel structures. Emre Ercan
knows English as a foreign language.
He has two published articles. Two papers were accepted by Gediz University and will be presented in June. He
completed four scientific research projects in Ege University and TBTAK and he is starting one new projects this year.
Research Assistant Emre ERCAN is a member of Turkish Chamber of Civil Engineers since 2001.
Bengi Arsoy Bengi Arsoy had Bachelor of Science degree in Civil Engineering from Dokuz Eyll University in 1987. She
had her Master degree in 1992 from same university and had her PhD degree in 2002 from Wayne State University. She
worked as research assistant for seven years in Faculty of Engineering and Architecture of Dokuz Eyll University, as
specialist for three years in City of Detroit, as lecturer for one year in Faculty of Engineering of Dokuz Eyll University. She
has been working as Assistant Professor since 2005 in Faculty of Engineering of Ege University. She is currently working on
use of fiber reinforced concrete in structural rehabilitation, strengthening and retrofitting of reinforced concrete structures.
785
Abstract. In conventional methods the structures are assumed fixed to the soil and soil-structure
interaction (SSI) is ignored. Many design codes have suggested that the effect of SSI can reasonably be
neglected for the seismic analysis of structures. In this study, SSI of reinforced concrete structures were
investigated with nonlinear analysis methods. Two equivalent reinforced concrete structures were modeled
with different soil conditions for finite element model and nonlinear dynamic and static analyses were
conducted. Analysis results were compared with the fixed structures. Consequently, the analysis results show
that SSI must not be ignored for structural analyses.
Keywords: Soil-Structure Interaction, Seismic Performance, Non-linear Analyses, RC Structures
1 Introduction
In structural analysis made due to classical methods, it is accepted that structures sits on soil as fixed. As a
result the interface of the structure and the soil the motions like rotation, settlement and rising do not take
place. In addition, in the case of resonance and application of the vertical component of earthquake to the
building, the responses are ignored. During earthquake, in case of ignorance of SSI, structures are subjected
to this type of forces. As a result, important hazards are seen, for this reason, structures should be modeled
together with soils under structures. Bettess and Zienkiewicz [1] in their study developed infinite element
model for static and harmonic loads, which was used by Medina [2] for the dynamic interaction of soil
structure interaction systems. For the solution of soil structure interaction problems, the researchers have
developed many methods depending on direct method and substructure method. The most important of this
was by Wolf and Song [3] who developed infinitely small sized finite element method for the solution of
soil structure interaction problems.
In this study, seismic behaviors of reinforced concrete structures in the different soil profiles are investigated
by nonlinear analysis methods. The finite element models of soil and structure are constituted. The soil
structure interaction is investigated with nonlinear static and dynamic analyses using SAP2000 V14 finite
element program. The RC structures having similar rigidity in different soil types are taken into account.
After the analysis, the capacity curves of the structures, damages of structural elements, max base sheardisplacement values, displacements and rotations of foundations are determined.
786
Displacement parameters of a structural system under dynamic loads are rigidity and damping ratio.
Therefore elasticity module, poissons ratio, density, shear wave velocity, damping ratio and rigidity are
taken account in soil-structure interface modeling.
3 Methods
3.1 The Properties of Structural Analyses
The selection of the type of analysis to be used within the framework of a performance-based design
procedure is also a topical issue, and nonlinear analysis is gaining popularity during the last few years, a
reason for this being that appropriate analytical tools are now available for performing both types of
nonlinear analyses, the incremental static pushover (NISP) and time-history (NTH). Even though NTH is the
most comprehensive method to determine the seismic demands of the structures, this method is complicated
and it takes time to complete the analysis. Seismic loads are applied to the building directly in this method.
Earthquake data should be selected carefully [5-10]. The seismic performances of existing structures are
determined with the NISP and NTH methods. In this study, the different acceleration records were used for
NTH method. The properties and acceleration records of Duzce and Kocaeli earthquakes are given in the
Table 1 and Fig. 2.
787
Date
Name
Mw
Velocity
Acceleration
Magnitude
(cm/s)
(g)
Duzce
1999/11/12
DZC180
7.1
60.0
0.348
Kocaeli
1999/08/17
ARC000
7.4
17.7
0.218
0.4
0.2
0.3
0.2
Acceleration (g)
Acceleration (g)
0.1
0.0
-0.1
0.1
0.0
-0.1
-0.2
-0.2
-0.3
Kocaeli
Duzce
-0.4
-0.3
0
10
15
Time
20
25
30
10
15
20
25
Time
Shear wave
Density
Poisson
kN/m3
Ratio
15
17.16
0.45
100
CH
50
20.30
0.40
100
Dense Sand
120
19.62
0.33
150
Soil layers
elasticity
MPa
Fill
velocity
m/sec
ML
13
17.16
0.45
60
SM/ML
70
17.16
0.33
150
CL or CL/ML
15
17.16
0.45
60
56
19.62
0.40
100
1200
19.62
0.45
500
CH/CL with
ML/SM layers
Clay Stone
788
Soil Profile-1
0m
0m
Soil Profile-2
3m
5m
8m
9m
11 m
15 m
15 m
ML
CL
SM-ML
Dense Sand
Soil profiles were modeled as 2D shell element in SAP2000 a commercial Finite Element program [12].
Dimensions of the shell element were chosen 210 meters (10 times of structure base width) to left and right
side and 15 meters depth. While nearby region to structure was meshed fine, far away region to the structure
was meshed coarse. Mesh sizes were chosen 1x1 m 40 m distance from the structure, 2.5x1 m 70 m distance
from the structure and 5x1 m 100 m distance from the structure. Index properties of soil profiles (Fig. 3)
were assigned to the shell elements. Connection of soil and structure was provided with 50 cm width rigid
foundation. 15 meter depth of the soil was assumed to be rock and this region was modeled as fixed end.
Dampeners that assumed the soil parameters assigned to the horizontal boundaries of the soil. These
dampeners provided reflection of earthquake waves back to the soil.
4 Numerical Modeling
In this study, while the soil structure model was constructed, existing RC buildings were used. In the
buildings, C20 concrete and S420 steel are used. The beam and column dimensions were 250x500 mm,
400x400 mm, respectively. The structure had 3 spans in X direction and 4 storeys. The span lengths are 3 m,
4 m and 3 m respectively. The storey heights were 4 m, 3 m, 3 m and 3m, respectively (Fig. 4). In
Nonlinear analyses 3 different type SSI were used.
: SFS
1. The structure having fixed support
: SSP-1
2. The structure in the soil profile-1
: SSP-2
3. The structure in the soil profile-2
In the soil structure interaction, damping ratio was taken %5 and soil was assumed isotropic and
homogeneous. The seismic behaviors of RC structures in the soil structure interaction were investigated
with NISP and NTH analyses. The acceleration records of 1999 Kocaeli and Duzce earthquakes were used
in the NTH method. Structural analyses were carried out with SAP2000 V14 program [12].
789
5 Results
Analyses of the SFS, SSP-1 and SSP-2 type of structures were done by NISP method and capacity curves of
the structures were determined (Fig. 5).
According to Turkish Earthquake Code-2007 SFS, SSP-1 and SSP-2 type of structures performance points
were determined by NISP method. The results are presented in Table 3. Number of damaged structural
elements at the performance point is presented in Table 4.
180
Rigidity Connections
Soil Profile-1
Soil Profile-2
160
140
120
100
80
60
40
20
0
0.0
0.1
0.2
0.3
0.4
0.5
Displacement (m)
Fig.5. The capacity curves of the structures with the incremental static pushover analysis
Table 3. Performance points of the structures that determined bye NISP method
SFS
SSP-1
SSP-2
95.526
60.087
59.175
Displacement (m)
0.033
0.219
0.157
SSP-1
SSP-2
Beam
Column
Beam
Column
Beam
Column
NTH analyses of the structures were made by using Duzce and Kocaeli Earthquakes and damage situation of
structural elements at maximum acceleration values are presented in Table 5 and Table 6.
SSP-1
SSP-2
Beam
Column
Beam
Column
Beam
Column
10
790
SSP-1
SSP-2
Beam
Column
Beam
Column
Beam
Column
Maximum displacement and rotation angle values at foundation-structure region were obtained from NTH
and NISP analyses. Results are presented in Table 7 and Table 8.
Table 7. Maximum displacement and rotation angle values at foundation-structure region gathered from
nonlinear static analyses
SFS
SSP-1
SSP-2
Displacement (m)
0.03578
0.0222
Rotation (rad)
0.02492
0.0212
Table 8. . Maximum displacement and rotation angle values at foundation-structure region gathered from
nonlinear time history analyses
Duzce Earthquake
Displacement
(m)
Rotation
(rad)
Kocaeli Earthquake
SFS
SSP-1
SSP-2
SFS
SSP-1
SSP-2
0.04200
0.01079
0.01735
0.00310
0.02840
0.01806
0.00599
0.00301
6 Conclusions
Many design codes have suggested that the effect of SSI can reasonably be neglected for the seismic
analysis of structures. In this study, SSI of two equivalent reinforced concrete structures were investigated
with NISP and NTH analysis methods. In consequence to the study;
Analysis results of neglecting and taking into account SSI have great differences that can not be
neglected.
Capacity curves of SFS, SSP-1 and SSP-2 types of structures were obtained by NISP analyses (Fig.
5). SFS type structure (neglecting SSI) has 3 times greater ductility than SSP-1 and SSP-2 type
structures (taking into account SSI)
Performance points were obtained by NISP analysis and results presented in Table 3. Base shear of
the SFS type structure was obtained 1.5-2 times greater than other structures. Displacements of SFS
type structures were obtained 5-6 times smaller than other structures. This indicates, taking SSI into
account makes the model more flexible.
Structural damages under Duzce Earthquake were presented in Table 5. Analysis results were
found similar with NISP analysis. It was found that under Kocaeli Earthquake structural elements
had no damage (Table 6). In the case of taking SSI into account caused damages on the structural
elements.
Displacement and rotation at foundation could not be observed in SFS type structures. As the SSI
was taken into account maximum displacement and rotation were occurred at soil-structure
connection region (Table 7 and Table 8). The effect of soil profile was found more operative in
NTH analysis method then NISP analysis method.
Consequently, the analysis results show that SSI affects the performance and earthquake response of the
structural model and must not be ignored for structural analyses.
791
References
1.
2.
3.
4.
5.
6.
7.
8.
9.
10.
11.
12.
13.
Bettes, P. and Zienkiewicz, O.C. (1977). Diffraction and refraction of surface waves using finite and infinite
elements. International Journal of Numerical Engineering 11:1271-1290.
Medina, F. (1980). Modeling of soil-structure interaction by finite and infinite elements'. Report No.
UCB/EERC-80/43, University of California, Berkeley. CA.
Wolf, J.P. and Song, C. (1994). Dynamic stiffness matrix of unbounded soil by finite element multi-cell
cloning. Earthquake Engineering and Structural Dynamics 23:233-250.
Garip, Z.., alar, N., Pala, M. and Ala, N. (2007). Soil-structure interaction analysis of rc buildings settled
in soil conditions of Adapazari. International Earthquake Symposium, Kocaeli, Trkiye.
Lawson, R.S., Vance V. and Krawinkler H.(1994). Nonlinear static push-over analysis-why, when, and how?
In: Proceedings 5th US NCEE. 1:283292 Chicago, USA.
Faella, G. and Kilar, V. (1996). Asymmetric multistorey R/C frame structures: push-over versus nonlinear
dynamic analysis. In: Proceedings 11th ECEE, Paris, France, CD-Rom.
Mwafy, A.M. and Elnashai A.S. (2001). Static pushover versus dynamic collapse analysis of RC buildings.
Engineering Structures 23:407424.
Sezer, F., Genolu, M. and Celep, Z. (2007), Betonarme Binalarn Deprem Gvenliinin
Deerlendirilmesinde Deprem Ynetmelii 2007 Kurallarna rnekle Kyaslamal Bir Bak, Altnc Ulusal
Deprem Mhendislii Konferans, 16-20 Ekim, stanbul.
Kappos, A.J. and Kyriakakis, P. (2000). A re-evaluation of scaling techniques for natural records, Soil
Dynamics and Earthquake Engineering 20:111-123.
Bardakis, V.G and Dritsos, S.E.(2007). Evaluating assumptions for seismic assessment of existing buildings.
Soil Dynamics and Earthquake Engineering 27:223233.
Pacific Earthquake Engineering Research Center web page http://peer.berkeley.edu
SAP2000, integrated finite element analysis and design of structures basic analysis reference manual,
Computers and Structures Inc. Berkeley (CA, USA).
TEC 2007, Specifications for buildings to be built in seismic areas. Turkish Earthquake Code 2007). Ministry
of Public Works and Settlement. Ankara. (Turkey).
BIOGRAPHIES
Ali Demir Ali Demir was born in Karabk in 1983. He graduated from Celal Bayar University in 2005. He completed his
graduate thesis in 2008. He is still a doctoral student in Celal Bayar University. He has been an research assistant in Celal
Bayar University for 6 years. He is studying concrete structures, performance evaluation of buildings, experimental study. Ali
Demir knows English as a foreign language.
He has three pressed, five national articles, eleven articles in national and international conferences. Four papers were
accepted by Gediz University and will be presented in June. He also has two accepted presentations and will be presented in
7UDMK in June. He completed different scientific research projects in Celal Bayar University, TBTAK and he has started
two new projects last year.
Research Assistant Ali Demir is a member of Turkish Chamber of Civil Engineers since 2005.
Emre Ercan Emre Ercan was born in zmir in 1977. He graduated from stanbul Techinical University in 2001. He
completed his master thesis in 2004 and doctorate thesis in 2010 in Ege University. He has been an research assistant in Ege
University since 2004. He is studying Impact-Echo, Operational Modal Analysis, FEM and steel structures. Emre Ercan
knows English as a foreign language.
He has two published articles. Two papers were accepted by Gediz University and will be presented in June. He
completed four scientific research projects in Ege University and TBTAK and he is starting one new projects this year.
Research Assistant Emre ERCAN is a member of Turkish Chamber of Civil Engineers since 2001.
Muhiddin Bac Muhiddin Bac was born in Konya. He graduated from Dokuz Eyll University in 1991. He completed
his graduated thesis in 1996 and his doctorate thesis in 1996. He started working as a research assistant since 1993. He has
been working as an assistant professor since 2006 in Celal Bayar University. He is studying concrete structures, performance
evaluation and finite elements method. He knows English as a foreign language.
He has five pressed and one accepted scientific articles, fifteen national articles. His three papers have been accepted and
presented in national conferences and one in international conference. He has two graduate and one doctorate students this
year. He has started scientific research projects in Celal Bayar University last year.
Assistant Professor Dr. Muhiddin Bac is a member of Chamber of Civil Engineers.
792
Abstract.Lattice Boltzmann Method (LBM) is a computational technique used to solve transport problems
in thermo-fluidic sciences. When applied to the solution of incompressible flows it shows a number of
advantages compared to classical finite volume or finite element based techniques. Its algorithmic structure
is very suitable for parallel programming on GPUs, which are today's state-of-the-art technology for parallel
computing. In this study, 2D lid-driven cavity problem, which is a standard benchmark problem for fluid
flow simulations, is solved using LBM with D2Q9 model and BGK collision approximation. Numerical
computations are performed on a workstation with an NVIDIA Tesla C1030 GPU. Parallelization potential
of a typical LBM code is investigated. Performance comparison with CPU- and GPU-based computing is
presented.
1 Introduction
Lattice Boltzmann Method (LBM) is a new generation numerical technique used to study transport problems in
thermo-fluidic sciences. It is well suited to simulate single and multi-phase flows of single and multicomponent fluids. Although it is commonly used for problems in high Knudsen number regimes for which
continuum assumption of Navier-Stokes solvers breaks down, it is also capable of simulating continuum flows
[1-5]. Its simple algorithmic structure is easy to code and very suitable for parallel programming [6].
Today's state-of-the-art technology for parallel computing is based on the use of Graphics Processing Units
(GPU). GPUs that are specifically designed for high performance scientific computing can offer tens of times of
more FLOP performance and over ten times more memory bandwidth compared to CPUs [6]. In 2007, GPU
manufacturer NVIDIA announced a new parallel computing architecture called CUDA, which popularized the
use of GPUs for general purpose computing. In the following years software vendors realized the great potential
of GPU based scientific computing and they started to migrate their CPU based numerical libraries to GPUs.
Jacket is such a commercial numerical library that provides GPU-based versions of many built-in MATLAB
functions. MATLAB and Jacket software are used in this study to implement LBM.
2D lid-driven cavity problem is solved using LBM with D2Q9 model and BGK collision approximation.
Numerical computations are performed on a workstation with a Tesla C1030 GPU. Performance comparison
793
with CPU-based and GPU-based computing is presented. Speedups exceeding 17 times are observed. It is seen
that computations using finer grids resulted in better parallel performance.
2 Lattice-Boltzmann Method
Although incompressible flow of a 2D lid-driven cavity, shown in Fig. 1, is commonly used as a benchmark
problem for new CFD codes, it is quite challenging considering boundary condition related singularities at its
top corners. Classical finite volume or finite element based solvers need excessively fine girds to resolve these
regions accurately and provide an oscillation free solution. A major challenge for solving incompressible flow
problems with these classical techniques is the weak coupling of pressure and velocity fields. Several remedies
to overcome this problem are suggested such as the use of staggered grids for finite volume solvers or the use of
different order approximation spaces for velocity and pressure fields for finite element solvers. However, all
these extra efforts bring complexities to computer codes. Several stabilization techniques are developed to
circumvent these complications, but they suffer from unphysical artificial dissipation.
(
B (
f
B (
(
ff
B (
Square
R
(
B (
In this study LBM is used to solve 2D lid-driven cavity problem. Difficulties of simulating incompressible
flows mentioned in the previous paragraph are not of a concern for LBM, which results in a very short and
clean code that is easy to understand and implement. A typical LBM code works in a time marching manner, in
which the following 4 sections are repeated inside a time loop: (1) collision, (2) streaming,
(3)
implementation of boundary conditions, and (4) calculation of macroscopic quantities. In this study, the popular
D2Q9 model is used [7]. D2 stands for 2 space dimensions, and Q9 is the number of lattice directions
associated with each node of the mesh, as shown in Fig. 2.
794
& e (1)
S
where term is a double integral that represents the collisions of particles. is the distribution function used to
model the number density of the particles in a control volume and in a range of velocities at any time S. e is the
speed of sound. BTE is an integro-differential equation which is hard to solve. It is commonly simplified using
the Bhatnagar-Gross-Krook (BGK) model, according to which the collision operator is replaced with the
following expression [8],
where { is collision frequency and " is equilibrium distribution function. BTE equation can now be
discretized in space and time as follows,
(3)
By the use of Chapman-Enskog expansion, i.e. single-relaxation time approach, collision frequency is given by,
{
( & ;
<
(4)
where is the nondimensional kinematic viscosity which is a function of relaxation time f]{. In the
common practice both the time step and the grid spacing are taken to be unity, which is also the case in
the current implementation.
Equilibrium distribution function of Eqn (3) is simplified by Rothman & Zaleski in the following form [9],
"
f &
&
7
(5)
which is the equation used to calculate the equilibrium distribution function at node } of the grid.
are the unit
vectors associated with each lattice direction. For example for the 5th direction of Fig. 2
& . and
are the macroscopic velocity and density at node } and is the lattice speed of sound. Weights used in D2Q9
model are,
f] for f g
f] for (6)
] for
2.2 Streaming Process
Streaming (updating) process transfers the recently calculated distribution functions to the neighboring nodes to
be used in the next time step. Fig. 3 demonstrates how the nodes communicate information with each other for
the simple case of a unidirectional flow from left to right discretized by a grid of 6 nodes (23). Arrows
represent 9 distribution functions and the one in the R -direction (1st direction) is highlighted and shown before
June 1 -4, Kusadasi, Aydin, Turkey http://iscse2011.gediz.edu.tr
795
Fig. 3. 1st distribution function before (left) and after (right) streaming.
(7)
For the moving top wall of the cavity, a modified version of the procedure described above given by Zhu and
He [11] is used.
(8)
which can further be used to calculate the macroscopic velocity at a node as follows
f
e
796
(9)
(a)
(b)
(c)
(d)
Fig. 4. Streamlines at different Reynolds numbers (a) t f((, (b) t f(((,
Using MATLAB's add-on library Jacket, LBM code is modified to be run in parallel on GPUs. Lid-driven
cavity problem is solved with a number of different uniformly spaced grids using both the serial version of the
code running on a single CPU (Intel Xeon E5620 Quad-Core 2.40GHz, 12MB Cache) and the parallel version
running on a GPU (NVIDIA Tesla C1030, 4 GB RAM). Results are shown in Table 1 and Fig. 6. Execution
times given in seconds in Table 1 are obtained by running the time marching LBM code only for
arepresentative number of time steps that is enough to collect meaningful data. That is actual runs take longer
than the amounts presented.
797
(a)
(b)
(c)
(d)
Fig. 5. Variation of horizontal velocity component at R ( for different Reynolds numbers
As seen in Fig. 6 for coarse grids the code runs slower on the GPU. The reason is that the amount of floating
point operations resulting from coarse grids are not enough to keep the GPU busy, but rather the memory
transfer to and from the GPU becomes the bottleneck and increases the total run time. As the grid gets finer,
parallelization on the GPU becomes more and more effective. For the finest mesh of 4 million nodes, which is
unnecessarily big for a 2D problem, but can be appropriate for a 3D problem, a speedup of more than 17 times
is obtained.
Table 3 - GPU vs CPU Time Comparison (in seconds)
Grid Size
GPU Time
CPU Time
Speedup
40 x 40
0.73
0.08
0.12
80 x 80
0.73
0.21
0.30
100 x 100
0.73
0.32
0.44
200 x 200
0.75
1.20
1.61
400 x 400
0.83
4.69
5.65
800 x 800
1.55
18.99
12.33
1200 x 1200
2.79
42.80
15.33
1600 x 1600
4.54
78.54
17.28
2000 x 2000
6.91
121.39
17.57
798
Current LBM implementation based on D2Q9 model has a memory requirement of 32xNN where NN is the
total number of nodes in the mesh. Even for the finest mesh tried, which has NN=4x106, memory requirement
is less than 1 GB, which fits nicely into the 4 GB memory of the NVIDIA Tesla C1030 card used in this study.
Even for D3Q15 model, which is commonly used for the solution of 3D problems, memory requirement of
LBM raises only to 52xNN, which is again smaller compared to the memory need of classical Navier-Stokes
solvers. Even though a single card seems to be enough for LBM simulation of large 3D problems, it is logical to
use multiple GPUs to increase parallelization potential of LBM and decrease run times. Jacket library has
support for multiple GPUs, which will be utilized in future studies.
4 Conclusion
2D lid-driven cavity benchmark problem is simulated using a MATLAB implementation of Lattice Boltzmann
Method. The code is parallelized to be run on GPUs by the use of the Jacket library. LBM is shown to provide
accurate results for this continuum flow, without the need of any stabilization techniques required for classical
incompressible Navier-Stokes solvers. Parallel version of the code provided a speed up of more than 17 times
for the finest mesh used with 4 million nodes with a memory usage on the GPU of less than 1GB. LBM has a
very simple algorithm with enormous potential for parallelization. Also as shown in this study its memory
footprint is very small. Coupled with the rapid coding and testing experience provided by MATLAB and
Jacket's easy to use GPU extensions, efficient flow solvers that can simulate realistic 3D flows on personal
desktop supercomputers with multiple GPUs can be developed. One disadvantage of LBM is the discretization
of complicated flow fields with arbitrary boundaries, which is a current field of active research with already
available promising solutions.
Acknowledgements
Financial support from METUNCC via the Campus Research Project (BAP-EN10) and from the Turkish
Scientific and Technical Research Council, Grant No. 110M750 is greatly appreciated.
799
References
1. Toschi F. and Succi S. (2005). Lattice Boltzmann method at finite Knudsen numbers. Europhys. Lett 69(4): 549555.
2. Nie X., Doolen G. D., and Chen S. (2001). Lattice-Boltzmann simulation of fluid flows. MEMS J. of Statistical
Phys. 107 (1-2): 279-289.
3. Ghazanfarian J., and Abbasi A. (2010). Heat transfer and fluid flow in microchannels and nanochannels at high
Knudsen number using thermal Lattice-Boltzmann method. Phys. Rev. E. 82 (2): 026307.
4. Verhaeghe F., Luo L.-S., and Blanpain B. (2009). Lattice Boltzmann modeling of
microchannel flow in slip
flow regime. Journal of Computational Physics. 228 (1): 147-157.
5. Zhou Y., Zhang R., Staroselsky I., Chen H., Kim W. T., and Jhon M. S. (2006). Simulation of micro and nano
scale flows via the Lattice Boltzmann method. 362 (1): 68-77.
6. Tubbs K. R. and Tsai F. T.-C. (2011). GPU accelerated Lattice Boltzmann model for shallow water flow and mass
transport. Int. J. for Numerical Methods in Engineering. 86 (3): 316334.
7. Mohammad, A. A., (2007). Applied LatticeBoltzmann Method for Transport Phenomena, Momentum, Heat and
Mass Transfer. Calgary: Sure Printing.
8. Succi S., (2001). Lattice Boltzmann Equation for Fluid Dynamics and Beyond, Oxford: Clarendon Press.
9. Rothman D.H,. and Zaleski S. (1997). Lattice Gas Cellular Automata, Cambridge: Cambridge University Press.
10. Bhatnagar P. L., Gross E. P., and Krook, M. (1954). Gaseous flow in microconduits with viscous dissipation.
Phys. Rev., 94: 511-525.
11. Zou, Q., and He, X., (1997). On pressure velocity flow boundary conditions for the lattice Boltzmann BGK model.
Phys. Fluids9 (6): 1591-1598.
12. Ghia U., Ghia K. N., and Shin C. T. (1982). High-Re solutions for incompressible flow using the NavierStokes
equations and a multigrid method. Journal of Computational Physics48 (3): 387411.
Biographies
S. Berat elik was born in Turkey (1985). Graduated from mechanical engineering program of Middle East Technical
University Northern Cyprus Campus (METU NCC) and currently attending graduate school in METU. Research interests
are Lattice Boltzmann Methods, parallel computing and microfluidics.
Cneyt Sert was born in Turkey in 1974. He received B.S. and M.S degrees in mechanical engineering from Middle East
Technical University (METU), Ankara, Turkey in 1996 and 1998, respectively. He received his Ph.D. degree in mechanical
engineering from Texas A&M University, College Station, TX, USA.
He is currently working as a faculty at the Mechanical Engineering Department of METU. His current research interest is
in the field of computational fluid dynamics by the use of finite element and spectral element formulations.
Barbaros etin was born in Bursa, Turkey in 1979. He received B.S. and M.S. degree in mechanical engineering from
Middle East Technical University, Ankara, Turkey in 2002 and 2005, respectively. He received his Ph.D. degree in
mechanical engineering from Vanderbilt University, Nashville, TN, USA.
He is currently working as a faculty at the Mechanical Engineering Program of METU-Northern Cyprus Campus. His
research interest lies in the fundamental and experimental investigations on transport phenomena in microchannel networks
and development of the LOC technology for biomedical, biological and chemical applications. He is also currently working
on the implementation of Lattice-Boltzmann method for fluid and heat transfer problems on graphical process units.
800
1,2,3,4
1
Abstract.In this study, weight optimization of symmetric and balanced laminated composites used in aircraft industry is
presented. Aircraft structure materials should
have high specific stiffness and strength. Carbon fiber reinforced polymer
matrix composites fulfill these needs. A stochastic search technique called genetic algorithm (GA) is utilized in the design
and optimization of the composite system. MATLAB Genetic Algorithm and Direct Search Toolbox is used to obtain optimal
design for different model problems. The objective of the problems is to minimize weight of the laminated composite
materials under two different failure constraints for minimum buckling load factor. Fiber orientation angle in each layer is
taken as a design variable and results are obtained for several load cases.
Keywords: optimization, genetic algorithm, aircraft structures, laminated composites, buckling analysis
1 Introduction
Fiber-reinforced composite materials are demanded by the aircraft industry because of their high ratios of
stiffness and strength to weight. Weight of an aircraft structure can be reduced significantly by using the
laminated composites. Further reduction is also possible by optimizing the composite structure such as fiber
orientations, ply thickness, stacking sequence, etc [1].
Many researchers attempted to make a better composite structure either by
reducing the weight or by
maximizing static strength of composite laminates for a given thickness. The problem of the multi-objective
optimal design of laminated composite structures has been studied by [2]. In addition, Pelletier and Vel have
presentedmultiobjective optimization of fiber reinforced composite laminates for strength, stiffness and minimal
mass [3]. Various design optimizations such as weight and deflection or weight and cost have been proposed by
[4]. Furthermore, Aydin and Artem have studied design of the dimensionally stable composite laminates
subjected to hygrothermal loading [5]. Layer optimization has been carried out for maximum fundamental
frequency of laminated composite plates under any combination of the three classical edge conditions [6]. The
problem for adjustment of residual stresses in asymmetric fiber reinforced composite has been investigated by
[7]. Lopez et al. have studied weight and cost optimization of laminated plates subjected to in-plane loads
considering different failure criteria [8]. Many researchers have used maximum stress and Tsai-Wu failure
theory to optimize their composite plates for various objectives. When the Tsai-Wu or maximum stress failure
criterion is used individually, an optimization algorithm is lead to false optimal designs because of the particular
features of their failure envelopes.
801
Unlike these studies, we have considered classical buckling theory to obtain objective function as buckling
load factor. In the present study, minimum weight design of carbon-epoxy laminated composite plates subjected
to the compressive in-plane loads under the constraints of laminate symmetry-balance and the first ply failure
criteria (maximum stress and Tsai-Wu) is considered.
2 Problem Statement
In-plane design of layered symmetric-balanced laminated composite plates subjected to compressive loads
Nx N y
,
with dimensions a and b are considered (Fig. 1).
This work consists of two main parts: mechanical analysis and optimization. The classical lamination theory
and the linear buckling analysis are used to determine the mechanical behavior of the laminated composite. In
optimization part, a stochastic search technique called genetic algorithms (GAs) is used to obtain the optimum
fiber orientation angles for minimum thickness.
The classical lamination theory is used to analyze the mechanical behavior of the composite laminate. The
plane stress condition is valid for each ply and the laminated composite is orthotropic. The relationship between
stress and strain components can be expressed as [1]
x
Q11 Q12
y = Q12 Q22
xy
k Q16 Q26
Q16
Q26
Q66
x
y
xy
(1)
k
Q
where k is the lamina number counted from the bottom. The transformed reduced stiffnesses ij are
Q
expressed in terms of principal stiffness terms ij related to some elastic properties of the material along the
E E
G
fiber directions such as elastic modulus ( 1 , 2 ), shear modulus 12 and Poissons ratio ( 12 and 21 ).
Bij
802
A = A 1
and we
*
y = A12
*
xy A16
where
N ij
Aij
n
*
A12
*
A22
*
A26
*
Nx
A16
*
A26 N y
*
A66
N
xy
(2)
[ ]
; j = 1,2,6 (3)
Principal stress components can be obtained using Eq. (1) and Eq. (3):
cos 2 k
1
= sin 2
k
2
cos sin
12
k
k
k
sin 2 k
cos 2 k
cos k sin k
2 cos k sin k x
2 cos k sin k y (4)
cos 2 k sin 2 k xy
k
k
buckling represents the failure buckling load divided by the applied load and is obtained as,
min
2
2
4
r 4
r s
s
2 D11 + 2( D12 + 2 D66 ) + D 22
a
a b
b
=
2
2
r
s
+
N
N
x y
a
b
(5)
D
where ij are the laminate bending stiffness matrices, r and s determine the amount of half waves in the x
and y directions.
The bending stiffness matrices in Eq. (5) are defined as:
1
Dij = [Qij ]k (hk3 hk31 ) i = 1,2,6 ; j = 1,2,6
3
(6)
It is not clear which value of r and s half waves is for the lowest critical buckling load. The laminated
composite is simply supported for all edges, so s = 1 is the necessary value. However, r includes several places
and depending upon flexural stiffness matrix ( D ij ) and the length to width ratio ( a / b ) [9].
803
Maximum Stress Failure Criterion.In this theory, failure of the lamina is assumed to occur whenever any
normal or shear stress component equals or exceeds the corresponding strength. This theory is described as
follows [8]:
2 YT
1 X T
1 X C or 2 YC
(7)
12 S12
1 , 2
12
XT , XC
YT , YC
2
F11 1 2 + 2 F12 1 2 + F22 2 2 + F21 12
+ F1 1 + F2 2 1
(8)
where Fij are parameters that are a function of the strength properties X T , X C , YT , YC and S12 [9]. The
principal material coordinates are shown in Fig. 2.
804
3. Optimization
Many engineering design problems are very complex and they are too difficult to solve by the traditional
optimization techniques. The genetic algorithm (GA) is a numerical optimization that based on the principles of
genetics and natural selections. The parameters of the problem are coded naturally as a DNA-like linear data
structure, a vector or a string. In order to evaluate the optimization results, GA uses three basic operators:
selection, crossover and mutation.
In this study, Matlab Genetic Algorithm Toolbox and Symbolic Math Toolbox are used to obtain the results.
Generally, the optimization tool options include following 9 fundamental processes 1) Population, 2) Selection,
3)Reproduction, 4)Mutation, 5)Crossover, 6)Migration, 7) Algorithm settings, 8)Hybrid function, 9)Stopping
criteria [5].
= 0.27 and mass density = 1605 kg / m 3 [8]. The strength properties of the
, Poissons ratio 12
composite are shown in Table 1. Thus, the optimization problem for minimum weight design can be stated as
G12 = 4173MPa
{ k , n} k {90,90} k = 1 n
Find
:
,
Minimize : Buckling load factor for minimum weight
Subject to: Maximum Stress and Tsai-Wu failure constraints
Table 4. The strength properties of the carbon-epoxy lamina [8]
XT
2062 MPa
XC
1701 MPa
YT
70 MPa
YC
240 MPa
S12
105 MPa
While using Genetic Algorithm toolbox, all the optimum designs have been obtained running the algorithm
at least 10 times and stopping it after maximum 6 generations for each case by using the parameters given in
Table 2.
805
Double vector
Population size
20
Creation function
Initial population
Use default
Initial scores
Use default
Initial penalty
10
Initial range
[-90;90]
Penalty factor
100
Selection function
2
Elite count
Stopping criteria
fmincon
Generation=100
Stall generation=50
Function tolerance=106
In the weight optimization, we have applied N x = N y = 500 N / mm in-plane loading to layered plates and it is
obtained that the given loading condition with the failure constraints is satisfied by 12 layered plate.
We have also investigated 2, 4, and 8 layered ([1 ], [1 ] s , [1 / 2 ] s ) plates subjected to different
loading conditions for maximum stress and Tsai-Wu failure criteria. Stacking sequences and the corresponding
minimum buckling load factor values are shown in Table 3.
For [ 1 ] case three different loadings N x = N y = 50,60,70 N / mm are investigated. Stacking sequences
and the corresponding minimum buckling load factor values are given in Table 3. It is seen that the first ply
failure occurs in maximum stress constraint with N x = N y = 70 N / mm . However, we obtained the stacking
sequence [ 49.7] when Tsai-Wu failure constraint is taken into consideration.
Table 3. Optimum weight and stacking sequence for the layered composites for different failure criteria
Ply
number
Maximum Stress
Weight
Nx
Ny
(kg)
(N/mm)
(N/mm)
0.321
0.642
1.284
12
1.926
For
[ 1 ]s
Stacking
Tsai-Wu
min
Sequence
Stacking
min
Sequence
50
50
[54]
1.006
[52.7]
1.135
60
60
[52]
1.009
[51]
1.114
70
70
[49.7]
1.086
100
100
[54]s
4.023
[52.7]s
4.542
125
125
[51.6]s
4.041
[50.6]s
4.427
150
150
[49.1]s
4.294
250
250
[902/35]s
2.957
[902/37.2]s
2.819
300
300
[902/27.8]s
2.824
[902/32]s
2.608
400
400
[902/78.8/16.2]s
5.007
[902/902/22.6]s
4.741
500
500
[89.5/53.7/02]s
5.644
sequences and the corresponding minimum buckling load factor values can be seen in Table 3. Similarly, the
first ply failure occurs in maximum stress constraint with N x = N y = 150 N / mm . The stacking sequence
[ 49.1]s is obtained for Tsai-Wu failure constraint. It is noticeable that stacking sequence and the corresponding
minimum buckling load factor values for the failure constraints are close to each other. In the first two cases,
806
genetic algorithm toolbox options are considered as default, because the practical optimum is obtained without
difficulty in every case.
In [ 1 / 2 ]s case, the loadings N x = N y = 250,300 N / mm are applied to the plate. Optimal lay-up
designs and minimum buckling load factor values for the loading types are presented in Table 3. In calculations
one of the most important factors that determine the performance of the genetic algorithm is the diversity of the
population. If the diversity is too high or too low, the genetic algorithm might not perform well. In order to
control the diversity and abstain from inexact solutions, initial range of the population must be set properly.
Therefore, for this case, optimum lay-up configurations and fitness function values are found by setting the
initial ranges to various values.
Finally, we investigated which plate having minimum number of plies is suitable for withstanding the
loading, N x = N y = 500 N / mm and we found that minimum 12 layered carbon-epoxy composite plate weighing
1.926 kg is convenient. If N x = N y = 600 N / mm loading is applied to the laminate, it fails for both of failure
constraints, maximum stress and Tsai-Wu.
In order to give a clue about the performance of genetic algorithm, an evaluation of the best values of the
fitness function at each generation for 2, 4, 8 and 12 layered cases with Tsai-Wu failure constraint is shown in
the Figure 3a-d, respectively. As seen in the figures, optimum solutions are obtained only in 5 generations since
the use of the failure theories as constraints narrows the search space of the algorithm.
(a)
(b)
(c)
(d)
Fig. 3. Genetic algorithm generation steps for (a) [1] plate subjected to Nx = Ny = 70 N/mm,
(b) [1]s plate subjected to Nx = Ny = 150 N/mm, (c) [1/2]s plate subjected to Nx = Ny = 300 N/mm, (d) [1/2/3]s
plate subjected to Nx = Ny = 500 N/mm.
807
5. Conclusion
In this study, an optimization process for weight minimization of carbon-epoxy composite laminates under
compressive in-plane loading is presented. A genetic algorithm is used as global optimization tool.
We have concluded with the following observations:
1. 12 layered carbon-epoxy composite plate weighing 18.9 kg is convenient to withstand 500 N/mm,
2. The minimum weight of the laminated composite plate depends on the choice of the failure criterion as
well as the load applied,
3. Although the different failure constraints are considered, the buckling load factor values are close to each
other,
4. Since the constraints used in this study significantly narrows the search space,
we get the results with fewer number of iteration.
References
1.
2.
3.
4.
5.
6.
7.
8.
9.
Kaw, A.K. 2006. Mechanics Of Composite Materials, Second Edition: CRC Press Taylor and Francis Group.
Spallino, R. and Rizzo, S. (2002). Multiobjective discrete optimization of laminated structures, Mechanics
Research Communications 29: 1725.
Pelletier, J. L. and Vel, S. S. (2006). Multiobjective optimization of fiber reinforced composite laminates for
strength, stiffness and minimal mass. Computers and Structures 84: 20652080.
Almeida, F.S. and Awruch, A.M. (2009) . Design optimization of composite laminated structures using genetic
algorithms and finite element analysis. Composite Structures 88: 443454.
Aydin , L. and Artem, H. S. (2009). Multi-objective genetic algorithm optimization of the composite laminates
as a satellite structure material for coefficient of thermal expansion and elastic modulus. 4th International
Conference on Recent Advances in Space Technologies, RAST '09. Istanbul, Turkey ,11-13 June : 114-119.
Apalak, M.K. , Yildirim, M. and Ekici R. (2008). Layer optimization for maximum fundamental frequency of
laminated composite plates for different edge conditions. Composite Science and Technology 68: 537550.
Hufenbach, W. , Gude, M. , Kroll, L. , Sokolowski, A. and Werdermann, B. (2001). Adjustment of residual
stresses in unsymmetric fiber-reinforced composites using genetic algorithms. Mechanics of Composite Materials.
37(1): 216-222.
Lopez , R.H. , Luersen, M.A. , Cursi, E.S. (2009). Optimization of laminated composites considering different
failure criteria.Composites Part B 40: 731740.
Gurdal Z., Haftka R. and Hajela P.1999. Design and optimization of laminated composite materials. New York:
Wiley.
808
Abstract. In the aerospace industry stretch forming is often used to produce skin parts which are manufactured by
AA2024 alloys due to its superior fatigue properties. In stretch forming process, the sheet is clamped at two sides
and stretched over a die such that the sheet gets the shape of the die. However for complex shapes it is necessary to
use expensive intermediate heat treatments in between, in order to avoid Lders lines and cracks to still achieve
large deformations.Traditionally, the forming steps with inter-stage treatment are defined using production
experience and are improved by trial and error. This is a costly and time consuming way. In recent studies the
material models based on the dislocation density have been designed to predict the time for heat treatment. These
models were used in computer programs so as to validate the results and then become the cheapest tool for
simulating stretch forming of Al2024. The accuracy of finite element analysis depends largely on these material
models that describe the work hardening during stretching; residual stresses and work hardening reduction during
heat treatments due to recovery and particle coarsening. In this model plastic kinematic hardening model is used.
1) INTRODUCTION
Stretch forming is a process in which a desired shape is obtained by stretching a sheet blank over a tool surface
while applying a tensile stress. It is a well suited process to the fabrication of parts which are convex in nature
[1]. This often makes stretch forming the process of choice in the aircraft industry for manufacturing of large
open shapes such as wing leading edges and engine cowlings [2].
An advantage of this process is that only one die is needed. The undeformed sheet is clamped with two CNC
controllable grippers which determine the path of deformation of the clamped sheet as you see in figure 1. This
2D motion is controlled by CNC machines and leads into four different steps. During the first step, the sheet is
draped around the die without any contact with the die by moving the grippers towards each other. In the second
step, the grippers are moved down until the sheet touches the die. During the third stage, the sheet is stretched
plastically around the die until the sheet and die are completely in contact. In the last stretching step, extra strains
are induced in the sheet by moving the grippers vertically downwards in order to minimize the effect of spring
back [3].
809
The most commonly used material for aircraft skin is the heat treatable aluminum alloy AA 2024. Aluminum
2024 (2024A) is one of the most popular high strength aluminum alloys. Due to Al2024s high strength and
excellent fatigue resistance it is commonly used on structures and components in the aircraft and transportation
industries. In the fabrication of significant double curved surfaces using heat treatable aluminum alloy sheet,
large strain gradients are induced across the part. That can lead to surface defects like Lder lines and orange
peel, causing rejection of the components. In order to avoid these failures and still achieve large deformations, a
multi stage forming approach is normally employed. Several forming steps with intermediate annealing
treatments are employed to return the sheet to a more ductile condition [2, 4].
Traditionally, the forming steps with inter-stage annealing and the die shape are defined using production
experience and are improved by trial and error. This is a costly and time consuming way that may lead to suboptimal solutions. Models of the stretch forming process are useful to achieve such an optimal process control.
The advantage of models is that they can be used before the tools are made, avoiding long and expensive trial
and error runs for modifications of the tools and for reducing the number of expensive intermediate annealing
steps. Models based on the finite element method are needed to gain fundamental knowledge of the stretch
forming process [5, 6].
810
Figure 2 The sheet over the die is in clamped situation, ready to start analysis
The initial shape of the sheet is curved and this shape is equal to the shape of the buckle mode of a compressed
sheet as shown in figure 2. It is assumed that this bend causes no stresses and plastic deformations in the sheet.
By modeling this initial shape it is ensured that the sheet deforms in the correct upward direction, when the sheet
is draped around the die in the first part of the stretch process [7]. The die is assumed to be rigid body. The sheet
has been meshed with quadrilateral mesh elements. The used trajectory of the grippers is shown in figure 3. The
positions at which an intermediate heat treatment is applied to the sheet are marked. In this study, simulations are
performed with an approach for heat treatments by assuming the work hardening effect is completely removed
by resetting plastic strains to zero. The die is stationary and modeled by 3D solid elements whereas the blank
consist of 2D shell elements.
The die was drawn in Catia V5 and then exported into another program called I-DEAS to mesh both the die and
the sheet. After meshing, the die and the sheet was exported into another program called eta FEMB to start
loading fundamental properties of analysis (load curves which is based on reference [7], gripper and material
model settings). LSDYNA solves analysis and resultant shape of first step deformation is imported into eta
FEMB to start arrangements of second step, which is after first heat treatment, and this time LSDYNA solves
811
second step analysis. This keeps going 5 times until the final shape is obtained. LSPOST shows the simulation
and results for each 5 step.
The material model for the die is rigid and the material model for the sheet is determined as plastic-kinematic
hardening model. This model is a combination of isotropic and kinematic hardening models with strain rate
dependency and failure. Isotropic and kinematic contributions may be varied by adjusting the hardening
parameter between 0 (kinematic hardening only) and 1 (isotropic hardening only). Strain rate is accounted for
using the Cowper-Symonds model which scales the yield stress by the strain rate dependent factor as shown
below:
where 0 is the initial yield stress, is the strain rate, C and P are the Cowper-Symonds strain rate parameters,
peff is the effective plastic strain, and Ep is the plastic hardening modulus which is given by
812
The strain and the motion of grippers are correlated to each other. In the Table 1 this relation is given
Table 1 Maximum Strain obtained on the surface of the part
Point 1
Point 2
Gripper Motion-distance
origin (mm)
Max. Strain
to
Point 3
Point 4
Point 5
1503
1371
1289
1008
892
0,037200
0,0320
0,02100
0,003500
0,0255
It is seen that the maximum strain evaluated at the points are not identical to each other as expected.
In addition to the strain magnitude the redundancy level on the total surface is also important. At the final stage
whole surface must be stretched. Since this is uniaxial deformation, it is achieved easily. In the following figure
6 strain distribution is given.
813
Point 4
Point 5
Thickness (mm), h
3,5
3,42
3,38
3,28
3,11
800
600
400
75
50
C index, h/R
0,004375
0,0057
0,00845
0,043733 0,0622
The proposed model needs to take both parameters into account. Therefore the strain and index values are
multiplied and the results are given in Table 3.
Table 3 Proposed Material Deformation Model Values
Point 1
Point 2
Point 3
Point 4
Point 5
Max. Strain
0,037200
0,0320
0,02100
0,003500
0,0255
C index, h/R
0,004375
0,0057
0,00845
0,043733 0,0622
Strain* C
0,153125
0,1824
0,17745
0,153067
0,1555
Although there are some differences, the proposed model can be used to determine the intermediate heat
treatment points.
4) CONCLUSION
The stretch forming process of AA2024 T3 aluminum alloys with intermediate heat treatment steps was
investigated to predict the effects of strain level at intermediate heat treatment points. In the study, there are 5
intermediate heat treatment points for which temperature and annealing time are 340 C and 20 minutes
respectively. The reason for application of heat treatment points is to reduce the work hardening due to
stretching. It is expected to return to the mechanical properties for AA2024 T3 heat treated condition. This
condition is the stretch forming starting point. It is already proved that 340 C heat treatment temperature and 20
minutes ageing time is enough to reach target.
The second important agreement is on the mechanism of elimination for work hardening. The recovery
mechanism was studied by some researches. The Scanning electron microscope and electrical resistivity
measurement tests concluded that there was no evidence of precipitate nucleation and growth mechanism. The
reason for this recovery is mainly coarsening of previously formed precipitates.
In our study simulations employed the same starting condition, i.e., AA2024 T3 HT. In the deformation path
there are 5 intermediate points. At each point deformation is quite different from each other. At the beginning
two point strains are high with respect to other three points. Therefore it can be concluded that the strain levels at
each point is not the only factor.
The heat treatment points for the alloyed sheet were determined by experiments including physical surface
controls to detect lder lines and orange peel defects. The surface defects are assumed to be occurred due to
micro surface cracking and level of bending applied at the corresponding heat treatment points. This is similar to
the spring winding. The wahl effect in the spring stress calculations is common multiplier to express the true
stress on the outer fiber of spring coil. It also expresses the surface micro cracking which lead to early failure of
springs. In the study sheet thickness is divided to the radius of curvature for the deformation at each five
intermediate points. The ratio is called as bending index. The strain level is then multiplied by the bending index
to see whether or not the five intermediate steps shows the same level of calculation. It can be calculated that the
results are close to each other according to the order of magnitude control.
814
Aluminum and its alloys as already known are strain rate sensitive. The stress strain behavior during stretch
forming depends on mostly work hardening behavior and strain rate. The forming of the aircraft parts such as the
one studied in this work do not formed at high strain rates. Therefore the only parameter is the work hardening
and its reduction during formation for complex and convex parts.
Although there are some materials a model trying to explain micro structural effects on the stress-strain behavior,
the forming characteristic of the geometry was constrained by redundancy (strain distribution) and the surface
defects caused by heterogeneous stress strain distribution. In the study the first parameter strain distribution is
related to the part geometry and the deformation method. The second parameter micro surface cracking is related
to micro structure and residual stresses. These parameters can be controlled by stretching with intermediate heat
treatment steps only.
5) RECOMMENDATIONS
The findings and conclusions obtain from this work are planned to be used to form a material model based on
only mechanical deformations. This model will only consist of strain level (%) and the bending index. The
model will be tested by means of the experiments on the different parts having complex shapes.
6) ACKNOWLEDGEMENTS
We wish to thank to Dr Mehmet Yeinmen his Company FORM2000 for his encouragements and help for this
study.
7) REFERENCES
1. McMurray RJ, ODonnell M, Leacock AG, Brown D (2009) Modelling the effect of pre-strain and interstage annealing on the stretch forming of a 2024 aluminum alloy. Key Eng Mater 410:421428
2. Straatsma EN, Velterop L (2004) Influence of heat treatments and straining steps on mechanical properties in
the stretch forming process. In: Muddle BC, Morton AJ, Nie JF (eds) 9th International conference on aluminum
alloys, Brisbane, Australia, pp 771776
3. Hol J (2009) Optimization of the stretch forming process using the finite element method. Master Thesis,
University of Twente
4. ODonnell M, Banabic D, Leacock AG, Brown D, McMurray RJ (2008) The effect of pre-strain and interstage annealing on the formability of a 2024 aluminum alloy. Int J Mater Form 1:253256
5. Jaspart O, Franois A, Magotte O, D Alvise L (2004) Numerical simulation of the stretch forming process
for prediction of fracture and surface defect. In: 7th ESAFORM conference on material forming, Trondheim,
Norway, pp 499502
6. Wisselink HH, van den Boogaard AH (2005) Finite element simulation of the stretch-forming of aircraft
skins. In: Smith LM, Pourboghrat F, Yoon JW, Stoughton TB (eds) Proceedings Of The 6th international
conference and workshop on numerical simulation of 3D sheet metal forming processes, Detroit, USA, pp 6065
7.Srihari Kurukuri, Alexis Miroux, Harm Wisselink, Ton van den Boogaard Simulation of stretch forming with
intermediate heat treatments of aircraft skins Int J Mater Form, pp 81-95
815
Abstract. In recent years, many scientists are focused on dynamic behavior of fluid conveying pipes. This study
also includes the dynamics on pipes, supported from two ends and due to both of internal and external flow. In
the scope of external flow, the constant-diameter pipe is considered in marine environment covering in different
sea motions as still waters, currents, waves, current and wave combinations. The internal flow relating on the
fluid through the pipe is assumed as ideal and unsteady flow. The pipe is statically represented as the EulerBernoulli beam with different support. The equation of motion on pipe is developed depending on boundary
conditions and its numerical analysis is performed by using the Galerkin method. Numerical results of vibration
analysis are given in graphical form, some dimensionless terms such as 1 , the slenderness, the kinematic
viscosity and the wave frequency are newly defined and employed first time solving the mentioned equation.
The figures contain the time-displacement and the Morison force-time graphics. In graphs, the effect of new
dimensionless parameters, 1 , the slenderness, the kinematic viscosity and the wave frequency on the Morison
force and amplitude of the pipe are presented.
Keywords: Non-linear analyses, Galerkin method, pipe conveying fluid, Morison forces, dynamic analyses
1. Introduction
In recent years, it is seen that many scientists have studied the dynamics of pipes conveying fluid. One of them,
Paidoussis [1] carried out some studies on the derivation of linear equations of motion of fluid conveying pipes
and the dynamics of pipe conveying fluid. Semler, et al. [2] are also derived geometrically nonlinear equations of
motion of fluid conveying pipes by developing their linear forms. Similar studies are performed as based on
studies of both scientists [3-4]. Snr [3] investigated that non-linear vibration of slightly curved pipes conveying
fluid due to constant velocity. He solved the equation of motion by using the Galerkin method. The periodic and
chaotic motions have been observed in the transverse vibrations of slightly curved pipe conveying fluid. A
linearized equation of motion which is a submerged and free-hanging riser conveying fluid is analyzed with
Galerkin method and frequency- and amplitude-dependencies of hydrodynamic drag is described to study both
the instability and the amplitude of the steady-state pipe vibrations [5]. The possibility of using the Coriolis
Effect as a pipe vibration controller is taken into account in the study denoted as reference [6].The Ddecomposition method is utilized for determining stability of pipes conveying fluid and the acting forces due to
the external surrounding water are modeled with the linearised Morison formula [7]. Morison equations are dealt
with in reference [8] for varying sea motions such as still waters, currents, waves, current-wave combinations.
The Morison equations considered for the mentioned sea motions have been also used [4]. At the scope of
reference [9], an equation of motion for planar vibration of a top-tensioned deep-water riser formulated and
816
stability of the straight equilibrium is carried out by employing the Floquet theory. In the paper denoted at [9] is
based on results of study published by Kuiper et al. [5] by taking into consideration the existence of
hydrodynamic drag in linear form. Galerkin method has been widely used in references mentioned at [3-5-9-10].
In this study, non-linear vibration of pipe conveying fluid is considered and Morison forces in still water,
current, wave and current-wave conditions are acted on pipe. Some dimensionless coefficients are newly
defined. The effects of these coefficients on the pipe conveying fluid and Morison forces are shown in the
displacement-time and Morison force- time figures.
2. Equation of Motion
In the derivation of governing equation on pipe conveying fluid, the internal forces due to unsteady flow on ideal
fluid and the hydrodynamic wave-current induced external forces around pipe are loaded over the EulerBernoulli beam to cause non-linear vibration of pipe in-plane [1, 2]. The equation of motion of non linear Euler
Bernoulli Beam is as follows,
&&
p ApY + EIY IV
EAp L 1 2
Y dx Y = 0
L 0 2
(1)
The non-linearity in equation (1) is due to extension of mid-stretching effect. Internal forces and external forces
are acted on equation (1) as follows [7],
&&
p ApY + EIY IV
EAp L 1 2
Y dx Y = f in (x , t ) + f out (x , t )
L 0 2
Here E is the modulus of elasticity of the pipe, I is the moment of inertia of pipe,
is the cross sectional area of pipe, L is the length of pipe,
(2)
is density of pipe, A p
y ,
( )
is the
()
d 2Y
fin (x , t ) = f A f
dt 2
f
Af
(3)
is density of fluid,
is the cross sectional of fluid. The total acceleration of the fluid mass can be
decomposed into a local, Coriolis and centrifugal acceleration;
f Af
d 2Y
&&
&
d &
= f A f Y + vY = f A f Y + 2vY + v 2Y + v&Y
2
dt
dt
(4)
Here, v is the internal flow velocity. If Eq.4 is replaced in Eq.2, we obtain the equation of motion of
pipe conveying fluid as follows,
EAp L 1 2
&&
&
Ap + f A f )Y + EIY IV + 2 f A f vY + f A f v 2Y + f A f v&Y
Y dx Y = f out (x , t )
L 0 2
(5)
For the sake of convenience, the equation is made dimensionless by introducing the following terms
[1, 3, and 4];
817
12
p A p = m, f A f = M , x =
x
EI
, t =
L
m+M
t
Y
ML2
M
L4 fout
,Y = ,v=
, f out =
v, =
2
L
R
EI
m+M
REI
(6)
These terms include the radius of gyration of the pipe cross section (R) and the ratio of fluid mass to
the fluid-pipe integrated mass ( ). Substituting dimensionless quantities in (6) into Eq. (5), the
dimensionless form can be obtained as follows,
1
Y&& + Y v + 2 vY& + v 2Y + v&Y Y 2 dx Y = f out
( )
(7)
External forces in Eq. (7) are considered for four different flow conditions around pipe. These are still
water, current, wave and wave-current conditions. The mentioned forces are expressed by the Morison
Equation including drag and inertia forces.
&&
& &
fout = C A AI Y C D AD Y Y
(8)
This dimensional external force can be rewritten in (6). The non-dimensionalized force is therefore
obtained as;
2
f out = C A 1Y&& C D 1 Y& Y&
(9)
D 2
4
.
1 =
m+M
&&
&
&
fout = C A AI Y + C D AD U Y U Y
(10)
By performing the similar procedure as mentioned above, the non-dimensionalized force is therefore
written as;
2
f out = C A 1Y&& + C D 1 U 2 Y& U 2 Y&
(11)
m+M
L
and =
.
D
EI
818
UD
, ;
&&
&
&
fout = CM AI u& C A AI Y + C D AD u Y u Y
(12)
The similar procedure is used for defining the external force by considering the Keulegan-Carpenter
2 u
instead of the Reynolds number (Re).
, wave frequency
D
C
C
(13)
= M 1 u& C A 1Y&& + D2 1u
u Y& D 1Y& u Y&
2
f out
(m + M )L4
EI
&&
&
&
fout = C M AI u& C A AI Y + C D AD u U Y u U Y
(14)
In this situation, the external forces surrounding pipe is non-dimensionalized by taking into
consideration both wave and current conditions together with Re and KC.
f out =
CM
2
2
+ CD
1Y& u 2 2U 2Y&
(15).
This equation covers the motion of oscillating pipe as the Bernoulli-Euler beam under the current and
wave-induced external forces and the internal forces of fluid transmitted by the pipe
3. Solution Method
The equation of fluid conveying pipe surrounding water with wave and current is governed as non-linear integrodifferential equation including the time-coordinate domain partial derivatives. Some partial derivatives as
known, especially in this study are analytically unsolvable. In such cases, the Galerkin Method is generally used
for solving these partial differential equations which are dependent on time and coordinate variables. It converts
time and coordinate domain partial differential equations to ordinary differential equations.
The solution of partial differential equation for Galerkin Method is assumed as follows[3,5,9,10];
Y ( x, t ) = qi (t ) i ( x )
(16)
i =1
qi (t ) , i = 1,2,..., N , are the generalized coordinates and i ( x ) are the eigenfunctions of the supported
(hinged-hinged, clamped-clamped, clamped-hinged) beam.
Taking the appropriate derivatives and substituting them into Eq.(7), the residual can be obtained as
follows [3].
819
= R ( x, t )
&
(
)
(
)
(
(
)
(
)
)
(
)
(
)
v
q
t
x
q
t
x
dx
q
t
x
f
i =1
i
i
i
i
i
i
out
(17)
R.Y (x )dx = 0
j
j = 1,2,..., N
(18)
where, Y j ( x ) is eigenfunction of the vibrating beam. The eigenfunctions are said to be orthogonal for
the boundary conditions. The boundary conditions for hinged-hinged, clamped- clamped, clampedhinged supports are respectively as follow [3];
(19)
(20)
(21)
In these cases, the identical eigenfunctions of a bar in transverse vibration can be obtained. For
hinged- hinged supports, the eigenfunctions are [3];
i (x ) = 2 sin (i x )
(22)
i ( x ) = cosh (i x ) cos(i x )
cosh (i ) cos(i )
(sinh (i x ) sin (i x ))
sinh (i ) sin (i )
(23)
tan (i ) = tanh (i )
(24)
The lowest two roots are obtained from the numerical solutions of Eq.(24) as 1 = 3.9266 and
2 = 7.0686 .
Then, for clamped-hinged supports, the eigenfunctions are the same as the eigenfunctions for
clamped-clamped but i are roots of
cos(i ) cosh (i ) = 1
(25)
For i = 1 and i = 2 , the obtained roots from the numerical solution of Eq. (25) are 1 = 4.7300 and
2 = 7.8532 .
4. Numerical Results
In numerical analysis of the considered equation of motion, the displacement and Morison forces changing with
time are separately studied by taking the three different cases including still water, wave and current. The
variation of drag and inertia coefficients in still water, dimensionless slenderness in current, wave frequency in
wave are analyzed with respect to the displacement and Morison force differentiations.
820
CD=0.9
2E-006
CD=1.4
CD=1.4
CD=1.6
CD=1.6
1E-007
MORISON FORCE
DISPLACEMENT
1E-006
-1E-007
-1E-006
-2E-007
-2E-006
10
TIME
10
TIME
(a)
(b)
Figure. 1. For hinged-hinged pipe in still water, (a) Time history with different values of
with different values of
In Fig.1, It is said that changing in C D does not effect the considered system. This is valid for other
support conditions of pipe because drag coefficients are independent from the mentioned conditions.
In second stage, effects of CM varying in same condition and their graphics after the mentioned
analyses are submitted.
CM=1.2
2E-007
1.2E-005
CM=1.4
CM=1.2
CM=2.1
CM=1.4
CM=2.1
8E-006
1E-007
MORISON FORCE
DISPLACEMENT
4E-006
-4E-006
-1E-007
-8E-006
-1.2E-005
-2E-007
0
10
10
TIME
TIME
(a)
(b)
Figure 2. For hinged-hinged pipe in still water, (a) Time history with different values of
As seen in Fig.2, changes of mass coefficients effect both amplitude and frequency of pipe system. It also
causes the small differentiations of the Morison forces with respect that of system.
821
6E-006
b1=0.4
b1=0.4
b1=0.9
b1=0.9
b1=1.3
b1=1.3
4E-006
1E-007
MORISON FORCE
DISPLACEMENT
2E-006
-2E-006
-1E-007
-4E-006
-6E-006
-2E-007
0
10
TIME
10
TIME
(a)
(b)
Figure 3. For hinged-hinged pipe in still water, (a) Time history with different values of
with different values of 1 . In both (a) and (b) figures,
The little fluctuations of amplitude and frequency relating on displacement of pipe are induced by the
changing as shown in Fig.3. 1 . The peak point of this motion is especially observed in case of the high density
of water surrounding thin-walled pipe transmitting the low density of fluid.
x=10
0.02
x=35
0.0012
x=10
x=50
x=35
x=200
x=50
x=200
0
MORISON FORCE
DISPLACEMENT
0.0008
0.0004
-0.02
-0.04
-0.06
-0.0004
0
10
TIME
10
TIME
(a)
(b)
Figure 4. For hinged-hinged pipe in current, (a) Time history with different values of
It is known that the slenderness of submarine pipelines are rather small because of being large structures
according to the ratio of length/diameter. Therefore, the slender pipes influence to increase the amplitude and
frequency of motion.
822
0.006
W=8
W=14
W=8
W=21
W=14
W=21
0.004
MORISON FORCE
DISPLACEMENT
0.002
-0.002
-2
-0.004
-0.006
-4
10
TIME
10
TIME
(a)
(b)
Figure 5. For clamped-clamped pipe in wave condition, (a) Time history with different value of
time with different value of . In both (a) and (b) figures,
Fig.5 shows that when wave frequency increases, the pipe displacement and the acted Morison force increase.
We can see clearly in this figure that the pipe is moving with the wave frequency.
Hereby, in this section, effects of newly defined dimensionless parameters are shown on pipe displacement and
Morison forces.
1 is
For the pipe conveying fluid, 1
As first finding of this study based on developing the governing equation of motion as beam behavior,
smaller than one all along when the beam element is surrounded by fluid.
might be bigger or smaller than one depending on external fluid density. When it is considered that either
density of external water is higher than that of fluid in pipe or equality their densities, 1 is smaller than one
depending on the thickness of pipe. Additionally, in case of being low-density internal fluid with high-density
external water, thin-walled pipe, and this number may be bigger than one.
Second finding is that the range of pipe slenderness changes approximately within 10 to 50 only for submarine
pipelines and it is seen that such slenderness does not affect its displacement or frequency. When this highslenderness pipe is almost equal to 200, their displacements and frequencies are however brought out.
Third finding is pointed out that the dimensionless kinematic viscosity of external water effected by neither the
modulus of elasticity of pipe nor fluid density. When steel and polyethylene submarine pipes and the external
fluid composed of fresh water, salt water and mud are taken into consideration, this dimensionless number are
calculated close by 10-9 .
Forth finding is that the dimensionless wave frequency. If the wave frequency increases, pipe displacement and
effected Morison force increase.
The last finding, it is regarded that the dimensionless flow velocity up to about 300 almost causes the linear
variation between the Reynolds number Re and drag coefficient. While changing these numbers in
823
300<Re<3x105, the drag coefficients become also settled about 1.2. When it reaches to 3x105, the drag
coefficient also approaches to 0.25. At this point, Re is called as the critical flow velocity. Finally, this study
especially answers the questions regarding on the constant drag coefficient or uniform flow velocity.
References
1) Paidoussis, M. P. and Li, G. X., 1993, Pipes conveying fluid: a model dynamical problem, Journal of Fluids and
Structures, 7, 137-204.
2) Semler, C., Li, G. X., and Paidoussis, M. P., 1994,The non-linear equations of motion of pipes conveying fluid, Journal of
Sound and Vibration, 169, 577-599.
3) Snr,B.G, 2010,Bifurcation and chaos of slightly curved pipes, Mathematics and Computational Applications, Vol. 15
No.3, 490-502.
4) Gkku, ., Eren, A., Snr, B.G., 2009,Visko-elastik mesnetlenmi Askdaki Deniz Borularnn etrafnda oluan dzenli
dalga kkenli akn non-lineer dinamik analizi, TBTAK,Manisa (Turkish).
5) Kuiper, G.L., Metrikine, A.V., Battjes, J.A.,2007,A new time-domain drag description and its influence on the dynamic
behaviour of a cantilever pipe conveying fluid,Journal of Fluids and Structures 23 ,429445.
6) Semercigil, S.E, Turan, .F, Lu, S., 1997,Employing fluid flow in a cantilever pipe for vibration control,Journal of Sound
and Vibration , 205(1), 103-111.
7) Kuiper, G.L, Metrikine, A.V., 2004,On stability of a clamped-pinned pipe conveying fluid, HERON, Vol. 49, No. 3.
8) Chakrabarti, S.K, 1987, Hydrodynamic of offshore structures, Springer-Verlag ,United States.
9) Kuiper, G.L., Brugmans, J., Metrikine, A.V., 2008, Destabilization of deep-water risers by a heaving platform, Journal of
Sound and Vibration 310, 541557.
10) Pakdemirli, M., Ulusoy, A. G., Ceranoglu, A.,1994,Transverse vibration of an axially accelerating string, Journal of
Sound and Vibration, 169 (2), p.179.
11)Smer, B. Mutlu, Fredsoe, Jorgen, 1997, Hydrodynamics around Cylindrical Structures, World Scientific, Singapore.
Biographies
Smeyye Snr Smeyye Snr was born in Fethiye/Mula in 1986. She graduated from departments of mathematics
and physics at Sleyman Demirel University in Isparta in 2008. She is still a graduate student in Celal Bayar
University in Manisa. She is studying fluid structure interaction, mathematical methods, and dynamic analyses of
structure. Smeyye Snr knows English as a foreign language.
She has two articles in national and international conferences. Three papers were accepted by Gediz University and
will be presented in June.
mit Gkku mit Gkku was born in Sivas in 1962. He graduated from Akdeniz University in 1983. He
respectively completed his graduate and doctoral dissertations in 1987 and 1993 at Middle East Technical University
and Dokuz Eyll Universty. He has been professor in Celal Bayar University for ten years. He is studying coastal and
harbor engineering, fluid structure interaction, steel structures. mit Gkku knows English as a foreign language. He
was earned eight gifts from multiple institutions.
He has one pressed, four national and three international articles, thirty five in national and international
conferences, twenty two written books or sections of books. One paper was accepted by Gediz University and will be
presented in June. He completed different scientific research projects in Celal Bayar University, TBTAK and other
institutions.
B. Gltekin Snr B. Gltekin Snr was born in Korkuteli/Antalya in 1971. He graduated from Dokuz Eyll
University in zmir in 1994. He respectively completed his graduate and doctoral dissertations in 1996 and 2004 at
Celal Bayar University in Manisa and Dokuz Eyll Universty. He has been assistant professor doctor in Celal Bayar
University for two years. He is studying mechanical engineering, fluid structure interaction, dynamic analyses of
structures, mathematical methods, coastal and harbor engineering. B. Gltekin Snr knows English as a foreign
language.
He has one pressed, an international article, eleven in national and international conferences. Three paper was
accepted by Gediz University and will be presented in June. He completed different scientific research projects in
TBTAK.
824
Abstract. It is preferred that parts made of more durable and lighter than parts used present days have been
used in automotive industry recently. The advantage of using light weighted components is to reduce fuel
consumption and contribute to a cleaner environment as well. For these purposes, it is manufactured
aluminum, magnesium and titanium instead of steel. In addition, for lightness in automotive designs, Al and
alloys which is found widely in nature have been successfully used. Aluminum based composite materials
have been used in radiator, condenser, evaporator, heater and air channel cooler. Lighter automotive parts
have been manufactured from composite made by using with different Al materials. However, it is needed to
several special methods in manufacturing these materials.
These materials manufactured by brazing in the controlled atmosphere have exposed to vibration, sudden
loading and temperature changes in working conditions. Therefore, it is important that the part to be designed
should have the strength and quality to meet the possible loading conditions. From Al based materials,
AA3003 fins 0.15 mm. thickness and AA5754 sheet parts were joined by brazing under protective gas
atmosphere. Impact test was applied to joined samples in this way. The samples having low amplitude,
normal amplitude, high amplitude fins have been subjected to a test at energy of 60 joule in Fractovis Plus
Test instrument. Also, the solid model of samples is consisted of Pro/E software in order to make
theoretically modeling of this Test applied samples, and so as to search their behaviors under dynamic loads,
ANSYS 11.0 WB integrated with this software has been used.The experimental results and theoretical results
obtained were found to be acceptable and close to eachother.
Keywords: Al composite, impact test, ANSYS
1 Introduction
Composite materials are preferred to use in the structures which needs high strength/weight ratio. The stressstrain characteristics and deformation states are more complicated than the metal structures. The faults in
composite materials have negative effects on life-time. Since even a small defect can decrease the life-time of
these materials, it is hard to calculate the life-time of these materials theoretically. Also, it is hard to predict the
location and the degree of damage.
Aluminum is a material which is extensively used in industry. It is preferred in manufacturing of ship hull,
propellers, radiators, and the engines of the vehicles. Therefore, it has an important place in general economy.
Today, almost 50 kg aluminum is used for manufacturing of a car. In this way, 100 kg materials of iron, steel,
and copper are saved. When we compare an aluminum used car with aluminum unused car, the car with
aluminum consumes 1500 liter fuel lesser. Aluminum is also used in the construction of the bridges to decrease
the dead load weight (the weight of the bridge itself) and to increase the useful load-carrying capacity.
825
Additionally, 70 percent of an airplane consists of aluminum material because of the lightness and endurance of
this material. Its conductivity, high corrosion resistance, and the capability to be alloyed with other materials are
other reasons for preferring this material in such an extensive field.
Corrugation is a widely used method in manufacturing of honeycomb structures of high thickness and high
density which work under high temperature (Figure 1). In this method, a metal sheet is corrugated and then
stacked into a block. The sheets are then joined by brazing together. Then, the core is sliced into the desired
thickness, and the corrugated layers are adhesively brazed to face sheets [1-5]. Blocks with corrugated ribbon
plates are kept at pre-defined temperature for pre-defined duration. Blocks are sliced in desired thickness. During
the brazing procedure, the base materials do not melt. Some of the elements in the base materials and brazing
metal diffuse to each other. Honeycomb-structured composite plates have a place in engineering structures with
an increasing rate of usage today. Damages in the aluminum cell composite plates are generally as honeycomb
cell wall buckling around the load contact surface, regional cell crushing, and shear under increasing load
between the cells at the intersection of the surface cover and the cells [6-12].
2 Experimental Study
2.1 Specimens
The specimens were produced in Seco/Warnick brazing oven under nitrogen gas. The specimens of which flux
was used for brazing were kept in the oven for 3.5 4 hours. The filling alloy and flux which were used in
production of the specimens as follows: Brazing filling alloy: AA4343, Flux material: SENFLUX 035.
Radiator specimens consisted of inner fin and bottom and top plates. Inner fin was produced separately, and
bottom and top plates were bought. Inner fins and outer plates were washed to obstruct the quality problems
before assembling. In this way, they were purified from the unwanted layers and chemical remains. Inner fin and
bottom and top plates were dried to let filling alloy and flux adhere on the material surfaces perfectly, and to
prevent the humid surfaces from evaporating and forming oxide layers. Liquid flux with 25% water solution was
sprayed on the aluminum material, which was coated with 10% clad for brazing process, to protect the material
against the corrosive effects and to let clad spread easily. After the drying and flux processes, the materials were
constricted in accordance with the radiator profile. The materials which were used for the specimens are as they
are seen in Table 1.
826
AA5754
AA3003
AA4343
Spectral analysis values of AA5754, AA3003, AA4343, and flux are as they are seen in Table 2, Table 3, Table
4, and Table 5, respectively.
AA5754
Al
Si
Fe
Cu
Mn
Mg
Cr
Zn
Ti
96.148
0.11
0.26
0.006
0.134
3.24
0.097
0.003
0.001
95.2
Si
Fe
Cu
Mn
Mg
Cr
Zn
Ti
Bi
Zr
0.85
0.4
0.05
1.6
0.05
0.05
1.5
0.05
0.05
0.13
Si
Fe
Cu
Mn
Zn
Min.-Max
6.8-8.2
<0.8
<0.25
<0.1
<0.2
Act.
7.454
0.191
0.001
0.006
0.020
Flux
Al
Fe
Ca
LOH
0.1
2.5
Mechanical properties of AA5754 and AA3003 and physical properties of flux are shown in
Tables 6 and 7, respectively.
Table 6. Mechanical properties of AA5754 and AA3003
Tensile
Yield
Strength
Strength
(MPa)
(MPa)
AA5754
235
AA3003
205
Elongation Hardness
(%)
(HB)
173
12
65
160
0.5
827
Density at 20 C (g/cm3)
2.8
565-572
4.5
V type fins of the aluminum radiator fins were used in tests. Figures 2 and 3 show the V type fins. V type fins in
high, low, and normal amplitudes were joined with 1.5 mm bottom and top plates in the oven by using brazing
method (Figure 4).
Fig. 2. V type radiator fin with low amplitude (a), normal amplitude (b), high amplitude (c)
Fig. 3. Brazed V type fins in low, normal, and high amplitudes (from top to bottom) with 1.5 mm aluminum top and bottom
plates
3
20
0,25
no
adjustable
76
19,8425
3,8153
0
0
3,49
Figure 4 illustrates the moment at which the striker hit on the specimen.
828
829
Fig. 7. The specimens with V type fins in low (a), normal (b), and high (c) amplitudes after the tests for 60J
Low
Normal
High
(mm)
amplitude
amplitude
amplitude
Ansys Results
14,22
14,88
16,19
16,82
20,03
18,16
15,4
25,7
10,8
As seen in the impact test results, the highest deformation occurred on the specimens with the fin of normal
amplitudes also the lowest deformation occurred on the specimens with the fin of low amplitudes under dynamic
loads. According to the finite element analyses results higher deformation occurred on the specimens with the fin
of high amplitudes than on the specimens with the fin of low amplitudes under static load. These results showed
that the fin amplitude has an important role on the strength of the materials.
5. Conclusions
Damageoccurring inaluminumsheets, the load-contact areaaround thelocalcrushing and in fins occurring
deformation. Applied sudden loads span of fin at the interface shape and size of the effect of the damage was
investigated theoretically and experimentally. The experimental results and theoretical results obtained were
found to be acceptable and close to eachother. In addition, samples of the reaction forces applied impact force of
830
60 j, in fin of low amplitudes 4200.93 kN, in fin of normal amplitudes 5438.24 kN and in fin of high amplitudes
6607.53 kN.
After the impact tests in the examination of the brazing locations in macro dimension. It was seen that these
locations were not deformed and there were no breaking or separation between two brazed materials. This result
showed the high quality joining of brazing method for the complex components in serial manufacture. Dynamic
loading test was made for the impact test samples with the same plate thickness of the applied energy is
increased the amount of deformation increased. This test is also applied by the device by changing the energy
value of the measurements were 30 and 60 joule. In this case the amounts determined in accordance with the
deformation of the samples. In addition, macro images, the composite inner region damage was investigated.Vtype fins are exposed to energy 60J, respectively the size of large deformations, the specimens with the fin of
high, normal, low amplitudes were observed.
References
1- Y.M. Jen. C.W. Ko and H.B. Lin. Effect of the amount of adhesive on the bending fatigue strength of adhesively bonded
aluminum honeycomb sandwich beams. International Journal of Fatigue. 31(3). 455462. 2009.
2- M.A. Hazizan and W.J. Cantwell. The low velocity impact response of an aluminum honeycomb sandwich structure.
Composites: Part. B. 34(8). 679687. 2003.
3- M.O. Kaman. M.Y. Solmaz. K. Turan. Experimental and Numerical Analysis of Critical Buckling Load of Honeycomb
Sandwich Panels. Journal of Composite 44 (21). 2010.
4- J.K. Paik. K. Anil. A.K. Thayamballi. and Kim. G.S. The strength characteristics of aluminum honeycomb sandwich
panels. Thin-Walled Structures. 35(3). 205231. 1999.
5- M. He and W.. Hu. A study on composite honeycomb sandwich panel structure. M. Materials and Design. 29(3). 709
713. 2008.
6- A.R. Othman and D.C. Barton. Failure initiation and propagation characteristics of honeycomb sandwich composites.
Composite Structures. 85(2). 126138. 2008.
7- I.M. Daniel and J.L. Abot. . Fabrication. testing and analysis of composite sandwich beams. Composites Science and
Technology. 60(12-13). 2455-2463. 2000.
8- P. Bunyawanichakul. B. Castani and J.J. Barrau. Non-linear finite element analysis of inserts in composite sandwich
structures. Composites: Part B. 39(7-8). 1077-1092. 2008.
9- L. Aktay. A.F. Johnson and B.H. Kroplin. Numerical modelling of honeycomb core crush behaviour. Engineering
Fracture Mechanics. 75(9). 26162630. 2008.
10- S. Belouettar. A. Abbadi. Z.Azari. R.Belouettar and P. Freres. Experimental investigation of static and fatigue behaviour
of composites honeycomb materials using four point bending tests. Composite Structures. 87(3). 265273. 2009.
11- T. Fiedler and A. Ochsner. Experimental analysis of the flexural properties of sandwich panels with cellular core
materials. Materialwissenschaft und Werkstofftechnik. 39(2). 121-124. 2008.
12- Y.P. Siow and V.P.W. Shim. An experimental study of low velocity impact damage in woven fiber composites. Journal
of Composite Materials. 32. 1178-1202. 1998.
Biographies
N. Sinan Kksal graduated from Mechanical Engineering Department at Dokuz Eyll University Engineering and
Architecture Faculty in 1990. He received B.Sc. in Mechanical Engineering, M.Sc. and Ph.D. in Mechanical Engineering at
Celal Bayar University. In 1994, he joined Department of Mechanical Engineering at Celal Bayar University as a research
assistant. Then he became a lecturer at the same department in 2004, and an assistant professor there in 2009. He is engaged
in the research of fracture mechanics, refractory materials, wear and friction, welding technology, manufacturing methods
and artificial neural networks. He has published articles in national and international journals. He has delivered presentations
at national and international conferences.
lyas Trkmen 23 February 1986. was born in Izmir. Turkey. Completed primary and secondary education in Izmir. He has
completed undergraduate education in Manisa and still continues study master. Undergraduate degree was in Mechanical
Engineering. Celal Bayar University Faculty of Engineering. Also Graduate study at the Institute of Science in Mechanical
Engineering in the same university is continuing. The study areas; welding methods. plastic forming. and composite
materials.
831
1,2
taskiranoz@itu.edu.tr, 2ergene@itu.edu.tr
Abstract. In this study, diesel spray injected into a constant volume combustion chamber was modeled with
KIVA3V R2 code. In order to take into account effect of nozzle geometry, one dimensional nozzle flow
model was used with KIVA3V R2 . This model calculates nozzle discharge coefficient and updates the initial
SMR and droplet velocities. Droplet break up was modeled by KHRT sub-model and droplet collisions were
modeled by a new trajectory-based sub-model. To determine the collision outcome, impact parameter
calculation was used in the new collision model. Diesel fuel is represented by a diesel oil surrogate
mechanism which consists of 71 species and 323 reactions. The numerical spray results are validated by data
of macroscopic spray characterization obtained by constant volume experiments.
1 Introduction
Many researchers are intensively interested in spray characteristics, ignition and combustion process of diesel
spray to reduce harmful emissions and fuel consumption of diesel engines. To enable better fuel mixing process,
the prominent number of experimental and numerical studies focused on new injection strategies and new nozzle
designs, since the nozzle geometries affect the flow behaviour inside the nozzle and thus spray atomization
outside of it. Future engine problems and optimizations could be dealt with numerical modeling programs.
However, numerical modeling of internal engine has complexity of moving boundary, turbulence, two-phase
flow and combustion chemistry. KIVA, which is a non-commercial CFD program for engine research, describes
the spray dynamics and combustion process with its related sub-models. Open source code of KIVA enables to
improve sub-models for better modeling outcomes of spray and combustion modeling. In KIVA code, equations
of motion for the representation of both gas and liquid phase are based on the conservation laws of energy and
momentum. They are solved by using Eulerian and Lagrangian formulation for gas and liquid phases
respectively. Eulerian gas phase equations have source terms coming from the vaporization of liquid phase and
droplet gas interactions. Conservation equations of chemical species have source terms due to change of
concentration by chemical reactions.
In this study, KIVA3V R2 version was used in numerical calculations. The effects of the nozzle geometry on
fuel spray injection are not taken into account in the standard code. One dimensional nozzle flow model of Sarre
et al. [1] was implemented into the KIVA3V R2 version to model effect of the nozzle geometry. Inlet rounding
of the injector nozzle could be taken into account by using this model. Droplet collision model of the standard
832
code was changed with a trajectory based collision model to obtain more realistic collision outcomes. After that,
diesel spray injection in a constant volume combustion chamber is numerically modeled and compared with
experimental data.
2 Numerical Modeling
Todays diesel engines use common rail injection systems which enable to inject fuel at high injection
pressures to get better air-fuel mixture formation. High injection pressures attracted attentions on injector nozzles
because of their role on spray formation. It was shown that the nozzle type and nozzle geometry (L/D ratio) have
influence on spray formation [2,3]. In numerical calculations, effect of nozzle geometry has to be taken into
account, but resolving the flow inside the nozzle needs Eulerian formulation which brings problem of primary
break-up of fuel. Instead of Eulerian formulation, simple models are used to simulate the flow inside nozzle.
KIVA code uses Lagrangian framework for the liquid phase while the continuous gaseous phase is modeled
by the standard Eulerian conservation equations [4]. In the Lagrangian framework, there are phenomenological
sub-models that describe the break-up, collision and evaporation of droplets for spray modeling. One
dimensional flow model of Sarre et al. [1], which is based on theoretical analyses and previously reported
experimental results, describes possible flow types inside injector nozzle. To investigate effect of nozzle
geometry, they have implemented one dimensional flow model in two different CFD codes. The reported results
showed that the numerical results could be improved by taking into account nozzle geometries and injection
conditions [1]. One dimensional nozzle flow model of Sarre et al. [1] assumes that the flow inside the nozzle is
turbulent for a first approximation. Model uses experimental and theoretical loss coefficients to calculate nozzle
discharge coefficient. Flow velocity and pressure at contraction point are calculated by using Nurricks [5]
expression and Bernoulli equation respectively. If the pressure at contraction point is lower than fuel vapor
pressure, then flow is assumed to be fully cavitating. At this condition, nozzle discharge coefficient is calculated
by using equation of cavitation number and contraction coefficient [1]. According to nozzle discharge
coefficient, velocity at the nozzle exit and initial droplet SMD are calculated.
One dimensional nozzle flow model was used with KIVA3V R2 code for the numerical calculations
presented in this study. In order to test the nozzle flow model two different nozzle inlet configuration was used
for calculations. Ratio of the inlet rounding radius to nozzle diameter (r/D) were chosen for sharp inlet, SI, 0.01
and for rounded inlet, RI, 0.1. The injection duration and injection profile were the same for both SI and RI
cases. The computed discharge coefficient and effective injection velocity are presented in Fig.1. Inlet rounding
has significant increasing effect on nozzle discharge coefficient. At the beginning of injection, the flow is
turbulent and when the velocities increase, the resulting cavitating flow decreases the discharge coefficient.
Increasing the rounding ratio decreases the inclination of cavitation and as a result of this, average discharge
coefficient increases. Initial SMD of injected droplets varies according to effective nozzle discharge area. Nozzle
discharge area reduction has increasing effect on injection velocity. As the flow contraction increases, effective
flow area reduces and thus effective velocity increases.
833
0.80
600
Velocity [m/s]
1.00
Cd
0.60
0.40
SI
0.20
0.00
0.00
RI
0.20
0.40
Tim e [m s]
400
200
0
0.00
0.60
Injection velocity
Eff. inj. velocity-SI
Eff. inj. velocity-RI
0.20
0.40
Tim e [m s]
0.60
Fig.1. Sharp inlet (SI) and Rounded inlet (RI) outputs obtained by using one dimensional nozzle flow model of Sarre et
al.[1], (a) Nozzle discharge coefficient, (b) Injection velocity profiles for KIVA input
After the injection of fuel spray, droplet dynamics become important for air-fuel mixing process. Since the
droplet diameter distribution, vaporization and penetration are highly effected by break-up model, it is the most
effective sub-model among the others. Kelvin Helmholtz and Rayleigh Taylor model, KHRT, based on KelvinHelmholtz instability growing on the droplet and instable RT waves grow at the backward direction of the liquid
was combined and used as a hybrid model by Beale and Reitz [6]. After that, KHRT breakup model intensively
used in modeling diesel spray simulations [7].
Droplet collision modeling is another important research area in spray modeling calculations. Recent
experimental researches and numerical models on spray collision are discussed in the Posts study [8] and
Munnannurs thesis [9]. The standard collision model of the KIVA code has the disadvantage of being mesh
dependent since there is a requirement that the parcels must be in the same cell to be taken into the collision
calculation [10]. It was shown in Nordins thesis [10] that standard KIVA model gave mesh depended results
and he proposed a trajectory based model for droplet collision, which reduced mesh dependency. In this study, a
new collision model, which was capable of calculating impact parameter of intersecting droplets was used. In
Fig. 2, two intersecting droplet is represented with subscript 1 and 2. It is assumed that droplet-2 is going to
collide with droplet-1 with a relative velocity U at the point P1t. Trajectory of droplet-2 intersects droplet-1at
the points P1t and P1d. Lagrangian tracking of particles enable to calculate these intersecting points and
intersecting time. Positive two roots (tc1 and tc2) of the following quadratic equation give intersecting times. If
the intersecting time is reached in the calculation time step collusion occurs and the result is droplet coalescence
or grazing collision, which is determined by using the impact parameter. In the new time step all intersecting
droplets are re-calculated for colliding conditions.
x + 2 ( ux ) tc + u tc2 = ( r1 + r2 )
2
(2 )
The impact parameter of the colliding droplets, b, could be obtained from geometric relation.
b=
( r1 + r2 )2 ( u (td tc ) 2 )
(2 )
834
P1d
P1t
U
X1
r1
r1 + r2
X2
r2
Instead of using a stochastic method, the new collision model calculates the collision of droplets with their
velocity and position vectors. It eliminates the mesh dependency. However, the drawback of this method is its
dependence on droplet number in calculation. Even today computer capabilities it is still impossible to use real
droplet number in calculations. Intersecting trajectories of droplets increase as the number of representative
droplets increase in the calculations.
The spray modeling gets more intricate when the turbulent spray combustion is involved in calculations. It is
known that thousands of chemical reactions and hundreds of species are involved in the auto-ignition and
combustion process of diesel fuel. However these complex phenomena could be simplified and modeled by
numerical models. The turbulence and chemistry interaction was utilized in a succesful manner by using
Partially Stirred Reactor model, PaSR [10]. Golovitchev et al. [11] adopted PaSR model into the KIVA code. By
using reduced chemical mechanism, it was shown that turbulent combustion of diesel spray very well modeled
by PaSR [10,11]. In this study, diesel fuel was represented by a surrogate mechanism which consists of 71
species and 323 reactions and turbulent combustion was modeled with partially stirred reactor model.
3 Numerical results
Numerical calculations are carried with a constant volume mesh that has 56 mm bore and 86 mm height. The
results of the calculations were compared with experimental results that were obtained with a constant volume
combustion chamber heated internally up to 825 K at 3.5 MPa. The details of the experimental setup and results
can be found in [12]. In order to verify numerical code the dimensions of a hydro-grinded nozzle (0.24 mm
outlet diameter and 0.12 mm rounded inlet radius geometry) was given as initial conditions for numerical code.
Spray penetration of experimental and numerical results at 300 K ambient temperature are given in Fig.3 (a) and
(b). In Fig 3 (a) 0.6 ms injection duration and 120 MPa injection pressure was applied at 3 MPa ambient
pressure, and in Fig 3 (b) 0.6 ms injection duration and 120 MPa injection pressure was applied at 2 MPa
ambient pressure conditions. Diesel surrogate fuel, KH-RT breakup model and new collision model are used in
the calculations. Though there are some small deviations from the experimental data, it is seen that numerical
and experimental results show very good agreement.
835
50
50
Distance [mm]
Distance [mm]
40
30
20
3 MPa, 0.6 ms, 120 MPa
Numerical Result
10
40
30
20
10
0
0.0
0.2
0.4
0.6
Tim e [m s]
0.8
0.0
0.2
0.4
0.6
Tim e [m s]
0.8
Fig. 3. Numerical and Experimental results of spray tip penetration (Experimental results are given with triangle sign:
Ambient Pressure, Injection Duration, Injection Pressure) (a) 3 MPa ambient pressure condition (b) 2 MPa ambient pressure
condition
In Fig.4, experimental images obtained by shadowgraphy technique for 1.5 ms injection duration and 50 MPa
injection pressure at 3 MPa ambient pressure condition and numerical images of the same conditions are
presented together in Fig. 4. The models used in numerical calculation could predict spray penetration and cone
angle compatible with experimental results. Numerical calculation presents additional data for spray
investigations such as distribution of droplet size, fuel vapor, temperature or combustion products. Collecting
experimental data of these parameters is very hard and sometimes impossible for the liquid dense regions. In Fig.
4. the numerical spray images are colored with droplet radius. At the upstream region droplets have
approximately same diameter size with nozzle and they get smaller due to droplet break-up as they penetrate to
downstream part.
[cm]
Fig. 4. Experimental and numerical images of spray development obtained at 3 MPa ambient pressure with 1.5 inj. duration
and 50 MPa inj. pressure, the first figure on left is obtained at 0.03 ms and last one on the right at 0.6 ms, the time interval
between the images is 0.0625 ms
For the autoignition calculations, the numerical sub-models and injection conditions are summarized in Table
1. The experiments were done with 11.3 and 6.5 mg diesel fuel, at 3.5 MPa pressure and 830 K temperature
conditions. Similar to Tao et al. [13], the ignition delay is determined by the 100 K increase of maximum
temperature in the volume. As shown in Fig. 5, auto-ignition started at 1.2 ms after the start of injection for 11.3
mg diesel fuel simulation. The corresponding experimental images could be seen at the right side of the
numerical results presented in Fig. 5. The autoignition started at fuel rich regions and multiple points of
836
downstream section of the spray. The spray ignition continues as a diffusion flame at the spray boundary. High
injection rates keep the flame at downstream region; prevent to get closer to nozzle tip. Both the numerical and
experimental results are consistent with Decs [14] explanations of diesel spray combustion.
350 K
Injection mode
Turbulence model
Standart k-
0.1 m2/s2
0.01 m
Evaporation model
Breakup model
KHRT
Collision model
New model
Combustion model
Turbulence/Chemistry interaction
PaSR model
[K]
Fig. 5. The temperature distribution of the numerical results, 11.3 mg fuel with P=3.5 MPa and T=830 K initial conditions.
The conditions tested in Fig. 5 were applied for a sharp inlet nozzle for numerical calculation. The ignition
delay duration reduced to 1.0 ms since the sharp inlet nozzle geometry produced smaller droplets due to
reduction at effective nozzle discharge area. When the ambient temperature increased to 850 K for the same
conditions in Fig.5, ignition delay duration reduced to 1.0 ms, as expected.
Smaller amount of fuel shortens the ignition delay, because of less cooling effect of injected fuel. The
simulation of 6.5 mg fuel with 0.6 ms injection duration is presented with experimental images in Fig.6. The
initial chamber conditions are the same as given in Table 1. The autoignition occurs at 0.9 ms which is the same
moment of the experiments. The self-ignition location is predicted very close to the experimental images.
837
Fig. 6. The temperature distribution of the numerical results, 6.5 mg fuel with P=3.5 MPa and T=830 K initial conditions.
4 Conclusion
One dimensional nozzle flow model calculates the discharge coefficient, initial SMR, nozzle discharge area
and initial velocity of nozzles that have different inlet geometries. It is shown that the initial droplet size and
initial velocity are better simulated by using nozzle flow model. The collision model was improved by impact
parameter calculation which enables to predict outcome of colliding droplets. The temperature distribution of the
temporal results of numerical calculations showed that the auto-ignition of the diesel fuel could be predicted with
detailed mechanism of diesel surrogate fuel. A good agreement was achieved between the numerical and
experimental results by means of auto-ignition delay, ignition location and flame propagation.
References
1.
2.
3.
4.
5.
6.
7.
8.
9.
10.
11.
12.
13.
14.
Sarre K.C., Kong S.C, Reitz R.D., (1999). Modeling the effects of injector nozzle geometry on Diesel sprays. SAE
Paper, 1999-01-0912.
Suh H.K., Lee C.S., (2008). Effect of cavitation in nozzle orifice on the diesel fuel atomization characteristics,
International Journal of Heat and Fluid Flow, 29:10011009
Benajes J., Molina S., Gonzlez C., Donde R., (2008). The role of nozzle convergence in diesel combustion. Fue,
87:18491858.
Amsden, A.A., (1997). KIVA-3V: A block structured KIVA program for engines with vertical or canted valves,
LA-13313-MS.
Nurick W.H., (1976). Orifice cavitation and its effect on spray mixing, ASME Journal of Fluids Engineering, 98,
681-687
Beale, J.C. and Reitz, R.D., (1999). Modeling spray atomization with the Kelvin-Helmoltz/Rayleigh-Taylor Hybrid
Model, Atomization and Sprays, 9, 623-650.
Baumgarten, C., (2006). Mixture Formation in Internal Combustion Engine, 85-210 Springerlink, ISBN 3-54030836-9_4
Post S.L., Abraham J., (2002). Modeling the outcome of dropdrop collisions in Diesel sprays. International
Journal of Multiphase Flow 28:9971019.
Munnannur A., Reitz R.D., (2007). Droplet collision modeling in multi-dimensional engine spray computations.
Ph.D. Thesis, University of WisconsinMadison.
Nordin, N., (2000). Complex Chemistry Modeling of Diesel Spray Combustion, PhD Thesis, Chalmers University
of Technology, Sweden.
Golovitchev, V.I., Nordin, N., Jarnicki, R., Chomiak, J., (2000). 3-D Diesel Spray Simulations Using a New
Detailed Chemistry Turbulent Combustion Model, SAE Technical Paper, 2000-01-1891
Takran .O, Ergeneman M., (2011). Experimental study on Diesel spray characteristics and autoignition process,
Journal of Combustion, Article ID 528126.
Tao, F., Golovitchev, V.I., Chomiak, J., (2000). Self-ignition and early combustion process of n-heptane sprays
under diluted air conditions: numerical studies based on detailed chemistry, SAE Paper, 2000-01-2931
Dec, J. E., (1997). A conceptual model of D.I. Diesel combustion based on laser-sheet imaging, SAE Paper,
970873.
838
Biographies
zgr Ouz Takran Dr. Takran received his BSc. degree (1999) from Naval Academy MSc. (2004) and Ph.D (2011)
degrees in Mechanical Engineering from Istanbul Technical University. His research focuses on internal combustion engines
and modeling of fuel spray.
Metin Ergeneman Prof. Ergeneman was graduated from Istanbul Technical University, Mechanical Engineering
Department in 1966. From 1970 to 1973, he worked as a researcher in German Aviation and Space Research Organization.
He received his Ph.D degree in 1985. He became an Associate Professor in 1988 and was appointed to professorship in 1997.
His main interests are internal combustion engines and alternative fuels. He is currently a Mechanical Engineering Professor
in Automotive Sciences in Istanbul Technical University.
839
1,
Abstract. Turkish boron minerals are traditionally concentrated by wet methods and the efficiency of such
ways are generally not satisfactory. In addition to the low efficiencies, such as 30-35 % B2O3, solid plant
tailings with a B2O3 content of about 14-20% pose serious environmental threat. It is clear that a dry
concentration method such as calcination should be preferred over the wet ones. In this study concentrating a
tincal ore from Krka Mine using calcination was investigated. The final results obtained from laboratory
experiments gives better values than already practiced wet method
Keywords: tincal ore, calcination, concentration
1 Introduction
Boron is widely distributed in the form of borates but is never found in the elemental form in nature and
minerals are borax, colemanite, ulexite, tincal, kermite, Major end uses for borates include; fiberglass,
ceramics,agriculture and fertilizer, flame retardant .It is also used for wood, plywood, textile products, cotton,
paper and cellulose, corrosion inhibitor, wood preservatives and pesticides, metallurgy, pharmaceuticals and
cosmetics and nuclear applications.
Turkey is the world's largest supplier of borate mineral products as well as the second largest supplier of
refined borates. Ulexite and colemanite ores are mined at Bigadic, Espiye and Kestelek while tincal (sodium
borate ore) is extracted at the open pit operation at Krka. Minerals are relined into boric acid at the Bandrma
and Emet plants. Tincal is relined into sodium borate chemicals at the Krka and Bandrma facilities [1].
Concentration of boron minerals are generally based on washing with water to separate clay minerals from
boron. In Turkey, in addition to the low efficiencies, the adjacent pond which currently holds hundreds of
thousands of solid plant tailings with a B2O3 content of about 14-20 % pose serious environmental threat. When
all these shortcomings are considered, it is clear that a dry concentration method such as calcinations should be
preferred over the wet ones. As a matter of fact, there are industrial plants in US which employ calcinations as an
effective concentration method [2].
840
2 Experimental
2.1 Materials
An ore sample from Etibank Krka Borax Processing Plant was used for the experiments. B2O3 content of
each screen fraction dependency of grain size are given in Table 2.1.It was found out that B2O3 content increase
toward finer particle size fractions.
Size
Fraction
Weigh
Grade
(% )
Cumulative
(mm)
50.00-25.00
5.92
18.02
5.92
25.00-12.50
21.58
23.92
27.50
12.50-9.51
8.63
23.54
36.63
9.51-4.76
14.93
22.71
51.06
4.76-3.00
10.81
25.17
61.87
3.00-2.00
5.01
22.60
66.88
2.00-1.00
12.50
27.55
79.39
1.00-0.50
9.00
28.30
89.38
0.50-0.00
11.62
29.91
100.00
TOTAL
100.00
25.17
841
(b)
(a)
Table 3.1 Concentrate grad.and recov. obtained at optimal particle size and temp.
Size fraction
-1.00+ 0.00
-3.00+ 1.00
-4.76+3.00
-9.51+4.76
-12.50+9.51
-25.00+12.50
-50.00+25.00
Optimum temp.
375
375
375
475
475
475
525
Grade (%
44.2
50.3
54.4
56.0
51.4
49.2
48.1
842
Recovery
74.1
90.1
89.1
88.3
82.0
90.4
63.2
(a)
(b)
Fig.3.1.The effect of particle size on B2O3grade(a) and recovery(b)
3.2. Effect of the temperature
As seen fig.3.2 and table 3.2 it can be said that there is a linear relation between heat and B2O3 grade for coarse
material, while for fine grains, heat increase makes an adverse effect on concentrate grade. To reach a desired
level of grade, low degrees of heating for fractions of fine materials and proportionally high degrees of heating
for fractions of course materials is required. On the basis of the achieved results it can be stated that the optimum
size fraction -9.51+1.00 mm at 3750C, concentrate with B2O3content of above 50% can be obtained. It can be
seen that the recovery decreases from finer particles to fine particles at 3250C.The recovery generally increases
towards coarse particles at 3750C and, above 90% recovery could be obtained at the -4.76+3.00 mm size
fraction. The highest recovery had been obtained at the -9.51+4.76 mm size fraction at 4250C.
843
(a)
(b)
Fig.3.2. The effect of temperatures, at various sizes, on B2O3grade(a) and recovery (b)
Table 3.2 Concentrate grades and recoveries obtained at optimal temperatures
Temperature (0C)
Grade (% B2O3)
Recovery (%)
325
375
-3.00+1.00
-4.76+3.00
50.2
54.4
76.3
89.1
375
-9.51+4.76
56.0
88.3
425
-12.50+9.51
49.2
90.4
475
-12.50+9.51
47.3
68.6
844
When B2O3 grade and recovery are carefully thought about simultaneously, for a concentrate of 60.0% B2O3
and a recovery of over 90%, the ore should be crushed down to minus 9.51mm for 15 min. calcination periods a
result these tests which were carried out to determine optimal calcinations time show that 15 min calcinations
time was optimal for the size fraction -9.51+3.00 mm.
(a)
(b)
Fig.3.3. The effect of calcinations time, at various sizes, on B2O3 grade (a)and recovery(b)
845
Table.3.3 Concentrate grades and recoveries obtained at optimal calcinations time and size fractions
Calcination time
Grade (% B2O3)
Recovery (%)
(min)
5
10
-4.76+3.00
-4.76+3.00
49.1
56.2
93.4
96.0
15
-9.51+4.76
59.4
92.6
20
-9.51+4.76
58.4
89.3
25
-9.51+4.76
59.7
94.2
(a)
846
(b)
Fig.3.4.The effect of feed rate, at various sizes, on B2O3grade(a) and recovery(b)
Table.3.4 Concentrate grades and recoveries obtained at optimal feed rate and size fractions
Grade (% B2O3)
Recovery (%)
5
10
-9.51+0.00
-9.51+0.00
54.0
57.2
87.2
94.2
15
-9.51+0.00
49.3
87.2
20
-9.51+0.00
48.1
80.4
25
-9.51+0.00
43.7
81.3
4 Conclusions
A semi-pilot scale rotary furnace was utilized for the experiments and 4250C calcinations temperature,15
minutes calcinations time, minus 9.51 mm particle size and 3 kg/h feed rate were determined as optimum
conditions. On the other hand, as can be seen from the results of plants(in Krka) and laboratory experiments,
summarized Table 4.1 for comparison, concentrate and tailing ( B2O3) grades in calcinations method represent
better values than already practiced wet method.
Table 4.1 Comparison of the results of wet method and calcinations method
Calcination Method
The Existing
Plant
24-26
24-26
57.2
32-34
Tailing grade(B2O3)
Recovery (%)
9.1
14
94.2
82-84
To obtain concentrates, from the calcined product of the rotary furnace, an air classifier was utilized and the
results are given in Table 4.2
847
Weight
Product
Grade (B2O3)
(%)
Yield (%)
Concentrate
68.52
50.46
93.20
Tailing
31.48
8.01
6.80
Total
100.00
37.10
100.00
One of the superior aspects of calcinations method is its environmental safety. Maximum allowable limit of
B2O3 content in agricultural irrigation water is 0.4%.The wet process tailings contain 14% B2O3.Tailing
disposal is a major problem at boron plants in Turkey. B2O3 grade of concentrate is higher and tailings will
contain less B2O3 compared to wet methods[3,4].The low solids content of such plant tailings require large
ponds with considerable impermeability problems. In calcinations process the tailings are in solid form therefore
the waste disposal problem is minimized [5].
Acknowledgement
The authors are grateful to Etibank Krka Borax Processing Plant for provision of samples and information on
plant operation; and special thanks are also due to Prof. Dr. ner pekolu, Prof. Dr. Ali Akar and Prof. Dr.
Mevlt Kemal for their valuable discussions.
References
1..DPT BeYllk Kalknma Plan zel htisas Komisyonu Endstriyel Hammaddeler Raporu,(1994).
2.US Bureau of Mines,Mineral Industry Surveys,Boron in 1993(Washington).
3..Aytekin,Y.,Akdag,M.and Batar,T.,Tinkal Cevherinin .Patlatma Yoluyla Zenginletirilebilirliinin ve bu
yntemin bilinen mevcut yntemler yerine ikamesinin aratrlmas, TBiTAK MAG 838, (1992), zmir.
4.elik,M.S.;Uzunolu,H.A.
and
Aslan,F.Decrepitation
properties
ofsome
boron
minerals,
Powder
Technology,(1994).
5.Aytekin, Y.;Akdag, M. and., Batar, T., Expansion experiments on Etibank Krka Ore at TBTAK
Invest.Center,Gebze(1991).
848
Abstract. In this study, dynamic analyses of a string that is under the effect of a tensional force and designed
as a visco-elastic material according to fractional Kelvin-Voigt Model are investigated. In this fractional
visco-elastic model, viscose behavior of the material is defined by fractional derivatives.The governing
equation is solved by using the perturbation technique. Analysis of fractional viscosity by multiple times
scale in these type problems is a new subject in the literature. In this method, first of all, linear solutions of
the problem which are natural frequencies and mode shapes are found. By using these results and solvability
conditions, non-linear frequency values and time-displacement graphs are obtained. Effects of fractional term
on non-linear frequency and time history graphs are investigated.
Keywords: visco-elastic string, fractional, multiple time scale
1 Introduction
Structures can be divided into some groups such as cables, strings, beams, plates, etc. String, which is selected
structure in this paper, is used in different type of structures such as various bridges, fitness centers, and roof
types with big spans. Modern structures have becoming more complex. So the demands of materials have
become more various. Many new materials have been produced with the development of material science.
Lately, classic Hook law, which is known as an elastic model, has started to fail to satisfy the expectations while
modeling the behavior of developed new materials. As a result of this, many new visco-elastic models have been
found to explain the behavior of the material such as Kelvin-Voigt, Maxwell, Zener, Burgers models etc. KelvinVoigt model is the most favorite one among these linear visco-elastic models. Dynamic analysis of the viscoelastic string is important due to common usage in engineering applications. Due to this reality, it is necessary to
improve new structural theories for the analyses of structures modeled with new material models.
In recent years, many researchers are interested in studying the applications of the fractional order system to
physics and engineering. Because, fractional derivatives are useful for describing viscoelastic features of
advanced materials or dissipative forces in structural dynamics. [1]
The seeds of fractional derivatives were planted over 300 years ago. Since then many great mathematicians (pure
and applied) of their times, such as N. H. Abel, M. Caputo, L. Euler, J. Fourier, P. S. Laplace, G. W. Leibniz, A.
V. Letnikov, J. Liouville, B. Riemann, M. Riesz, and H. Weyl, have contributed to this field. During the second
half of the twentieth century, fractional calculus has been applied to almost every field of science, engineering,
and mathematics. Some of the areas where fractional calculus has made a profound impact include viscoelasticity and rheology, electrical engineering, electrochemistry, biology, biophysics and bioengineering, signal
and image processing, mechanics, mechatronics, physics, and control theory. [14] However, many real dynamic
systems are better characterized using a non-integer order dynamic model based on fractional calculus or,
849
differentiation or integration of non-integer order. The concept of fractional calculus has remarkable potential to
change model and control the nature around us.
In this study, our interest is non-linear dynamic analyses of the fractional viscoelastic string under the effect of
tensional force. The mathematical model of this beam is taken in Ref.[5]. In fractional viscoelastic model,
damping forces is described by fractional derivative. All available studies [69] are concentrated on axially
moving strings constituted by the differential relationships such as the Kelvin model [6], the Maxwell model [7]
and the standard linear solid model [8] and the integral relationship, the Boltzmann superposition principle [9].
The paper given in Ref.[5] adopts the fractional differentiation relationship to constitute the axially moving
string. . In this fractional visco-elastic model, viscose behavior of the material is defined by fractional
derivatives. It is difficult to find out the analytical solution of nonlinear fractional-order system. So in most of
literature in the fractional calculus, approximate and numerical solutions are important to analyze the nonlinear
fractional-order equations. There are a lot of numeric and semi analytical methods such as perturbation method
and serial solutions respectively. Also, perturbation method is an asymptotic serial solution method. Multiple
times scale, pedestrian expansion and Lindstedt-Poincare method etc are well known perturbation methods. In
addition to this, multiple times scale has two applications that direct and discretize. in Ref.[5], the researchers are
used Galerkin method, approximate solution and finite difference as solutions techniques. In this study, the
multiple time scale methods direct application, approximate solution technique, is used to solve nonlinear
fractional differential equation for the reason that the direct application of multiple time scale method gives more
reliable results than other approaches. In this method, firstly, linear solutions of the problem which are natural
frequencies and mode shapes are found. By using these results and solvability conditions, non-linear frequency
values and time-displacement graphs are acquired. Effects of fractional term on non-linear frequency and time
history graphs are examined.
which denotes time in dynamical problems and is a positive small parameter. r is the
fractional differential operator that denotes the th derivative of the related function with
respect to t and it is defined in the Caputo sense as[1, 3]
r RS
where is gamma function.
NN
J
O
N
( f(2)
Two boundary conditions are enough to solve the governing equation. Its assumed that the
The method of multiple time scale is used to determine approximate solution of Eq. (1).
850
o: shows different time scales and the fast and slow time scales are
oO S o m(5)
Then, the derivatives with respect S for first, second and fractional derivatives transform into
%
O & m & p
O & m
O
Substituting these expansions into Eq. (1) and Eq. (3) and separating at each order of 8 ,
O ( S O f S (
8qrO 7 EE 7grO r O & tO OE OEE & OEE rO OE & OE rO OE OEE
g
g
( S f S ((8,a,b)
is obtained.
uw ,x
yyyy
O R oO o n 8 1: t uw ,x & 1
s: R(9)
:t
and their conjugates. Also here s: and 6: are mode shape functions and natural frequencies
of string, respectively.
If this solution is written into Eq. (7), s: must satisfy the following equations and boundary
conditions;
s( sf (
s: R |g \} ~R(11)
To determine the first order solution of m, Eq. (9) is substituted into right hand side of Eq.
(8.a) and the Eq.( 8.a) is rearranged according to secular and non-secular terms,
851
The structure of 1: is appointed by using solvability condition that the secular terms are
equal to zero.
r 1: & O 1: yyyy
1: ((13)
where,
O
?x uw
uw
O is a complex number so it can be divided into two part which are O & Oa using below
relation
& \}
(15)
7 g6:
x 7 g6: \}
?
are obtained.
If Eq. (13) is rewritten by using Eq. (16-17) and 1: is written in polar form,
are obtained.
Eq. (19) is substituted into (18) and below equations can be obtained to separate real and
imaginary terms.
:E &
:E &
: ((20)
: ((21)
: o
x ,
(22)
x
: o 7
x
x
} o (23)
R S _
8S
x
x
u x x u
x
t w
& ee s: R(24)
From Eq.(24), we cannot separate the nonlinear natural frequency term due to fractional
derivation. For justification, we are taking into account for Eqs. (1 and 12). Using Eqs.(1 and
852
12), we find model equation and solution procedure justification, respectively. In justification
we take f and we find
(25)
When uniform axially speed of the axially moving string is taken zero, Eqs. (11 and 33) in
Ref.[10] turn into above equations.
is constant.
When is constant and a0 increases, Fig.1 shows that, the string displacement and frequency
increase. We know that vibration frequency of string is proportional to displacement
amplitude ( a0 ) from equation of motion of solution by time scale method.
853
and a0
is constant.
When a0 is constant and increases, it is seen in Fig.2 that, the string displacement and vibration
frequency of the string decrease.
3 Conclusion
Mathematic model of a string which is designed by fractional visco-elastic model is a partial differential
equation that has cubic non-linear terms. Direct application of a time scale method is selected to solve the
mathematic model. In this method, first of all, linear solutions of the problem which are natural frequencies and
mode shapes are found. By using these results and solvability conditions, time-displacement graphs are obtained.
Effects of fractional term on non-linear frequency and time history graphs are investigated. The non-linear
frequencies are changed with the amplitude of the vibration. For special situation, value of the fractional viscose
term is taken 0 and 1 and results are compared by common solutions.
References
1.
Rossikhin, Yuriy A.; Shitikova, Marina V. (2010). Application of fractional calculus for dynamic problems of solid
mechanics: novel trends and recent results. Applied Mechanics Reviews, Vol.63/010801: 1-52.
2.
Das, Shantanu (2008). Functional fractional calculus for system identification and controls. Berlin. SpringerVerlag.
3.
Ge, Zheng-Ming; Hsu, Mao-Yuan; (2007). Chaos in a generalized van der Pol system and in its fractional order
system. Chaos, Solitons and Fractals 33: 17111745.
4.
Ge, Zheng-Ming; Zhang, An-Ray. (2007). Anti control of chaos of the fractional order modified van der Pol
systems. Applied Mathematics and Computation 187: 11611172.
854
Chen, Li-Qun; Zhao, Wei-Jia; Zu, Jean W. (2004). Transient responses of an axially accelerating visco-elastic
string constituted by a fractional differentiation law. Journal of Sound and Vibration 278: 861871.
6.
Zhang, L.; Zu, J.W. (1999). Nonlinear vibration of parametrically excited moving belts, part 1: dynamic response.
American Society of Mechanical Engineers, Journal of Applied Mechanics 66(2): 396402.
7.
Zhao, W.J.; Chen, L.Q. (2002). A numerical algorithm for non-linear parametric vibration analysis of a viscoelastic moving belt. International Journal of Nonlinear Science and Numerical Simulations 3(2): 139144.
8.
Fung, R.F.; Huang, J.S.; Chen, Y.C.; Yao, C.M. (1998). Nonlinear dynamic analysis of the visco-elastic string with
a harmonically varying transport speed. Computers and Structures 66(6): 777784.
9.
Chen, L.Q.; Zu, J.W. (2003). Parametric resonance of axially moving string with an integral constitutive law.
International Journal of Nonlinear Science and Numerical Simulations 4(2): 169177.
10. Chen, Li-Qun; Chen H.; Lim C.W. (2008). Asymptotic analysis of axially accelerating visco-elastic strings.
International Journal of Engineering Science 46: 976-985.
Biographies
B. Gltekin Snr B. Gltekin Snr was born in Korkuteli/Antalya in 1971. He graduated from Dokuz Eyll
University in zmir in 1994. He respectively completed his graduate and doctoral dissertations in 1996 and 2004
at Celal Bayar University in Manisa and Dokuz Eyll Universty. He has been assistant professor doctor in Celal
Bayar University for two years. He is studying mechanical engineering, fluid structure interaction, dynamic
analyses of structures, mathematical methods, coastal and harbor engineering. B. Gltekin Snr knows English
as a foreign language.
He has one pressed, an international article, eleven in national and international conferences. Three paper was
accepted by Gediz University and will be presented in June. He completed different scientific research projects
in TBTAK.
Nurhan Hacolu Nurhan Hacolu was born in zmir in 1988. She graduated from departments of
mathematics Ege University in zmir in 2010. She is respectively still a graduate student and a student in Celal
Bayar University and Anadolu University in Manisa and Eskiehir. She is studying mathematical
methods,mechanical engineering and dynamic analyses of structure. Nurhan Hacolu knows English as a
foreign language.
Two papers were accepted by Gediz University and will be presented in June.
Smeyye Snr Smeyye Snr was born in Fethiye/Mula in 1986. She graduated from departments of
mathematics and physics at Sleyman Demirel University in Isparta in 2008. She is still a graduate student in
Celal Bayar University in Manisa. She is studying fluid structure interaction, mathematical methods, and
dynamic analyses of structure. Smeyye Snr knows English as a foreign language.
She has two articles in national and international conferences. Three papers were accepted by Gediz University
and will be presented in June.
855
Abstract. The aim of this study is to determine stress evaluation of root canal wall and instruments during
shaping curved canals with ProTaper and HeroShaper systems by using three-dimensional finite element
analysis. The Ni-Ti rotary instruments selected according to their different cross-sectional geometries. The
geometries of the selected files were created for each system by using mechanical design and analyze
program I-deas11. The three-dimensional FEM model within a curved canal was established by integrating a
simulated canal and natural tooth. A micro-focus CT scanner was used to scan the models. After integration
and modification a final three-dimensional model with a smooth surface and mostly centered canal was
established. The stresses on the instrument during simulated shaping of a root canal were analyzed
numerically by using a three-dimensional finite element package, LS-DYNA, taking into account the
nonlinear mechanical behavior of the nickel-titanium material. Stress distribution in the instruments and in the
canal walls was recorded during simulated shaping. More stress concentrations were determined in the
HeroShaper cross-section. It was observed there were stress concentration towards the outer aspect of the
curvature in the apical portion of the canal by the shaping with ProTaper file system but contrarily with
HeroShaper file system, stresses were more concentrated with having lower values in the middle portion of
the tooth model. Because of positive rake angle of HeroShaper files, cut more efficiently and generated less
stress on the canal walls than ProTaper files. Stress values of all files were found in the transformation phase
of stress-strain diagram.
Keywords: Curved canals, finite element analysis, HeroShaper, ProTaper, stress
1 Introduction
The three dimensional finite element method which is one of the most advanced simulation techniques in solid
mechanics, is used for dentistry and endodontic therapies. It is used as a tool for the analysis of determining
stress evaluation on the canal walls of the root and design of the instruments used during shaping curved canals
by the file models of ProTaper and HeroShaper systems.
In this study, ProTaper (Dentsply, Maillefer, Ballagues, Switzerland) and HeroShaper (Micro-Mega, Besanon,
France) new generation Ni-Ti rotating tools which were enhanced in recent years are used. The aim is
evaluation of stresses on the canal walls and on the tools by using three dimensional finite element method
during widening the skewed canals. The geometric properties of files and the shapes of root canals, the velocity
values, and the frictional stress modeling could not be used with the true values of these during analyze in
committed studies until today.
856
Creation of an accurate model was aimed in our study with using exact values of file geometries, canal shapes
and structures, frictional stresses, proper velocity. The created model is compared to the clinical specifications to
see if the model is realistic or not.
2 Literature Review
2.1 Root Canal Treatment
The need for some manner of root canal preparation prior to root canal filling has long been recognized as an
essential step in endodontic treatment. Initially, root canals were manipulated primarily to allow placement of
intracanal medicaments, with little attempt to remove completely the organic contents of the root canal system.
In time, the concept of modifying root canal preparations to facilitate the placement of root canal fillings became
part of accepted endodontic practice, but the methods employed for these procedures remained, for the most part,
unrelated both to the true anatomy of root canal systems and to the physical nature of the materials with which
the root canals werepresumed to be filled. Cleaning and shaping the root canal refers to the removal of all
organic substrate from the root canal system and the development of a purposeful form within each canal for
reception of a dense and permanent root canal filling. This process is achieved with instruments of specific
design; namely, endodontic broaches, reamers and files, and certain rotary instruments [1, 2, 3].
1_The root canal preparation should develop a continuously tapering funnel from the root apex to the coronal.
Access cavity, but the degree of taper required is less than for gutta percha fillings in most instances.
2_An optimal silver cone preparation should exhibit an apical collar of two, three, or four millimeters, in which
the dentinal walls are nearly parallel rather than a configuration whose cross-sectional diameter narrows at every
point apically.
3_As in gutta percha preparations, the silver cone preparation should occupy as many planes as are presented by
the root and root canal under treatment. They will exhibit flow.
4_ There should be absolutely no movement or transportation of the periapical foramen during canal preparation.
5_ As is the case with gutta percha, the apical opening should be kept as small as possible.
857
5. Create sufficient spaces during canal enlargement to receive intracanal medicaments and to accommodate
small amounts of periapical exudates should subclinical inflammation follow canal preparation. [1]
3 Methods
The three dimensional finite element method which is one of the most advanced simulation techniques in solid
mechanics, is used for dentistry and endodontic therapies. It is used as a tool for the design and analysis of
determining stress evaluations on root canal wall and on instruments during shaping curved canals with ProTaper
and HeroShaper systems.
The design of a teeth and Ni-Ti rotary drilling files is such a complex process that requires close co-operation
between engineers and dentists. Two Ni-Ti rotary instruments are selected which are HeroShaper and ProTaper
file systems. The main difference between these two file systems is their different cross-sectional geometries.
The geometries of the selected files were created for each system by using three-dimensional finite element
analysis.
The purpose of our study is to determine the stresses occurred on the skewed root canals and on the tools by
using three dimensional finite element methods during shaping the tooth by the Ni-Ti rotary file systems. By
means of using the finite element methods as a tool, threedimensional filing processes will be modeled and stress
values will be predicted.
858
859
Because of the fact that LS DYNA output could be generated by I-deas, each of the files and tooth models were
read on FEMB. A drilling motor was created to generate rotation for the files. Nonmoving areas of the tooth
model are constrained by creating and fixing a Node Set.
860
4 Findings
In our study, the stress distributions on the file models belong to ProTaper and HeroShaper Ni-Ti rotary tool
systems, and on the
slopped mezio bukkal root canal were investigated. File models were driven at 250 rpm
average velocity. Stress distributions on the file models and on the root canal walls were evaluated by three
dimensional finite element method.
4.1 Shaping of root canals middle 1/3 portion by ProTaper Rotary File System:
The modeled S1 file is the first shaping file. At clinical usage, S1 type file is used on the middle 1/3 portion of
the root. Thus, S1 type file model was used to shape the middle 1/3 portion of the root in our simulation.
The first contact point of the file model to the root canal wall is displayed, and the stress values are shown in the
Figure 2.
4.2 Shaping of root canals apical 1/3 portion by ProTaper Rotary File System:
Shaping process of the apical 1/3 part of the root canal was generated by using S1 and S2 files along the whole
canal.
When S1 file were used along the canal to shape the apical 1/3 part of it, high stress values occurred about 350
and 500 Mpa. After the new root canal geometry was shaped by S2 file, almost same stress values were found. It
is determined that stress concentrations were intensive on the files cutting corners which had contacts with the
dentin.
After the shaping process of the root canal was done, finishing process simulation was created. For this process,
F1 type file model which is the one of the finishing files was used along the root canal. As a result, stress values
which concentrated on the apical and coronel areas, were found between 350 and 500 MPa.
861
Figure 3: The stress distribution of the 04 angular, number 20 HeroShaper file model.
By using the #25.04 type file, a new shaping simulation was created. The stress results of this step were
determined between 350 and 500 Mpa. Stress concentrations were noticed on the middle 1/3 portion of the root,
and on the skewed region of the canal.
The last file model of HeroShaper , the #30.04 file used for the last shaping process. As a result of this process,
stress concentrations occurred on the apical 1/3 area of the root canal, and on the canal walls.
5 Conclusion
In this study, the model of the Ni-Ti files are penetrating the canal, enlarging it and reaching the apical while
rotating 250 rpm velocity by using three dimensional finite element methods was successfully performed.
Each model of the ProTaper and HeroShaper rotating file systems are operated in the canal model. Thus,
compatible stress values with respect to clinical conditions were gained.
As a result of our study, there are more stress concentrations on Hero Shaper file model with respect to ProTaper
file model. The files of HeroShapers sectional areas and positive cutting angles causes more stress creation. It
was observed that there were stress concentration towards the outer aspect of the curvature in the apical portion
of the canal with ProTaper file system but contrarily with HeroShaper file system, less stress concentrations
appeared in the middle portion of the tooth model.
HeroShaper file models causes less stress concentration on the canal walls with respect to ProTaper files.
HeroShaper files have high cutting efficiency, thus dentin resistance becomes lower. Because of positive rake
angle of HeroShaper files, cutting processes are more efficient and they generated less stress on the canal walls
than ProTaper files. All Stress values of the files were found in the transformation phase of stress-strain diagram.
With the ProTaper file models, stress composition increased towards the slope especially in apical areas. This
shows us that ProTaper files are more feasible for apical transportations.
862
6 References
[1] Schilder H. (1974). Cleaning and shaping the root canal system. Dent Clin North Am,18(2):269.
[2] Walton RE. (1992). Current concepts of canal preparation. Dent Clin North Am, 36:309-325.
[3] Weine FS. (1996). Endodontic Therapy. 5 th ed, St Louis, Missouri, Mosby, p.305-394.
[4] Himel VT, McSpadden JT, Goodis HE. (2006). Instruments Materials and Devices.In: Pathways of The
Pulp. Cohen S, Hargreaves KM. (Eds), 9th ed, Mosby, p.233-289.
[5] Shaw AM, Sameshima GT, Vu HV. (2004). Mechanical stress generated by orthodontic forces on apical root
cementum: a finite element model. Orthod Craniofacial Res, 7:98-107.
[6]Cheng R, Zhou XD, Liu Z, Yang H, Gao QH, Hu T. (2009). Finite element analysis of the effects of three
preparation techniques on stress within roots having curved canals. Int Endod J, 42:220-226.
[7] Wang GZ. (2007). A finite element analysis of evolution of stress-strain and martensite transformation in
front of a notch in shape memory alloy NiTi. Materials Science and Engineering A, 460-461:383391.
863
ugur.turkan@ediz.edu.tr, 2 mustafaguden@iyte.edu.tr
Abstract.The effect of alkali surface treatment on CaP deposition defined as bioactivity of an open-cell Ti6Al4V foam (60%
porous and 300500 m in pore size), prepared by means of the space holder method using 94 (P1) and 66 (2) m average
particle size powders, was investigated using the response surface methodology in a simulated body fluid (SBF) solution up
to 14 days. The response surface methods used in this study gives the most significant parameters on the CaP deposition.
Based on the analysis of variance (ANOVA), the most significant factor for alkali treated P1 (p<0.1261) and P2 (p<0.065) is
the temperature. The samples after the bioactivity test was investigated with scanning electron microscopy and found to be
nearly fully coated at 1M NaOH solution treatment at 20C for 12.5 h. This finding was approved to an optimum surface
coating parameters for the alkali treatment of Ti6Al4V foam samples after 14-days of SBF immersion.
1 Introduction
Titanium and its alloys are widely used in biomedical applications because of their good biocompatibility
combined with good mechanical properties [1]. The native oxide layer, which provides biocompatibility, forms
on the surface of Ti spontaneously [2]. The clinical success of the implants is not only associated with the
biocompatibility but also osseointegration, which is described as the direct structural and functional connection
between living bone tissue and the surface of an artificial implant. This biological fixation is considered as a
prerequisite for a successful implantation. Bonelike apatite coating is usually applied to the surface of Ti in order
to overcome the fixation problems [3]. The applied chemical surface treatment methods such as alkali treatment,
nitric acid treatment and acid etching further improve bonelike apatite coating on the implant surface; hence let
the formation of a bioactive surface [4-9]. It is noted that the most widely used chemical surface treatment
technique in order to improve the bioactivity of Ti metal is the alkali treatment. The alkali surface treatment is
usually performed in a 10M NaOH solution at 60C for 24 h. Some researchers applied a subsequent heat
treatment after alkali treatment to provide further improvement in bonelike apatite coating [10, 11]. In addition,
Ti and Ti6Al4V implants are generally subjected to a passivation treatment to minimize the corrosion in
biological environment. The ASTM-F86 standard of the surface passivation refers to a 20-40% nitric acid
containing solution treatment. Although, a variety of studies were conducted on the nitric acid passivation of Ti
surfaces; few of these have been carried out on the biomimetic CaP deposition ability of the nitric acid surface
treatment. It is also noted that there have been few studies on the optimization of surface treatment with respect
864
to the formation of CaP coating on the surface of Ti metal. Wei et al. [12] conducted an optimization study on
AT Ti6Al4V alloy. The optimization parameter selected included the molarity of solution, surface treatment
temperature and the heat treatment temperature. An optimum heat treatment temperature of 600C was proposed
based on the bonding strength of the apatite to the surface. In this study, the optimum conditions of alkali surface
treatment for the biomimetic CaP coating (bioactivity) of open cell Ti6Al4V foams were investigated using the
response surface methodology (RSM).
2.2 Methodology
RSM is a useful technique for modeling the effect of variables on the response. The objective of this
methodology is to optimize and identify the most significant variables. The most widely used Box Bhenken
design cube technique in RMS is shown in Figure 1. In this technique, there are three variables (A, B and C) and
the responses are to be determined experimentally shown on the cube as the points. The points are selected at the
mid points of the maximum and minimum values of variables and also at the center of the cube. The numbers at
each data point represent the number of experiments to be performed.
The selected three design variables and their centers and variations are tabulated in Tables 1 for alkali treatment,
respectively. The selected variables are as follows X1: molarity (M), X2: temperature of the solution (T), and
X3: time (t), duration of the surface treatment. The Box Bhenken design was implemented using Design Export
Software. The range of molarity, temperature and time of alkali treatment were selected between 1-10M, 20-60
oC, and 1-24 h, respectively.
865
Table 1.Factors and levels of the experimental design for alkali treatment.
Factor
Centre
Variation
( X 1)
Molarity
5.5
4.5
( X 2)
Temperature
40
20
( X 3)
Time
h.
12.5
11.5
For each response, 15 experiments were performed, 12 of them were at the corners and 3 of them at the
center of the cube. These data points were used to estimate the coefficients of a reduced second order model,
which provided a link between the response (CaP deposition mass) and the variables. The proposed response
function can be represented as,
(1)
866
Y=-9.00552+0.710981X1+0.377461X2+0.319534X3+0.002222X1X2+0.004348X1X30.00489X2X3-0.06831X12-0.00358X22-0.00687X32
(2)
and
Y=0.41918+0.069847X1-0.02846X2-0.09299X30.003333X1X2+0.001449X1X3+0.003043X2X3+0.001852X12+0.00053X22+0.000662X32
based on the ANOVA, the most significant factor for AT P1 (p<0.1261) and P2 (p<0.065) is
the temperature.
Table 2. Factors and response of AT P1 and P2 foam.
Run Order
Molarity
(M)
Temperature (C)
Time
m (mg),
(h)
P1
m (mg), P2
10
20
12.5
5.5
20
5.5
40
12.5
3.5
5.5
40
12.5
4.1
0.2
5.5
40
12.5
2.8
40
24
0.7
0.9
20
12.5
0.2
5.5
20
24
10
60
12.5
1.5
0.7
10
60
12.5
0.9
1.9
11
5.5
60
4.5
12
10
40
24
2.8
1.2
13
5.5
60
24
2.8
14
10
40
1.2
15
40
867
Analyses of variance (ANOVA) of AT P1 and P2 foam are further tabulated in Table 3. The R2 values of the
quadratic models of AT P1 and P2 foam equal to 0.77 and 0.91, respectively. About 70% of the variability in the
data of P2 foam and 90% of the variability in the data of P1 foam response are represented by the models.
R-squared=0.77 (0.66)*
R-squared=0.91 (0.87)*
P2 Foam
P1 Foam
Sum of Squares
df
Mean Square
F Value
p>F
X1-Molarity
1.71125
1.71125
1.026031
0.3576
X2-Temperature
5.61125
5.61125
3.364395
0.1261
X3-Time
0.605
0.605
0.362746
0.5733
X1X2
0.16
0.16
0.095933
0.7693
X1X3
0.2025
0.2025
0.121415
0.7417
X2X3
5.0625
5.0625
3.035375
0.1419
X12
7.065641
7.065641
4.236419
0.0946
7.585641
7.585641
4.548201
0.0861
X32
3.046410
3.046410
1.826567
0.2345
5.899606
0.1484
X2
Residual
8.339167
1.667833
Lack of Fit
7.4925
2.4975
Pure Error
0.846667
0.423333
Cor Total
37.164
14
Effects
Sum of Squares
df
Mean Square
F Value
p>F
X1-Molarity
0.10125
0.10125
0.557851
0.4887
X2-Temperature
3.645
3.645
20.08264
0.0065
X3-Time
3.00125
3.00125
16.53581
0.0097
X1X2
0.36
0.36
1.983471
0.2181
X1X3
0.0225
0.0225
0.123967
0.7391
X2X3
1.96
1.96
10.7989
0.0218
X12
0.005192
0.005192
0.028608
0.8723
X22
0.166731
0.166731
0.918627
0.3819
X32
0.028269
0.028269
0.155753
0.7094
0.41369
0.7630
Residual
0.9075
0.1815
Lack of Fit
0.3475
0.115833
Pure Error
0.56
0.28
Cor Total
10.184
14
868
Figure 2. The response surface 3D plots of molarity-temperature of AT (a) P1 and (b) P2 foam.
Figure 3. The response surface 3D plots of molarity-time of AT (a) P1 and (b) P2 foam.
869
Figure 4. The response surface 3D plots of temperature-time of AT (a) P1 and (b) P2 foam.
The CaP growth mechanism on alkali and heat treated Ti metal in SBF was previously explained [1517]. The alkali treatment of Ti metal is usually performed in a 10M or 5M NaOH solution at various
temperatures. Liang et al. [7] investigated the apatite formation on AT porous Ti. The alkali treatment was
carried in a solution of 0.5 to 10M, at 60C for 24 h. It was shown that AT samples in 0.5 and 1M solution did
not induce CaP formation when the samples were soaked in SBF for 28 days. However, the samples treated with
5 and 10M solution formed continues layer of CaP on the surface. In another study, Ti6Al4V specimens were
soaked in 3, 5, 10 and 15M NaOH solution at 60 and 80C for 1, 3 and 7 days [12]. The treated samples were
heat treated at 500, 600 and 700C for 1 h. It was shown that 15M NaOH treatment at 60C for 3 days without
any heat treatment resulted in the formation of a dense sodium titanate layer on the surface. The apatite
formation on the surface of all heat treated specimens were observed after soaking in SBF for only 1 day except
the specimens heat treated at 700C. It was reported that the optimum treatment conditions for Ti6Al4V
specimens were 5M NaOH solution treatment at 80C for 3 days and subsequent heat treatment at 600C for 1 h.
The surfaces of the AT P1 and P2 foam specimens were microscopically analyzed after bioactivity test.
The following experimental conditions were observed to induce a uniform coating layer on the surface and
interior part of pores of both AT P1 and P2 foam specimens: 1M NaOH solution treatment at 20C for 12.5 h.
These findings contradict with the result of Liang et al. [7], in which no CaP precipitation on Ti was observed in
1M solution at 60 oC for 1 h. The difference between two results simply arises from the duration of the
treatment. The alkali treatment in 5.5M solution at 40C for 12.5 h also resulted in CaP coating on the surface
and interior of pores of P2 foam, while only CaP precipitates were seen microscopically in cells of AT P1 foam.
The ANOVA table of AT P2 foam quite agrees with the SEM observations of CaP layer. The treatment
temperature and time are more significant than the molarity. Figures 5(a-d) show the SEM micrograph of AT 14day SBF immersed P1 and P2 foam samples at optimum treatment conditions. The observations clearly show
that both foam flat surfaces and interior of cell are uniformly coated with CaP layer.
(a)
(b)
870
(c)
(d)
Figure 5. SEM micrograph after the response surface design of optimum alkali treatment condition
(a) flat surface of P1 foam, (b) interior of the P1 foam, (c) flat surface of P2 foam and (d) interior of the P2
foam.
References
1.
2.
3.
4.
5.
6.
7.
8.
9.
10.
11.
12.
13.
14.
15.
16.
17.
Katti, K.S., (2004).Biomaterials in total joint replacement. Colloids and Surfaces B-Biointerfaces. 39(3). 133-142.
Callen, B.W., et al., (1995).Nitric-acid passivation of TI6AL4V reduces thickness of surface oxide layer and
increases trace element release. Journal of Biomedical Materials Research. 29(3). 279-290.
Narayanan, R., et al., (2008).Calcium phosphate-based coatings on titanium and its alloys. Journal of Biomedical
Materials Research Part B-Applied Biomaterials. 85B(1). 279-299.
Faure, J., et al., (2009).Morphological and chemical characterisation of biomimetic bone like apatite formation on
alkali treated Ti6Al4V titanium alloy. Materials Science & Engineering C-Biomimetic and Supramolecular
Systems. 29(4). 1252-1257.
Jonasova, L., et al., (2004).Biomimetic apatite formation on chemically treated titanium. Biomaterials. 25(7-8).
1187-1194.
Kim, H.M., et al., (1996).Preparation of bioactive Ti and its alloys via simple chemical surface treatment. Journal
of Biomedical Materials Research. 32(3). 409-417.
Liang, F.H., L. Zhou, and K.G. Wang, (2003).Apatite formation on porous titanium by alkali and heat-treatment.
Surface & Coatings Technology. 165(2). 133-139.
Lu, X., Z.F. Zhao, and Y. Leng, (2007).Biomimetic calcium phosphate coatings on nitric-acid-treated titanium
surfaces. Materials Science & Engineering C-Biomimetic and Supramolecular Systems. 27(4). 700-708.
Vanzillotta, P.S., et al., (2006).Improvement of in vitro titanium bioactivity by three different surface treatments.
Dental Materials. 22(3). 275-282.
Kim, H.M., et al., (1997).Effect of heat treatment on apatite-forming ability of Ti metal induced by alkali treatment.
Journal of Materials Science-Materials in Medicine. 8(6). 341-347.
Lee, B.H., et al., (2002).Surface modification by alkali and heat treatments in titanium alloys. Journal of
Biomedical Materials Research. 61(3). 466-473.
Wei, M., et al., (2002).Optimising the bioactivity of alkaline-treated titanium alloy. Materials Science &
Engineering C-Biomimetic and Supramolecular Systems. 20(1-2). 125-134.
Bayraktar, D. and A.C. Tas, (1999).Chemical preparation of carbonated calcium hydroxyapatite powders at 37
degrees C in urea-containing synthetic body fluids. Journal of the European Ceramic Society. 19(13-14). 25732579.
Jalota, S., S.B. Bhaduri, and A.C. Tas, (2007).Osteoblast proliferation on neat and apatite-like calcium phosphatecoated titanium foam scaffolds. Materials Science & Engineering C-Biomimetic and Supramolecular Systems.
27(3). 432-440.
Takadama, H., et al., (2001).TEM-EDX study of mechanism of bonelike apatite formation on bioactive titanium
metal in simulated body fluid. Journal of Biomedical Materials Research. 57(3). 441-448.
Takadama, H., et al., (2001).An X-ray photoelectron spectroscopy study of the process of apatite formation on
bioactive titanium metal. Journal of Biomedical Materials Research. 55(2). 185-193.
Kim, H.M., et al., (2003).Surface potential change in bioactive titanium metal during the process of apatite
formation in simulated body fluid. Journal of Biomedical Materials Research Part A. 67A(4). 1305-1309.
871
Department of Mechanical Engineering, Celal Bayar University, 45140, Muradiye, Manisa, Turkey
3
Department of Mathematics, King Saud University, P.O. Box 2455, Riyadh 11451, Saudi Arabia
Abstract-Boundary layer equations are derived for the Sisko fluid. Using Lie Group theory, symmetry analysis
of the equations is performed. Partial differential system is transferred to an ordinary differential system via
symmetries. Resulting equations are numerically solved. Effects of non-Newtonian parameters on the solutions are
discussed.
Keywords: Non-Newtonian Fluid, Sisko Fluid, Boundary Layer Theory, Lie Group Analysis
1 INTRODUCTION
Due to the inadequecy of Newtonian fluid model which predicts a linear relationship between the shear stress
and velocity gradient, many non-Newtonian fluid models were proposed to explain the complex behavior.
Usually, the stress constitutive relations of such models inherit complexities which lead to highly nonlinear
equations of motion with many terms. To simplify the extremely complex equations, one alternative is to use
boundary layer theory which is known to effectively reduce the complexity of Navier-Stokes equations and
reduce drastically the computational time. Since there are many non-Newtonian models and new models are
being proposed continuously, boundary layer theory for each proposed model also appear in the literature. It is
beyond the scope of this work to review vast literature on the boundary layers of non-Newtonian fluids. A
limited work on the topic can be referred as examples [1-10].
Some of the recent work on Sisko fluids is as follows: Khan et al. [11] presented an analytical solution using
homotopy analysis method for the flow of a magnetohydrodynamic Sisko fluid through a porous medium. Sajid
and Hayat [12] investigated wire coating analysis by withdrawal from a bath of Sisko fluid. Wang et al. [13]
studied numerically peristaltic flow characteristics of a Sisko fluid in a symmetric or asymmetric channel.
Akyldz et al. [14] qualitatively analyzed the equation modeling thin film flow of a Sisko fluid on a moving belt
and found a series solution using homotopy analysis method. Abelman et al. [15] numerically investigated the
unsteady rotating flow of a Sisko fluid bounded by a suddenly moved infinite flat plate. Molati et al. [16] studied
unsteady unidirectional flow of a Sisko fluid bounded by a suddenly moved plate. Hayat et al. [17] considered
unsteady flow of a Sisko fluid over a moving wall with suction or blowing. Khan et al. [18] presented a
numerical study for the unsteady flow of a magnetohydrodynamic Sisko fluid in annular pipe. Heat transfer was
also incorporated into the annular pipe problem [19]. To the best of authors knowledge, boundary layer analysis
and symmetry reductions of Sisko fluids do not exist in the literature.
872
In this work, boundary layer theory is developed for the Sisko fluid, a non-Newtonian fluid model which
combines the features of viscous and power law models. A complete symmetry analysis of the boundary layer
equations is presented. Using one of the symmetries, the partial differential system is transformed into an
ordinary differential system. Resulting ordinary differential system is numerically solved by a finite difference
algorithm. Effect of non-Newtonian parameters on the velocity profiles are shown in the graphs.
T = pI + S
(1)
where
n 1
1
2
S = a + b
tr ( A 1 ) A 1
2
A 1 = L + LT ,
(2)
L = grad V
(3)
V is the velocity vector and A1 is the first Rivlin Ericksen tensor. a and b are the material constants. The model is
a combination of viscous and power-law models. For a=0, the model exhibits power-law behavior whereas for
b=0, the flow is Newtonian. Steady-state, two dimensional, incompressible equations of motion including mass
conservation can be written as
u * v *
+
=0
x * y *
u *
(4)
*
p * S xx S xy
u *
* u
v
=
+
+ *
+
y
x * x *
x *
y *
u *
(5)
*
p * S xy S yy
v *
* v
v
+
+ *
+
y * x *
y
x *
y *
(6)
where x* is the spatial coordinate along the surface, y* is vertical to it, u* and v* are the velocity components in
the x* and y* coordinates. The shear stress components are inserted into the equations of motion and the usual
boundary layer assumptions are made i.e. x*~O(1), y*~O(), u*~O(1), v*~O(). The highest order terms are
retained and the momentum equations become
u *
u *
p *
2u *
u *
*
bn
u * * + v * * = * + a
+
2
x
y
y *
y
x
0=
n 1
2u *
y *
p *
y *
(7)
(8)
from which dependence of pressure on y* is eliminated. Dimensionless variables and parameters are defined as
follows
873
p*
bnV n2
y*
u*
v*
x*
a
x= ,y=
=
,u =
,v = , p =
, =
L
L
V
V
V 2 1 VL 2
Ln
(9)
whereL is a characteristic length and V is a reference velocity. Expressing the pressure in terms of outer velocity,
the final dimensionless boundary layer equations become
u v
+
=0
x y
(10)
u
dU
u
u
2u
u
+v
=U
+ 1 2 + 2
dx
x
y
y
y
n 1
2u
y 2
(11)
v(x,0)=0,
u(x,)=U(x)
(12)
For 2=0 or n=1, the equations reduce qualitatively to those of Newtonian fluid.
X = 1 ( x , y , u , v )
+ 2 ( x, y , u , v)
+ 1 ( x, y , u , v )
+ 2 ( x, y , u , v )
x
y
u
v (13)
A straightforward and tedious calculation (See [20, 21] for details) yields
1 = 3ax + b ,
2 = ay + c( x) , 1 = au , 2 = c( x)u av
(14)
(3ax + b)
d
(UU ) + a (UU ) = 0
dx
(15)
Symbolic packages developed to calculate symmetries fail to produce the above results due to the arbitrary outer
velocity function and hand calculation becomes inevitable for precise results. There are two finite parameter Lie
Group symmetries represented by parameters a and b, the former corresponding to scaling symmetry and the
latter translational symmetry in the x coordinate. There is an additional infinite parameter Lie Group symmetry
represented by c(x) function.
Usually boundary conditions put much restriction on the symmetries which may lead to removal of all
the symmetries. In our case however, some of the symmetries remain stable after imposing the boundary
conditions. For nonlinear equations, the generators should be applied to the boundaries and boundary conditions
also [20]. Applying the generator to the boundary y = 0 yields c(x)=0. Applying the generator to the boundary
conditions do not impose further restrictions and hence the symmetries valid for the equations and boundary
conditions reduce to
1 = 3ax + b ,
2 = ay , 1 = au , 2 = av
874
(16)
Selecting parameter a in the symmetries, the associated equations which define similarity variables
are
dx dy du dv
=
=
=
3x y
u v
(17)
y
x
1/ 3
v=
u = x1/ 3 f ( ) ,
g ( )
x1 / 3
(18)
From (15) with b=0, U(x)=x1/3. Substituting all into the boundary layer equations yields the ordinary differential
system
f f + 3 g = 0
(19)
f 2 f f + 3 gf = 1 + 3 1 f + 3 2 ( f ) n1 f
(20)
f (0 ) = 0 ,
g ( 0) = 0 ,
f ( ) = 1
(21)
4. NUMERICAL RESULTS
Equations (19) and (20) are numerically integrated using a finite difference scheme subject to the
boundary conditions (21). In Figure 1, f function and in Figure 2, g function related to the x and y components of
the velocities are drawn for different 1 parameters. Boundary layer becomes thicker for an increase in 1. There
is an increase in y component of velocity which is negative over the whole domain excluding the surface for an
increase in 1 as can be seen from Figure 2. A similar trend is observed for parameter 2, although not as
influential as in the previous case (See Figures 3 and 4).
Figure 1- Effect of 1 parameter on the similarity function f related to the x component of velocity (2=1, n=2).
875
Figure 2- Effect of 1 parameter on the similarity function g related to the y component of velocity (2=1, n=2).
Figure 3- Effect of 2 parameter on the similarity function f related to the x component of velocity (1=1, n=2).
Figure 4- Effect of 2 parameter on the similarity function g related to the y component of velocity (1=1, n=2).
876
5. CONCLUDING REMARKS
Boundary layer equations of Sisko fluid is derived for the first time. Lie group theory is applied to the
equations. Equations admit two finite parameter Lie Group transformations and an infinite parameter Lie Group
transformation. The infinite parameter Lie Group transformation is not stable with respect to usual boundary
layer conditions. Using the scaling symmetry which is one of the finite parameter transformations, the partial
differential system is transferred into an ordinary differential system. Resulting ordinary differential system is
solved numerically using a finite difference scheme. Effects of non-Newtonian parameters on the boundary
layers are discussed in detail. It is found that an increase in the parameters 1 and 2 causes the boundary layer to
thicken.
Acknowledgements-This work is completed during mutual short visits of Tasawar Hayat to Turkey and
Mehmet Pakdemirli to Pakistan. Funding supports of TUBITAK of Turkey and HEC of Pakistan are highly
appreciated.
6. REFERENCES
[1] Acrivos A., Shahand M. J. and Petersen E. E., (1960). Momentum and heat transfer in laminar boundary layer flows of
non-Newtonian fluids past external surfaces, AIChE Journal, 6 312-317.
[2] Schowalter W. R., (1960). The application of boundary-layer theory to power-law pseudoplastic fluids: Similarity
solutions, AIChE Journal 6 25-28.
[3] Bizzell G. D., Slattery J. C., (1962). Non-Newtonian boundary layer flow, Chemical Engineering Science 17 777-781.
[4] Srivastava A. C., (1958). The flow of a non-Newtonian fluid near a stagnation point, ZAMP, 9 80-84.
[5] Beard D. W. and Walters K., (1964). Elastico-viscous boundary layer flows, Proceedings of Cambridge Philosophical
Society 60 667-674.
[6] Astin J., Jones R. S. and Lockyer P., (1973). Boundary layers in non-Newtonian fluids, Journal Mechanique 12 527-539.
[7] Rajagopal K. R., Gupta A. S. and Wineman A. S., (1980). On a boundary layer theory for non-Newtonian fluids,
International Journal of Engineering Science 18 875-883.
[8] Rajagopal K. R., Gupta A. S. and Na T. Y., (1983). A note on the Falkner Skan flow of a non-Newtonian fluid,
International Journal of Non-Linear Mechanics 18 313-319.
[9] V. K. Garg and K. R. Rajagopal, Flow of a non-Newtonian fluid past a wedge, Acta Mechanica, 88 (1991) 113-123.
[10] Massoudi M. and Ramezan M., (1989). Effect of injection or suction on the Falkner Skan flows of second grade fluids,
International Journal of Non-Linear Mechanics 24 221-227.
[11] Khan M., Abbas Z. and Hayat T., (2008). Analytical solution for flow of Sisko fluid through a porous medium,
Transport in Porous Media 71 23-37.
[12] Sajid M. and Hayat T., (2008). Wire coating analysis by withdrawal from a bath of Sisko fluid, Applied Mathematics
and Computation 199 13-22.
[13] Wang Y., Hayat T., Ali N. and Oberlack M., (2008). Magnetohydrodynamic peristaltic motion of a Sisko fluid in a
symmetric or asymmetric channel, Physica A 387 347-362.
[14] Akyldz F. T., Vajravelu K., Mohapatra R. N., Sweet E. and Van Gorder R. A., (2009). Implicit differential equation
arising in the steady flow of a Sisko fluid, Applied Mathematics and Computation 210 189-196.
877
878
Dokuz Eyll University, Izmir Vocational School, Buca 35150 Izmir, Turkey
Celal Bayar University, Dept. of Mechanical Engineering, Muradiye 45140 Manisa, Turkey
3
Mula University, Department of Mathematics, Ktekli 48000 Mula, Turkey
1
nurcan.kurt@deu.edu.tr, 2 mehmet.cevik@bayar.edu.tr, 3 m.sezer@mugla.edu.tr
Abstract.The free and forced vibration of the single degree of freedom system is a fundamental topic in
mechanical vibrations. A new practical collocation method for the numerical solution of this system is
proposed based on the Chebyshev polynomials. In particular the first kind Chebyshev polynomials based on
powers of the independent variable are applied as the basis polynomials. Particular and general solutions of
the differential equation are determined by this method. The method is illustrated by a numerical application
and the results obtained are compared with those of the exact solution.
Keywords: Chebyshev polynomials, vibration, collocation method
1 Introduction
The study of the free and forced vibrations of damped spring-mass systems having single degree of freedom is
fundamental to the understanding of advanced subjects in mechanical vibrations. In many cases, a complicated
system can be idealized as a single degree of freedom spring-mass system. Therefore, solving the equations of
motion of this system would serve for many other more advanced problems. There are various useful methods
for calculating solutions of a spring-mass-damper system excited by a harmonic force. Method of undetermined
coefficients, frequency response method, geometric method and Laplace transform method are given in
textbooks [1,2]. Recently, the same problem is solved by Taylor polynomial matrix method [3].
The present study introduces a practical collocation method based on the Chebyshev polynomials. Chebyshev
polynomials are a well known family of orthogonal polynomials on the interval [1,1] of the real line. These
polynomials present, among others, very good properties in the approximation of functions. Therefore,
Chebyshev polynomials appear frequently in several fields of the mathematics, physics and engineering. Some
examples are thermoconvective problems [4], high-order ODEs [5], numerical modeling of micro- and macrobehavior of viscoelastic porous materials [6], interaction problem in magnetoelectroelastic solids [7] and
vibration analysis of functionally graded material sandwich plates [8]. Chebyshev spectral method was employed
for the solution of nonlinear partial differential equations [9] and dimensional reduction of nonlinear delay
differential equations [10]. The Chebyshev collocation method used in the present study is a relatively practical
one since the solution is performed on a matrix basis. This method is utilized for the solution of linear
differential equations [11], high order linear differential equations with variable coefficients [12], and linear
integro-differential equations [13]. Our purpose, in this study, is to develop the Chebyshev matrix and
collocation methods given for solving the differential equation of the single degree of freedom vibration system.
879
(1)
k =0
(2)
k =0
where aik, bik, cik, i are suitable constants and, Pk(x) and g(x) are analytical functions defined in the range [1,1].
The approximate solution of (1) is expressed as the truncated Chebyshev series
y(x) =
' a n Tn (x)
n =0
, 1 x 1
(3)
'
where denotes a sum whose first term is halved. Tn(x) denotes the Chebyshev polynomial of the first kind of
degree n [14,15], defined by
1 x 1
y(k) (x) =
n =0
(4)
where
1
T(k) (x) = T0(k) (x) T1(k) (x) T2(k) (x) K TN(k) (x) A = a 0
a1 K a N
(0)
(0)
(0)
Note that y (x) y(x) , Tn (x) Tn (x) and T (x) T(x) .
The powers of x can be expressed in terms of the polynomials Tn(x) as follows [14,15]
n 2n
n 2n + 1
x 2n = 22n +1 '
x 2n +1 = 22n '
T2 j (x)
T2 j+1 (x)
j= 0 n j
j= 0 n j
By using relations (5) and taking N the truncated limit, one obtains the matrix equation
X(x) = T(x) DT
(6)
such that
1
0
0
0
0
0
0 L
0
1
0
0
0
0
0 L
1 2
0
12
0
0
0
0 L
0
34
0
14
0
0
0 L
D =
38
0
12
0
18
0
0 L
58
0
5 16
0 1 16
0 L
0
5 16 0 15 32
0
3 16
0 1 32 L
M
M
M
M
M
M
M
O
X(x) = x 0
(5)
x1
In order to find the relation between X(x) and its derivatives X(k)(x), one should write
880
(7)
where
( ) (x ) (x )
(1)
X (x) = x 0
(1)
1 (1)
2 (1)
(x )
N
0
1
(1)
0
B
=
M
0
0 L
0 L
2 L
M O
0 L
0
0 0
0 0
M M
N 0
0
and similarly
( )
, k = 0,1,2,K , m N
(8)
can be derived. On the other hand, from the matrix relations (6) and (8), we have
( ) ( DT )
-1
, k = 0,1,2,K , m N
th
Consequently, substituting the expression (9) into (4), the solution function (3) and its k
converted into the matrix form
( ) ( DT )
k
-1
(9)
derivative are
, k = 0,1,2,K , m N
(10)
x i = cos
Ni
,
N
i = 0,1, 2,..., N
(11)
i = 0,1, 2,..., N
k =0
Pk Y(k) = G
(12)
k =0
where
y k (x 0 )
g(x 0 )
g(x )
k
1
(k) y (x1 )
=
G
Y
=
,
M
O
M ,
M
L Pk (x N )
g(x N )
y (x N )
Substituting the collocation points xi in the relation (10) yields the system of matrix equations, briefly
0
Pk (x 0 )
0
Pk (x1 )
Pk =
M
M
0
0
L
L
( ) ( DT )
Y(k) = X BT
0
0
(13)
where
X(x 0 ) 1 x 0
X(x )
1 1 x1
=
X=
M M M
X(x N ) 1 x N
K x 0N
x12 K x1N
M O M
N
x 2N K x N
Consequently, from (12) and (13), the fundamental matrix equation for Eq.(1) is obtained as
x 02
881
Pk X ( BT )
k =0
( DT )
A=G
(14)
The corresponding matrix equation for the mixed conditions (2) can be determined, using the relation (10), as
follows:
m 1
k =0
( DT )
A = [i ] ,
i = 0,1,..., m 1 (15)
4 Method of Solution
The fundamental matrix equation (14) for Eq.(1) corresponds to a system of (N+1) algebraic equations for the
(N+1) unknown coefficients a0, a1,, aN. Briefly, Eq.(14) can be written as
WA = G or [ W; G ]
(16)
where W = w pq =
Pk X ( BT )
k =0
( DT )
p, q = 0,1,K , N
A = ( W ) 1 G
By means of Eq.(15), matrix form for the mixed conditions (2) can be written briefly as
Ui A = [ i ]
or
[ Ui i ] ;
i = 0,1,..., m 1
(17)
where
Ui =
m 1
k =0
( DT )
u i 0
u i1 K u i N
In order to obtain the general solution of Eq.(1) under the conditions (2), the last m rows of the matrix (16)
are replaced by the rows matrix (17) and the required augmented matrix is constructed
w 00
w
10
N m,0
W; G = u 00
u10
M
u m 1,0
w 01
w11
M
w N m,1
u 01
u11
M
u m 1,1
K
K
w 0N
w1N
M
K w N m,N
u 0N
K
u1N
K
M
L u m 1,N
;
g(x 0 )
;
g(x1 )
M
M
; g(x N m )
0
;
1
;
M
M
m 1
;
(18)
1
Using (18) one obtains A = ( W) G and the Chebyshev coefficients an , n = 0,1,, N of the general solution
are determined.
Finally, the homogeneous solution is determined by taking the difference of the general and particular
solutions determined above, because the general solution is the sum of the particular solution plus the
homogeneous solution in a linear system.
5 Illustrative Example
In this section, a numerical example is given to illustrate the properties of the method. A spring-mass-damper
system with a mass of M=60kg, damping coefficient of C=20kg/s and spring stiffness of K=2000N/m subject to
an excitation force of amplitude F0=20N and frequency ~=3rad/s is considered with the initial conditions y(0) =
&
= 0 . The matrix operations are performed using the software package MATLAB 6.5.1. [16].
0.02m and y(0)
Let us now consider the second order differential equation of the single degree of freedom spring-massdamper system
882
&& + Cy(t)
& + Ky(t) = F0 cos t
My(t)
where N=6 and P0 (t)=K=2000, P1 (t)=C=20, P2 (t)=M=60, g(t) = F0 cost = 20cos3t
General Solution. The equation can be expressed in matrix form
2
Pk X ( BT )
k =0
( DT )
A=G
2000
980
800
1280
440
100
880 ; 1. 4147
20
1760
60
1040
100
160 ;
20
W;G = 2000
2000
1020
720
1280
600
300
880 ; 1.4147
0
0
1
0
0.02
1
1 ;
1
0
1
0
0
5
0 ;
0
3
% =W
% and the result is determined as
% -1G
This system has the general solution A
0,015
homogeneous
0,005
-0,005
-0,015
0
0,2
0,4
0,6
0,8
Fig. 1. Time response of the system obtained by the Chebyshev collocation method for N=10.
The time response of the system obtained by the method of undetermined coefficients (exact solution) is
given as [1]
883
0,015
homogeneous
0,005
-0,005
-0,015
0
0,2
0,4
0,6
0,8
Evaluating Figures 1 and 2 together, it is observed that general and particular solutions obtained by both methods
coincide each other; accordingly, the homogeneous solutions obtained by both methods are also identical in the
interval 0 ~ t ~ 1.
0,04
0,02
N=6
N=7
N=8
N=10
Exact
-0,02
-0,04
0
0,2
0,4
0,6
0,8
In general, the truncation limit N should be increased to have a better approximation. In Figure 3, it is obviously
seen that the Chebyshev collocation solution approaches the exact solution as the truncation limit N is increased.
884
The method proposed in this paper can be applied also to a wide class of ordinary differential equations and
integral equations of the second kind. The method can be developed for the shifted Chebyshev polynomials
Tn*(x) of the first kind and other kinds of Chebyshev polynomials.
References
1.
2.
3.
4.
5.
6.
7.
8.
9.
10.
11.
12.
13.
14.
15. Mason, J.C, Handscomb, D.C. 2003. Chebyshev Polynomials. Boca Raton: CRC Press.
16. MATLAB 6.5.1. Users Manual, 2003. The MathWorks, Inc
885
Abstract.One of the most important factors that determine implant success is the achievement and
maintenance of implant stability. This numerical study was carried out using three dimensional finite
element method (FEM) and aimed to investigate static behaviors of dental implant under different load
directions. Solid models of implants were built up in Solid Works and then transferred to a mesh model in
FEM (ANSYS) to perform a stress analysis. The distribution of the equivalent stresses and total deformations
in the dental implant components (abutment, implant and bone) is presented.According to all these
investigations, as with natural tooth restorations, having plastic deformations on xy-axis posterior tooth
region and yz-axis anterior tooth region at implant supported restorations explain us that natural tooth show
parallel mechanical behaviors with implant supported restorations.
Keywords: Finite element analysis, Dental implant, Prosthesis design, Mandible, Stress distributions
1 Introduction
In prosthetic applications, understanding of the stress/strain which effects on implant-mandible and implantsoft tissue is one of the most important factors that have an impact on the achievement of treatment. Making
investigations on the mechanical interference between implant and mandible is really difficult because of its
complex geometry and dependence to many parameters (materials, age, type of bone and biocompatibility).
Understanding of biologic movements of jaw structure and abutments plays an important role on improving the
quality of implants which has used at restorations and developing more successful restoration applications. Due
to these reasons, it is necessary to have biomechanical models and make investigations in order to have reliable
studies on restoration treatments. Meanwhile, new methods were developed to investigate complex prosthetic
behaviors that can make biomechanical simulations. One of the most important methods which have used
commonly to analyze mechanical properties is Finite Element Methods (Finite Element Analysis, FEA). This
method provides really successful analyze results for the solution of biomechanical problems. With the use of
these FEA softwares, we can make new models, define deformations and analyze stress/strain that caused by
applied forces.
In most of studies, FEA is used as a successful method to make analyses on mechanical interferences
between implant-mandible and implant-soft tissue. When the appropriate conditions are used to take better
results, it gives opportunity to make investigations on most of the physical parameters such as material properties
that is used for implant production, stress/strain/load applied over the implant, occurred deformations and
stress/strain/load transfer between implant and mandible.[5], [3], [12], [9], [7], [6], [8], [10]. The system consist
886
of different materials, 2 - 3 dimensional complex geometries and various boundary conditions which is
impossible to do clinically can be made easily with the use of FEA.
Osseointegrated implants and load transfer mechanisms that are affected by implant shape, geometrical and
mechanical properties of placement were studied by FEA and they are analyzed the influence of implant
diameter and length on stress distribution and overload risk of clinically confirmed bone loss at the implant neck
in mandibular and maxillary regions [2], [9].Strain and stress distributions in endodontically restored with
different post, core and crown materials resulted in the homogeneous stress and strain distributions are analyzed
by FEA [1], [10]. The effects of different inter-implant distances were studied on stress distribution in the bone
around the endosseous titanium implants under vertical, oblique and horizontal loads in the posterior mandibular
edentulousim using FEA [11]. 3-D finite element analysis is used to investigate of the human mandible during
clenching [4].
Although there are many studies on implant and the interference between implant and mandible, the
investigations which contains 3D model of all mandible are inadequate. In order to have a successful implant
treatment, it is necessary to make correct treatment plan and have proper mandible for implementation of
implant. The models which have investigated with FEA help to apply correct treatment plan and give significant
information as bone structure, applied implant size/shape and selection of materials. The purpose of present
study is to investigate the forces, in a type-4 bone structured 8 implant fixed mandible restoration with the use of
FEA, that effects during functional movements of jaw.
And then, this 3D model is transferred to ANSYS software to analyze stress distributions and total
deformation which occur on human mandible, implant and prosthesis.
887
Fig. 2. The Finite Element Models of the mandible and prosthesis with applied forces from left, right and front parts and
loading conditions.
The numbers of nodes and elements for implant and mandible are listed in Table 1. Type-4 bone structured 8
implant fixed mandible restoration was meshed to materials and nodes which has similar physical properties
(Fig. 2).
Table 1. The numbers of nodes and elements for implants and mandible
Implant and Mandible
Elements
Nodes
266967
469908
The mandible is accepted as combined with bone, tissue, and 8 implants. In order to analyze stress
distributions on each element, tooth materials are accepted as isotropic, homogeneous, symmetric and elastic.
These materials were assumed as it transfers the applied force linear and homogeneous. The material properties
which were used during analyses are listed in table 3.
The model that shows applied forces and boundary conditions of 8 implant supported mandible is given in
Fig. 2. The surfaces which has red colour symbolize the areas where forces were applied.
3 Analysis Results
In this study the distribution of equivalent stress and deformations in dental implant prosthesis were
presented with applying loadings from different locations on 8 implants used mandible structures. The loadings
which were used as mastication forces, poisson ratios, elasticity modules, tensile strengths and other mechanical
888
properties were taken from literature values. On the other hand, 3 different types of implants were used in this
work for different mandible locations and properties. The implants were designed as standards and its material
properties are listed in Table 3. Finite Element Model of the implant and mandible model is shown at Figure 2.
Additionally, after meshing process the force was applied from right, left and frontal surfaces of the prosthesis.
The framework dimensions depended on the room needed for the artificial teeth. The dimensions of dental
implants which were used for analyses are listed in Table 2.
Table 2. The dimensions of dental implants
Lower
Diameter(m
m)
Type-1
(molars)
Type-2
(premolars
)
Type-3
(incisors)
Great
diameter(mm
)
Pitch(mm)
Total
length(mm)
3.2
0.5
15
2.6
15
17
Titanium
Cortical Bone
Spongy Bone
Modulus of Elasticity(E)
110,000
14,700
1,370
Type-4 bone structured 8 implant fixed mandible restoration was designed after these steps. In order to
analyze effects of mastication forces to human mandible and implant fixture, we applied the loadings which was
equal to 300 N. Thus, our 3D models consist of 266967 elements and 469908 nodes for 8 implants located
mandible.
889
Equivalent Stress
8-F
8-R
8-L
Total Deformations
Fig. 3. Total deformations and equivalent stresses that occurred by the effect of forces from left, right and
frontal surfaces.
Figure 3 shows the total deformations and equivalent stresses that are obtained with the load of 300 N
mastication forces from right, left and front sides of prosthesis. The normal stress values on x-y-z directions were
observed and listed in Table 3. According to these results, it can be seen that the left side of the structure has
more capacity to absorb stresses depending on its bone quality. The stress increases due to the reaction of contact
between the implant and the bone area. The load case according to z direction is the most dangerous. Indeed, the
maximal stress for this case is located at the upper zone of the contact region. One can also note that for the
loading according to the axes x and y, the lower zone of contact between the implant and bone is subjected to
very weak level of stresses.
The shear stresses that can be seen at Table 4 are observed from xy-yz and xz planes under the effects of 300
N mastication forces. The distribution values of shear stresses on the analysis that was made for 8 implants from
right, left and front locations are listed in Table 4. Indeed, the significant stress is located on the upper parts of
the implants. When we compare also from the point of shear stresses it is significant that the risky results were
obtained from the locations where we applied forces from front side.
The directional deformations that occurred on the prosthesis and mandible-implant fixtures can be seen from
Table 4. When we compare the results that illustrated at Table 4, it can be seen that the maximum deformations
were obtained on the right sides due to the quality of mandible.
The von Mises equivalent stresses that obtained from titanium parts and mandible is summarized in Table 4.
It appears that the molar configuration leads to the highest stress values for loading from front side and
890
framework. As expected, the minimum values of stresses were taken from the left side of 8 implant design.
According to the analysis results, it can be observed that the maximum total deformations were concentrated on
implant abutment connection regions at right sides of 8 implants fixture (Table 4). At the same time, the
minimum deformations were observed on the left parts of the models which had designed with 8 implants.
Table 4. The values of deformations and stresses
Normal
Stress(MPa)
x
8-L
y
z
x
8-R
y
z
x
8-F
y
z
43.325
55.078
80.552
54.709
64.412
108.72
99,51
134,56
186,32
Directional
Deformation
(mm)
0.015874
0.025978
0.0034505
0.013142
0.035823
0.0036544
0.016889
0.060201
0.0080334
Shear
Stress(MPa)
xy
yz
xz
xy
yz
xz
xy
yz
xz
23.514
52.327
36.11
31.955
92.726
47.42
47.973
137.95
82.931
Total
Deformation
(mm)
Equivalent
stress (MPa)
0.042209
114.95
0.054825
178.00
0.096013
279.88
4 Discussions
In this study, the stress results which was obtained from the functional movements of 8 implant supported
type-4 bone tissued mandible, were found different from each other. The dental materials were accepted as
isotropic, homogeneous, symmetric and elastic in order to make analyses easily. On the other hand, because of
these materials were assumed as having elastic properties under stress, these materials transfer the applied forces
linear and homogenous, the stress that occurred against applied force was accepted linear on the model. The
reason why the stresses were minimum at x-axis can be explained like; there were dominant stresses on the
mandible region where implants are located as an impact against applied forces. Because of the alongside
implants, that locates on y-direction, were supporting each other against the force that applied on dental implant;
the stress at x-direction is expected to be more than y direction. (Itoh et al., [7]) was reported in one of his study
that the stress which applied on 3 implants was concentrated on the middle implant. Comparatively to these
stresses, implant and abutment has caused to plastic deformations on their connection surfaces. It can be seen
from table 4 that the forces which was affected to implants has tended minimum deformation on z-direction and
maximum on y-direction. The deformations that were inversely proportional with the stresses which occurred on
x,y and z directions, can be explained with having force transfer on y-direction less than z-direction.
When we applied forces to the jaw, the stresses are occurring at the opposite direction of applied force in
order to keep it on its first position. The stresses will be maximum at the points where the reaction forces and the
stresses keep mandible movements. Due to these reasons, the maximum deformations were observed along xaxis on corpus, y-axis on mentus and z-axis on ramus regions. Total maximum deformations were calculated on
the mentus area where the reaction forces were reached maximum. According to all these investigations, as with
natural tooth restorations, having plastic deformations on xy-axis posterior tooth region and yz-axis anterior
tooth region at implant supported restorations explain us that natural tooth has shown parallel mechanical
behaviors with implant supported restorations. Due to these reasons, we should be careful on crown - tooth root
ratio, occlusal arrangements and buccolingual teeth diameter forming at implant supported restorations. The
lack of peri implants on basal implants makes the indication group for fixed restorations even broader.[13]
However, when we take into consideration that natural mastication stresses are typically between 250 N and
450 N, the implant designs for human mandible can be used safely. And also this study gives us certain
information about the stresses which occur depending on the bone quality where we implement the implant.
891
References
1.
2.
3.
4.
5.
6.
7.
8.
9.
10.
11.
12.
13.
Aykul, H. Toparli, M. Dalkiz, M. 2002. A Calculation of Stress Distribution in Metal-Porcelain Crowns By Using
Three-Dimensional Finite Element Method. Journal of Oral Rehabilitation 29(4), 381-386.
Baggi, L. et al. 2008. The influence of implant diameter and length on stress distribution of osseointegrated
implants related to crestal bone geometry: A three dimensional finite element analysis. The Journal of Prosthetic
Dentistry 100, 422-431.
Brunski, J.B. 1988. Biomaterials and Biomechanics in Dental Implant Design. Int J Oral Maxillofac Implants 3, 8597.
Choi, A.H. Ben-Nissan, B. and Conway, R.C. 2005. Three-dimensional modeling and finite element analysis of
the human mandible during clenching. Australian Dental Journal 50 (1), 42-48.
Danza, M. et al. 2009. 3D finite element analysis to detect stress distribution: spiral family implants. J Maxillofac
Oral Surg 8(4), 334339.
Himmlova, L. et al. 2004. Influence of implant length and diameter on stress distribution: A finite element analysis.
The Journal of Prosthetic Dentistry 91(1), 20-25.
Itoh, H. et al. 2004. Biomechanical comparison of straight and staggered implant placement configurations. Int J
Periodontics Restorative Dent. 24(1), 47-55.
Li, T. et al. 2010. Optimum selection of the dental implant diameter and length in the posterior mandible with poor
bone quality- A 3D finite element analysis. Appl. Math. Modelling, doi: 10.1016/j.apm.2010.07.008.
Motoyoshi, M. et al. 2009. Bone stress for a mini-implant close to the roots of adjacent teeth - 3D finite element
analysis. International Journal of Oral Maxillofacial Surgery 38, 363368.
Sorrentino, R. et al. 2007. Three-dimensional finite element analysis of strain and stress distributions in
endodontically treated maxillary central incisors restored with different post, core and crown materials. Dental
Materials 23, 983993.
imsek, B. et al. 2006. Effects of different inter-implant distances on the stres distribution around endosseous
implants in posterior mandible: A 3D finite element analysis. Medical Engineering & Physics 28, 199213.
Wua, T. et al. 2010. Design of acustom angled abutment for dental implants using computer-aided design and
nonlinear finite element analysis. Journal of Biomechanics 43, 19411946.
Kopp, S., Bienengrber, V. and Ihde, S. 2009. Basal implants as a solid base for immediately loaded full arch
bridges Prace Pogldowe, Dental Forum /1/2009/XXXVII/
892
asagbas@nku.edu.tr, 2tunca_nt@msn.com
Abstract.In this study, a new strategy to estimate annual wind energy output for a given wind turbine was
purposed. Annual wind speed of three different stations located in Mediterranean sea region of Turkey and
output power curve was estimated. A predictive model was developed based on fuzzy logic AHP and artificial
neural networks. These proposed methods were tested and it is shown that this new strategy has better accuracy
in the wind speed estimation than the conventional methods.
1 Introduction
The wind speed annual average of a given region varies from year to year, resulting in various annual energy
outputs for a given wind turbine. However, due to unpredictable nature of wind speed from time totime and
from location to location it is difficult to estimate theutilization factor of wind farms. The accurate long-term
andshort-term forecasting of wind speed is vital for wind power generationsystems efficiency. Therefore, to be
able to estimate wind energy potential of a region, wind speed variations in that region should be studied over
several years. The power generatingefficiency of a wind turbine can be significantlyincreased if the turbines
operation is controlled based on theinformation of wind and wind changes at the turbine location [1, 2]. Energy
output estimation for wind turbines of different power ranges has been the subject of a number of papers. Most
of these studies have used the wind speed pattern of a region to estimate the wind speed distribution of that
region and then, by knowing the wind turbine specifications, the energy output of that turbine is estimated. In
this paper, a new method is developed to estimate the annual average wind energy output of two different
turbines in Mediterranean Sea region of Turkey, using fuzzy modeling techniques and artificial neural networks.
To develop such a model, the output power curve of the wind turbine and the annual average wind speed
records of 32 years of three different regions namely Mersin, Anamur and Silifke are required.
2 Literature
There are various strategies for wind speed prediction that can be classified into two categories: statistical
methods that can be subdivided into numerical weather prediction and artificial intelligence techniques that have
subdivisions such as neural networks and fuzzy logic. Large number of wind speed measurements for fuzzy
logic system, the dimension of fuzzy rule base is large and it consumes more computational time. Several
studies have applied AHP during renewable energy planning, such as the utilization of solar energy technologies
in Jordan, sustainable development of rural energy in China, renewable energy planning in Korea, evaluation of
renewable energy plants. Kim and Lee [3] proposed a time-series prediction method using a fuzzy rule-based
893
system. In order to solve the fuzzy logic drawback in non-stationary systems they proposed a method of
building fuzzy rules for prediction which utilized the difference of consecutive values in a time series. In
references [4, 5] a fuzzy model has been suggested for the prediction of wind speed and the produced electrical
power at a wind park. They trained their model using a genetic algorithm-based learning scheme at various
stations in and around the wind park. Use of neural networks for prediction of wind speed and wind turbine
power production has been the subject of some articles. In the past many researchers tried to choose the best
structure for layer number, neuron number of different layers, and preprocessing and activation functions of
ANN in order to get the most accurate forecasting. Early applications neural networks in forecasting chaotic
time series have been performed by Lapedes and Farber [6]. They proposed an ANN along with feed forward
and error back-propagation algorithm in wind speed prediction. Song [7] developed an ANN-based
methodology to perform one-step ahead prediction, which is found to be good when the wind data does not
oscillate violently. It is shown that as long as the wind speed variation is not extremely brisk, the proposed
algorithm is able to predict the wind speed with a fairly good precision. The proposed predicting scheme does
not assume the precise knowledge of the wind model. Alexiadis et al. [8] also found that their ANN predictor is
about 10% better than persistence model for one-step ahead prediction. The work is focused on the operation of
power systems with integrated wind parks. Artificial neural networks models are proposed for forecasting
average values of the following 10 min or 1 hour. Also, spatial correlation of wind speeds and its use for
forecasting are investigated. Kalogirou [9] used the neural network to predict the mean monthly wind speed in
regions of Cyprus. He developed two networks, one with 5 inputs for prediction of mean monthly wind speed in
a determined region and another with 11 inputs for prediction of mean monthly wind speed in nearby locations
and he found that the proposed neural networks perform successfully. Mabel and Fernandez [10] developed the
artificial neural network model for prediction of mean monthly energy production of a wind farm in Muppandal/
India.
894
The network possesses three input neurons, representing, the monthly average relative humidity, the monthly
average atmospheric temperature and the monthly average atmospheric pressure. Finally, the network has one
output layer with one neuron, representing the total wind speed output of the considered stations. The second, the
analytic hierarchy process was applied to select the best location for wind power generating. The analytic
hierarchy process (AHP) is a simple, mathematically based multi-criteria decision-making tool to deal with
complex, unstructured and multi-attribute problems AHP proposed by Saaty (1980) shows the process of
determining the priority of a set of criteria and the relative importance of a multi-criteria decision making
problem among them. AHP is criticized for its inability to reect the human cognitive process because it does
not cope with the uncertainty and ambiguity, which occurs in decision-makers judgments. It is difcult to
respond to the preference of decision-makers by assigning a specic number. A control hierarchy, as depicted in
the rst part of Fig. 2, contains strategic criteria, the very basic criteria used to assess the problem, and the three
merits, Potential (P), environmental (E) and Technical (T). The control hierarchy is used to calculate the relative
importance of thethree merits.
The Selection of the most Suitable Location for Wind Power Generation
Potential (P)
Environmental (E)
Technic (T)
*Wind speed
*Seismic belt
*Road distance
*Wind power
*Forest area
*Transformer distance
*Radar facility
Mersin
Silifke
Anamur
Fig. 2. The control hierarchy for wind power generation system selection.
Experts are asked to pairwise compare the strategic criteria toward achieving the overall objective. The scores of
pairwise comparison of from each expert are transformed into linguistic variables by the transformation concept
listed in Table 1.
Table. 1. Membership functiona of triangular fuzzy members
IMPORTANCE OF ORAL
FUZZY SCALE
OPPOSITE SCALE
Equally important
(1, 1, 1)
(2, 3, 4)
(3, 4, 5)
(5, 6, 7)
Fully important
(8, 9, 9)
895
4 Findings Conclusions
In the present study, authors proposed a new strategy in wind speed prediction based on fuzzy logic and artificial
neural network. Fuzzy modeling techniques and artificial neural networks are used to develop a model to
estimate annual energy output for a given wind turbine in three different regions. The performance value of
each location on each criterion was calculated by synthesizing the results from all the experts by the arithmetic
average method. These performance values are further transformed into a number between zero to one by
dividing the performance value of a locations on a criterion by the largest performance value among all locations
on the same criterion. The priorities of criteria and sub-criteria were given in Figure 3.
The above performance values of locations and the priorities of criteria were synthesized to obtain the overall
performance of each location under each merit. Under the benets merit, the most important criterion is
potential, with a very high benet priority of 0.485. The most important sub-criterion is wind power and wind
speed. The performances of locations under three merits were calculated as shown in Figure 4.
896
Anamur location is expected to be the best site mainly because it has the best performance under three merits.
Even though Silifke location ranks the second under the opportunities merit, its beter performance under the
other merits guarantees its rst rank under the synthesizing methods. Also note that Mersin location
hasaperformanceworse thanthe otherstations. Hovewer, some networks and techniques are introduced to estimate
the annual wind energy output of a wind turbine in any region which should be easy to use and has satisfactory
accuracy.The predicted wind speed and wind power values were shown in Table 2.
Table 2.The predicted wind speed and wind power values for all stations
WIND SPEED (m/s)
MERSIN
3.45
25,15
SILIFKE
3.52
26,71
ANAMUR
4.17
44,41
According to Table 2, the best choice in considered locations is Anamur station which has the nominal power
capacity of 44.41 (W/m2). The power capacity of Silifke and Mersin stations are close.
References
1.
2.
3.
4.
5.
6.
7.
8.
9.
10.
11.
12.
13.
14.
15.
Monfared, M. Rastegar, H. Kojabadi, H.M. (2009). [A new strategy for wind speed forecasting using artificial
intelligent methods]. Renewable Energy, 34, 845848.
Jafarian, M. Ranjbar, A.M. (2010). [Fuzzy modeling techniques and artificial neural networks to estimate annual
energy output of wind turbine]. Renewable Energy, 35, 20082014.
Kim, I. Lee, S.H. (1999). [A fuzzy time series prediction method based an consecutive values]. In: Proceedings of
the IEEE international fuzzy system conference, Seoul, Korea; 2, 703707.
Damousis, I.G. Dokopoulos, P. (2001). [A fuzzy expert system for the forecasting of wind speed and power
generation in wind farms]. In: 22nd IEEE Power Engineering Society international conference on power industry
computer applications, 6369.
Damousis, I.G. Alexiadis, M.C. Theocharis, J.B. Dokopoulos, P. (2004). [A fuzzy model for wind speed
prediction and power generation in wind parks using spatial correlation]. IEEE Transactions on Energy
Conversion, 19(2), 352361.
Lapedes, A. Farber, R. (1987). [Nonlinear signal processing using neural networks: prediction and system
modeling]. Technical report LA-UR-87-2662 Los Alamos, NM: Los Alamos National Laboratory.
Song, Y. D. (2000). [A new approach for wind speed prediction]. Wind Engineering, 24(1), 3547.
Alexiadis, M. Dokopoulos, I. Sahsamanoglou, H. Manousardis, I. (1988). [Short-termforecasting of wind speed
and related electircal power]. Solar Energy, 63(1), 6168.
Kalogirou, S.A. (2004). [Artificial neural networks in renewable energy systems applications:a review].
Renewable and Sustainable Energy Review, 5, 373401.
Mabel, M.C. Fernandez, E. (2008). [Analysis of wind power generation and predictionusing ANN: a case study].
Renewable Energy, 33, 986992.
Wang, L.X. Mendel, M. (1992). [Generating fuzzy rules by learning from examples]. IEEETransactions on
Systems, Man and Cybernetics, 22(6), 14141426.
Wang, L.X. (1996). [A course in fuzzy systems and control]. Upper Saddle River, NJ:Prentice-Hall Inc.
Lee, S.H. Kim, I. (1994). [Time series analysis using fuzzy learning]. In: Proceedings of theinternational
conference on neural information processing, Seoul, Korea; 6, 15771582.
Export choice inc.
Chen, H.H. Kang, Y. Lee, H.I. (2010). [Strategic selection of suitable projects for hybrid solar-wind
powergeneration systems]. Renewable and Sustainable Energy Reviews, 14, 413-421.
Biographies
Aysun SABA Dr. Sagbas is an Associate Professor in Industrial Engineering Department at Namk Kemal University,
Turkey. She received his Ph.D.(2003) in Industrial Engineering from Cukurova University, Turkey. Dr. Sagbas's major
research areas are optimization and modeling, statistical quality improvement, experimental design, artificial intellegent.
Tunca KARAMANLIOLU He was born in Sultanhisar Aydn / TURKEY at 1985. He finished University of Mersin,
Department of Mechanical Education with 83 points. Masters education at the University of Mersin began in 2009
continues.
897
Abstract. This paper details a numerical simulation of the flow field inside a three dimensional bypass graft to artery
connection under steady flow conditions with different graft connection angles The solution of the Navier-Stokes equations
was obtained using a commercially available finite difference code (ANSYS-FLUENT).
feasibility of simulating flow with this code inside a complex three-dimensional graft geometry at different flow rates.
Critical flow regions such as separation, stagnation and high shear stress are detected. The location of low shear stress and
high shear stress are changed by changing graft angles. Smooth flow phenomena is found inside curved bypass graft to
artery connection.
Introduction
In this study the flow phonemena inside a bypass graft to artery connection is analysed with different graft
angles under steady flow conditions. The velocity profiles and shear stresses at different flow locations are
found numerically. Due to the ever increasing speed in computers, numerical solutions are become important
technique to determine the flow patterns and wall shear stress (WSS) distributions of arterial flows. Critical flow
regions are found in this connection. Low shear stress regions are found at separation points and also stagnation
points. The results are expected to find new graft designs and model development. New graft models will offer
least atherosclerosis formation and less blockage in connection regions. New graft designs will be manufactured
and used first in animal studies and humans
Literature review
In cardiovascular system, Bypass graft implantation is a widely used surgical procedure for those patients with a
severe stenosis of the coronary arteries. Although the surgical technique is well established, the need for a better
comprehension of the fluid dynamic and structural behavior of the reconstructed portion of the coronary
circulation still seems important. Indeed, a significant number of graft conenction fails postoperatively due to the
development of intimal hyperplasia, which occurs preferentially at the site of the distal anastomotic junction.
Intimal thickening (hyperplasia) has been cited as a major cause of vascular graft failure [Echave et al. 1979],
and hemodynamics have been implicated in the localization of intimal hyperplasia [Sottiurai et al. 1989, White et
al. 1993, Bassiouny et al. 1991]. The two main techniques currently employed to investigate the flow patterns
898
and WSS distributions inside vascular grafts are experimental measurements inside in-vitro models and
numerical simulations of the flow field. Figure 1 shows a typical by pass graft configuration with blood entering
the distal anastomosis from the graft and exiting out both the distal outlet segment (DOS) and the proximal outlet
segment (POS). Figure 2 indicates the nomenclature employed herein. The surgically constructed geometry can
vary in many ways including: ratio of host artery-to-graft vessel diameter, graft angle with host artery, and hood
length.
Graft
DOS
POS
Figure 1.
Sketch showing blood flow bypassing the occlusion and exiting the distal anastomosis through both
the proximal and distal outlet segments (POS and DOS)
Hood
Graft
Toe
Heel
(Apex)
DOS
POS
Floor
-x/D
Figure 2.
+x/D
899
900
901
902
References
1.
2.
3.
4.
5.
6.
7.
8.
9.
Bassiouny H.S., White S.S., Glagov S., Choi E., Giddens D.P., Zarins C.K., 1992, "Anastomotic intimal
hyperplasia: mechanical injury or flow induced," Journal of Vascular Surgery, Vol. 15, No. 4, pp. 708717.
Echave V., Koornick A.R., Haimov M., 1979, "Intimal hyperplasia as a complication of the use of the
polytetrafluoroethylene graft for femoral-popliteal bypass," Surgery, Vol. 86, No. 6, pp. 791.
Fei D.Y., Thomas J.D., 1994, "The effect of angle and flow rate upon hemodynamics in distal vascular
graft anastomoses: A numerical model study," Journal of Biomechanical Engineering, Vol. 116, pp.
331-323.
Loth F., 1993, "Velocity and Wall Shear Measurements Inside a Vascular Graft Model Under Steady
and Pulsatile Flow Conditions," Ph.D. Dissertation, Georgia Institute of Technology.
Loth F., Jones S.A., Bassiouny H.S., Giddens D.P., Zarins C.K., Glagov S., 1993, "Laser Doppler
velocity measurements inside a vascular graft model under steady flow conditions," Bioengineering
Conference Proceedings, Edited by N.A. Langrana, M.H. Friedman, E.S. Grood, BED Vol. 24, pp. 4851, presented at the ASME Summer Annual Bioengineering Conference, Breckenridge, CO, June 2429, 1993.
Ojha M., Ethier C.R., Johnston K.W., Cobbold S.C., 1990, "Steady and pulsatile flow fields in an endto-side arterial anastomosis model," Journal of Vascular Surgery, Vol. 12, pp. 747-53.
Perktold K., Helfried T., Rappitsch G., "Flow dynamic effect of the anastomotic angle: a numerical
study of pulsatile flow in vascular graft anastomoses models", Technology and Health Care, 1994, Vol.
1, pp. 197-207.
Sottiurai V.S., Yao J.S.T., Batson R.C., Sue S.L., Jones R., Nakamura Y.A., "Distal Anastomotic
Intimal Hyperplasia: Histopathologic Character and Biogenesis," Annals of Vascular Surgery, Vol. 3,
No. 1, pp. 26-33, 1989.
White S.S., Zarins C.K., Giddens D.P., Bassiouny H.S., Loth F., Jones S.A., Glagov S., 1993,
"Hemodynamic patterns in two flow models of end-to-side vascular graft anastomoses: effects of
pulsatility, flow division, Reynolds number and hood length," Journal of Biomechanical Engineering,
Vol. 115, pp. 104-111.
Acknowledgements
This work is supported by the Scientific Research Fund of Fatih University under the project number
P50091001_2.We would like to thank Fatih university for the support given to our project.
903
Abstract - This work obtains an explicit analytical solution for nonlinear Poisson- Boltzmann equation by the
homotopyperturbation method and perturbationiteration method, which does not require a small parameter in
the equation under study, so it can be applied to both weakly and strongly nonlinear problems. Results of these
two methods were compared for the problem of nonlinear Poisson Boltzmann equation.
1. Introduction
Perturbations methods derive approximate analytical solutions for mathematical models. Differential equations
can be solved approximately with these techniques. The direct expansion method (straightforward expansion)
can not yield solutions for most of the cases. An alternative methods such as LindstedtPoincare technique,
Renormalization method, Averaging methods, Method of Multiple Scales [1] have been developed within time.
Perturbation method is that a small parameter is needed in the equations or the small parameter should be
introduced artificially to the equations.
Methods of iteration-perturbation [3, 4] are alternative attempts in the literature to validate solutions for strongly
nonlinear systems. In the proposed iteration-perturbation methods, the nonlinear terms are linearized by
substitution of iterative solution functions from previous iteration results. Usually, in iteration-perturbation
methods [5, 6], the equations are cast into an alternative form before applying the iteration procedure.
In the study, different perturbation-iteration algorithms depending upon the number of Taylor expansion terms
were proposed. Using the iteration formulas derived by perturbation-iteration algorithm, new solutions of
PoissonBoltzmann equations were obtained.
2. Problem Formulation
Recently Oyanader and Arce [2] suggested a more effective, accurate, and mathematically friendly solution for
prediction of the electrostatic potential, commonly used on electro kinetic research and its related applications. In
order to obtain an explicit mathematical expression for the electrostatic potential, we have to solve the following
nonlinear Poisson-Boltzmann equation:
u=2sinh(u),
(1)
where u is the dimensionless electrical potential, the dimensionless inverse Debye length. The nonlinear term
on the right-hand side of Poisson-Boltzmann is related to the free charge density. A very common simplification
invokes the Debye-Hckel approximation usually written as
904
sinh(u)u.
(2)
(3)
Such approximation is valid for the case when -1 u 1, In this paper, we will suggest an alternative approach
for an explicit analytical solution for Poisson-Boltzmann by homotopy perturbation method [2] and perturbation
iteration method.
F(u , u, ) = 0
(4)
with u=u(t) and the perturbation parameter. Only one correction term is taken in the
perturbation expansion:
u n +1 = u n + ( u c ) n
(5)
Upon substitution of (5) into (4) and expanding in a Taylor series with first derivatives only
yields
F( u n , u n ,0) + Fu (u n , u n ,0) ( u c ) n + Fu ( u n , u n ,0) (uc ) n + F (u n , u n ,0) = 0
(6)
where ( ) denote differentiation with respect to the independent variable and Fu =
Fu =
F
,
u
F
F
, F =
. Reorganizing the equation as
u
( uc ) n +
F +
Fu
(u c ) n =
Fu
Fu
(7)
variable coefficient linear second order differential equation. Starting with an initial guess u0, first (uc)0 is
calculated from (7) and then substituted into (5) to calculate u1. The iteration procedure is repeated using (7) and
(5) until a satisfactory result is obtained. This iteration algorithm may produce similar results with the homotopy
perturbation method. See[7].
905
Example 1.
We consider the mathematical pendulum with initial conditions; the equation can be
expressed as
u + 2 sin(u ) = 0,
(8)
u (0) = A, u (0) = 0.
(9)
(10)
(11)
where is an artificially introduced small parameter. Terms in iteration formula (7) are
Fu (u0, u 0 ,0) = w ,
1 2 3
2
2
F (u0, u 0 ,0) = w u0 u0
6
F (u0, u 0 ,0) = u0 + w2 u 0 ,
Fu (u0, u 0 ,0) = 1,
3
( uc ) n + w 2 ( u c ) n = 2 w2 u0 2u0 0
6
n = 0, 1, 2 L
(13)
For the initially assumed function, one may take a trivial solution which satisfies the given
initial conditions,
u 0 = A cos(wt )
(14)
)
(15)
w = 1
1 2
A
8
(16)
906
2
1 18 A2
(17)
( )
Example 2.
(18)
(19)
Here, is a modified inverse Debye length. Equation (13) is different from Debye Hchel
approximation (2) in that the former is valid for the whole solution domain, while the latter
for 1 u +1.
We construct the equation in the following form
1
u w 2 u [(2 w 2 )u + 2 u 3 ] = 0,
6
(20)
where is an artificially introduced small parameter. Terms in iteration formula (7) are
F (u 0 , u 0 ,0) = u 0 w 2 u 0 ,
Fu (u 0 , u 0 ,0) = 1,
Fu (u 0 , u 0 ,0) = w 2 ,
1
3
F (u 0 , u 0 ,0) = 2 w 2 u 0 + 2 u 0
6
2
2
( u c ) n w ( u c ) n = w u 0 u 0
n = 0, 1, L
(21)
3
( uc ) 0 + w2 ( u c ) 0 = 2 w 2 u0 + 2u0 (u0 + w 2u0 ) n = 0, 1, 2 L
6
907
For the initially assumed function, one may take a trivial solution which satisfies the given
boundary conditions
u0 =
A cosh wx
.
cosh w
(23)
Substituting u 0 into (22) and collecting the cosh wx term, and eliminating the cosh wx term
altogether, we have
w = 1+
A2
8 cosh 2 w
(24)
cosh 3w
A cosh(wx)
A32
u1 (t ) =
cosh wx
+
cosh 3wx
2
3
cosh w
192w cosh w
cosh w
(25)
Where w is solved from (24). The w and T are the same as in ref[7] .
References
[1] A.H. Nayfeh, Introduction to Perturbation Techniques, John Wiley and Sons, New York 1981.
[2] M. Oyanader and P. Arce, A new and simpler approach for the solution of the electrostatic potential
differential equation.Enhanced solution for planar, cylindrical and annular geometries, Journal of Colloid and
Interface Science 284 (2005), no.1, 315-322.
[3] J.-H.He, A coupling method of a homotopy technique and a perturbation technique for non-linear problems,
International Journal of Non-Linear Mechanics 35 (2000), no. 1, 37-43
[4] J. H. He, Iteration Perturbation Method for Strongly Nonlinear Oscillations, Journal of Vibration and
Control 7: 631-642 (2001).
[5] M. Pakdemirli, Perturbation - Iteration method and comparisons with variational iteration method, The 1st
International Symposium on Computing in Science & Engineering (ISCSE 2010) June, 3-5, 2010, in Kuadas,
Aydn, Turkey.
[6] Y. Aksoy, M. Pakdemirli, New Perturbation Iteration solutions for Bratu-type equations, Computers and
Mathematics with Applications, 59 (8): 2802-2808 (2010).
[7] Li-Na Zhang And Ji-Huan He, Homotopy perturbation method for the solution of the electrostatic potential
differential equation, Mathematical Problems in Engineering, Volume 2006, Article ID 83878, Pages 1-6 (2006)
908
Abstract.A general cubic nonlinear vibration model is considered. Parametric vibrations of the cubic
nonlinear system are analysed. The principal parametric resonances and three-to-one internal resonances of
parametric excitation are obtained. Approximate analytical solutions of the equation of motion are found by
using the method of multipe scales (a perturbation method). The amplitude and phase modulation equations
are derived. The solution algorithm is applied to the problem of nonlinear vibrations of pipe conveying fluid.
Natural frequencies of pipe are found.
1 Introduction
Cubic nonlinear terms may appear in algebraic, differential or integral forms in vibration models of
continuous systems. The nature of nonlinearities does not permit exact solutions. To find the approximate
solution of system in general sense, a perturbation method is suitable. To understand the effects of quadratic and
cubic nonlinearities on the solutions of continuous systems, an operator notation suitable for perturbation
analysis was developed [1]. That analysis was involved only one mode. The analysis was generalized to infinite
modes [2]. Parametrically excited linear system expressed by arbitrary linear operators were also analysed [3].
Forced vibrations of an arbitrary cubic nonlinear system was employed to compare results of different versions
of method of multiple scales [4]. For system modelled with more than one partial differential equation, a solution
procedure was developed [5]. Using the same model, arbitrary internal resonances of vibration of continuous
system were discussed [6]. Three to one internal resonances of generalized cubic nonlinear vibration model were
analysed [7]. The new operator notation were adopted to understand effects of nonlinearities by others also [8] [9]. General solution procedure for cubic nonlinear vibration model including additional linear and cubic
operators with time derivatives were intercorporated into the model [10] -[11].This enables analyzing a more
general class of continuous systems such as gyroscopic systems, which are encountered in axially moving media.
In that works, primary and three to one internal resonances with external excitation were analysed respectively.
In this work principal parametric resonance with three to one internal resonances of parametrically excited
system are obtained. Firstly, a general model is developed. Then method of multiple scales [12] is used to solve
the system. Amplitude and phase modulation equations are found. Steady state solutions and their stability are
discussed in a general sense. The solution algorithm is applied to nonlinear vibration model of pipe conveying
fluid [13-14]. Natural frequencies of pipe are found.
909
2 General Model
The general model and boundary contidions are
&& + L1 ( w) + L 2 ( w& ) + L 3 ( w& ) = L 4 ( w) cos t + {C1 ( w, w, w) + C 2 ( w& , w, w)} (1)
w
B1 ( w ) = 0
x = 0 B2 ( w) = 0
x =1
(2)
In equation (1), w( x, t ) represents deflection, x is time and t is time. L1 , L 2 L 3 and L 4 are linear differential
and/or integral operators. L 4 include parametric term.
is
a physical
3 Perturbation Solution
Approximate solution is assumed of the form
w( x, T0 , T1 ; ) = w0 ( x, T0 , T1 ) + w1 ( x, T0 , T1 ) + ...
(3)
In equation (3), T0 = t represents fast time scale and T1 = t is slow time scale. Time derivates are
d2
d
= D02 + 2 D0 D1 + ......
= D0 + D1 + .....
2
dt
dt
(4)
Inserting Eqs. (3)-(4) into Eqs. (1) and (2) and separating at each order of
O ( 0 ) : D02 w0 + L1 ( w0 ) + L 2 ( D0 w0 ) = 0
B1 ( w0 ) = 0 at x = 0
O ( 1 ) :
and B 2 ( w0 ) = 0
(5)
at
x =1
(6)
D02 w1 + L1 ( w1 ) + L 2 ( D0 w1 ) = 2 D0 D1 w0 L 2 ( D1 w0 ) L 3 ( D0 w0 )
+ L 4 ( w0 ) cos T0 + C1 ( w0 , w0 , w0 ) + C2 ( D0 w0 , w0 , w0 )
B1 ( w1 ) = 0
at
x=0
and B 2 ( w1 ) = 0
at
x =1
(7)
(8)
(9)
In equation (9), An and An are complex amplitudes and their conjugates respectively. Yn ( x)
(and Ym ( x) ) satisfies the following equation and boundary conditions,
L1 (Yn ) + in L 2 (Yn ) n2 Yn = 0
n=1,2,
910
(10)
B1 (Yn ) = 0
at
x = 0 and B 2 (Yn ) = 0
at x = 1
(11)
1
1
1
An ei (n + )T0 L 4 (Yn ) + An ei ( n + )T0 L 4 (Yn ) + Am ei (m )T0 L4 (Ym )
2
2
2
(12)
B1 ( w1 ) = 0
at
x = 0 and B 2 ( w1 ) = 0
at x = 1
(13)
where n and n are detuning parameters. c.c. and N.S.T. represent complex conjugates and
non-seculer terms respectively. Solution of equation (12) possesses
w1 ( x, T0 , T1 ) = n ( x, T1 ) ei n T0 + m ( x, T1 ) ei m T0 + c.c. + W ( x, T0 , T1 )
(14)
(15)
911
=0
(16)
Where
1
k1n =
i n Yn L 3 ( Yn ) dx
, k1m =
i m Ym L 3 (Ym ) dx
2 i n Yn Yn dx + Yn L 2 ( Yn ) dx
(17)
2 i m Ym Ym dx + Ym L 2 (Ym ) dx
Y {C (Y , Y , Y ) + C (Y , Y , Y ) + C (Y , Y , Y )
n
k2 n =
2 i n Yn Yn dx + Yn L 2 (Yn ) dx
Y {C (Y
m
, Ym , Ym ) + C1 ( Ym , Ym , Ym ) + C1 (Ym , Ym , Ym )
k2m =
k3 n =
(18)
2 i m Ym Ym dx + Ym L 2 (Ym ) dx
1
1
2
Yn L 4 (Yn ) dx
0
2 i n Yn Yn dx + Yn L 2 (Yn ) dx
k4 mn =
1
2
, k4 nm =
1
2
L 4 (Ym ) dx
2 i n Yn Yn dx + Yn L 2 (Yn ) dx
L 4 (Yn ) dx
(21)
2 i m Ym Ym dx + Ym L 2 (Ym ) dx
Y {C (Y
n
, Yn , Yn ) + C1 (Yn , Ym , Yn ) + C1 (Yn , Yn , Ym )
k5 nm =
Y {C (Y
1
(20)
(19)
(22)
2 i n Yn Yn dx + Yn L 2 (Yn ) dx
2 i n Yn Yn dx + Yn L 2 (Yn ) dx
912
(23)
Y {C (Y , Y , Y ) + C (Y , Y
m
, Yn ) + C1 (Ym , Yn , Yn ) + C1 (Yn , Yn , Ym )
(24)
2 i m Ym Ym dx + Ym L 2 (Ym ) dx
1
Y {C (Y , Y , Y ) + i C (Y , Y , Y )} dx
m
k7 mn = 0
(25)
2 i m Ym Ym dx + Ym L 2 (Ym ) dx
All coefficient have real and imaginary parts ( k j = k j R + i k j I ). Polar forms of complex
amplitudes are
An =
1
i T
an (T1 ) e n ( 1 )
2
1
2
are Am = am (T1 ) ei
(T1 )
Eqs. (15) and (16) finally gives the amplitude and phase modulation equations
1
an = an k1n R + an3 k 2 n R + an (k3 n R cos n k3 n I sin n ) + am (k 4 nm R cos nm k 4 nm I sin nm )
4
1 2
1
+ an am (k5 nm R cos( n + nm ) k5 nm I sin( n + nm )) + an am2 k6 nm R = F1 (an , am , nm , n )
4
4
1
1
am = am k1m R + am3 k 2 m R + an (k 4 mn R cos nm + k4 mn I sin nm ) + an2 am k6 mn R
4
4
1 3
+ an (k7 mn R cos( n + nm ) + k7 mn I sin( n + nm )) = F2 (an , am , nm , n )
4
1
2
(27)
(28)
2am
(k4 nm R sin nm + k4 nm I cos nm )
an
1
1
an am (k5 nm R sin( n + nm ) + k5nm I cos( n + nm )) am2 k6 nm I = F3 (an , am , nm , n )
2
2
(29)
1
4
am
(k4 nm R sin nm + k4 nm I cos nm )
an
a
1
1
1
an am (k5 nm R sin( n + nm ) + k5 nm I cos( n + nm )) am2 k6 nm I k1m I + am2 k2 mI n (k4 mnR sin nm
4
4
4
am
1
1 an3
(k7 mnR sin( n + nm ) k7 mnI cos( n + nm )) = F4 (an , am , nm , n )
k4 mn I cos nm ) + an2 k6 mn I
4
4 am
(30)
Where
nm = ( n n )T1 n + m ,
n = nT1 2 n , nm = n + nm
913
(31)
F1
a
n
F2
an
F3
an
F
4
an
F1
am
F1
nm
F2
am
F2
nm
F3
am
F3
nm
F4
am
F4
nm
F1
,
n
F2
n
F3
n
F4
n an = an 0 , am = am 0
= , =
nm
nm 0
(32)
n0
= n = 0
a n = a m = nm
w ( x, t ; ) = an cos [ t n ] Yn R sin [ t n ] Yn I
2
2
(33)
+ am cos [3 t n ] + nm Ym R sin [3 t n ] + nm Ym I + O( )
2
2
4 Application
In this section the solution algorithm is applied to a specific problem. Application problem
represents nonlinear vibrations of pipe conveying fluid [13-14]
1
w ( 0, t ) = w (1, t ) = w ( 0, t ) = w (1, t ) = 0
(34)
(35)
In equation (34), x is the coordinate along the axial direction, t is the time,P is the axial
tension force ( P = P0 + P1 cos t ), w is the transverse displacement. is damping coefficient.
k1 is dimensionless parameter related to pipe geometry.
(36)
C1 ( w, w, w) = k1 w w2 dx , C2 ( w& , w, w) = 0
L4 ( w) = P1 w ,
(37)
] ei
4n ( x)
(38)
}
Figure 1 shows first natural frequencies versus fluid flow velocities for different P0 values. Frequencies decrease
by increasing the mean velocity values. Natural frequencies increase by increasing axial tension force.
914
915
5 Conclusion
A general model of cubic nonlinear system is obtained. Parametric vibrations of system are analysed. Internal
resonances are obtained in analytical sense. The general solution algorithm is adopted to nonlinear vibration
model of pipe conveying fluid. Natural frequencies are found. Stability conditions of system is analitacally
showed.
References
1.
2.
3.
4.
5.
6.
7.
8.
9.
10.
11.
12.
13.
14.
Pakdemirli, M. (1994). A comparison of two perturbation methods for vibrations of systems with quadratic and
cubic nonlinearities, Mechanics Research Communications 21 (2) 203-208.
Pakdemirli, M. and Boyac, H. (1995). Comparision of direct-perturbation methods with discretizationperturbation methods for nonlinear vibrations, Journal of Sound and Vibration 186 837-845.
Pakdemirli, M. and Boyac, H. (1997). The direct perturbation method versus the discretization perturbation
method: Linear systems, Journal of Sound and Vibration 199 (5) 825-832.
Boyac, H. and Pakdemirli, M. (1997). A comparison of different versions of the method of multiple scales for
partial differential equations, Journal of Sound and Vibration 204 (4) 595-607.
Pakdemirli, M. (2001). Vibrations of continuous systems with a general operator notation suitable for perturbative
calculations , Journal of Sound and Vibration 246 (5) 841-851.
Pakdemirli, M. (2001). A general solution procedure for coupled systems with arbitrary internal resonances,
Mechanics Research Communications 28 (6) 617-522.
Pakdemirli, M. and zkaya, E. (2003). Three-to-one internal resonances in a general cubic non-linear continuous
system, Journal of Sound and Vibration 268 543553.
Lacarbonara, W. (1999). Direct treatment and discretizations of non-linear spatially continuous systems, Journal of
Sound and Vibration 221 849866.
Nayfeh, A.H. (1998) Reduced order models of weakly nonlinear spatially continuous systems, Nonlinear Dynamics
16 105125.
zhan, B. B. and Pakdemirli,M. (2009) A general solution procedure for the forced vibrations of a continuous
system with cubic nonlinearities: Primary resonance case, Journal of Sound and Vibration325 894906.
zhan, B. B. and Pakdemirli,M. (2010) A general solution procedure for the forced vibrations of a system with
cubic nonlinearities: Three-to-one internal resonances with external excitation, Journal of Sound and Vibration 329
2603-2615
Nayfeh, A. H. (1981). Introduction to Perturbation Techniques. New York: John Wiley.
Padoussis, M. P., and Li, G. X. (1993) Pipes Conveying Fluid: A Model Dynamical Problem, Journal of Fluids
and Structures, 7 (2) 137-204.
Panda, L.N.,and Kar, R.C. (2008) Nonlinear dynamics of a pipe conveying pulsating fluid with combination,
principal parametric and internal resonances, Journal of Sound and Vibration 309 375-406
916
selim.solmaz@gediz.edu.tr, 2baris.yagci@gmail.com3
Abstract.In this paper we analyze a generic ocean power converter buoy model that is composed of a twomass spring-damper system. The damper in this model is considered to be part of the power take-off (PTO)
mechanism, which is acting as a hydraulic pump providing an adaptive damper reaction force. In doing so, a
Kalman Filter is used to estimate the periods and the amplitudes of the incoming waves in the near future with
reference to the current wave state. We demonstrate the approach using historical ocean wave data. We then
describe how a Kalman Filter based estimator can be used to estimate various sea states.
Keywords: Wave Energy Conversion, Kalman filtering, Estimation, non-stationary noise, spectral analysis.
Introduction
Wave energy converters (WEC) represent a new research field in the context of renewable energy generation
systems. It is well known that the power density (i.e., the total mechanical energy per unit area) of ocean waves is
quite high, particularly as compared to the wind power density. Once such potential is harnessed, it could be the
solution to the ever-increasing renewable energy demand. Indeed this potential is well recognized and many
researchers and commercial enterprises are known to be active in this specific research area.
There are various types of ocean energy generation systems found in the literature. Existing systems can be
categorized into 3 groups: shore-based oscillating-water-column type generators, tidal power generators, and
offshore buoy-type ocean wave generator systems. While either systems offer great potential for unlimited energy
generation, the largest potential lies with the offshore heaving-buoy type ocean power generators since the wave
height and activity increases with water depth.
Buoy type generators usually include a power take-off (PTO) system that uses an electro-mechanical or an
electro-hydraulic mechanism to convert the heaving motion of the buoy for driving the embedded electric
generators. Some of the common mechanisms of the PTO depend on the relative displacement of parts of the
energy conversion device. It is important to maximize the relative displacement of the parts of such a device to
increase the power converted. In turn, the effectiveness of the PTO mechanism is maximized once the natural
frequency of the PTO is same as that of the oncoming waves. While it is possible to adapt oscillation frequency
of the PTO using different means, such modifications require an estimation of certain features of the oncoming
1
2
917
Literature Review
There are many examples for the use of Kalman Filtering in the context of marine engineering applications. For
example, ship motion estimation and control is a very mature area [1,2] and has similar constraints to the waveforecasting problem [3].
In recent years, Kalman Filtering has been applied to wave energy analysis and design problems. A recent
publication on this very specific topic is [4], which implements a Kalman Filter based estimator as we suggest.
However, the specific model and the type of Kalman Filter used in this paper is different than our paper. Their
adaptation scheme for use in PTO control is different than our approach. Specifically, here Kalman Filtering is
used to estimate the incoming wave amplitudes and the forces and adjust the applied force on the system by
adapting the damping coefficient between the two masses to obtain the maximum power take off. Another recent
918
paper in a related subject is [5], which again implement a Kalman Filter predictor for the ocean waves. In this
paper a sinusoid wave model is assumed and an Extended Kalman Filter was suggested based on the linearization
of the process model. Again this paper represents a different technique than our approach.
In terms of statistical methods utilized for wave feature analysis, we use Capon's method, which is based on the
Capon-Geronimus algorithm presented in [6,7,8,9]. This algorithm estimates the frequency spectrum of the data
corrupted by noise. As the data set gets larger, the estimation converges to the point spectrum. In the limit, the
amplitude of the spectrum converges to the amplitude of the underlying sinusoids in the corrupted signal. The
convergence result of the method makes it a practical tool for distinguishing peaks due to sinusoids in the
spectrum from the peaks that may occur due to noise process. This method is applied on actual ocean wave data
to demonstrate its fitness. Using synthetic harmonic waves as an input to such a system may create an unrealistic
favorable condition.
The wave frequency and height are important parameters in the design of the wave energy conversion design
methodology discussed here. The nominal values of the spring constant and the damping coefficients are
determined after determining the dominant ocean wave characteristics at CDIP buoy off of Humboldt Bay. For
this work the wave data for the whole year 2009 is scanned to obtain data for 52 days, at a 7-day sampling rate.
The overall data covers the seasonal changes in wave characteristics and the analysis of the data gives an idea on
how quickly they change. The data that pertains to each day includes the whole 24 hours of the day. This high
resolution data helps understand the daily variations of wave height and frequency. The seasonal and daily wave
characteristic changes dictate how quickly the system should adapt itself to different sea conditions.
Here Capon's method is used for spectral analysis since it provides an estimate of sinusoid frequencies corrupted
by noise. Other spectral methods could have been used instead but the convergence criteria of the spectrum to
sinusoid frequencies to the sinusoid amplitudes gives a higher level of confidence in picking dominant spectral
peaks. As more number of points is used to calculate the spectrum, the estimate converges to the point spectrum
of the sinusoid process at the sinusoid frequencies and to zero everywhere else. Therefore, artificial spectral peaks
due to noise processes would die down as larger data sets are used for spectrum calculations.
When a short time window is used for calculations the dominant frequency on the spectrum is an estimate of the
ocean wave frequency in the time window [9]. In this particular application Capon's method is used on 5 minute
shifting segments of data for over 24 hours period of time for the selected 52 days. The maximum and minimum
wave frequencies are determined for each day. The collection of these extremums and their average values over
the whole year plotted in Figure 1. The star on the vertical red lines indicates the average daily wave frequency
and the extremities of the line indicate the minimum and the maximum. The horizontal axis starts with January
8th and the average frequency is 1.41 rad/s on this day according to the measurement data. Towards the middle of
the year, by 22nd week, the waves have highest average dominant frequency of 2.86 rad/s. The lowest average
frequency seen during the year is 1.24 rad/s on Jan 29, 2009.
919
Figure 1 Frequency variation CDIP Humboldt Bay Buoy over 2009 at selected days once a week
Another look in to data is given in the histogram in Figure 2. This histogram uses the same data in Figure 1
however all the data is aggregated over 52 days. The average wave frequency over the whole year is 1.8 rad/s and
the standard deviation is 0.34 rad/s.
Figure 2 Histogram of the dominant frequency variation CDIP Humboldt Bay Buoy, for 5 min. shifting windows at
selected days once a week over 2009.
Here we describe the model that we consider for the wave energy convertor device, which shall be utilized for
Kalman filter based state observer design. The model is composed of a 2-mass spring-damper oscillator that is
constrained to move in vertical heaving direction. The vertically oriented 2-mass spring damper system system is depicted in Figure 3. Assumptions regarding this model are as follows:
The buoyancy force on the bottom mass can be assumed to be constant since it is constantly submerged
underwater,
Buoyant forces on the top mass is time varying since its draft is changing due to the wave movement,
Vertical forces (buoyancy and gravitational forces) act on the motion axis of the 2-mass oscillator,
The only input the system can be taken as the wave force acting on top mass due to the motion the wave
front.
920
Utilizing the free body diagram with reference to Figure 3, and using Newtons 2nd law of motion, the equations
of motion for this model are given with following two second order ordinary differential equations (ODEs):
&z&1 =
&z&2 =
( w z1 )a 2 g w ha 2 g
k
c
,
( z1 z 2 )
( z&1 z& 2 ) g + w
+
m1
m1
m1
m1
a3 g
k
c
,
( z1 z 2 ) +
( z&1 z& 2 ) g + w
m2
m2
m2
(1)
(2)
where W is the water density, w is the displacement of the water surface measured from flat sea state, and g
represents the gravitational acceleration. Also, c denotes the damping coefficient, and k is the spring stiffness of
the link connecting the two masses. The terms including W in both equations are due to the buoyancy force.
While the equations of motion are derived assuming that both m1 and m2 are point masses, for the calculation of
buoyancy forces they are assumed to have the shape of a cube with side length a. The last term in (1) is the
buoyancy force acting on m1with changing draft h.
In order simplify this model, one can consider the fact that the buoys oscillate in the close vicinity of their
respective positions. Therefore, at steady state, the buoyancy force acting on each buoy must be equal toits weight
(so that they stay at the same average position). Based on this fact, we deduce that m1g = Wha2g in (1) and m2g =
Wa3g in (2), which after substitution results in the following simplified 2nd order ODEs
&z&1 =
(w z1 )a 2 g
k
c
,
( z1 z 2 )
( z&1 z& 2 ) + w
m1
m1
m1
(3)
k
c
( z1 z 2 ) +
( z&1 z& 2 ) .
m2
m2
(4)
&z&2 =
Further defining the states as x1= z1, x2= z2, x3= dz1/dt, x4= dz2/dt, and the state vector as x = [ x1, x2, x3, x4 ]T, one
can then express the equations of motion of this model with the following state space equations:
921
x&1 = x3
x& 2 = x 4
w ( w x1 ) a 2 g
k
c
( x3 x 4 ) +
x& 3 = ( x1 x 2 ) +
m1
m1
m1
k
c
( x1 x 2 ) +
( x3 x 4 )
x& 3 =
m2
m2
(5)
Finally, the state equations can be expressed in vector differential equation form as follows:
x& = Ax + Bw
y = Cx
(6)
where the system matrix A, the input matrix B, and the measurement matrix C are described as follows:
k + w a2 g
A=
m1
m2
B = 0 0
0
0
k
m1
k
m2
wa2g
m1
0
1
c
m1
c
m2
1
0
c
m1
c
m2
C = I 44
5
(7)
(8)
(9)
A classical (or simple) Kalman filter is an optimal state observer that recursively estimates the internal states of a
linear dynamical system from as series of noisy measurements. Internal states of a dynamical system is often not
fully measurable; in this case Kalman prediction is often used to get the best estimate of the system states given
the available measurements. With this background in mind, we aim to implement a Kalman filter to estimate
wave features with the view to optimize PTO.
For this purpose we first need to discretize the equations of motion of the second order vertical oscillator model.
Denoting the sampling interval for discretization as t the discrete time state space equations can be expressed as
follows
y ( k ) =C d x ( k )
(10)
where assuming t to be small, the discrete time system matrices Ad,Bd,Cd are given as below
Ad ( I + At )
B d Bt
Cd= C
(11)
Further assuming uncorrelated white noise processes (k) and (k) acting respectively on the state and
measurement equations, the discrete time system can be expressed as
922
x( k + 1) = Ad x(k ) + B d w(k ) + ( k )
y ( k ) =C d x ( k ) + ( k )
(12)
where (k) and (k) are zero-mean Gaussian random processes (i.e., white noise). These white noise processed
are also denoted as (k)~N(0,Q) and (k)~N(0,R), where Q and R are state and measurement error covariance
matrices with compatible dimensions. For the specific system described by (12), the discrete time Kalman filter
is given by the following recursive prediction and update equations:
Prediction Equations:
(13)
P ( k ) = Ad P ( k 1) AdT + Q
(14)
~
y (k ) = y (k ) C d x (k )
(15)
Update eqautions:
(16)
(17)
(18)
Conclusions
In this paper we analyzed measured wave data from a particular site in the US Pacific coast. A generic wave
generator model is derived and then presented in the form of a Kalman filter for predicting the wave state. The
estimated states for the Kalman filter shall be used for maximizing the amount of pumped hydraulic fluid as a
result of damper arm displacement in the scope of the wave generator model. Analysis of this and the
corresponding numerical implementation shall be an immediate future work for this study.
Acknowledgements
The authors are indebted to Coastal Data Information Program (CDIP), USA website for the wave data. Mr.
Bugra Cakmak was a great help in automating the data download with Matlab.
References
[1]
M.S.Triantafyllou, M.Bodson,and M.Athans (1983). Real-time estimation of ship motions using kalman filtering techniques.
J. P. Pinto, M. C. Bernadino, and A. Pires Silva (2005). A Kalman filter application to a spectral wave model. Nonlinear
923
Jrgen Hals, Johannes Falnes, and Torgeir Moan (2011). Con- strained optimal control of a heaving buoy wave-energy
F. Fusco and J.V. Ringwood (2010). Short-term wave forecasting for real-time control of wave energy converters. Sustainable
J. Capon (1969). High-resolution frequency-wave number spectrum analysis. Proceedings IEEE, 57:14081418.
T.T.Georgiou (2001).Spectral estimation via selective harmonic amplification. IEEE Transactions on Automatic Control,
46:2942, 2001.
[8]
C. Foias and A. E. Frazho (1988). A geometric approach to the maximum likelihood spectral estimator for sinusoids in noise.
S. L. Marple (1987). Digital Spectral Analysis with Applications. Prentice-Hall, Englewood Cliffs.
[10] B. Yagci and P. Wegener (2009). Determination of sea conditions for wave energy conversion by spectral analysis. In
Proceedings of the 8th European Wave and Tidal Energy Conference, Uppsala, Sweden.
Biographies
Dr. Selim Solmaz Dr. Solmaz was born in Izmir, Turkey in 1978. He received his BSc (High Honours) degree in
aeronautical engineering from Middle East Technical University, Ankara, Turkey in 2001 and MSc from Purdue University at
West Lafayette, School of Aeronautics and Astronautics in 2003. Dr. Solmaz received his PhD in 2007 from Hamilton
Institute, National University of Ireland-Maynooth. As part of his PhD project he spent a year as a visiting researcher at
Daimler Chrysler Research AG in Esslingen, Germany in 2003-2004. Upon his return to Ireland he conducted research work
on theoretical and applied problems arising from automotive vehicle dynamics estimation and control.
Dr. Solmaz was employed as a Postdoctoral Research Fellow at the Hamilton Institute, NUIM for during 2007-2010. During
this time he secured and completed two independent EU research grants (co-funded by Enterprise Ireland) with him as the
Principal Investigator (PI).
Dr. Solmaz is currently with Gediz University, Mechnanical Engineering Department, where he is employed as an Assistant
Professor. His current research interest includes vehicle dynamics control, vehicle state and parameter estimation, stability
theory, switched systems, mechatronics, robotics, renewable energy generation, and human-machine interfaces. Dr. Solmaz
has authored and co-authored about 30 peer reviewed journal and conference articles, 2 research project reports and 2 patents.
He is a past member of IEEE and AIAA.
Dr. Bar Yac Mr. Yac is an applied expert in the field of estimation and system control. In his career, he has applied
estimation in structural vibrations and energy conversion from renewable sources. More recently, he has been working on
dynamic system modeling and automation. He holds an MSc from Middle East Technical University, Ankara, Turkey and a
PhDfrom Purdue University, West Lafayette, USA.
924
ZF Lemfrder Aks Modlleri San. Tic. A.. AOSB 10003 Sok. No:13 ili, zmir, Turkey
musa.ozcelik@zf.com.tr, 2kuralay@deu.edu.tr
Abstract.A double wishbone suspension system is applied on non-driven rear axle of passenger car. ADAMS
(Automatic Dynamic Analysis of Mechanical Systems) software is utilized to perform the multi body
dynamic analysis aim to calculate force and moment on joint connections of sub-component according to the
thirteen different load cases which are derived from the standard driving maneuvers. In order to correlate
force and moment results of ADAMS software the vector theory is also performed on double wishbone
suspension system. Sequentially the strength analysis performed by linear material behavior using Nastran
software. Results of the strength analysis are set to the upper limit of the topology optimization response. The
topology optimization is performed by finite element method using Optistruct software. Optistruct responses
effective designs for optimum part geometry then new design iterations are confirmed by linear durability
analysis. As a result, the upper control arm component of double wishbone suspension system is redesigned
according to the optimization responses so mass of control arm is decreased forty five percent due to first
iteration
Keywords: Multibody dynamics, finite element analysis, control arm, topology optimization.
1 Introduction
The suspension system comprises the interface between the vehicle body and the road surface [4]. Rigid axles
have been using in passenger cars, light and heavy commercial vehicles since early eras of automotive industry.
Developing of automotive industry caused using of independent suspension systems firstly in non-driven axles
secondly in driven axle. The chassis of passenger car must be able to handle the engine power installed. Ever
improving acceleration, higher peak and cornering speed, and deceleration lead to significantly increased
requirements for safer chassis [2]. Not only independent suspension systems have to be resistant but also
independent suspension has to be lighter. Nevertheless independent suspension systems are prevalent resulting
from advantages like, low weight, little space requirement and no mutual wheel influence. Low weight is
important for good road-holding, especially on bends with uneven road surface [2].
Un-sprung mass should be optimum which has influence on road-holding and cause production cost optimum.
Nowadays one of the main targets of the automotive companies is producing best cost vehicle in order to be
competitive for market. In addition lighter vehicles are needed for decreasing fuel consumption.
Suspension system components should be designed lighter enough at the same time not to cause stiffness,
strength and fatigue life decreases. By trying different designs and regions to aim of the optimum design of a
925
part could increase development process therefore optimization software usage become widespread with the aim
of decreasing development time. In this study MSC ADAMSTM software is utilized to calculate kinematic
point forces and moments and Altair OptistructTM software is utilized to perform topology optimizations.
F f = m.g (
lf
l
hc a
. )
l g
(1)
The lateral acceleration generated by centrifugal effect of cornering causes lateral dynamic wheel loads
transfers. These load transfers are added to outer wheels and subtracted from inner wheels. Some dimensional
parameters are shown in figure 2. Additionally front and rear roll stiffness are used for load transfer calculations;
1030N.m/deg for front roll center and 920N.m/deg for rear roll center.
926
The following formulations are applied on vehicle model to calculate wheel load transfers. FRzM and FRzL
are generated by the effect of the lateral acceleration of vehicle mass. Total wheel load transfer can be calculated
as FR . Finally this load should be used by subtracting from one side and adding to other side [1].
K Tf
FRzM = m.a.h
K +K
Tr
Tf
t
r
(2)
lf
FRzL = m.a
l +l
f r
hcr
t
r
FR = FRzL + FRzM
(3)
(4)
927
defined as non-rigid properties. It is important that this stiffness is defined between rigid arms. Finally assembly
of suspension system is created by assembling sub-systems so parameters related to whole vehicle are defined
like, toe angle, camber angle, drive shaft offset, wheel radius, sprung mass and wheel base.
Fig. 2. Dynamic wheel loads are applied on contact point or wheel center to the ADAMS suspension model.
X
Y
Z
928
Fig. 3. Optimization analysis path start with current design and finally redesigned according to the optimization software
responses.
x1
x
X = 2 x design space which minimize f ( X )
M
x n
(5)
g j(X ) 0 ,
j = 1,2,..., m
(6)
f j(X ) = 0 ,
j = 1,2,..., p
(7)
929
- Constraints
- Design variable
In order to perform topology optimization analyses Altair Optistruct software is utilized with 4 nodded
tetrahedron elements. Calculation of optimization spends much more time then linear analysis to calculate all
iteration. After having feasible design software calculates an optimization response about optimum design.
Optimization response can export as iso surfaces of optimized model so it can be used in computer aided design
(CAD) software for designing new geometry.
New finite element model has created according to new CAD model. Then same process performed for
optimum design. After analyzing 13 load cases results should checked with first design.
268
Fig. 4. Optimization response created by Optistruct (left). Some non-used regions are deleted and important walls are used
for increasing stiffness. Structural analysis of optimum design (right). Comparison between first design and optimum design
according to von Misses stress shows that optimum design has less stress then first design and also 45% lighter.
5 Conclusions
The mass of upper control arm which is used in passenger car's non-driven rear suspension system decreased
about 45% due to first iteration without any stress increase. To this end topology optimization performed by
using Optistruct software. This software created design spaces related to 13 different load cases which not
exceed stress limitation. Thus for only one component 45% lighter and economical geometry designed by this
study.
References
1.
2.
3.
4.
5.
Blundell M. and Harty D. (2004) The Multibody Systems Approach to Vehicle Dynamics (1st ed.). London:
Elsevier Butterworth-Heinemann.
Reimpell J., Helmut S. and Jrgen W. B. (2001) The Automotive Chassis : Engineering Principles (2nd ed.).
Oxford: Butterworth - Heinemann.
Martin P. Bendsoe (1995) Optimization of Structural Topology, Shape and Material (1st ed.). Berlin: SpringerVerlag.
Ersoy M. and Heissing B. (2011) Chassis Handbook (2nd ed.). Berlin: Springer Fachmedien.
Rao S.S. (1996) Engineering Optimization (3rd ed.). Indiana: A Wiley-Interscience Publication.
930
Biographies
Musa zelik The author was born in Akehir in 1985. He was graduated from Dokuz Eyll University in 2008 as a
Mechanical Engineer and still continues Master of Science education in Design and Production Department of Dokuz Eyll
University.
He studied on heavy commercial vehicle steering systems in bachelors level and Master of Science thesis is about vehicle
dynamics and optimization of suspension system parts.
N. Sefa KURALAY was born in Afyon in 1953. He received his B.Sc. in mechanical engineering from Ege University,
Izmir. He also earned a Ph.D. in mechanical engineering from Hannover University, Germany. He has been at Dokuz Eyll
University since 1986, when he joined as an assistant professor in mechanical engineering. From 1986 to 1988 he continued
as an associate professor. Since 1994 he has been full professor for mechanical engineering.
His current research interests include vehicle dynamics and simulation, 3-D kinematic analysis and design of vehicle
suspensions. He is also interested in reconstruction of traffic accidents and machine design.
Dr. Kuralay is a member of UCTEA Chamber of Mechanical Engineers.
931
Abstract. In this work, cold flow simulations of a six cylinder nine lt diesel engine have been performed to
find out flow development and turbulence generation in the piston-cylinder assembly. Various turbulence
models have also been tested and results have been compared. In this study, the goal is to understand the flow
field and the combustion process in order to be able to suggest some improvements on the in-cylinder design
of an engine. Therefore cold flow simulations of a six-cylinder nine-liter diesel engine have been performed to
find out flow development and turbulence generation in the piston-cylinder assembly. Moreover, the
interaction of air motion with high-pressure fuel spray injected directly into the cylinder has also been carried
out. A Lagrangian-drop Eulerian-fluid based model has been applied to the in-cylinder spray-air motion
interaction in a heavy-duty CI engine under direct injection conditions. A comprehensive model for
atomization of liquid sprays under high injection pressures has been employed.
Further investigation has also been performed in-cylinder design parameters in a DI diesel engine that result in
low emissions by effect of high turbulence level. The results are widely in agreement qualitatively with the
previous experimental and computational studies in the literature.
Study is in progress to suggest changes about in-cylinder design parameters in a DI diesel engine that results in
low emissions by the effect of high turbulence level.
Keywords: CFD, Turbulence models, Diesel engine.
1 Introduction
The intake and compression strokes are very important processes which influence the pattern of air flow
structure incoming cylinder during intake stroke and generate the condition needed for the fuel injection during
the compression stroke.
Recent research shows that it is possible to decrease the emissions considerably by modifying the geometrical
sub-systems of the engine that affect the turbulence generation and spray formation.
In direct-injection Diesel engines, near TDC at the end of the compression stroke inside the cylinder the flow
conditions are critical for the power stroke. Studies to find out the conditions just before the power stroke
determines the mixture and combustion efficiency which is crucially important for an engine.
The purpose of this study is to find out in-cylinder flow characteristics of a six-cylinder, nine-liter CI engine
before the fuel injection and also properly simulate the atomization process in order to improve the combustion
efficiency and emission parameters. A complete intake and compression stroke, also fuel injection at the end of
the compression stroke simulated in 3D. Both naturally aspirated and turbo-charged modes of operations were
studied.
932
2 LITERATURE REVIEW
Fluid flows in engines are always turbulent and are highly anisotropic due to complex geometry, wall effects,
flow rotation (swirl), internal separation, and so on. For IC engine flow, additional complications are involved
because of the rapid compression and expansion strokes produced by the piston motion. Among these flow
phenomena, swirl, squish and tumble motions are essential and desirable engine design features. For directinjection IC engines, these motions are used to enhance the airfuel mixing for emission control and better
efficiency. Moreover, Reynolds or Ensemble-averaged Navier-Stokes equations yield an unknown Reynolds
stress tensor with its six components which necessitates the use of a turbulence model.
Therefore, to properly simulate flow field inside various types of engines, flow characteristic needs to be
resolved by a robust and versatile turbulence model, without having to perform ad hoc modifications for each
engine geometry.
Two-equation, k- models are commonly used for engineering predictions of turbulent flows which provide
computational expediency and ease of implementation in finite-volume numerical codes. Their accuracy is
generally acceptable only for prediction of mean flow quantities. In particular, the standard k- model of
Launder and Spalding [1-3] has been criticized by many researchers as being inaccurate for prediction of recirculating flows and complex shear layers, due to its over dissipative nature.
In an attempt to improve the predictive ability of the standard k- models, a number of alternatives have been
offered. Among them the RNG k- model [4, 5], anisotropic k- model of Speziale [6], Morel & Mansour
version of the k- model [7], and the k model of Wilcox [8] are well known. Gul [9], has also offered a threeequation turbulence model in order to predict the compressed turbulence in IC engines.
In this work various turbulence models including the Chens k- model [10] have been tested and the RNG k-
model has been chosen for the rest of the simulations.
3 METHODS
Real engine geometry of FORD 9-litre six-cylinder diesel engine were modeled for a single cylinder together
with exhaust and inlet manifolds. Simulations were extended for both ports to get proper boundary conditions at
the valve gaps.
3-D CFD codes, 1-D engine codes and multi-objective optimization codes have been utilized to reach an inside
into the in-cylinder flows and optimize the design of piston-cylinder assembly to achieve low level of emissions.
Although computer requirements for such a simulation are quite high, recent computer technology allows us to
solve unsteady flow problems in a complex domain with improved memory and parallel processing facilities.
Surface meshes were created for the CAD model of the solid geometry after extracting the occupied volume by
the fluid. Then, piston and valve motions were defined in the code. After specifying the discretization scheme
and the algorithm, turbulence models were determined.
The program is based on pressure-correction method and uses the PISO algorithm [1, 2]. For the momentum and
turbulence equations Monotonic Advection and Reconstruction Scheme (MARS) have been employed, for the
energy first-order upwind differencing (UD) and for the density, central differencing (CD) schemes have been
used.
Fine and course grids were tested with 1,000,000 and 60,000 elements respectively. Fine grid results are
presented in the paper. Typical execution time for 3900CA with fine meshes takes about 40 hours on a
workstation with four processors running in parallel.
933
4 FINDINGS-CONCLUSION
In this study four turbulence models were examined for a single cylinder of heavy duty diesel engine. In this part
it was examined that the effects of in-cylinder flow capabilities especially cold flow with respect to the different
turbulence models. For each configuration, these four turbulence models were assessed (k-/High Reynolds, k/RNG, k-/Chen, k-/Speziale/High Reynolds and k-/Standart/High Reynolds models). k-/RNG model has
best results for this simulation.
934
935
936
Figure 6.Vector fields at different crank angles during the compression stroke
REFERENCES
1.
Issa, R.I. 1986. Solution of the implicitly discretised fluid flow equations by operator-splitting, J.
Issa, R.I., Gosman, A.D., and Watkins, A.P. 1986. The computation of compressible and
incompressible recirculating flows by a non-iterative implicit scheme, J. Comp. Phys., 62, pp. 6682.
3.
Launder, B.E., and Spalding, D.B. 1974. The numerical computation of turbulent flows, Comp.
Yakhot, V., and Orszag, S.A. 1986. Renormalization group analysis of turbulence-I: Basic theory, J.
Yakhot, V., Orszag, S.A., Thangam, S., Gatski, T.B., and Speziale, C.G. 1992. Development of
turbulence models for shear flows by a double expansion technique, Phys. Fluids, A4(7), pp. 15101520.
6.
Speziale, C. G. 1987. On nonlinear k-l and k- models of turbulence, J. Fluid Mech., 178, pp. 459-475.
7.
Technical Paper Series, 820040,International Congressand Exposition, Detroit, Mich., February 22-26,1982.
8.
Wilcox, D.C. 1998. Turbulence Modelling for CFD.2nd edition, DCW Industries, Inc.
9.
M.Z. Gl, Prediction of In-Cylinder Flow By Use of a Multiple Time Scale Turbulence Models,
Chen, Y.S., and Kim, S.W. 1987. Computation of turbulent flows using an extended k- turbulence
937
Biographies
Hasan KTEN (1983) is a research assistant at CFD in mechanical engineering department of Marmara University. Hasan
Kten obtained his BSc with honor of degree in mechanical engineering also graduated with double major in electricalelectronical engineering, Sakarya University. Hasan Kten obtained MSc Degrees from Marmara University, Istanbul,
Turkey, and his doctorate program continues in mechanical engineering department, Marmara University. His MSc thesis
discussed combustion models within a multidimensional framework applied to heavy duty compression ignition engines. His
research areas include CFD, combustion, controlling injection systems and optimization.
Mustafa YILMAZ is an Assistant Professor of mechanical engineering at the Marmara University Istanbul and is the Head
of Studies for the study area of CFD, internal combustion engine (ice) and numerical methods at the mechanical engineering
department, Marmara University, Turkey. Mustafa Yilmaz holds PhD degree in mechanical engineering from Marmara
University, Turkey. He has extensive experience in CFD and ice issues. His area of research interest is performance ice
engines and CFD, with a particular focus on the automotive industry.
M. Zafer GL is a professor and head of mechanical engineering department at the Marmara University Istanbul and is the
head of studies for the study area of thermo-fluids, CFD, internal combustion engine (ice) and at the mechanical engineering
department, Marmara University, Turkey. Gul holds MSc from Istanbul Technical University and a PhD degree in
mechanical engineering from University of Manchester Victoria, UK. He has extensive experience in CFD and ice issues,
both as an academic and as a consultant to international companies. His area of research interest is performance ice engines
and CFD, with a particular focus on the automotive industry.
938
kamilozdin@hitit.edu.tr
Abstract.In this study, an analytical elasticplastic residual stress analysis is carried out on thermoplastic composite beams
of arbitrary orientation, supported from two ends under a transverse uniformly distributed load. The composite layer consists
of stainless steel fiber and thermoplastic matrix (LDPE, F.2.12). The material is assumed to be perfectly plastic during the
elasticplastic solution. The intensity of the uniform force is chosen as a small value; therefore, the normal stress component
yis neglected in the elasticplastic solution. The expansion of the plastic region and the residual stress component of xare
determined for orientation angles of 0,15, 30 and 45. The intensity of the residual stress component of xis found to be
maximum at the lower and upper surfaces; however, the intensity of the residual stress component of xyis a maximum on or
around the x axis of the beam.
Keywords: Thermoplastic composites; residual stress analysis, simply supported beam, transverse uniformly distributed
load, analytical solution.
1. Introduction
Among fiber-reinforced composites, thermoplastic matrix composites are gaining popularity due to their
many advantages. In relation to thermoset composites systems, they offer improved interlaminar fracture
toughness, increased impact resistance, and higher solvent resistance. In addition to their competitive mechanical
properties, thermoplastic composites do not require complex chemical reactions to be processed and can be
formed without lengthy curing process. These features offer engineers a potential for cost-effective forming
processes and the capability of integrating material processing and structure design into a single step. Residual
stresses in the matrix are particularly important because they may lead to premature failure. Prediction and
measurement of residual stresses are therefore important in relation to production, design and performance of
composite components. Jeronimidis and Parkyn [1] obtained residual stresses in carbon fibre-thermoplastic
matrix (ICI plc.) laminates. Structural polymer composite materials used for aerospace and automotive
applications may exhibit time-dependent creep or relaxation behavior because of the viscoelastic behavior of
polymer matrix and/or fiber materials. It has been found that creep deformation shows similarities to plastic
deformation. Chung et al. [2] used the same potential function to describe plastic flow and time-dependent creep
deformation in AS4/PEEK thermoplastic composites. Domb and Hansen [3] developed a numerical model for
prediction of the process-induced thermal residual stresses in thermoplastic composite laminates.
Residual stresses in the matrix are particularly important because they may lead to premature failure.
Prediction and measurement of residual stresses are therefore important in relation to production, design and
939
performance of composite components. Akay and zden [4] measured the thermal residual stresses in injection
moulded thermoplastics by removing thin layers from specimens and measuring the resultant curvature or the
bending moment in the remainder of the specimens. Akay and zden [5] investigated the influence of residual
stresses on the mechanical and thermal properties of injection moulded ABS copolymer and polycarbonate.
The finite element technique gives an excellent elastic-plastic stress analysis in composite structures [68]. Aykul and Kaya [9] investigated the elasticplastic analysis of thermoplastic composite cantilever beams.
In the present study, an elastic- plastic stress analysis is carried on simply supported steel fiber
reinforced thermoplastic composite beams under uniformly distributed load. The distribution of the residual
stress of x is determined along the sections of the beams for 0, 15, 30 and 45 orientation angles. The
transverse stress component y is neglected during the solution since the magnitude of the uniform distributed
force is chosen to be small. The composite material is assumed to be perfectly plastic for the simplicity of the
solution.
2. Elastic Solution
The composite simply supported beam is subjected to a transverse uniformly distributed load q, as
shown in Fig. 1. The angle between the first principal axis of the composite fibers and the x axis is . For the
plane-stress case the equation of equilibrium is given as [10],
a 22
4F
x 4
2 a 26
+ ( 2 a 12 + a 66 )
4F
x 3y
4F
x 2y 2
2 a 16
4F
x y 3
+ a 11
4F
y 4
= 0
(1)
where F is a stress function. The constants in Eq. (1) are given by Jones [11] as,
x a 11 a 12 a 16 x
y = a 12 a 22 a 26 y
a a
a 66 xy
xy 16 26
(2)
where:
a 11 = a 11 cos
a 12 = a 12 (sin
a 22 = a 11 sin
+ ( 2 a 12 + a 66 ) sin
4
4
+ cos
a 16 = ( 2 a 11 a 12 a 66 ) sin cos
a 26 = ( 2 a 11 a 12 a 66 ) sin
+ a 22 sin
cos
cos
+ a 22 cos
(3)
( 2 a 22 a 12 a 66 ) sin
cos ( 2 a 22 2 a 12 a 66 ) cos
a 66 = 2 ( 2 a 11 + a 22 4 a 12 a 66 ) sin
a 11 =
) + ( a 11 + a 22 a 66 ) sin
+ ( 2 a 12 + a 66 ) sin
cos
cos
+ a 66 (sin
cos
3
sin
+ cos
1
1
1
; a 12 = 12 = 21 ; a 22 =
; a 66 =
E1
E1
E2
E2
G 12
(4)
F is chosen as a fifth order polynomial in order to satisfy the governing differential equation,
F=
x 2y3
y5
xy 4
x2y
xy 2
y3
x2
d5 +
f5 +
e5 +
b3 +
c3 +
d3 +
a2
6
20
12
2
2
6
2
(5)
e5 =
2 a 16
d 5 = md
a 11
(6)
940
2 a 12 a 66 + 2 a 16 m
d
3 a 11
= nd
;m =
2 a 12 a 66 + 2 a 16 m
2 a 16
;n =
a 11
3 a 11
(7)
= 0; y = c
= q / t; y = m c
x = L
(8)
= 0
xy
xy
= qL ; x = L
tdy
xy
tdy
(9)
= qL
x = mL
tdy = 0 ; x = m L
(10)
= 0
tydy
2F
= x 2 yd
y 2
+ y
f 5 + xy
xy =
2F
xy
e 5 + xc
+ yd
= xy 2 d 5
y3
2F
d
=
2
3
x
+ yb
+ a2
y3
e 5 xb 3 yc 3
3
(11)
(12)
3q
3 qL 2
9 qn
qm
q
3q
3 qm
3 qn
, d5 =
,
,
,
,
,
b
e
f
c
d
=
=
=
=
=
+
3
5
5
3
3
2t
4 tc
4 tc
20 tc
4 tc 3
4 tc 3
4 tc 3
4 tc 3
(13)
qm
3q
L2 x 2 y +
4 tc 3
4 tc
xy
3y2
1 2
c
3q
xy
4 tc 3
3 qn
x
4t
y3
3y
3
5c
c
qm
3q
y3
4 tc 3
4 tc
q
3q
q
y
+
4 tc 3
4 tc
2t
qm
4 tc
(14)
(15)
3. Elasticplastic solution
The equations of equilibrium for the plane-stress case,neglecting the body force, are given as,
xy
0 ;
xy
(16)
If the length of the beam is very large in comparison with its height and q is chosen to be a small value,
y can be neglected in comparison with x and xy. Steel fiber reinforced thermoplastic composites have the
same yield points in tension and compression; therefore the TsaiHill theory is used as a yield criterion in this
study. The equivalent stress in the first principal material direction is given as,
2
1
X
Y
2
2
2
2
X
S
2
2
941
2
12
= X
(17)
where X, Y are the yield points in the 1st and 2nd principal material directions, respectively and S is the
shear yield strength in the 12 plane. It is assumed that the yield strength (Z) in the 3rd direction is equal to Y
and yield points of pure shear in the 13, and 23 planes are equal to S. The stress components in the principal
material directions are written as,
1 = X cos 2 + 2 xy sin cos
(18)
Substituting
1, 2
and
12
y =0
= f ( y)
xy
x
=
0 .
edges the shear stress is equal to zero and, for a perfectly plastic material, x can be taken as a constant C at the
upper and lower surfaces, where C is the yield point of the composite in the x direction. When the plastic zone is
expanded for a perfectly plastic material,
(17) gives the yield point in the x direction for the orientation angle
xy
,
(19)
X1 =
cos
sin
cos
sin
Y 2
sin
cos
In order to satisfy both the governing differential equation and the boundary conditions in the elastic
region of the elastic-plastic part, the stress function F is chosen as,
F =
(20)
d5 2 3
f
g
b
a
x y + 5 y5 + 3 y3 + 3 x2 y + 2 y2
6
20
6
2
2
Substituting the stress function into the governing differential equation gives the following relation,
(2 a
12
+ a 66 2 yd
2 a 12 + a 66
+ 6 a 11 yf 5 = 0 f 5 =
3 a 11
d 5 = rd
(21)
2F
y 2
= d
y + f5 y
xy =
2F
xy
+ g
y + a
+ d
y;
= xy 2 d 5 xb 3
942
2F
x2
= 0
(22)
(23)
y = h1
xy
= 0; y = h2
xy
= 0
(24)
h2
xy
(25)
tdy = qx
h1
The resultant of
; y
(26)
X 1 t (c h1 ) + X 1 t (c h 2 ) +
X 1 t (c h1 )
x t dy
= 0
(27)
c + h1
c + h2
+ X 1 t (c h 2 )
+
2
2
h1
L2 x 2
x tydy = q
(28)
From the boundary conditions, the unknown parameters are determined as,
X
h1 = h
a 2 = 0 , b 3 = d 5 h1 ,
L2 X
+ q
2t
X 1
qr
+
3
5t
d5 =
4 th 1
(29)
X
2
, g 3 = 1 d 5 x 2 rd 5 h 1
/3
h1
(30)
F = d5
x2y2
y5
xy 4 3 xy 2
xy 2
y3
x2
y2
+ f5
+ e5
b
+ c3
+ d3
+ a2
+ b2
6
12
2
2
6
2
2
20
(31)
Substituting the derivative values into Eq. (1), gives the following equation,
[(2 a 12
+ a
66
)2
4 a 16 e
+ 6 a 11 f
]y
4 a 16 d
+ 2 a 11 e
]x
= 0
In order to satisfy the equation, each term of x and y must be equal to zero. Hence,
e5 =
2 a 16
d
a 11
= md
; f
2 a 1 2 a 66 + 2 a 16 m
= nd
3 a 11
943
(32)
2y
= d 5 x 2 y + f5 y
y2
2 f
= d 5 xy
x y
xy
+ e 5 xy
+ c3 x + d 3 y + b2
(33)
e5 3
y b3 x c 3 y
3
(34)
If Eq. (32) is put into Eqs. (33) and (34), the stress components are found as,
= d5x
y + nd
= d 5 xy
xy
md
3
+ md
xy
+ c3x + d 3 y + b2
(35)
b3 x c3 y
The parameters d5, b3, c3, b2, h1, h2 are determined from the boundary conditions. Using the boundary conditions in
Eqs. (24)(25), they are obtained as,
c 3 = d 5 x (h 2 h 1 )
d5 =
md
3
(h
2
1
x
(h 1 + h 2
6
h1 h 2 + h 2
)3
); b
qx
t
m
+
(h 1 + h 2
12
= d 5 h1 h 2 +
md 5
(h 1 h 2 )h 1 h 2
3x
(36)
(37)
)3 ( h 2
h1 )
From the boundary conditions in Eqs. (26)(28) the other parameters are determined as,
d5 =
X 1 (h1 h2 )
(h h2 ) d x 2 + h1 h2 nd h 2 + h 2 md x
; b2 = 1
1
5
5
2
5
3
3
2
2
n (h1 h2 )(h1 + h2 ) mx (h1 + h2 )
4
6
d3 =
2X1
2
d 5 x 2 h 2 1 h1 h 2 + h 2 nd 5 + ( h1 h 2 ) md 5 x
h1 + h 2
(38)
(39)
12 qx
t [2 x + m (h 2 h 1
)]
(40)
u =
2 xX 1 t + 3 qxn +
( 2 xX
t + 3 qxn
2 X 1 mt
1
)+ 8 X
m 2 tqx
; h 2 = h1 + u
944
(41)
nd s
(2h1 + u ) 8(h1 + u )4 + 8h1 4 + 2h13 + (h1 + u ) + 2h1 (h1 + u )3 12h1 2 (h1 + u )2
60
2
md 5
2
X
(2h1 + u )3 u X 1c 2 + 1 h1 2 + h1u + u 2 + q L x = 0
+
12
3
2t
(42)
Finding h1 by using the NewtonRaphson method gives the other unknown parameters.
The magnitude of the residual stress component of x are shown in Fig. 6 and 7 for 30 and 45 orientation
angles, respectively. For both cases, It is the greatest at the middle section and becomes smaller at the sections
futher away from the middle section. As shown in the figures, when the orientation angle is further increased, the
residual stresses component decreases.
5. Conclusions
The results given below are concluded in this investigation:
1. Plastic yielding begins earlier at the lower surface for 30o and 45o orientation angles, whereas; it begins earlier at the upper
surface for 0o and 15o orientation angles.
2. The magnitude of the residual stress component of
x.
5. The beam length must be chosen so as to be upper and lower limits.
945
E2
G12
(MPa)
(MPa)
(MPa)
38000
1300
360
12
0.25
Axial Strength
Transverse Strength
Shear Strength
(X)(MPa)
(Y) (MPa)
(S) (MPa)
37
13
Table 2. The half length of the beam for the plastic collapsing
Orientation Angles
0
15
30
45
Lower sides
115
103
71
52.5
Beam L (mm)
Upper sides
95
94
76
57.5
15
30
45
X1 (MPa)
37.0
29.62
16.04
9.33
100
95
75
55
h1
h2
(mm)
(mm)
(mm)
x p
x e
x r
( MPa)
(
x p
x e
x r
xy r
upper
upper
upper
lower
lower
lower
on the
surface
surface
surface
surface
surface
surface
x axis
4.91
4.91
-37.00
-21.49
-15.50
37.00
21.49
15.50
0.025
4.93
4.93
-37.00
-21.39
-15.60
37.00
21.39
15.60
-0.027
10
4.99
4.99
-37.00
-21.06
-15.93
37.00
21.06
15.93
-0.091
15
5.08
5.08
-37.00
-20.52
-16.47
37.00
20.52
16.47
-0.175
20
5.20
5.20
-37.00
-19.76
-17.23
37.00
19.76
17.23
-0.285
25
5.35
5.35
-37.00
-18.78
-18.21
37.00
18.78
18.21
-0.422
4.04
4.04
-29.62
-21.25
-8.36
29.62
21.25
8.36
0.025
4.07
4.27
-29.62
-21.34
-8.27
29.62
21.34
8.27
0.114
10
4.16
4.57
-29.62
-21.21
-8.40
29.62
21.21
8.40
0.160
15
4.31
4.91
-29.62
-20.87
-8.74
29.62
20.87
8.74
0.136
20
4.51
5.31
-29.62
-20.31
-9.30
29.62
20.31
9.30
0.034
25
4.75
5.76
-29.62
-19.53
-10.08
29.62
19.53
10.08
-0.133
1.68
1.68
-16.04
-15.59
-0.45
16.04
15.59
0.45
0.025
1.81
1.97
-16.04
-15.60
-0.44
16.04
15.60
0.44
3.991
10
2.16
2.48
-16.04
-15.39
-0.64
16.04
15.39
0.64
4.384
15
2.65
3.12
-16.04
-14.97
-1.07
16.04
14.97
1.07
3.146
20
3.20
3.84
-16.04
-14.33
-1.71
16.04
14.33
1.71
1.859
25
3.80
4.60
-16.04
-13.47
-2.56
16.04
13.47
2.56
0.862
3.28
3.28
-9.33
-15.61
6.27
9.33
15.61
-6.27
0.025
0o
15o
30o
45o
946
3.40
3.40
-9.33
-15.50
6.16
9.33
15.50
-6.16
0.364
10
3.74
3.74
-9.33
-15.17
5.84
9.33
15.17
-5.84
0.405
15
4.24
4.24
-9.33
-14.63
5.30
9.33
14.63
-5.30
0.179
20
4.86
4.86
-9.33
-13.87
4.54
9.33
13.87
-4.54
-0.152
25
5.55
5.55
-9.33
-12.89
3.56
9.33
12.89
-3.56
-0.488
Fig. 1. Simply supported composite beam under a transverse uniformly distributed load
947
References
1.
Jeronimidis G, Parkyn AT. Residual stresses in carbon fibre-thermoplastic matrix laminates. J Compos Mater 1998;
22(5): 401-15.
2. Chung I, Sun CT, Chang IY. Modelling creep in thermoplastic composites. J Compos Mater 1993; 27(10): 100929.
3. Domb MM, Hansen JS. The effect of cooling rate on free-edge stress development in semi-crystalline
thermoplastic laminates. J Compos Mater 1998; 32(4): 361-85.
4. Akay M, zden S. 1994. Measurement of residual stresses in injection moulded thermoplastics. Polym Test 1994;
13: 323-54.
5. Akay M, zden S. Influence of residual stresses on mechanical and thermal properties of injection moulded
polycarbonate. Plast Rubber Compos Process and Appl 1996; 25(3): 138-44.
6. Owen DRJ, Figueiras JA. Anisotropic elasto-plastic finite element analysis of thick and thin plates and shells. Int J
Numer Methods Eng 1983; 19: 541-66.
7. zcan R. Elasto-plastic stress analysis in thermoplastic composite laminated plates under in-plane loading. Compos
Struct 2000; 49: 201-8.
8. Aykul H, Ata C, Akbulut H. Elastic-Plastic Stress Analysis and Plastic Zone Propagation in Fiber Reinforced
Thermoplastic Laminates Using Finite Element Method. J Reinf Plast Compos 2001; 20(15): 1301-25
9. Aykul, H., Kaya, S. An elastic-plastic and residual stress analysis steel reinforced hermoplastic composite
cantilever beam, Mater Design 2010; 31: 3474-3481.
10. Lekhnitskii SG. Anisotropic plates theory. Gordon and Breach Science Publishers; 1968
11. Jones RM. Mechanics of composite materials. Taylor & Francis, 1999
948
Chapter 5
FUZZY LOGIC
949
Abstract:This paper considersdesign of a stable Takagi-Sugeno fuzzy control system via linear matrix
inequalities (LMIs) with constraint on the input/output & initial state independence. Firstly, the necessary
condition for the closed loop systems stability was given using LMIs. In addition to this condition the
following three theorems were given concerning input & output constraint and initial state independence.
Using these theorems, fuzzy control algorithm was designed for a concrete plant. The model was then
implemented in MATLAB/Simulink and LMI Toolbox.
Key words: T-S fuzzy model, LMIs, PDC, I/O constraints, initial state independence.
Introduction
Design of a fuzzy control system can be realized in two different methods, namely Mamdani models and TakagiSugeno (T-S) fuzzy models. Rule-based fuzzy systems include Mamdani models considered in [1] and [2], and
model-based fuzzy systems include Takagi-Sugeno (T-S) fuzzy models in [4]. This paper will only consider T-S
fuzzy models [2,5,6] due to their important features such as optimality and robustness which do not exist in
Mamdani models [7,8].
A general T-S model employs an affine model with a constant term in the corresponding part for each rule. If
each rule is regarded as a part of the system and a controller is designed to stabilize these corresponding parts, it
will be found that the overall system will be stabilized.In [3], this controller structure called parallel distributed
compensation (PDC) is introduced. This structure utilizes a fuzzy state feedback controller which mirrors the
structure of the associated linear T-S model. The idea is that for each local linear model, a linear feedback
control is designed. The resulting overall controller, which is nonlinear in general, is a fuzzy blending of each
individual linear controller. Applications of T-S model together with PDC controller have achieved many
successes in real systems [11,14].
In order to derive the T-S fuzzy model from given nonlinear system equations, local dynamics in different state
regions are represented by linear models. The overall model is the combination of these local models obtained by
fuzzy blending. The fuzzy control design is realized via PDC scheme [9][10]. The PDC controller is primarily
used to derive a control rule from the corresponding rule of the T-S fuzzy model so as to compensate for it. The
resulting overall fuzzy controller, which is a fuzzy mixture of individual linear controllers, is nonlinear in
general. This concept was used by [10] to design fuzzy controllers that ensure the stability of fuzzy systems.
950
In many control systems, it is vital to consider the limiting factors such as saturation of actuators. In actuators,
the amplitudes and rates of change of signals are bounded to certain maximum levels. A particular control
system that performs properly in the presence of input/output and state constraints should be designed [14]. It is
possible that a controller will generate control signals that cannot be followed by the physical system: for
instance, trying to rotate a DC motor shaft with excessive speed. This attempt may produce undesired behavior
of the system. Therefore, an additional constraint is needed for the stabilization approach of T-S fuzzy model.
The analysis of the stability of T-S fuzzy systems with constraint on the input/output and initial state
independence is not only of theoretical interest, but also of practical value.
In this paper we will consider a T-S model with PDC controllers with constraint on the input/output and initial
state independence and its model designed in MATLAB/Simulink and LMI Toolbox.
2
A T-S fuzzy system is described by fuzzy IF-THEN rules that represent locally linear input-output relations of a
system. The overall fuzzy model of the system is obtained by fuzzy blending of linear system models. The ith
rule of this continuous fuzzy system is of the following form:
w z Ai x Bi u
i 1 i
(2.1)
w z
i 1 i
wi z M ij ( z j )
(2.2)
j 1
M ij is the membership function of the jth fuzzy set in the ith fuzzy rule. Let,
i z
wi z
(2.3)
w z
i 1 i
where i z is the normalized weight for each rule. Then (2.1) may be expressed as:
r
x i z Ai x Bi u
(2.4)
i 1
Since wi z 0 and
w z 0 , we have z 1 and i z 0
i
i 1
i 1
951
Designing a fuzzy controller for the control of a nonlinear system described by a T-S fuzzy model, PDC provides
us to easy and straightforward procedure.
In PDC design, for each rule of the fuzzy model we construct a control rule separately. The designed fuzzy
controller and the T- S fuzzy model share the same fuzzy sets in premise parts. There are linear controllers in
consequent parts of the control rules. As a matter of fact the ith rule of the controller is:
w z K x z K x
u
w z
r
i 1 i
r
i 1 i
(3.1)
i 1
The design problem is calculating the local feedback gains K i in consequent parts.
Stability Analysis with Constraint on the Input/Output and Initial State Independence
The stabilization of a fuzzy control system has been investigated extensively in the non-linear system stability
[10]. The main issue is selecting K i coefficient matrices to stabilize the non-linear system. Tanaka and Sugeno
derived the stability conditions corresponding to a quadratic Lyapunov function.
When we replace (3.1) in (1.4), the control system can be represented as:
x i z j z Ai - Bi K j x
r
(4.1)
i 1 j 1
x i z i z Ai - Bi Ki x i z i z Gij
(4.2)
(4.3)
i 1
i 1 i j
Where
Theorem 1 [10]: the closed loop fuzzy control system can be stabilized via the PDC controller if there is a
common positive definite matrix P which satisfies the following Lyapunov inequalities:
( Ai Bi Ki )T P P( Ai Bi Ki ) 0, i 1, 2,...r
Gij T P PGij 0, i j r
(4.4)
When we pre-multiply and post-multiply both sides of inequalities (4.4) and change variables as follows:
Y P 1
X i K iY
(4.5)
we obtain
N i YAi T AY
- Bi X i - X i T Bi T 0
i
(4.6)
where Lij Bi X j B j X i
952
According to the basic LMI lemma [11] we can convert (4.6) into equivalent linear matrix inequalities. Than we
can rewrite (4.6) in the following form:
Ni
0
0
0,
I
Oij
0
0
0
I
(4.7)
If there exists a common positive solution for LMIs in (4.7) than stability is guaranteed. In accordance with
theorem 1, a practical realization was given in [10]. In this paper, constraint on the input/output and initial state
independence was introduced in addition to [10].
Theorems 2-4 represent input/output constraints and initial state independence conditions, respectively.
Theorem 2 [10]: Assume that the initial condition ( ) is known. The constraint ( )
times
if the following LMIs hold:
T
1
Y
x 0
0,
Y
x 0
X i
X iT
0
2 I
is enforced at all
(4.8)
Theorem 3 [10,15]: Assume that the initial condition x 0 is known. The constraint y (t ) 2 is enforced at all
times t 0 the following LMIs hold:
T
1
Y
x 0
0,
x
0
Y
CiY
YCi T
0
2 I
(4.9)
Theorem 4 [10]: Assume that x(0) , where x(0) is unknown but the upper bound, , is known. Then,
T
1
1
x 0
0,
Y
x(0)i
x 0
x(0)iT
0
2 I
(4.10)
The design approach is illustrated by considering the problem of balancing an inverted pendulum controlled by a
dc motor via a gear train [12].
The state variables and plant output are as follows:
d
x1 p , x2 p wp , x3 I a , y x1
dt
The state equations describing the plant can be written as:
x2
x1
0
x ( g / l )sin x ( NK / ml 2 ) x 0 u
1
m
3
2
x3 ( Kb N / La ) x2 Ra / La x3 1 / La
(5.1)
where K m is the motor torque constant, K b is the back emf constant and N is the gear ratio. Acceptable plant
parameters are g 9.8 m / s2 , l 1m , m 1kg , N 10 , Km 0.1 Nm / A , Kb 0.1Vs / rad , Ra 1 and
953
x2
x1
0
x1
x 9.8sin x x 0 u , y 1 0 0 x
2
1
3
2
x3 10 x2 10 x3 10
x3
(5.2)
(5.3)
Here,
1
0
0
0
A1 9.8 0
1 , B1 0 ,
0 10 10
10
C1 1 0 0
0
0 1
0
A2 0 0
1 , B2 0 ,
0 10 10
10
C2 1 0 0
sin x1 t / x1 t , x1 t 0
M 1 x1 t
1,
x1 t 0 (5.4)
M 1 and M 2 are fuzzy sets defined as:
M 2 x1 t 1 M 1 x1 t
The fuzzy model completely represents the Dynamics of the nonlinear system under x1 t conditions.
~=
Saturation
0
Constant
Relational
Operator
1
Product
w1
Product2
w2
sin(u[1])
w1
Fcn2
x1
x1
Fcn3
x2
==
Fuzzy Operations
0
x3
Constant1
1
xo s
Product3
1/(u[1])
w2
A1* u
Relational
Operator1
Scope
~=
Integrator
B1*K1* u
Constant2
Relational
Operator2
1-(u[1])
[1 0 0]
Fcn1
Constant
==
A2* u
0
Constant3
B2*K2* u
Relational
Operator3
Product1
0
Constant4
(a)
(b)
Fig. 1. Simulink implementation of the two-rule T-S fuzzy control system (a) Fuzzy Operations block (b)
954
For calculating control law and solve the linear matrix inequalities, a MATLAB code was written in LMI
Toolbox [13]. The system response and control output for initial conditions x1 (0) 1 , x2 (0) 0 and x3 (0) 0 is
shown in Fig. 2, where K1 and K 2 were determined as:
-5
0.5
-10
0
0.5
x2
-5
-0.5
-10
u2
-1
0
u1
-15
10 0
10
Fig.2. System response (x1, x2) and control output (u1, u2)
6
Conclusion
This paper presents fuzzy Lyapunov approach to the stabilization of Takagi-Sugeno fuzzy models with
constraint on the input & output and initial state independence. This model can be handled as linear matrix
inequality problem which can be solved using LMI Toolbox in MATLAB. As stated earlier; asymptotic stability
is of great importance to the research of fuzzy control systems. This approach gives less conservative results and
very good effects are obtained for an inverted pendulum controlled by a dc motor via a gear train. Future work
might consider optimality in terms of various performance indices and robustness properties based on T-S fuzzy
model.
References
1.
Zadeh, L. A. (1973). Outline of a new approach to the analysis of complex systems and decision processes. IEEE
Trans. Systems, Man, and Cybernetics, vol. 3, no. 1, pp. 28-44.
2. Mamdani, E. H. (1977). Application of fuzzy logic to approximate reasoning using linguistic systems. Fuzzy Sets
and Systems, vol. 26, pp. 1182-1191.
3. Wang, H. O., Tanaka, K., and Griffin, M.F. (1995). Parallel distributed compensation of nonlinear systems by
Takagi-Sugeno fuzzy model. In Proc. FUZZ-IEEE/IFES95, pp. 531-538.
4. Takagi, T. and Sugeno, M. (1985). Fuzzy identification of systems and its applications to modeling and control.
IEEE Trans. Syst., Man, Cybern., vol. SMC-15, pp. 116-132, Jan.
5. Pedrycz, W. and Reformat, M. (1997). Rule-based modeling of nonlinear relationships. IEEE Trans. Fuzzy
Systems, vol. 5, no. 2, pp. 256-269.
6. Yager, R. and Filev, D. 1994. Essentials of Fuzzy Modeling and Control. NY: J. Wiley & Sons.
7. Assilian, S. and Mamdani, E.H. (1974). An Experiment in Linguistic Synthesis with a Fuzzy Logic Controller. Int.
J. Man-Machine Studies, vol. 7(1), pp. 1-13.
8. Kickert, W. J. M. and van Nauta Lemke, H. R. (1976). Application of a Fuzzy Controller in a Warm Water Plant.
Automatica, vol. 12(4), pp. 301-308.
9. Tanaka, K. and Sugeno, M. (1992). Stability analysis and design of fuzzy control systems. Fuzzy Sets Syst., Vol.45
No.2, pp.135-156.
10. Tanaka, K. and Wang, H. O. 2001. Fuzzy Control Systems Design and Analysis: A Linear Matrix Inequality
Approach. New York: John Wiley & Sons, Inc.
11. Hong, S. K. and Langari, R. (1998). Synthesis of an LMI-based fuzzy control system with guaranteed optimal H
performance. Proc. FUZZ-IEEE98, pp. 422-427.
955
Biographies
Manafeddin Namazov was born in Azerbaijan in 1963. In 1986, he graduated from Saint Petersburg Electrotechnical
University "LETI" (ETU, The Faculty of Computing Technologies and Automation, Russian Fed.). He received his PhD
degree in computer science from the St. Petersburg Institute for Informatics and Automation of the Russian Academy of
Sciences (SPIIRAS) in 1993. He is currently an Associate professor in the Dept. of Electrical & Electronics Engineering,
Cumhuriyet University (Turkey).
His current research interests include fuzzy control, automatic control, and automatic control of robot
manipulators.
Burak Tekgun - was born in Sivas, TURKEY in 1985.He received his BS degree from the Electrical & Electronics Eng.
Dept. of Cumhuriyet University, TURKEY in 2008. Mr. Tekgun is currently a masters student in the field of Control Theory
and is a research assistant at the same institution.
His research interests include fuzzy logic control, optimal control and robust control, and his recent publications are
concentrated on fuzzy logic control of nonlinear dynamic systems.
Ibrahim Emre Celikkale was born in Izmir, TURKEY in 1984.He received his BS degree from the Celal Bayar
University (Mechanical Eng.), TURKEY in 2006, and received his masters degree from the University of Pennsylvania
(MEAM), PA, USA in the field of mechatronics in 2009. Mr. Celikkale is currently a PhD Student at Cumhuriyet University,
TURKEY in the field of Machine Theory and Dynamics and is a research assistant at the Electrical and Electronics
Engineering Dept. of the same institution.
His research interests include dynamics and control of robotic systems, fuzzy logic control, optimal control and
robust control, and his recent publications are concentrated on fuzzy logic control.
956
1,2,3
Abstract. In this paper, it is aimed to achieve fuzzy logic control of a simple 2-link robot manipulator via
Robotics and Fuzzy Logic Toolboxes of MATLAB. Blocks from these toolboxes were blended in Simulink
environment in order to obtain the simulations. The robot block available in the Robotics Toolbox was used
to create the model by modifying the block parameters. Next, a PD-controller was designed and optimized,
whose results, in turn, acted as a guide for generating the fuzzy rule-base of the fuzzy controller. The fuzzy
controller was then designed and implemented into the system using the Fuzzy Logic Toolbox. Finally, the
fuzzy control system that has satisfactory reference tracking performance was simulated for various inputs.
Key words: Fuzzy control, Robotics Toolbox, 2-link robot manipulator.
Introduction
These days, robot manipulators are widely used in manufacturing, and tuning of the controllers is an actual
problem in the control of these systems. Classical controller schemes require tuning for each reference and
disturbance signals. Using the advantageous features of fuzzy controller scheme, control of plants with a wider
range of reference and disturbance signals can be handled [12,17].
There are several approaches for robot manipulator control that are well defined in literature [2,3,6,7]. However,
one of these approaches, namely, computed-torque control, provides a quite satisfactory performance
particularly when the robot arm parameters are known fairly accurately [7]. This method was first introduced in
Pauls 1972 report [1], and the name was first used in Markiewicz and Bejczy [2,3,15].
A fundamental problem in robot control is to accomplish the task of tracking a desired trajectory with adequate
precision. In classical robot manipulator control, tuning of the controllers are achieved by ranging the controlled
variables. However, in real applications, control systems are subject to external effects such as noise and
disturbance. In order to suppress the undesired effects of these inputs, the fuzzy control algorithm, which allows
for control in a wider interval of the controlled variables, is implemented.
In contrast with the conventional methods, fuzzy control schemes based on the Fuzzy Set Theory, proposed by
Zadeh [8], provides a more effective method for the analysis and control of relatively complex, mathematically
difficult to model, nonlinear, time-varying systems. One of the most significant strengths of fuzzy control is the
ability to be successfully applied to complex systems such as Mamdanis steam engine and boiler combination
[9], Kickerts warm water plant [10] and Ostergaards heat exchanger system [11] using experience, or control
engineering knowledge without a mathematical model of the plant [12,13].
Modeling the robot manipulator, design and analysis of the fuzzy control algorithm and acquisition of the
simulation results using the strong computational tools listed above are of primary interest to this research. The
main advantage of using such computational tools is the ability to address the issue of synthesis and analysis of
957
such complex systems in a very easy and compact manner. In this paper, we first introduce the model of the
robot manipulator. Then, a PD-controller is implemented and optimized for the desired operation regime. Next,
the design procedure of a fuzzy controller based on optimized PD parameters, which performs fuzzy tuning is
described. This paper focuses on the design of fuzzy logic control of a simple 2-link robot manipulator via
simultaneous application of Simulink, Robotics Toolbox and Fuzzy Logic Toolbox of MATLAB.
2
The robot manipulator is modeled so as to obtain a solution to the inverse dynamics problem in which the
torques applied by the actuators is computed in order to follow a prescribed trajectory. In other words, the goal
of this model is to determine the torque regimes required to achieve given position, velocity and acceleration
objectives for each link of the robot manipulator. Manipulator dynamics deals with the equations of motion in
order to obtain such a solution [4]. In a general sense, this problem is formulated as follows [5];
(2)
where(t) is the torque vector, q is the vector of generalized coordinates, D is the inertial matrix, h is the Coriolis
and centrifugal terms, and c is the gravitational term. The 2-link robot manipulator of interest is illustrated in Fig.
1 [5].
Fig. 1. 2-link robot manipulator used as the plant of the fuzzy control system [6]
Equations of motion for the robot manipulator can be calculated as follows [5]:
4
1
1
1
m l 2 m2l 2 m2C2l 2
m2l 2 m2l 2C2
1 3 1
3
3
2
1
1
1
1
2
2
m2l 2 m2l 2C2
m2l 2
3
2
3
2 2
2
2 m2 S2l 2 m2 S2l 1 2
2 2
m
S
l
2 2
1
2
1
1
m
glC
2
12
2
(3)
An effective computational tool for the simulation and analysis of robotic control systems is the Robotics
Toolbox for MATLAB, which is introduced by Corke [4]. Among various methods for deriving the dynamic
equations of motion [refs], the Robotics Toolbox utilizes the Newton-Euler formulation due to its computational
efficiency [18]. The overall robotic system was modeled in MATLAB/Simulink via blending the blocks from
both Simulink and Robotics Toolbox. The main components of the model include the robot object, inverse
958
dynamics (rne) and the reference trajectory (jtraj) blocks. Important features of these blocks are discussed
below.
The first block is called jtraj, which provides a reference trajectory to be followed by the robot manipulator. The
inverse dynamics (rne) block computes the torques required to accomplish the task of following the assigned
trajectory by using recursive Newton-Euler formulation. Finally, the 2-link robot manipulator is modeled based
on Denavit&Hartenberg (DH) Representation in the Robotics Toolbox of MATLAB. Input to this block consists
of the torques applied by the actuators, and it outputs the generalized coordinates and their derivatives for each
link.Kinematics and dynamics of the manipulator are compactly represented in a single matrix (dyn) [4], which
in turn, provides the Simulink robot object with the necessary parameters such as DH parameters and inertial
parameters, etc. A complete list of parameters that are passed to the robot object is given in Table 1.
Table 1. Column assignments for the dyn matrix of the Toolbox [4]
Symbol
i
Ai
i
Di
i
m
3
Description
Link twist angle [rad]
Link length
Link rotation angle [rad]
Link offset distance
Optional joint type; 0 for revolute, non-zero for prismatic
Link mass
Actual values
1 = 2 = 0
A1 = A2 = 1
1 = 2 = 0
D1 = D 2 = 0
1 = 2 = 0
m1 = m2 = 1
Computed-torque control provides easy-to-understand schemes for robot manipulator control. In general, these
schemes work well especially when the arm parameters are known fairly precisely [7]. In this section, a fuzzy
controller will be designed based on computed-torque control scheme to improve the reference tracking
performance of the 2-link robot manipulator discussed above. The design will be carried out in Simulink
environment and will involve blocks from the Fuzzy Logic Toolbox, as well.
The first step to fuzzy control design is achieved by utilizing a PD controller scheme to find the optimized
controller gains. The optimization procedure was carried out using the Signal Constraint block of Response
Optimization Toolbox of MATLAB instead of the well-known conventional methods such as Ziegler-Nichols
and hand tuning due to their shortcomings mentioned in [12,20]. Signal constraint block allows for graphically
constraining the response signal, and optimization of the model parameters that satisfy the performance
requirements, namely, rise time, settling time, overshoot, and steady-state error [16,21,22]. Optimized
parameters for our design that satisfy given conditions were computed after six iterations as shown in Figure 2.
959
Table 2. Optimal values for Kp and Kd parameters for each links controller
Kpopt
Kdopt
-4
Controller 1 1x10
70.4696
Controller 2 8.9944
4.2533
The membership functions of inputs/outputs of the Fuzzy inference system (FIS) are shown in Figure 3 [14].
(b)
(a)
(c)
Fig. 3. Membership functions of the input and output signals: (a) error of q1, (b) change of error of q1, (c) Fuzzy
value of Kp1
Table 3. Fuzzy rules of the FIS
NB
NS
ZE
PS
PB
Negative BIG
Negative SMALL
ZERO
Positive SMALL
Positive BIG
960
Ranges and locations of the membership functions are determined according to the error signal values obtained
from the previous design (with a PD controller). The rule-base of the FIS is then constructed after careful
examination of the error signals. Fuzzy rules employed in the FIS are given in Table 3.
Optimal Kp and Kd parameters are tuned according to the following formulas [19];
K p K popt K p
(3)
K d K dopt K d
In Table 3, fuzzy values of the output variable are shown according to the values of error and change of error. As
the fuzzy controller is designed, next step will be simulating the overall system to check its performance.
4
The fuzzy control system was designed and simulated in MATLAB Simulink in order to evaluate the reference
tracking performance of the system. For this purpose, Robotics and Fuzzy Logic Toolbox blocks were
implemented on the Simulink model. The functional block diagram of the overall system is illustrated in Figure
4.
e_q1
Kpd1
e_q2
de_q1
Kpd2
de_q2
Fuzzy PD Controller
q
qd
qd
qdd
qdd
Trajectory
Generator
2-link robot
2-link robot
q
2-link robot
tau
tau
qd
Animation of
Robot Manipulator
qdd
Inverse Dynamics
Robot Object
Clock
simout
To Workspace
Fig. 4. Simulink model of the 2-link robot manipulator system with fuzzy control
1
e_q1
e_q1
de_q1
uq1
Fuzzy Kp1
2
1
Kpd1
e_q1
uqd1
e_q2
de_q1
Fuzzy Kd1
3
e_q2
uq2
de_q1
de_q2
Fuzzy Kp2
4
de_q2
2
Kpd2
e_q2
uqd2
de_q2
Fuzzy Kd2
961
2
yx z
Z
2-link robot
-2
2
0
-2 -2
0.8
0.6
0.6
q2
0.8
0.4
0.4
0.2
0.2
0
0.15
0
0.15
0.1
0.1
qd2
qd1
q1
Output signals are then subtracted from the reference inputs, and the calculated error signals are sent to two
separate fuzzy controllers, in which the control signals are generated for the two links of the manipulator.
Control signals are finally fed back to the RNE block as acceleration inputs that will regulate the torques
accordingly.
Some simulation results are provided below to understand the dynamic behavior of the system. In these
simulations, input values of angular position are pi/4 for both links.
0.05
0
0
0.05
0
10
time
10
time
(a) (b)
Fig. 7. (a) Angular position (q1) and velocity (qd1) of link-1 versus time, (b) Angular position (q2) and velocity
(qd2) of link-2 versus time
As seen from the output curves, trajectory tracking performance of the fuzzy control system is satisfactory in
terms of time-domain characteristics such as overshoot, rise time, settling time, and steady-state error, even
though they are not readily observed because of the structure of the input signal. As a result, the fuzzy approach
considered in this paper was found to be relatively effective compared to the PD-controller scheme.
5
Conclusion
Dynamic model of a 2-link robot manipulator is created and a fuzzy logic controller is designed and
implemented in order to adjust the PD-controller parameters for improved reference-tracking performance.
Design procedures for all components of the overall system are explained in detail. The system is then simulated
in MATLAB/Simulink and the output curves are presented to emphasize the results. The fuzzy logic controller is
found to be efficient in terms of adjusting Kp and Kd parameters, which in turn decreased the tracking error, and
improved the reference tracking performance of the robot manipulator. Future work might be the design of a
model-based (Takagi-Sugeno) fuzzy controller scheme which includes several uncertainties.,
962
References
1.
2.
3.
4.
5.
6.
7.
8.
9.
10.
11.
12.
13.
14.
15.
16.
17.
18.
19.
20.
21.
22.
Paul, R.P., "Modeling, Trajectory Calculation, and Servoing of a Computer Controlled Arm," Technical Report
AIM-177, Stanford University Artificial Intelligence Laboratory, 1972.
Markiewicz, B. "Analysis of the Computed Torque Drive Method and Comparison with Conventional Position
Servo for a Computer-Controlled Manipulator," Jet Propulsion Laboratory Technical Memo 33601, March 1973.
Bejczy, A.K. "Robot Arm Dynamics and Control," Jet Propulsion Laboratory Technical Memo 33669, February
1974.
Corke, P.I. A Robotics Toolbox for MATLAB, IEEE Robotics & Automation Magazine pp.24-32, March 1996.
Fu, K.S.; Gonzalez,R.C.; Lee,C.S.G. ROBOTICS: Control, Sensing, Vision, and Intelligence, McGraw-Hill
Book Company ISBN 0-07-022625-3, 1987.
Tseng,C.S.; Chen,B.S.; Uang,H.J. Fuzzy Tracking Control Design for Nonlinear Dynamic Systems via T-S Fuzzy
Model, IEEE Transactions on Fuzzy Systems, Vol. 9, No. 3, pp. 381-392, June 2001.
Lewis,F.L.; Dawson,D.M.; Abdallah,C.T. Robot Manipulator Control - Theory and Practice, Second Edition,
Marcel Dekker, Inc. ISBN 0-8247-4072-6, 2004.
Zadeh,L.A. Fuzzy Sets, Information and Control. Vol.8, pp. 338-353, 1965.
Assilian, S. and Mamdani,E.H. An Experiment in Linguistic Synthesis with a Fuzzy Logic Controller, Int. J.
Man-Machine Studies, vol. 7(1), pp. 1-13, 1974.
Kickert, W.J.M. and van Nauta Lemke,H.R. Application of a Fuzzy Controller in a Warm Water Plant,
Automatica, vol. 12(4), pp. 301-308, 1976.
Ostergaard,J.J. Fuzzy Logic Control of a Heat Exchanger System, In Fuzzy Automata and Decision Processes,
North-Holland, Amsterdam, 1977.
Namazov, M. and Basturk,O. Fuzzy Proportional-Derivative Controller Design forPosition Control of a DC
Motor, Turkish Journal of Fuzzy Systems, Vol.1, No.1, pp. 36-54, 2010.
Namazov, M. and Basturk,O. Computer Simulation of a Stable Takagi-Sugeno Fuzzy Control System Designed
via Linear Matrix Inequalities, International Symposium on Innovations in Intelligent Systems and Applications,
pp. 244-249, June 2010.
Schacherbauer, A. and Xu, Y. Fuzzy Control and Fuzzy Kinematic Mapping for a Redundant Space Robot
Technical report CMU-RI-TR-92-12, Robotics Institute, Carnegie Mellon University, August, pp. 24-25, 1992.
Akamatsu, S.; Konishi, M.; Imai, J. Position Control of 2-Link SCARA Robot by using Internal Model Control,
MEM.FAC.ENG.OKA.UNI. Vol. 43, pp. 49-54, 2009.
Gunnersson,S.; Rousseaux,O.; Collignon,V. Iterative feedback tuning Applied to Robot Joint Controllers, 14th
IFAC World Congress, Beijing, P.R.China, 1999.
GreenA. and Sasiadek,J. Z. Dynamics and Trajectory Tracking Control of a Two-Link Robot Manipulator,
Journal of Vibration and Control pp.15-16, 2004.
Araghi,L.F. and Koorayme,M. H. Neural Network Controller for Two links- Robotic Manipulator Control with
Different Load, In Proc .of International MultiConference of Engineers and Computer Scientists 2009, Vol II,
IMECS 2009, Hong Kong, 2009.
Xiao-kan, W.; Zhong-liang, S.; Wanglei; Dong-qing, F. Design and Research Based on Fuzzy PID-parameters
Self-tuning Controller with MATLAB, In Proceedings of International Conf. on Advanced Computer Theory and
Engineering 2008 pp.996-999
Jantzen J. Foundations of Fuzzy Control, WS: John Wiley & Sons, Ltd., 2007, pp.79-83
Simulink Response Optimization Getting Started Guide, 3rdprinting, Natick: The Mathworks Inc., 2008
Veremey,
E.I.
and
Pogojev,S.B.
Nonlinear
Control
Design
Blockset
[Online]
Available:http://matlab.exponenta.ru/nonlinecondes/book1/preface.php
Biographies
* Biographies of the authors are the same as in the paper Design of a Stable Takagi-Sugeno Fuzzy Control System via LMIs
with Constraint on the Input/Output & Initial State Independence published in the proceedings of this symposium.
963
Abstract. In this study, it is aimed to apply neural network based fuzzy time series model for handling a
nonlinear problem due to advantages of artificial neural network and fuzzy time series models. A method is
improved which is based on Yu and Huarngs algorithm for forecasting fuzzy time series. The length of
intervals is decided by using an effective method which is called distribution-based length instead of using
constant length of intervals as 100 or 1000 by following previous studies. A C program is written to perform
this method and also, different number of hidden nodes is used for analyzing. A data set of Exchange rate of
Turkish Liras(TL) to United States Dollars(USD) during 2006-2010 years is used for comparison all models
performances.
Keywords: fuzzy time series, neural networks, c programming, forecasting.
1 Introduction
Fuzzy set theory was firstly proposed by Zadeh [1] in 1965. Then, fuzzy time series concept was introduced by
Song and Chissom [2], [3], [4]. After that, many fuzzy time series models such as first order models [5], [6],
high-order models [7], [8], hybrid models [9], seasonal models [10], [11], bivariate models [12], [13] and
multivariate models have been proposed. These fuzzy time series models have been applied to various problem
domains, such as enrollment [14], temperature [15] and stock index [16].
In forecasting, artificial neural networks are mathematical models that imitate biological neural networks.
Artificial neural networks (ANN) consist of some elements. Determining the elements of the ANN issue that
affect the forecasting performance of ANN should be considered carefully [17]. One of them is network
architecture. However, there are not general rules for determining the best architecture. So, much architecture
should be tried for the correct results. There are various types of ANN. The best three networks for the related
data of the recent study [18] are Generalized Regression Neural Networks (GRNN), Multilayer Perceptron
(MLP) and Radial Basis Function (RBF).
Fuzzy time series models have become increasingly important and have been used to solve nonlinear problems
in recent years. Also, ANN models have been popular because of their effects for handling this kind of problems.
However, ANN models give efficient performance for researchers by forecasting fuzzy time series. Hence, in
this study we applied neural network based fuzzy time series model for handling a nonlinear problem. We
improved a method based on Yu and Huarngs method [20] for forecasting fuzzy time series. Differing from
previous studies, we used various degrees of membership in establishing fuzzy relationships with various
numbers of hidden nodes and we performed different ANN models such as GRNN, MLP and RBF to improve
forecasting performance. We also used these models for forecasting without using fuzzy approach. We decided
the length of intervals by using an effective method instead of using constant length of intervals as 100 or 1000
by following previous studies. We decided the length of intervals by using an effective method which is called
distribution-based length for determining the length of intervals proposed by Huarng [19] and we wrote an
964
optimization program by using C programming language to perform this convenient method efficiently. To
demonstrate comparison between these models we used a data set of exchange rate of TL to USD during 20062010 years. All empirical results compared with each other. To show these things, the remainder of this paper is
organized as follows. Section 2 clearly reviews artificial neural networks. Section 3 briefly explains fuzzy time
series and improved method. Section 4 includes all empirical analysis. Finally, section 5 concludes the paper.
U be the universe of discourse, where U u1 , u2 ,..., un and ui are possible linguistic values of U ,
a
fuzzy
set
of
Ai Ai u1 / u1 Ai u2 / u2
linguistic
variables
Ai
of
is
defined
by
Y t which is a subset of real numbers be the universe of discourse defined by the fuzzy set fi t . If
denoted by F t 1 F t .
This proposed neural network-based fuzzy time series forecast method based on Yu and Huarngs method [20] but
has some changes, modifications and is summarized as follows:
965
d t 1, t obs t obs t 1
where obs t and
t 1 instead of
(3)
The differences may turn out to be negative. To ensure that all the universes of discourse are positive, different
positive constants are added to the differences for different parts of data:
d ' t 1, t d t 1, t const
(2)
For each part of the data (For example for the year), the minimum and maximum of all the differences,
Dmin and
for all t,
for all t,
(3)
ui Dmin D1 i 1 l , Dmin D1 il
i 1, 2,..., k
(4)
mi Dmin D1
2i 1 l
2
i 1, 2,..., k
(5)
u1 , u2 , u3 ,...
In this step, we used Huarngs approach [19] for determining the length of intervals which are called distributionbased length. It is calculated according to the distribution of the first differences of data. Its algorithm is clearly
summarized as below.
1. Calculate all the absolute differences between
differences.
2. According to the average, determine the base for length of intervals by following Table 1.
Table 1.Base mapping table.
Range
0.01-0.1
0.11-1.0
1.1-10
11-100
101-1000
Base
0.01
0.1
1
10
100
966
3. Plot the cumulative distribution of the first differences. The base determined in step 2 is used as interval.
4. According to the base determined in step 2, choose as the length of intervals the largest length that is smaller
than at least half the first differences.
We wrote an optimization program by using C programming language to perform this convenient method
efficiently.1
2. Fuzzification:
Next, d ' t 1, t can be fuzzified into a set of degrees of membership, V t 1, t ,
(6)
(7)
This method then applies GRNN, MLP and RBF neural networks to establish (or train) the fuzzy relationships. It
has one input layer, one or more hidden layers and one output layer. The numbers of input and output nodes are
equal to the number of degrees membership in V t 1, t . The number of hidden nodes is determined by neural
networks performance by using Statistica 7.0 software. With V t 1, t , it can be proceeded to forecast
V t , t 1 by means of the trained neural network. If the in-sample observations are used, the results are
referred to as those of the in-sample estimation. If out-of-sample observations are used, the results are referred to
as those of out-of-sample forecasting.
4. Defuzzification:
Weighted averages is applied to defuzzify the degrees of membership:
fd t 1, t
k 1
k 1
where
k
t 1,t
mk
k
t 1,t
(8)
fd t 1, t is the forecasted difference between t 1 and t ; tk1,t represents the forecasted degrees of
(9)
(40)
Program code is too long to show in this section and if it is necessary, it can be obtained by communication with authors.
967
5. Performance evaluation:
For performance evaluation, the root mean squared error (RMSE) is calculated as follows:
n
RMSE
forecast t obs t
t k 1
(51)
nk
where there are n observations, including k in-sample and n-k out-of-sample observations.
4 Empirical Analysis
In this section it is presented a time-series data which is Exchange Rate of TL to USD in Turkey [21]. It consists
of weekly closed prices of time series which starts 06.01.2006 and ends 31.12.2010, comprising n=261
observations.
The performance of the model is related with how close are the forecasted values and the observed values.
RMSE is used in order to compare the performances of obtained results of neural network-based fuzzy time
series models for three ANN types and ANN models.
In this study, we use the intelligent problem solver (IPS) module of the STATISTICA 7.0 for deciding the best
ANN model for given data set for both fuzzy approach and ANN. The corresponding program gives opportunity
for designer to construct million ANN models at a time and get the best of them. So, we run 100.000 ANN
models to get the appropriate one for a data set of exchange rate of TL to USD.
Neural network-based fuzzy time series with some changes for given ANN types and with Yu and Huarngs
method are used to analyze for this data set. Then, RMSE is calculated respectively for each model. The best
ones for each model according to RMSE values are shown in Table 2.
Table 2. RMSE values for fuzzy approach with different ANN models.
Models
Proposed method with MLP
Proposed method with RBF
Proposed method with GRNN
Yu and Huarngs method
RMSE
0,023146
0,023441
0,023211
0,024132
GRNN, MLP and RBF are used to analyze for this data set. Then, RMSE is calculated respectively for each
model. The best ones for each model according to RMSE values are shown in Table 3.
Table 3. RMSE values for ANN models.
Models
MLP
RBF
GRNN
RMSE
0,024372
0,024405
0,043209
And now we compare all methods that we used in this analysis. Table 4 shows the best RMSE values for fuzzy
approach and ANN models.
968
RMSE
0,023146
0,024372
5 Conclusion
We can summarize the results shortly as follows:
This study applied a neural network-based fuzzy time series model which is different from previous studies
because of using various degrees of membership with neural networks. So, by training and forecasting, more
information was considered.
According to Yu and Huarngs method, the length of intervals is chosen constant as 100 or 1000 by following
previous studies. We decided the length of intervals by using a well-known method which is called distribution
based length approach instead of based on other studies. Also, we wrote an optimization program by using C
programming language to perform this convenient method efficiently. So, in a short time, we determined the
length of intervals owing to C programming instead of working hard for a long time and the length of intervals is
not chosen randomly, so results are more reliable.
According to Yu and Huarngs method, number of hidden nodes is equal to sum of input and output nodes. We
used various numbers of hidden nodes which are important for forecasting while using neural networks and we
found better RMSE values not only using MLP but also using GRNN and RBF neural network architectures
according to Table 2.
We separated the observations into in-sample (for training) and out-of-sample (for forecasting) observations. The
ratio was 83%:17%.
We applied the best three networks which are called MLP, RBF and GRNN. We compared neural network based
fuzzy time series models and ANN models. Empirical results show that the neural network based fuzzy time
series model for MLP with proposed method outperformed the other best and most commonly used ANN models
according to Table 3 and 4 and Yu and Huarngs method according to Table 2 for demonstrating 100000
networks by using STATISTICA 7.0 software.
References
1.
2.
3.
L.A. Zadeh, (1965). Fuzzy Sets, Information and Control. 8, pp. 338-353.
Q. Song and B. S. Chissom, (1993). Fuzzy Time Series and its Models. Fuzzy Sets and Systems. 54, pp. 269-277.
Q. Song and B. S. Chissom, (1993). Forecasting Enrollments with Fuzzy Time Series, Part 1. Fuzzy Sets and
Systems. 54, pp. 1-10.
4. Q. Song and B. S. Chissom, (1994). Forecasting Enrollments with Fuzzy Time Series, Part 2. Fuzzy Sets and
Systems. 62, pp. 1-8.
5. S.-M. Chen, (1996). Forecasting Enrollments Based on Fuzzy Time Series. Fuzzy Sets and Systems. 81(3), pp. 311319.
6. J. Sullivan and W. H. Woodall, (1994). A Comparison of Fuzzy Forecasting and Markov Modeling. Fuzzy Sets and
Systems. 64(3), pp. 279-293.
7. S.-M. Chen, (2002). Forecasting Enrollments Based on High-Order Fuzzy Time Series. Cybernetics and Systems.
33(1), pp. 1-16.
8. K. Huarng and H.-K. Yu, (2003). An Information Criterion for Fuzzy Time Series to Forecast the Stock Exchange
Index on the TAIEX. In Ninth International Conference on Fuzzy Theory and Technology. pp. 187-189.
9. F.-M. Tseng and G.-H. Tzeng, (2002). A Fuzzy Seasonal ARIMA Model for Forecasting. Fuzzy Sets and Systems.
126(3), pp. 367-376.
10. P.T. Chang, (1997). Fuzzy Seasonality Forecasting. Fuzzy Sets and Systems. 90(1), pp. 1-10.
11. Q. Song, (1999). Seasonal Forecasting in Fuzzy Time Series. Fuzzy Sets and Systems. 107(2), pp.235.
969
Biographies
Ozer Ozdemir Ozer Ozdemir was born in Turkey in 1982. He received his B.Sc., M.Sc. degrees in statistics in the
Department of Statistics at Anadolu University, Turkey, respectively in 2005 and in 2008. He is currently doing his Ph.D. in
the same place. He has worked as a Research Assistant from 2006-2008 and as a Lecturer from 2008 in the Department of
Statistics at Anadolu University, Turkey.
He has published over 15 international conference papers and journals in his research areas. His research interests include
Simulation, Artificial Neural Networks, Fuzzy Logic, Fuzzy Modeling, Time Series, Computer Programming, Statistical
Software and Computer Applications in Statistics.
Memmedaga Memmedli Memmedaga Memmedli was born in Azerbaijan in 1947. He received a diploma at
BakuStateUniversity in 1971. He received a PhD diploma from RussianAcademy of Science in 1977. From 1985-2006, he
has worked as a Associate Professor; from 2007, as a Professor Dr. at Department of Statistics at Anadolu University,
Eskisehir, Turkey.
He has published many journals and conference papers in his research areas. His research interests are Game Theory,
Control Theory, Optimization, Artificial Neural Networks, Fuzzy Modeling and Nonparametric Regression Analysis.
970
Introduction
Fuzzy regression studies mainly have focused on fuzzy parametric regression models. The different types of
approaches have been used, including mathematical programming based methods, least-squares based methods
and other methods. In practical problems, expressing relation between explanatory variable and response
variable with a certain parametric model is a very important restriction, and it can cause incorrect results. Hence,
interest in nonparametric regression models have increased in recent years and for this purpose different type of
models have been developed. For example k-nearest neighborhood regression, local polynomial models, kernel
regression, different regression models with spline functions, wavelet regression, additive models and
generalized additive models, etc.
There are very few studies on fuzzy nonparametric regression models. It is necessary transform many
notations and approaches of nonparametric regression models to fuzzy form for creating fuzzy nonparametric
regression models. This procedure is not so easy. The first studies about fuzzy nonparametric models were made
by using neural networks, k-nearest neighbor, kernel and local linear regression approaches.
The aim of this study is to research fuzzy local polynomial nonparametric regression models and to evaluate
model performance which depends on the degree of polynomial and smoothing parameter (bandwidth) by using
simulation study.
2 Literature Review
Fuzzy regression researches which was firstly introduced by Tanaka et al. [1], belongs to fuzzy parametric
regression models and especially fuzzy linear regression problems. Many studies of this area are based on
mathematical programming [1-3], least squares [4-6] and neural networks methods [7-8]. There are also many
practical applications of fuzzy parametric regression analysis [9-11].
971
The first study of fuzzy nonparametric models is Ishibuchi and Tanakas study [12]. Authors have suggested
several fuzzy nonparametric regression models by using back propagation neural network algorithm in this
paper. Cheng and Lee [7] have used radial basis function neural networks in fuzzy regression analysis. The
developments of nonparametric regression models have been getting more attention on fuzzy version of these
models (see [13-17]). Aspects of this situation, Cheng and Lee [18] have examined fuzzy versions of k-nearest
neighbor and kernel estimation methods. Wang et al. [19] have considered fuzzy local linear regression model
and advantages according to fuzzy models of [18] of this approach have shown by using experimental way.
Cross-validation has used for bandwidth selection in these researches. But, Fan and Gijbels have proposed more
effective selection methods for local regression fitting. To improve fuzzy versions of these methods and use
these in fuzzy local regression models are very important issues. Bandwidth size and order of local polynomial
are associated with each other in local polynomial regression. So, analysis of this relation for fuzzy models is
another important issue.
In this paper, we considered the relationship between the smoothing parameter value and degree of
polynomial as a simulation study in fuzzy local polynomial nonparametric regression. Besides the local linear
models, local cubic models are also used in this simulation study. Fuzzy forms of different methods are used for
bandwidth selection. Performances of the models are evaluated in accordance with the selected bandwidth.
3 Methods
Let the univariate fuzzy nonparametric regression model is given by,
y ( x)
In
this
model,
is
( x) c( x), ( x), ( x)
term and
Let
Each
crisp
(6)
independent
variable
whose
domain
is
assumed
and
is a fuzzy error
)be a sample of the observed crisp inputs and LR fuzzy outputs for model (1).
output is a LR fuzzy number which represented by ( c y , y , y ). The functions c( x), ( x), ( x)
show center, right and left spreads of LR fuzzy number and they have continuous derivatives up to fourth order
in the domain D. Taking l y c y y and u y c y y , we can write this number like ( l y , c y , u y ) by
helping its lower, upper and center limits. Suppose that l ( x), c( x), u( x) functions are respectively the lower,
center and upper limits of LR - function which has derivatives up to pth order in the domain D . For a given
neighborhood of
l ( p ) ( x0 )
l ( x) l ( x0 ) l ( x0 )( x x0 ) ...
( x x0 ) p
p!
c( x) c( x0 ) c( x0 )( x x0 ) ...
c ( p ) ( x0 )
( x x0 ) p
p!
(2)
u ( p ) ( x0 )
( x x0 ) p
p!
Based on Diamond`s distance [4] with using pth order polynomial, following local estimators can be
defined by locally weighted least squares method with a certain kernel function K () :
u ( x) u ( x0 ) u( x0 )( x x0 ) ...
972
l( x ), l( x ),..., l
( p)
c( x ), c( x ),..., c
( p)
u( x ), u( x ),..., u
0
( p)
( x0 ) = H ( x0 , h)l y
( x0 ) = H ( x0 , h)c y
(3)
( x0 ) = H ( x0 , h)u y
where H ( x0 , h) X ( x0 )W ( x0 , h) X ( x0 )
T
X T ( x0 )W ( x0 , h) ,
l y1
c y1
u y1
1 ( x1 x0 ) ( x1 x0 ) 2 ... ( x1 x0 ) p
.
.
.
,
,
,
X ( x0 ) . . . . . . . . . . . . . .
ly . cy . u y .
1 ( x x ) ( x x ) 2 ... ( x x ) p
.
.
.
n
0
n
0
n
0
l
c
u
yn
yn
yn
(|
|)
(|
2
1 x2
1 z
e function and K h ( z ) K . Because the local regression
h h
2
estimate solves a least squares problem, ( x) is a linear estimate. That is, for each x there exists a weight
K () is selected Gaussian K ( x)
( x) li ( x) yi
[14]. Here,
i 1
(4)
n n matrix L with rows l T ( xi ) which maps the data to the fitted values:
( x1 ),..., ( xn ) Ly
T
(5)
It is used cross validation and AIC estimators for selection of bandwidth parameter. Instead of crisp number y
in these estimators as defined in formulas used fuzzy numbers (l y , c y , u y ) and using Diamond distance [4]
appropriate formulas can be obtained:
2
2
2
1 n (l yi lyi ) (c yi cyi ) (u yi u yi )
CV (h)
n i1
(1 Lii )2
2
2
2
1 n (l yi lyi ) (c yi c yi ) (u yi u yi )
GCV (h)
1
n i1
(1 trL) 2
n
(6)
(7)
n
1
2(trL 1)
(8)
AIC (h) log (l yi lyi )2 (c yi cyi )2 (u yi u yi )2
1
n trL 2
i 1 n
In this paper for model (1) taking p 1 and p 3 , the fuzzy local linear and fuzzy local cubic regression
models are investigated with help of appropriate (2) formulas. The bandwidth parameter
minimum conditions of CV (h) , GCV (h) ve AIC (h) functions.
973
h determined by the
Simulation Study
By repeating the simulation experiments 200 times, the two datasets were generated which have 100
dimension, using same functions with Cheng and Lee`s [1] study. The fuzzy response outputs and center, rightleft spreads are triangular fuzzy numbers that generated by Sample-1 and Sample-2. In simulation study, noise
obtained from uniform distribution on interval [0.5,0.5] for center and on interval [0.25,0.25] for symmetric
spread. The aim of this study is to compare fuzzy local linear and fuzzy local cubic models performances for
smoothing parameter selected by different methods.
Sample-1. Let
For
(9)
( )
(10)
]
]
(11)
] for each i. Assuming that the observed fuzzy outputs are symmetric triangular fuzzy numbers,
interval [
so the presentation of them will be as following,
(
(12)
For same
) )
(13)
( )
[
( )
(14)
(15)
(16)
In this study, it is examined that fuzzy local linear and fuzzy local cubic models are calculated for generated
data sets by helping of these functions. For selection of appropriate smoothing parameters (h-bandwidth)
optimum value, the cross-validation (CV), generalized cross-validation (GCV) and Akaike criteria have been
used in both models. 200 repetitions were made to reduce the impact of randomness in simulation calculations.
The Gaussian kernel used to generate the weight sequence in the applications that made with both data sets.
Performances of the models are compared with their
ASE (h)
1 n 2
1 n
d
(
(
x
),
(
x
))
ASE values which are appropriate for Gauss kernel are given by selecting optimum bandwidth with the help
of CV , GCV and AIC criteria of fuzzy local linear and cubic regression models for sample 1 in Table 1.
974
LCS
ASE
CV
0.51
0.2682013
GCV
0.28
0.2367753
AIC
0.41
0.2555963
CV
1.10
0.2639194
GCV
0.60
0.2490921
AIC
1.10
0.2637155
LCS
ASE
CV
0.21
0.2312725
GCV
0.10
0.1401306
AIC
0.18
0.2126562
CV
0.40
0.2294235
GCV
0.30
0.2132408
AIC
0.50
0.2417500
Since the curve which is appropriate for sample-2 data set is more fluctuating (curved) than the curve which is
appropriate for sample-1, proper bandwidths in Table-2 are quite smaller than Table-1s. Also, as we can see in
LLS model are effectively smaller than suitable h values for LCS
model. Both models show the best performances on optimum bandwidth which is selected with GCV function
in sample-1 and sample-2. While comparing ASE values which are appropriate for the same selection criteria
of LLS and LCS models, we see that bandwidth is effectively large in LCS model. For example,
both tables, h values which are selected for
appropriate bandwidths selected by CV for both models are 0.51 and 1.1 and ASE values are 02682013 and
0.2639194 in Table 1. Large bandwidths cause reduction of local fitting points and this also leads to a reduction
for calculations effectively.
975
of local linear models are not convenient; in this case, it is necessary to reduce bandwidth but it increases
computational complexity.
For future studies, it is aimed to select adaptive local smoothing parameters in fuzzy local polynomial
regression or to consider adaptive fuzzy local polynomial models. Also, it is important to make studies on
considering fuzzy forms of other nonparametric regression models.
References
1.
2.
3.
4.
5.
6.
7.
8.
9.
10.
11.
12.
13.
14.
15.
16.
17.
18.
19.
Tanaka H., Uejima S., Asai K. (1982), Linear Regression Analysis with Fuzzy Model, IEEE Transactions on
Systems, Man and cybernetics 12, 903-907.
Kao C., Chyu C.-L., (2002), A Fuzzy Linear Regression Model with Better Explanatory Power, Fuzzy Sets and
Systems 126, 401-409.
Nasrabadi M., Nasrabadi E. (2004), A Mathematical-Programming Approach to Fuzzy Linear Regression Analysis,
Applied Mathem. and Computation 155, 673-688.
Diamond P. (1988), Fuzzy Least Squares, Information Sciences 46 , 141-157.
DUrso P., Gastaldi T. (2002), An orderwise Polynomial Regression Procedure for Fuzzy Data, Fuzzy Sets and
Systems 130, 1-19.
Hong H., Song J.-K., Do H. (2001), Fuzzy Least Squares Linear Regression Analysis Using Shape Preserving
Operations, Information Sciences 138, 185-193.
Cheng C.-B., Lee E.S. (2001), Fuzzy Regression with Radial Basis Function Networks, Fuzzy Sets and Systems
119, 291-301.
Hong D., Hwang C. (2003), Support Vector Fussy Regression Machines, C
Hong Y.-Y., Chao Z.-T., Yang M.-S. (2004), A Fuzzy multiple Linear Regression Based Loss Formula in Electric
Distribution Systems, Fuzzy Sets and Systems 142, 293-306.
Lee H. T., Chen S. H. (2001), Fuzzy Regression Model with Fuzzy Input and Output Data for Manpower
Forecasting, , Fuzzy Sets and Systems 119, 205-213.
Sanchez J. D., Gomez A. T. (2003), Applications of Fuzzy Regression in Actuarial Analysis, Journal of Risk
Insurance 70, 665-699.
Ishibuchi H., Tanaka H. (1992), Fuzzy Regression Analysis Using Neural Networks, Fuzzy Sets and Systems 50,
257-265.
Fan J. and Gijbels I. (1996), Local Polinomial Modelling and Its Applications, Chapman & Hall/CRC.
Loader C. (1999), Local Regression and Likelihood. Springer
Hardle W. (1990) Applied Nonparametric Regression, CambridgeUniversity Press, New York.
Hastie T., Tibshirani R. (1990), Generalized Additive Models, Chapman&Hall, London.
S. N. Wood (2006), Generalized Additive Models (An Introduction with R), Chapman&Hall/CRC.
Cheng C.-B., Lee E. S. (1999), Nonparametric Fuzzy Regression k-NN and Kernel Smothing Techniques,
Computers and Mathematical with Applications 38, 239-251.
Wang N., Zhang W.-X., Mei C.-L. (2007), Fuzzy Nonparametric Regression Based on Local Linear Smoothing
Technique, An International Journal Information Scences, 177, 3882-3900.
Biographies
Memmedaga Memmedli Memmedaga Memmedli was born in Azerbaijan in 1947. He received a diploma at
BakuStateUniversity in 1971. He received a PhD diploma from RussianAcademy of Science in 1977. From 1985-2006, he
has worked as a Associate Professor; from 2007, as a Professor Dr. at Department of Statistics at Anadolu University,
Eskisehir, Turkey.
He has published many journals and conference papers in his research areas. His research interests are Game Theory,
Control Theory, Optimization, Artificial Neural Networks, Fuzzy Modeling and Nonparametric Regression Analysis.
Munevver Yildiz Munevvere Yildizwas born in Turkey in 1983. She received her B.Sc., M.Sc. degrees in statistics in
Department of Statistics at EgeUniversity and at Afyon Kocatepe University, Turkey, respectively in 2006 and in 2009. She
is currently doing her Ph.D. in the same department at AnadoluUniversity, Eskisehir, Turkey.
976
977
Abstract. In this study, a Trajectory Tracking Fuzzy Logic Controller (TTFLC) is designed for the nuclear
research reactor Triga Mark-II and its membership functions parameters are tuned by Partial Swarm
Optimization (PSO). In this work, PSO is preferred because, it is an algorithm that is simple, easy to
understand and few parameters require. Compared with other algorithms, PSO retained the global search
strategy used an easily operated speed-displacement model, and avoided complicated genetic operation.
TTFLC has also been preferred as a controller because of its use the only error and change in error as input
parameter. To substantiate the efficiency and effectiveness of work, Triga Mark-II research reactor
established at Istanbul Technical University in Turkey was considered.
Keywords: fuzzy logic, PSO, trajectory tracking, nuclear reactor control, Optimization of Membership
Function.
1 Introduction
Fuzzy logic provides a means of processing vague, inexact or incomplete information, expressed in linguistic
terms. This differs from set theory where a value or an object either belongs to a set or is excluded from it. In
fuzzy set theory an object can belong to a set fully or partially (i.e. to a certain extent). This vagueness is
captured by membership functions (MF). The extent to which an object belong to a set (called the degree of
membership or membership value ) ranges between 0 and 1. It is usual for MFs to be represented graphically;
the popular MFs are the triangular, trapezoidal and the sigmoid functions [1]. In this study, The MFs are
considered as trapezoidal. The Fuzzy Inferencing System (FIS) contains the MFs and a rule base together.
In the recently studies, PSO is used to optimize parameters of Fuzzy Controller. The authors are proposed
automatic adjustment of membership functions using PSO on application of parking a vehicle into a garage [2].
The other work used the particle swarm optimization technique to automatically tune the MFs of a Mamdanitype of fuzzy controller [3]. A method for optimizing membership functions of a fuzzy system is presented using
particle swarm optimization (PSO) algorithm in [4]. All these studies and more are illustrated that PSO is useful
to optimize fuzzy parameters.
In control of nuclear research reactors power, rod position, period and fuel temperature of the reactors were
used as the controllers input variables and they required fuzzification [5, 6]. Although they produced results
superior to those of classical controllers, these approaches have difficulties for the designers to get an insight on
how the designed membership functions and fuzzy rule base affects the performance of the control system.
These difficulties can be eliminated by using trajectory tracking fuzzy logic controller to represent all the
978
variables in a single domain. A method (TTFLC) including a trajectory has been proposed to eliminate the
fuzzification of such variables of the controllers by Baba in 2004 [7]. However the weights of the fuzzy rules and
MFs parameters were only determined by the user in this study. In design of controllers, if the fuzzy controller
is used, the choice of any parameters must be appropriate to the physical system so that simulation studies could
be closer to the real system. Experts generally select these parameters by trial and error method. This procedure
is time consuming and irksome. An expert trajectory was proposed for control of this reactor in [8]. A 2 DOF
planar robot was controlled by Fuzzy Logic Controller tuned with a particle swarm optimization in [9].
Numerous researches have been proposed a controller such as PID, FLC, adaptive control and optimal control
for the power control of ITU Triga Mark-II [10-17]. The work presented here is a continuation of earlier research
on control of this nuclear reactor. In our study, in order to find the optimum MFs parameters, a PSO algorithm
was designed. Hence, the parameters are found by PSO and insert automatically into the TTFLC. To substantiate
the efficiency and effectiveness of work, a series of simulations are carried out on ITU Triga Mark-II reactor.
The structure of the paper as follows: Structure of ITU Triga Mark-II research reactor in Section 2. The
trajectory used in this study is explained in Section 3. Partial Swarm Optimization is briefly explained in Section
4. Designing of a PSO-TTFLC is shown in Section 5. Simulation results regarding the developed control system
are shown in Section 6. The last Section contains the conclusions.
X AX BU
(1)
Where: X is a representative vector of prompt neutron density, precursor concentration, fuel temperature,
coolant temperature, and coolant velocity. U is a vector representing control variables. A and B are related
matrixes.
In this model, mass flow rate and inlet temperature of the coolant are assumed to be time independent.
However, the thermal conductivity and specific heat of the fuel and the thermal conductivity of the cladding and
the physical properties of the coolant are all dependent on temperature [19].
PT1 a3 (t t0 )3 a2 (t t0 ) 2 a1 (t t0 ) a0
979
(2)
when t1 t t 2 ;
PT2 b2 (t t1 ) 2 b2 (t t1 ) b0
(3)
PT3 c3 (t t2 )3 c2 (t t2 ) 2 c1 (t t2 ) c0
(4)
when t 2 t t s ;
when t t s ;
PT4 Ps
(5)
There are 14 unknown coefficients in the above equations. These coefficients can be calculated from the initial
conditions, the conditions at the crossing points, the final conditions, and the period conditions.
X i (k 1) X i (k ) Vi (k 1),
w wmax I
(7)
wmax wmin
I max
(8)
In these sets, position X stands for satisfaction utility preference value set, initial position Xi is the initially
generated satisfaction utility preference value set; V stands for speed, P stands for individual extreme value of
each particle, g stands for global extreme value, w is weight of inertia velocity, c1 and c2 are study factors, r1 and
r2 are random numbers in (0, 1). wmax and wmin are maximum and minimum values of w, respectively. I max is the
total number of iterations, I is the current iteration number.
5 Designing of a PSO-TTFLC
The first step in designing the PSO-TTFLC is to determine the input and output variables and their ranges of
TTFLC. The second step is to select the membership functions to be used in setting up the values for each input
and output. The next step is to construct the weighted fuzzy control rule base. The design of fuzzy logic mainly
depends on the construction of the membership functions and the fuzzy controllers. The last step is to apply the
PSO algorithm to find optimum MFs parameters. The basic block diagram of the trajectory tracking fuzzy
980
control of the research reactor is shown in Fig. 1. The objective is that the reactor power follows a specified path
with minimum error.
PSO Algorithm
Trajectory
Power
+
_
PSO
Parameters
The Best of
Parameters
Fuzzy
Interface
Fuzzifier
Optimization
Of
Parameters
Defuzzifier
Cost
Function
1-
de/dt
u +
_
Z-1
Weighted
Rule Base
Point Kinetics
Model
Pout
Thermo -Hydroulic
Model
Reactor Model
Fig. 2. (a) The MFs of Change of Error in Power (b) The MFs of Error in Power
The next stage involved combining the rules to obtain a final control action. To obtain a final control action, the
reactor power is measured at every sampling interval. The change in power error is calculated. The degree of
membership of each variable in every labeled group is determined, and the variables can belong to more than one
fuzzy set. The degree of fulfillment (DOF) of each and every rule is determined. The final control action is
calculated by weighing the action of each rule by its DOF.
Fifteen control rules were used in designing the TTFLC. These rules are shown in Table 1. A fuzzy control
action consists of situation and action pairs conditional rules based on IF and THEN statements are usually used.
TTFLC provided to work with the PSO algorithm by making required changes in the designed controller. Thus,
PSO has the task of overcoming such difficulties in selecting suitable membership functions. Object of PSO
algorithm is to find optimum parameters of the membership functions which provide minimum performance
criterion value. Integral Time Absolute Error Criterion (ITAE) was selected as a performance criterion and can
be written as;
ITAE
ts
t | PT (t ) PR(t ) |
(9)
t t0
Where, t is the time vector, PR(t) and PT(t) are reactor and trajectory output power.
981
P.E.
C.P.E.
NB
NB
NB
NS
NS
NS
ZE
ZE
N
Z
P
N
Z
P
N
Z
Action
#Rul
e
9
10
11
12
13
14
15
-.7
-.8
-1
-.4
-.015
-.6
-.3
.0
P.E.
C.P.E.
ZE
PS
PS
PS
PB
PB
PB
P
N
Z
P
N
Z
P
Action
.3
-.3
.456
.8
-.05
.986
1
6 Simulation Results
In this section, simulation results regarding the PSO algorithm and the developed control system are presented.
PSO algorithm was started with parameters given in the table 2. This table also contains the simulator
parameters.
The results from this study are encouraging, and surpass previous attempts at optimizing MFs parameters.
After running the PSO optimization for the parameters shown in Table 2, the final gbest performance criterion
(ITAE) is 33374. For this system, this value is very successful. This value can be seen from Fig. 3 showing how
the best particles performance value decreased throughout the simulation:
982
The MFs tuned by PSO is given Fig.4. The MFs parameters found by PSO are applied to the TTFLC and the
results shown in Figure 4 are obtained. Since the error is too small, In this figure and in some others, the
trajectory path seems to be covered by output power.
Fig. 4. Simulation result for the best MFs parameters found by PSO
Reactor outputs and errors between trajectories and outputs are given in fig. 4-A. Control signals and rod
positions are also given in fig. 4-B. It is clearly observed from these figures that reactor outputs are perfectly
tracked the defined trajectories. Control signal and rod position gradually changed so there is no destructive
signal to send to the electro mechanic apparatus of the rod.
To test the effect of different initial power levels, the simulations has been started up for 250kW desired power
level for the periods of 20-50-22 seconds and for 500W, and 5kW levels of initial power. As shown in figures 5
and 6, PSO-TTFLC is able to bring reactor power to desired level from different initial levels of power.
To test the effect of different desired power levels, the simulations has been started up for 1kW initial power
level for the periods of 20-50-20 seconds and for 100kW and 200kW levels of desired power. The results of the
simulation are shown in Figs. 7 and 8.
7 Conclusions
This paper introduced the PSO based TTFLC controller to control the Triga Mark-II research reactor
established at Istanbul Technical University. The fuzzy membership functions were determined using PSO
983
algorithm. In applications, criterion value was optimized successfully by PSO. To test the parameters of
membership functions found by PSO, various initial and desired power levels are applied to TTFLC to control
reactor model. The simulation results show that the reactor power tracks the given trajectory successfully and
reaches the desired power level through the optimized parameters of membership functions. As a result, PSO
could optimize the parameters of membership functions for this controller and the PSO-TTFLC could control the
system successfully under all conditions within the acceptable error tolerance.
References
[1] Babuska R and Verbruggen H. B.: An Overview of Fuzzy modeling for Control. Control Engineering Practice, vol
4 No 11, 1996 pp1593 1605.
[2] A.A.A. Esmin, A.R. Aoki, and G. Lambert-Torres. Particle Swarm Optimization For Fuzzy Membership Functions
Optimization. IEEE SMC MP2Q5 2002.
[3] Gu Fang, Ngai Ming Kwok and Quang Ha. Automatic Fuzzy Membership Function Tuning Using the Particle
Swarm Optimisation. IEEE Pacific-Asia Workshop on Computational Intelligence and Industrial Application 2008.
[4] Elijah E. Omizegba and Gbijah E. Adebayo. Optimizing Fuzzy Membership Functions Using Particle Swarm
Algorithm. Proceedings of the 2009 IEEE International Conference on Systems, Man, and Cybernetics.
[5] Bernard, J. A. Use of a rule-based system for process control. (1988). IEEE Control System Magazine, 8(5), 313.
[6] Ruan, D., & Van der Wal, A. J. Controlling the power of a nuclear reactor with fuzzy logic. (1998). Information
Sciences, 110, 151177.
[7] Baba, A. F. Fuzzy logic controller. (2004). Nuclear Engineering International, 49, 3638.
[8] Coban R., Can B., An expert trajectory design for control of nuclear research reactors, Expert Systems with
Applications 36 (2009) 1150211508
[9] Bingl Z., Karahan O., A Fuzzy Logic Controller tuned with PSO for 2 DOF robot trajectory control, Expert
Systems with Applications 38 (2011) 10171031.
[10] Can, B., Yavuz, H., Akbay, E.: The Investigation of Nonlinear Dynamics Behavior of ITU Triga Mark-II Reactor,
Eleventh European Triga Users Conference, The Germen Cancer Research Institute Heildelberg, Germany, (1990)
11-16
[11] Can, B.: The Optimal Control of ITU Triga Mark II Reactor, The Twelfth European Triga Users Conference, NRI
Bucuresti-Pitesti, Romania, 22-25, (1992)
[12] Erkan, K., Can, B.: Self-Tuning Control of ITU Triga Mark-II Reactor, First Trabzon International Energy and
Environment Symposium, KTU, Turkey (1996) 29-31
[13] Baba, A.F., Can, B., Gok, I., Akbay, E.: Design of a Fuzzy Logic Controller for the ITU Triga Mark-II Reactor,
Proceedings of the Nuclear Plant Instrumentation Control And Human-Machine Interface Technologies, The
Pennsylvania State University, USA, (1996)
[14] Erkan, K., Butun, E., Can, B.: Triga Mark-II Reactor Controller Design Using Genetic Algorithm, Fuzzy Logic and
Intelligent Technologies for Nuclear Science and Industry, World Scientific, Singapure, (1998)
[15] Can, B., Baba, A.F., Erdal, H.: Optimal Controller Design by Using Three Zones Trajectory for ITU Triga Mark-II
Reactor, Journal of Marmara University for Pure and Applied Sciences, Vol.15, Istanbul, Turkey, (1999) 187-196
[16] Akbay, E., Can, B, Baba, A.F.: Fuzzy Logic Control for ITU TRIGA Mark-II Reactor, 3rd International
Symposium on Intelligent Manufacturing Systems, Sakarya, Turkey, (2001)
[17] Coban, R., Can, B.: An Expert Trajectory Design for Control of Nuclear Research Reactors, Expert Systems with
Applications 36 (2009) 1150211508
[18] Instrumentation System (1976). Operation and maintenance manual. USA: General Atomic Co. Omatu, S., Khalid,
M., & Yusof, R. (1996). Neuro-control and its applications: Advances in industrial control. London: SpringerVerlag.
[19] Can, B., Yavuz, H., Akbay, E., The Investigation of Nonlinear Dynamics Behavior of ITU Triga Mark-II Reactor,
Eleventh European Triga Users Conference, The Germen Cancer Research Institute Heildelberg, Germany,
September 11-13, 1990.
[20] J. Kennedy, R.C. Eberhart, Y. Shi, Swarm Intelligence, Morgan Kaufman Publishers, San Francisco, 2001.
984
1 Introduction
Data mining as a modern technology has been used for knowledge discovery for recent years due to the wide
availability of vast amonts of data and the imminent need for turning such data into significant information and
knowledge [1]. The rapid growth and integration of databases provides scientists, engineers, practitioners and
business people with a huge new resource that can be analyzed to make scientific discoveries, optimize
production control and industrial systems, provide business management, and uncover financially valuable
patterns [2,3]. The persistent need for transforming data into useful information and knowledge, researchers and
practitioners have adopted pre-established algorithms from statistics, machine learning, neural networks, and
databases and have also developed new methods targeted at large data mining problems [4,5].
The purpose of our study is to predict the profit of the 500 major industrial enterprises of Turkey of year 2009 by
using data mining techniques. Sales revenue, assets, equity, export, and number of employee are used as fuzzy
inputs and profit as fuzzy output.
2 Literature Review
Profit is the difference between revenue from selling output and the cost acquiring the factors necessary to
produce it [6]. The firm seeks to maximize the difference between its total revenue ant is total economic costs as
large as possible [7]. Profit is the signal people receive which tells entrepreneurs whether they are successful or
not, thus it is the ultimate goal for firms [8]. Above average profitability is probably to protect the firm from
corporate predators, and will certainly foreclose the need for costly and painful adjustments in strategy and/or
985
structure [9,10]. According to the neo-classical microeconomic theory, the firm behavior is usually characterized
by profit maximization [11]. Venkatraman and Ramanujam emphasised that profitability had reflected by ratios
such as return on investment, return on sale, and return on equity [12]. They had also pointed out that marketshare position was a determinant of profitability. Cefis and Ciccarelli stated that innovative activities affect the
performance of firms [13].
Data mining approach for performance analysis of enterprises for profitability will be a renewal for financial
applications.
3 Methods
3.1 Clustering
Clustering deals with partitioning the data set into homogenous groups with respect to proper similarity measures
in such a way that patterns likely to be similar to each other in the same cluster, and as dissimilar as possible in
different clusters. The partitions of a data set can be either hard as in the k-means clustering algorithms, or fuzzy
as in the fuzzy c-means (FCM) clustering algorithm [14,15]. FCM algorithm is a simple and yet the most widely
used and extended fuzzy clustering method of all fuzzy clustering approaches. FCM algorithm partitions a
collection of n vectors ( X x1, x2 ,...,xn p ) into c fuzzy groups such that the weighted within-groups
sum of squared error objective function is minimized. The objective function and constraints for FCM is defined
as
c
Min: J m (u, v)
u ijm d 2 ( i , x j )
(1)
i 1 j 1
i uijm x j
j 1
uijm
(2)
j 1
and
uij 1
(3)
l 1
FCM algorithms runs by using Eq.(2) and Eq.(3) iteratively. Since the assumption of a known or a previously
fixed number of clusters is not realistic for many data analysis problems, cluster validity analysis is used to
determine the number of clusters of FCM clustering algorithm [16,17].
3.2 Classification
Discovering regularities in complex data is an important research topic. Many problems in medical,
astronomical, and financial applications involve a large number of data that need to be classified. Classification
can be thought as the base of ability to knowledge acquisition [1,5,14]. Classification has crucial applications as
decision support in banking, finance, manufacturing, astronomy, medicine, etc. A classifier may be represented
using a Bayesian network, a neural network, a decision tree, etc.
986
H S (T ) Pi H S (Ti )
(4)
i 1
where Pi represents the proportion of records in subset i . Then, the information gain is calculated as follows:
gain(S ) H (T ) H S (T )
(5)
which means the increase in information produced by partitioning the training data T according to this
candidate split S .
(6)
where c is the number of rules in system model; x j is the j th input variable, j 1,..., nv where nv is the total
number of variables; X j is the domain of x j ; A ji is the linguistic label associated with input variable, x j in
rule i with membership function Aji ( x j ) : X j [0,1] ; y is the output variable; Bi is the linguistic label
associated with the output variable in the i th rule with the membership function Bi ( y) : Y [0,1] ; AND is the
logical connective used to aggregate membership values of input variables for a given observation in order to
find the degree of fire of each rule; THEN is the logical implication connective; ALSO is the logical
connective used to aggregate model outputs of fuzzy rules; isr is introduced by Zadeh and it represents that the
definition or assignment is not crisp, it is fuzzy.
987
3.5 Defuzzification
The last step of the fuzzy inference method is to obtain a crisp output. In order to achieve this aim,
defuzzification methods are used. Defuzzification is a mapping from a space of fuzzy control actions defined
over an output universe of discourse into a space of crisp control actions [1]. This process is necessary because
in many practical applications, crisp control action is required for decision making. A defuzzification strategy is
aimed at producing a nonfuzzy control action that best represents the possibility distribution of an inferred fuzzy
control action. Unfortunately, there is no systematic procedure for choosing a defuzzification strategy. Among
many defuzzification methods, we handle most widely used ones.
4 Application
In this study, ICI500 enterprises, listed by ICI (Istanbul Chamber of Industry) every year including 500 major
industrial corporations of Turkey according to their sales from production values, are handled in order to analyze
their financial performance. The location of the enterprises throughout Turkey is shown on a map (Fig. 1).
Optimal
clusters
Sales
Equity
Assets
Export
Employee
Profit
number
of
5
4
5
6
7
5
The fundamental problem in developing any type of fuzzy systems is in determining the shape, size, and the
number of fuzzy sets used to represent the inputs and outputs of the proposed system. After completing the
preprocessing stage, i.e. dealing with outliers, filling empty data using appropriate methods, etc., membership
functions for the variables sales revenue, assets, equity, export, number of employee, and profit are formed by
988
applying FCM algorithm to each of the variables. The number of linguistic variables, i.e., number of clusters for
FCM algorithm, is determined with Fukuyama-Sugeno cluster validity index by changing the number of clusters
2 through 10. The optimal results for each variable are given in Table 1 and the membership functions of
Equity and Export are given in Fig 2.
1,0
0,6
0,4
0,2
0,0
0,0
Variable
too_low
low
normal
high
v ery _high
v ery _v ery _high
0,8
Membership degree
0,8
Membership degree
1,0
Variable
too_low
low
normal
high
0,6
0,4
0,2
0,2
0,4
0,6
Equity_norm
0,8
0,0
0,0
1,0
0,2
0,4
0,6
Export_norm
0,8
1,0
Table 3. Sum square errors for triangular and Gaussian membership functions for various defuzzification
methods.
Defuzzification
Triangular membership
Gaussian membership
17)
method
function (x10
function (x1017)
Centroid
9.8728
9.5912
Bisector
9.7941
9.6876
MOM
11.683
11.1688
LOM
19.663
19.654
SOM
8.2943
8.2959
According to various firing strengths and defuzzification methods, the classification accuracy of the models are
compared. In addition, sum square error of the results are calculated for two different triangular and Gaussian
membership functions each defuzzified with 5 different methods (Table 3) and they are compared with SSE of
regression analysis which is found to be 1,379x1018. It is observed that using fuzzy inference system in
prediction of profit is approximately 100 times better than using regression analysis.
989
5 Conclusion
In this study, after completing the preprocessing stage, i.e. dealing with outliers, filling empty data using
appropriate methods, etc., membership functions for the variables sales revenue, assets, equity, export, number
of employee, and profit are formed by applying FCM algorithm to each of the variables. Moreover, the number
of linguistic variables, i.e, number of clusters for FCM algorithm, is determined by using various cluster validity
indices. Then, we apply decision tree analysis and determine the rules for fuzzy inference system. According to
various firing strengths and defuzzification methods, the classification accuracy of the models are compared. In
addition, sum square error of the results are calculated for all the models and they are compared with SSE of
regression analysis. It is observed that using fuzzy inference system in prediction of profit is approximately 100
times better than using regression analysis. We can conclude that using data mining approach for performance
analysis of enterprises for profitability will be a renewal for financial applications.
References
[1] Jang J-S. R.; Sun C-T.; and Mizutani E. (1997). Neuro-Fuzzy and Soft Computing, Prentice Hall.
[2] Kennedy R.L.; Lee Y.; Roy B.V.; Reed C.D.; and Lippman R.P. (1995). Solving Data Mining Problems through Pattern
Recognition, Prentice-Hall.
[3] Larose D.T. (2004). Discovering Knowledge in Data, Wiley-Interscience.
[4] Olafsson S.; Li X.,; and Wu S. (2008). Operations research and data mining, European Journal of Operational
Research 187:14291448.
[5] Witten I.H.; and Frank E. (2005). Data Mining: Practical Machine Learning Tools and Techniques, Morgan Kaufmann.
[6] Carden A. (2009). Profit and Production, The Quarterly Journal of Austrian Economics 12 (2): 13-26.
[7] Liu S-T. (2007). A computational method for the maximization of long-run and short-run profit, Applied Mathematics
and Computation 186 (2):11041112.
[8] Delis M.D.; Koutsomanoli F.A.; Staikouras C.K.; and Gerogiannaki K. (2009). Evaluating cost and profit efficiency: a
comparison of parametric and nonparametric methodologies, Applied Financial Economics 19:191202.
[9] Geroski, P.; and Jacquemin, A. (1988). The Persistence of Profits: An International Comparison, Economic Journal
98:375-379.
[10] Love J.H.; Roper S.; and Du J. (2009). Innovation, ownership and profitability, International Journal of Industrial
Organization 27:424-434.
[11] Kuosmanen T.; Kortelainen M.; Sipilinen T.; and Cherchye L. (2010). Firm and industry level profit efficiency analysis
using absolute and uniform shadow prices, European Journal of Operational Research 202 (2): 584594.
[12] Venkatraman N. and Ramanujam V. (1986). Measurement of Business Performance in Strategy Research: A
Comparison of Approaches, The Academy of Management Review, 11(4): 801-814.
[13] Cefis E.; and Ciccarelli M. (2005). Profit differentials and innovation, Economics of Innovation and New Technology 14
(1):43-61.
[14] Nasibov E., and Ulutagay G. (2007). A New Unsupervised Approach for Fuzzy Clustering, Fuzzy Sets and Systems
158(19): 2118-2133.
[15] Bezdek J.C. (1973). Fuzzy Mathematics in Pattern Classification, PhD Thesis, Cornell Univ., NY.
[16] Bezdek J.C. (1974). Cluster validity with fuzzy sets, Journal of Cybernetics 3:58-73.
[17] Fukuyama Y., and Sugeno M. (1989). A new method for choosing the number of clusters for the fuzzy c-means
method, In Proc. 5th Fuzzy Systems Symposium 247-250.
[18] Abelln J.; and Masegosa A.R. (2010). An ensemble method using credal decision trees, European Journal of
Operational Research 205:218226.
[19] Zadeh L.A.. (1965). Fuzzy Sets, Information and Control 8:338-353.
[20] Celikyilmaz A.; and Turksen I.B. (2009). Modeling Uncertainty with Fuzzy Logic, Springer.
990
Biographies
Gzde ULUTAGAY Dr. Gzde Ulutagay was born in Izmir in 1979. She received her B.Sc. and M.Sc. degrees from
Department of Statistics in Ege University, Izmir, Turkey in 2001 and 2004, respectively. She received her PhD degree from
Department of Statistics, Dokuz Eylul University, Izmir, Turkey in 2009. Her research interests include fuzzy cluster
analysis, data mining, fuzzy sets theory, optimization, and multivariate statistics.
Dr. Ulutagay is currently working as an Assistant Professor in the Department of Industrial Engineering, Izmir University,
Izmir, Turkey. She is the Vice Dean of Faculty of Engineering.
Fatih ECER- Dr. Fatih Ecer received the B.Sc. degree in Mathematics from Dokuz Eylul University, Izmir, Turkey in 2000
and M.Sc. degree in Business Administration from Afyon Kocatepe University, Afyonkarahisar, Turkey. He received her
PhD degree in Business Administration from Afyon Kocatepe University, Afyonkarahisar, Turkey. His research interests are
fuzzy sets, multiple criterion decision making, and optimization.
Dr. Ecer is currently working as an Assistant Professor in the Department of International Trade and Finance, Afyon
Kocatepe University, Turkey. Hes the Vice Head of department.
Efendi NASIBOV Dr. Efendi Nasibov received the B.Sc. and M.Sc. degrees in Applied Mathematics Department from
Baku State University, Azerbaijan in 1983, and the Ph.D. in Mathematical Cybernetics and Dr.Sc. in Computer Science
degrees from Institute of Cybernetics Academy of Science of Azerbaijan in 1987 and 2003, respectively. His research
interests are in the application of data mining, fuzzy sets theory, optimization and decision making problems, and software
engineering.
Prof. Nasibov is currently working as a full Professor and Head of Department of Computer Science, Dokuz Eylul
University, Izmir, Turkey.
991
1 Introduction
Hydraulic motors are preferred at heavy industrial conditions owing to its characteristic features, high size-power
proportion, loaded startup ability, durability and resistance for environmental conditions. Though there are some
control algorithms, it is hard to control motor finely because of the high torque of motor. Velocity changes,
especially startup and stop periods, causes tensions excessively on the systems that motor runs. Consequently, it
is important to control that how motor accelerates or decelerates as well as its steady state reference velocity.
Controlling acceleration speed of the motor prevents the unwanted tensions on the system so it is a significant
ability for controllers. Although there are many studies done about hydraulic motor control, no one is about
trajectory tracking speed control. The most similar study in literature is about speed control trajectory tracking
system with pneumatic motor [1]. In this study Fuzzy Logic and Genetic algorithm are used to control the
system. One of other studies is about to control velocity of hydraulic motor at a constant reference [2]. Control
algorithm is tested on simulation by mathematic model. Another study is about to control velocity of hydraulic
motor too [3]. The difference in this study is control algorithm. Control algorithm is designed to handle uncertain
parameters that effects velocity. Also this study is done with simulations about mathematic models. An
alternative study is about controlling velocity of hydraulic motor against disturbance inputs [4]. Difference on
this study is that algorithm is tested on a real nonlinear system.
In our study, speed of hydraulic motor is controlled by two different algorithms, Trajectory Tracking PID
Controller (TTPIDC) and Trajectory Tracking Fuzzy Logic Controller (TTFLC), to find out the best control
algorithm for hydraulic motor. The experimental system has many nonlinear components such as servo hydraulic
valve, compressor and encoder. These nonlinear components make the control more difficult and complex. As a
result of nonlinearity, TTPIDC algorithm fails at tracking trajectory. Especially on start up and stop periods
992
where the characteristic of the system is most nonlinear, TTPIDC algorithm failed. On the other hand TTFLC
algorithm uses the simplicity of Fuzzy Logic system and manages to track the trajectory without any steady state
errors [6, 7]. Finely designed fuzzy rules managed the system not to get out control on nonlinear regions.
2 Hardware
In our study, Festo hydraulic systems of education set were used. Main components of hydraulic system,
compressor, motor, servo valve are shown in figure1.
3 Control Software
Control software was designed on Visual Basic platform. Control software user interfaces are shown in figure 3.
993
4 Control Algorithms
The two control algorithms are designed and applied to make a comparison. The general control block
diagram is shown in figure 4.
994
995
Test with 60 second rise and fall times shows that Fuzzy Logic algorithm controls the motor smoothly than PID.
Parameters are set to make steady state error zero. Chosen parameters manage this but cause oscillations and PID
algorithm is affected more by this situation.
Fig. 10. Fuzzy Logic Algorithm 40 Second Fig. 11. PID Algorithm 40 Second Rise and
Rise and Fall Time.
Fall Time.
Shorter rise and fall times are getting control more difficult. Faster rises and falls means that controller have to
be faster too. With constant control parameters, little distortion can be observed.
Fig. 10. Fuzzy Logic Algorithm 10 Second Fig. 11. PID Algorithm 10 Second Rise and
Rise and Fall Time.
Fall Time.
996
Fuzzy Logic algorithm can handle 10 Second rise and fall time trajectories. The other test algorithm, PID, fails at
this trajectory. Advantages of Fuzzy Logic on nonlinear systems make it better on this study.
References
1.
Safak C., Topuz V., Baba F. A. (2010). Pneumatic Motor Speed Control by Trajectory Tracking Fuzzy Logic
Controller. In Indian Academy of Science Sadhana Vol. 35, Part1, February 2010, pp. 75-86.
2. Hasanifard G., Zarif M.H. and Ghareveisi A.A. (2007). Nonlinear Robust Backstepping Control of an
Electrohydraulic Velocity Servosystem. In 2007 Mediterranean Conference on Control and Automation, Huly 2729, Athens-Greece
3. Ito K. (2004). Robust Nonlinear Control of Water Hydraulic Servomotor System with Parameter Uncertainties. In
8th International Conference on Control, Automation, Robotic and Vision Kunming, China, 69-th December 2004
4. Ito K., Takahashi H., Ikeo S. and Takahashi K. (2006). Robust Control of Water Hydraulic Servo Motor System
Using Sliding Mode Control with Disturbance Observer. In SICE-ICASE International Joint Conference 2006
Oct.18-21, Bexco, Busan, Korea
5. Katrancioglu S., Sava K., Erdal H. (2010). A modular and low-cost data acquisition card design with multitasking
support. In Procedia - Social and Behavioral Sciences, Volume 2, Issue 2, 2010, Pages 5266-5270
6. Baba A. F. (2004). Fuzzy Logic Controller. In Nuclear Engineering International, May 2004
7. Baba A. F., afak C. and Kutlu .(2003). The Position Control of Pneumatic Stimulated Manipulator Arm with the
Fuzzy Logic Controller Method. In 3rd International Advanced Technologies Symposium, 40-47, Ankara, 2003
8. Righettini, P. and Giberti, H. (2002). A non-linear controller for trajectory tracking of pneumatic cylinders. In IEEE
2002
9. Lee S. Y. and Cho H. S. (2002). A Fuzzy Controller for an Electro-Hydraulic Fin Actuator Using Phase Plane
Method. In Control Engineering Practice 21 August 2002
10. Fu Y., Zhang W. (2010). The Deadband Analysis of the Electro-hydraulic Actuator of the Motor-pump
Coordinated Control. In International Conference on Mechatronics and Automation August 4-7, 2010 Xian, China
11. Tang M., Xiao S. (2010). The Performance Improvement in Electro-Hydraulic Servo System with PDF Control. In
IEEE 2010
Biographies
Sevan KATRANCIOLU was born in Sivas, in 1987. He received the BS degree in electronics and computer education
from Marmara University, Faculty of Technical Education, Istanbul, Turkey, in 2010. He is still a student of the Institute of
Science and Technology at Marmara University. He is currently research assistant at Okan University, Information Systems
and Technologies Department. His research interests include programming languages, mobile programming, distance
learning, distance control, control card design.
Erdal YILMAZ was born in Adyaman, in 1986. He received the BS degree in electronics and computer education from
Marmara University, Faculty of Technical Education, Istanbul, Turkey, in 2010. He is still a student of the Institute of
Science and Technology at Marmara University. He is currently an associate of IT Experts at Okan University. His research
interests include programming languages, nonlinear systems control, education set design, distance learning module design
and mobile programming.
A.Fevzi BABA received the BS degree in electrical and electronics education from Marmara University, Faculty of
Technical Education, Istanbul, Turkey, in 1986 and MSc, PhD degrees from the Institute of Science and Technology at
Marmara University in 1989 and 1995, respectively. He completed the teacher-training course in electronic engineering at
Purdue University School of Technology, West Lafayette, IN, in 1990. He is currently an associate professor and Head of
Department of Control System. His research interests include intelligent control, fuzzy logic, and engineering technology
education.
997
Abstract. The notion of fuzzy soft function between fuzzy soft sets is mentioned by Aygunooglu et al [1] and
Kharal et al [2]. In this study, some results on this function is given.
Keywords: soft set, fuzzy soft set, fuzzy soft function
1 Introduction
To solve complicated problems in economics, engineering, and environment, we cannot successfully use
classical methods because of various uncertainties typical for those problems. There are three theories: theory of
probability, theory of fuzzy sets, and the interval mathematics which we can consider as mathematical tools for
dealing with uncertainties. But all these theories have their own difficulties. Besides, a theory, Soft Set Theory,
is introduced by Molodtsov to indicate and then solve these kinds of problems [3]. In his pioneer paper he had
shown that soft set theory has a rich potential for applications in many directions. Then Maji et al. [4] presented
a significant paper on set theoretical operations on soft set theory. The authors defined equality of two soft sets,
subset and super set of a soft set, complement of a soft set, null soft set, absolute soft set and soft binary
operations like AND, OR. Since soft set theory has an interaction between fuzzy set theory, researcher has
considered fuzzy generalization of soft sets. Fuzzy soft sets was introduced as a fuzzy generalization of soft sets
by Maji et al. [5]. In [6], Ahmad et al. revised Maji's definition of fuzzy soft intersection and defined arbitrary
fuzzy soft union and intersection, then proved De Morgan Inclusion and De morgan Laws in fuzzy soft set
theory. In [7], X.Yang et al. gave a brief application of fuzzy soft sets. Some properties of fuzzy soft functions
which is mentioned by Aygnolu et al [1] and Kharal et al [2]. In [1,2], Aygnolu et al. and Kharal et al.
presented concepts as image and preimage by using fuzzy soft functions. In this study, some results on fuzzy soft
function is given.
2 Preliminaries
Let U be an initial universe, E be the set of all possible parameters which are properties of the objects in U,
P(U) be the set of all subsets of U, and F(U) be the family of all fuzzy sets in U.
Definition 1. [3] Let A be a subset of E. A pair ( F , A) is called a soft set over U where F : A P(U ) is a
mapping.
Definition 2. [5] Let A be a subset of E. A pair ( f , A) is called a fuzzy soft set over U where f : A F (U ) is
a mapping.
998
Example 3. Let U x, y be universe, E 1,2,3,4,5 be parameter set and A 1,2 E . From Definition
1., ( F , A) 1 x ,2 U is a soft set over U . From Definition 2., ( f , A) 1 x0.5 ,2 x0.2 , y0.6 is a fuzzy
soft set over U .
Definition 4. [5] Let A, B E and ( f , A),( g , B) be two fuzzy soft sets over a common universe U. We say that
( g , B)( f , A) .
Definition 6. [5] Union of two fuzzy soft sets ( f , A) and ( g , B) over common universe U is the fuzzy soft set
hc ( x) g c ( x),
f ( x) g ( x),
c
c
if c A B
if c B A , x U .
if c A B
Definition 7. [1] The intersection of two fuzzy soft sets ( f , A) and ( g , B) over common universe U is the
fuzzy soft set (h, C ) , which is denoted by (h, C ) ( f , A) ( g , B) , where C A B and for each c C ,
hc ( x) f c ( x) g c ( x) for all x in U .
Definition 8. [6] A fuzzy soft set ( f , E ) over U is said to be a null fuzzy soft set and is denoted by if and
only if for each e E, fe 0 , where 0 is the membership function of the null fuzzy set over U , which takes
value 0 for all x in U .
Definition 9. [6] A fuzzy soft set ( f , E ) over U is said to be an absolute fuzzy soft set and is denoted by U if
and only if for each e E, f e 1 , where 1 is the membership function of the absolute fuzzy set over U , which
takes value 1 for all x in U .
Definition 10. [1] The complement of a fuzzy soft set ( f , A) is the fuzzy soft set ( f c , A) , which is denoted by
x and
999
( f ) ( y ) x ( y )
( ) A
f ( x )
, ( y ) , ( ) A
, otherwise
for every y U 2 .
Similarly the inverse image of each ( g , B) FS (U 2 ; E2 ) is denoted by ( , ) ( g , B) ( g , ( B)) and the
1
g ( ) ( ( x)) , ( ) B
0
, otherwise
( g ) ( x)
1
for every x U1 .
In [2], Kharal et al have following results for the image of a fuzzy soft set and for the preimage of a fuzzy soft
set.
Theorem 12. [2] Let ( , ) be a fuzzy soft function from FS (U1; E1 ) to FS (U 2 ; E2 ) , ( f , A) and ( g , B) in
FS (U1; E1 ) and
(1) ( , )() ,
(2) ( , )(U 1 ) U 2 ,
(3) ( , ) ( f , A)( g , B) ( , )( f , A)( , )( g, B) ,
In general, ( , )
( f k , Ak )
( , )( f k , Ak ) ,
( f k , Ak )
( , )( f k , Ak ) ,
FS (U1; E1 ) and
(1) ( , )1 () ,
(2) ( , )1 (U 2 ) U 1 ,
(3) ( , )1 ( f , A) ( g , B) ( , )1 ( f , A) ( , )1 ( g , B) ,
In general, ( , )1
( f k , Ak )
( , ) 1 ( f k , Ak ) ,
1000
In general, ( , )1
( f k , Ak )
( , ) 1 ( f k , Ak ) ,
( f , A) FS (U1 ; E1 ) , we have
(, )( f , A)
( , )( f , A)c ,
where 1 ( y) , 1 ( ) A , y U 2 , ( A) .
Proof. Since ( , )( f , A) f , ( A) ( f )c , ( A) and
c
( , )( f , A)
( f )c ( y) 1 ( f ) ( y) 1 f ( x)
x 1 ( y ) 1 ( ) A
and
( f c ) ( y )
c
f ( x) (1 f ( x))
1
1
1
( y ) ( ) A
x ( y ) ( ) A
1 f ( x) ,
x 1 ( y ) 1 ( ) A
for all y U .
Hence ( f )c ( y) ( f c ) ( y) . So, the claim is true.
U1 x, y ,U 2 a, b and : U1 U 2 , ( x, a),( y, b) ,
E1 1,2,3 , E2 4,5,6 and : E1 E2 , (1,4),(2,4),(3,5) . Let
( f , A) 1 x0 , y0.7 ,2 x0.2 , y0.4 be a fuzzy soft set over U with A 1,2 E1 . The image of ( f , A)
under the fuzzy soft function ( , ) , ( , )( f , A) ( f , ( A)) we will use following computations:
f (t ) f (t ) f1 (t ) f 2 (t )
t 1 ( a ) 1 (4) A
tx 1,2
tx
f1 ( x) f 2 ( x) 0 0.2 0.2
( f ) 4 (a)
f (t ) f (t ) f1 (t ) f 2 (t )
1
t ( b ) (4) A
t y
t y 1,2
f1 ( y ) f 2 ( y ) 0.7 0.4 0.7
( f ) 4 (b)
Membership degree is equal zero for all remained parameters and each members in the universe U 2 . Thus, we
obtain ( , )( f , A) ( f , ( A)) 4 a0.2 , b0.7 . So, ( , )( f , A) (( f )c , ( A)) 4 a0.8 , b0.3 .
c
1001
On the other hand, the complement of ( f , A) is ( f , A)c ( f c , A) 1 x1 , y0.3 , 2 x0.8 , y0.6 . The image of
( f , A)c under ( , ) is ( , )( f , A)
following;
c
c
c
c
f (t ) f (t ) f 1 (t ) f 2 (t )
1
1
t
1,2
t
t
( a ) (4) A
c
c
f 1 ( x) f 2 ( x) 1 0.8 1
( f c ) 4 (a)
c
c
c
c
f (t ) f (t ) f 1 (t ) f 2 (t )
1
1
t
1,2
t
t
( b ) (4) A
( f c ) 4 (b)
Thus
we
(, )( f , A)
obtain
( , )( f , A)c ( , )( f c , A) f c , ( A) 4 a1 , b0.6 .
So,
( , )( f , A)c .
Conclusions
To solve a social problem given by using a fuzzy soft set, one can want to use the information that another fuzzy
soft set has. So, the notion of a function between that two fuzzy soft sets is very important. From the view of this
point, in this paper, some results about such a function is given in detail.
References
1.
2.
3.
4.
5.
6.
7.
A. Aygnolu, H. Aygn, Introduction to Fuzzy Soft Groups, Comput. Math. Appl. 58 (2009) 1279-1286.
A. Kharal, B. Ahmad, Mappings on Fuzzy Soft Classes, Advances in Fuzzy Systems, 2009 (2009).
D. Molodtsov, Soft set Theory-First Results, Comput. Math. Appl. 37 (1999) 19-31.
P.K. Maji, R. Biswas and A.R. Roy, Soft Set Theory, Comput. Math. Appl. 45 (2003) 555-562.
P.K. Maji, R. Biswas and A.R. Roy, Fuzzy Soft Sets, Journal of Fuzzy Math. 9 (2001) 589-602.
B. Ahmad, A. Kharal, On fuzzy soft sets, Advances in Fuzzy Systems, 2009 (2009).
X. Yang, D. Yu, J. Yang, C. Wu, Generalization of soft set theory: From crisp to fuzzy case, Fuzzy Inform. Engin.
40 (2007) 345-354.
Biographies
Bekir TANAY He was born in Ar, Turkey, in 1970. He received M.Sc. degree in 1998 from Institute of Science in
Mugla University, Turkey and Ph.D. degree from Institute of Science in Ege University, Turkey. He is a faculties in the
Mugla University, Faculty of Science, Department of Mathematics.
His research interests are Topology, Algebraic Topology, Soft Set Theory, Fuzzy Soft Set Theory.
M. Bur KANDEMR He was born in Ankara, Turkey, in 1980. He received M.Sc. degree in 2008 from Institute of
Science in Mugla University, Turkey, and he is in a graduate programme for his Ph.D. degree in Instute of Science in Mugla
University, Turkey. He is a faculties in the Mugla University, Faculty of Science, Department of Mathematics.
His research interests are Topology, Algebraic Topology, Soft Set Theory, Fuzzy Soft Set Theory.
1002
Abstract. Identification of splice sites depends on the precise recognition of the boundaries between exons
and introns of a given DNA sequence. This problem can be formulated as a classification of sequence
elements into exon-intron (EI), intron-exon (IE) or None (N) boundary classes. In this study, Fuzzy
Weighted Position Specific Scoring Method (FWPSSM) is proposed to solve this classification problem. This
method uses a position-specific scoring matrix constructed via nucleotide base frequencies. A fuzzy
membership function is used in order to tune the importance of the positions on.
Keywords: Fuzzy Logic, Position Specific Scoring Matrix, Fuzzy Weighted Position Specific Scoring Method.
1 Introduction
Gene is the principal unit of the inheritance which is an ordered sequence of nucleotides. The genes of
eukaryotic organisms include coding and non-coding regions. These coding regions within the genes which are
encoding the proteins are called exons. On the other hand, the non-coding regions which are another part of
DNA that regulates gene expression by removing from the messenger Ribonucleic acid (RNA) in a splicing
process are called introns.
The identification of genes is a major task of bioinformatics. However, gene structure needed to be predicted
for the accurate identification of genes. Recognizing splice sites in eukaryotic genes play an important role in
identifying gene structure. The accurate identification of the splice sites depends on correct estimation of exonintron structures. Splice sites are the locations in DNA which separate coding regions (exons) from non-coding
regions (introns).
The splice site upstream of an intron is called the donor splice site and one that is downstream of an intron is
the acceptor splice site. Sequences which have GT consensus dinucleotide in the splice sites are classified as
EI, which have AG consensus dinucleotide in the splice sites are classified as IE, and which have not any
of the consensus dinucleotide in the splice sites are classified as None. The splice sites which consist of certain
consensus dinucleotide are known as canonical splice sites (Bursets et al., 2000).
In the preliminary work by Nasibov and Tunaboylu, the weighted position specific scoring method (WPSSM)
was proposed. In this work, the optimal weights and threshold parameters are calculated by the genetic algorithm
(Nasibov and Tunaboylu, 2010a, b).
In the current study, we propose the Fuzzy Weighted Position Specific Scoring Method (FWPSSM) for the
classification problem of the splice sites. This method consists of three phase: learning phase, identification
phase and validation phase. In the learning phase, we used position specific scoring matrix and fuzzy patterns to
tune the optimal weights. Then the optimal weights calculated from the learning phase are applied to each
sequence in the test set in the identification phase. This way, each sequence classified into EI, IE, and None
classes. Finally in the validation phase, performance parameters such as sensitivity, specificity, accuracy rates
are calculated.
The difference between WPSSM and FWPSSM methods is only calculation way of the optimal weights.
1003
2 Methods
The proposed method is based on the position specific scoring matrix and fuzzy membership function which is
utilized for optimization of the positional weights parameter. Methods used in FWPSSM are covered in a detail
below.
2.1 Fuzzy Importance of the Positions
The smooth transitions among variables are characterized by membership functions that give fuzzy sets
flexibility in modeling of the problems (Jang et al., 1997).
If X is a collection of objects denoted generically by x , then a fuzzy set A in X is defined as a set of ordered
pairs:
A {( x, A ( x)) | x X } ,
where A ( x) is called the membership function (MF) for the fuzzy set A . The MF maps each element of X to a
membership grade between 0 and 1.
In this study, we deal with the symmetrical triangular MF to determine fuzzy importance of the positions. The
membership degrees of each position are calculated by the following membership function depending on the
parameter, as follows:
i max 1 (i / n) 0.5 / ,0 .
(1)
When the parameter goes to the infinity ( ) , the fuzzy MF in (1) is transformed to the uniform MF,
which represents the lack of differences between positions importance. In this situation, membership degrees
(importance) of the positions are calculated as:
1
, i 1,..., n .
n
(2)
1004
a) PSSM EI , PSSM IE and PSSM None which are calculated for EI, IE, and None classes by using position
specific scoring matrix.
b) x* ( x1* , x2* ,..., xn* ) sequence which is needed to be classified, where xi* {A, C, T , G}, i 1,..., n .
Output: Class of the given sequence x* ( class( x* ) ).
Step 1. Scoring vectors belonging to EI, IE, and None classes of the sequence x* are calculated by using
PSSM EI , PSSM IE and PSSM None matrices, respectively:
SEI ( x* ) ( PSSM EI [ x1* ,1], PSSM EI [ x2* , 2],..., PSSM EI [ xn* , n]) ,
SIE ( x* ) ( PSSM IE [ x1* ,1], PSSM IE [ x2* , 2],..., PSSM IE [ xn* , n]) ,
SNone ( x* ) ( PSSM None [ x1* ,1], PSSM None [ x2* , 2],..., PSSM None [ xn* , n]) ,
where PSSM[.] [ xi* , i] is a scoring value of xi* residue in i th position for the class[.] .
Step 2. Scores of the given sequence x* , SEI ( x* ) , SIE ( x* ) and S None ( x* ) belonging to EI, IE, and None
classes are calculated by using the scoring vectors:
n
sEI ( x* ) wi S EIi ( x* ) ,
i 1
n
sIE ( x* ) wi S IEi ( x* ) ,
i 1
n
i
sNone ( x* ) wi S None
( x* ) ,
i 1
class( x* )
arg
{ EI , IE , None}
End.
Validation phase. Now, let X {xk , k 1,..., N} , be a test set of sequences with n (n 60) residue where
xik {A, C, T , G}, i 1,..., n . The classes of test sequences are predicted as explained in the identification phase.
1005
Error rates (ERs) are computed according to true/false classification results via the following formula:
ER
#( FP FN )
#(TP TN FP FN )
(3)
Also overall error rate (OER) is computed with the below formula:
OER 1
i {EI , IE , N }
i {EI , IE , N }
(TPi TNi )
(4)
The performance parameters such as specificity (Sp), sensitivity (Sn), and accuracy rates are calculated via
following formulas:
Sn
# TP
#(TP FN )
(5)
Sp
# TN
#(TN FP)
(6)
Accuracy
#(TP TN )
#(TP TN FP FN )
(7)
3 Application Results
This method has been applied to the human splice site data from UCI Repository of machine learning
databases (Asuncion and Newman, 2007). This data set consists of 3190 sequences with 60 nucleotides. It
contains 767 EI, 768 IE, and 1655 None sequences. These sequences consist of A, C, T, G
nucleotides and also D, N, S, R ambiguity characters.
Cross validation method is used to find the overall ability of FWPSSM classification.10-fold cross-validation
method is applied to this data set. In this method, the dataset is broken into 10 sets and each fold contains 319
sequences. One fold is selected as test set and the rest folds are selected as learning set. This process is repeated
10 times and then the mean accuracy is taken.
1006
0.02
Triangular
Uniform
0.019
Weights
0.018
0.017
0.016
0.015
0.014
0.013
10
20
30
Positions
40
50
60
EI %
IE %
2.32
2.48
2.73
2.73
2.92
2.70
N
%
3.04
3.13
3.17
Overall %
2.70
2.84
2.87
Sn
Sp
0.9650
0.9625
0.9634
0.9770
0.9761
0.9753
Accuracy
0.9730
0.9716
0.9714
As it is seen from the tables 3 and 4, triangular membership function has better results than uniform
membership function and WPSSM.
4 Discussion
There are many soft computing methods in the literature for the classification problem of the splice sites. Some
of these algorithms are; PNN (Plausible Neural Network), ID3 (a decision tree algorithm), BP (Back
propagation), LVQ (a combination of procedures of LVQ), FS+LVQ (a LVQ learning with feature selection), 1nn (1-nearest neighbor neural network), FS1NN (1-nearest neighbor neural network with previous feature
selection), k-nn (k-nearest neighbor neural network), FSKNN (k-nearest neighbor neural network with previous
1007
feature selection) (Li et al., 2007), BRAIN (Rampone, 1998), KBANN (Knowledge Based Neural Network)
(Noordewier et al., 1991), and Hierarchical Multi Classifier (Lumini and Nanni, 2006).
When we compare the classification error rates of our FWPSSM with many methods proposed in the literature,
the FWPSSM have showed lower error rates than many known methods (Table 5). The methods which have the
higher results than our proposed FWPSSM are Hierarchical Multi Classifier Method (HM), Linear Support
Vector Machine (Rank SVM) and Support Vector Machine (SVM).
Table 5. Error rates of different methods
Methods
HM
Rank SVM
SVM
FWPSSM
WPSSM
Subspace
PNN
ID3
BRAIN
KBANN
BP
FS+LVQ
FSKNN
LVQ
K-NN
FS1NN
1-NN
EI %
IE %
N %
0,76
0,86
1,65
2,32
2,73
1,68
5
7,6
5,74
-
1,59
1,86
1,99
2,73
2,7
6,25
4
8,5
10,75
-
1,6
1,71
1,9
3,04
3,17
1,61
4
4,6
5,29
-
1.317
1.477
1.847
2,7
2,86
3,18
3,33
3,33
4.333
6,9
7,26
14,92
22,23
22,34
27,77
28,14
33,62
Finally, the methods depending on positional weights can be usually used for the fixed-length sequences. On
the other hand, variable-length sequences are usually used in the classification of the protein sequences and in
this situation, the sequences are encoded by overall frequencies instead of positional frequencies (Nasibov and
Kandemir-Cavas, 2008, 2009). But we hope that the proposed FWPSSM approach also can be successfully
applied to the analysis of the protein sequences, when fixed-length sequential data are handled.
References
1. Asuncion,
A.,
Newman,
D.J.
(2007)
UCI
Machine
Learning
Repository
(http://www.ics.uci.edu/~mlearn/MLRepository.html). Irvine, CA: University of California, School of Information
and Computer Science.
2. Burset, M., Seledtsov, A., Solovyeva, V.V. (2000) Analysis of canonical and non-canonical splice sites in
mammalian genomes. Nucleic Acids Research, 28(21):4364-4375.
3. Jang, JSR., Sun, CT., Mizutani, E. (1997) Neuro-Fuzzy and Soft Computing. Prentice-Hall, Inc. Simon &
Schuster/A Viacom Company Upper Saddle River, NJ07458.
4. Li K., Chang D., Rouchka E., Chen Y.Y. (2007) Biological sequence mining using Plausible Neural Network and
its application to exon/intron boundaries prediction. IEEE Symposium on Computational Intelligence in
Bioinformatics and Computational Biology. 165-169.
5. Lumini, A., Nanni, L. (2006) Identifying splice-junction sequences by hierarchical multiclassifier. Pattern
Recognition Letters, 27:13901396.
1008
Biographies
Sezin TUNABOYLU is graduated from the Department of Statistics at EgeUniversity in 2008. Now, she is master degree
student at the Department of Statistics at DokuzEyllUniversity since 2008. She is interested in bioinformatics especially
DNA sequences.
Efendi NASIBOV received the B.Sc. and M.Sc. degrees in Applied Mathematics Department from Baku State University,
Azerbaijan in 1983, and the Ph.D. in Mathematical Cybernetics and Dr.Sc. in Computer Science degrees from Institute of
Cybernetics Academy of Science of Azerbaijan in 1987 and 2003, respectively. His research interests are in the application
of data mining, fuzzy sets theory, optimization and decision making problems, and software engineering.
Prof. Nasibov is currently working as a full Professor and Chair of the Department of Computer Science,
DokuzEylulUniversity, Izmir, Turkey.
1009
Abstract. In this study, fuzzy logic graphical user interface software for evaluation of wooden material
combustion performance is realized. The most accurate monitoring of measurements obtained by the
combustion is provided. Operations on parameters that obtained from the result of combustion can be made
faster. Test performance is determined with fuzzy logic by using data obtained during the combustion process
and errors are reported to the user. Unexpected situations that experts could not notice can be determined
easily by this software during the combustion. It has been observed that this user interface software prevents
the data-loss and gives better results with sensitive measurements. Since repetition of the experiment is
reduced, especially time, work and energy savings are provided. This fuzzy logic user interface software with
these features can be used easily in wooden material combustion R&D centers, academic studies on this issue
and companies in this area.
Keywords: fuzzy logic software, real time, process control
1 Introduction
Wooden material is the oldest one in various building materials used by people. Brought technical innovations of
our age and despite of competing with the large number of new material, it keeps its importance in many fields
because of its superior properties today [1]. Many studies have been done with different combustion mechanisms
on the combustion properties of wooden material. When the importance of the combustion properties of wooden
material is taken into consideration in literature, combustion mechanisms cannot meet the needs of current
developing technology. Computer-aided combustion mechanism is needed to minimize the manual measurement
and human-induced errors and to increase the sensitivity of the parameters obtained by the combustion. There is
a need for software for developed computer-aided wooden material combustion mechanism. For this purpose, in
the light of technological advances, fuzzy logic graphical user interface software for evaluation of wooden
material combustion performance is designed.
Today, studies of the combustion are in the form of reading and recording data observations without the aid of a
computer. This case causes not only humanly errors but also measurement errors. Detecting errors beforehand
provides manpower and especially time saving. This software can be used in the field of wooden material
protection technology and in determining the combustion properties of wooden material. It shed lights on the
works about the protection of historical buildings against fire. Protection of wooden which is a combustible
organic material against fire and maintenance of our historical culture will be provided for the protection of
1010
historical wooden houses in case of a fire. It will contribute to the sustainability of our culture by protecting the
wooden material against the combustion danger.
Many studies have been done on the process control systems. The cases of sintering plant using fuzzy logic in
process control systems were investigated [2]. In one study, thermal comfort care application was studied with
fuzzy logic process control in measurement laboratories [3]. In another study, experimental investigations were
made for fuzzy logic process control [4]. Design, implementation and real-time operation of advanced process
monitoring techniques for wastewater treatment plants were presented with studies [5]. Distributed real-time
process control was designed for heating and power plant [6]. A model based on fuzzy logic classifier was
created in order to determine the values of modulus of elasticity and modulus of rupture of flakeboards. Using
these values, input and output values and rule base of fuzzy logic classifier were created. With the fuzzy logic
classifier model prepared in Matlab Simulink, modulus of elasticity and rupture values for flakes mixture ratios
were predicted [7]. Real time control of DC-DC buck converter with fuzzy logic controller was designed [8].
In this study, we realized fuzzy logic graphical user interface software for evaluation of wooden material
combustion performance.
1011
1012
1013
Fuzzy inference system is a part that makes real-time inference according to the measurement results coming
from data acquisition card. Firstly, start recording data button must be pressed in fuzzy inference system
interface shown in Figure 4 menu. After three seconds from this process, the start fuzzy inference button is
activated. Real time fuzzy inference system is activated by pressing this button. Input variables can be seen
under the inputs group and output variable can be seen under the outputs group. Rule base can be seen under the
rule base group. By clicking the pictures in these groups, shifting can be done on the related variable features
screen. Click on the triggered rules link to see the triggered rules during real time inference and click on the
results link to see the instant results of fuzzy inference.
It is necessary to press set input output option to screen for set input and output variables in fuzzy inference
menu. Input or output variable properties are shown under the input output group. Input or output variable name,
number of membership function and minimum and maximum value ranges are determined in this section. After
setting the variable structure of these features taking into account the characteristics of the same group, the save
operation is performed by pressing the save button. After this process the necessary properties are assigned to the
relevant variables. Properties of membership functions are stated for variables to be added under the membership
functions group. Graphical representation of the membership functions is given under the group of membership
functions graph. As an example, upper_temp input variable membership functions are given in Figure 5.
1014
Acknowledgements
This work is supported by The Scientific and Technological Research Council of Turkey (TUBTAK) under
project number 110O548.
References
1.
2.
3.
Bozkurt, Y. ve Gker, Y. 1981. Orman rnlerinden faydalanma. stanbul: . Orman Fakltesi Yaynlar.
Dinis, CM; Cuntan, CD; Iagar, A; Popa, GN; Anghel, SR. (2009). Aspects Regarding the Fuzzy Logic in the
Process Control From the Sintering Plants. In Proceedings of the 13th WSEAS International Conference on
Systems.Wisconsin, USA, 277-282.
Baus, ZL; Nikolovski, SN; Maric, PZ. (2008). Process control for thermal comfort maintenance using fuzzy logic.
Journal of Electrical Engineering 59 (1): 34-39.
1015
Fileti, AMF; Antunes, AJB; Silva, FV; Silveira, V; Pereira, JAFR. (2007). Experimental investigations on fuzzy
logic for process control. Control Engineering Practice 15 (9): 1149-1160.
Wade, MJ; Sanchez, A; Katebi, MR. (2005). On real-time control and process monitoring of wastewater treatment
plants: real-time process monitoring. Transactions of the institute of measurement and control 27 (3): 173-193.
Zhang, XH; Zhao, LH; Liu, Y; Ma, Y; Jiang, WJ; Wu, G. (2008). Design of Distributed Real-Time Process Control
for Heating and Power Plant. Progress in Safety Science and Technology 7: 420-422.
Yapici F., zcifci A., Akbulut T. and Bayir R. (2009). Determination of modulus of rupture and modulus of
elasticity on flakeboard with fuzzy logic classifier. Materials & Design 30: 2269-2273.
Ozcan, C., Uysal, A. and Bayir, R. (2009). Real time control via Matlab/Simulink of DC-DC buck converter with
fuzzy logic controller. In Proceedings of 1st International Fuzzy Systems Symposium. Ankara.
Kalogirou, S. A. (2003). Artificial intelligence for the modeling and control of combustion processes: a review.
Progress in Energy and Combustion Science 29, 515-566.
Sivanandam, S.,N., Sumathi, S., Deepa, S., N. 2007. Introduction to Fuzzy Logic using MATLAB. New York:
Springer.
Lee CC. (1990). Fuzzy logic in control systems: fuzzy logic controller. III, IEEE Trans. Syst., Man, Cybern.
20(2):404-435.
Zadeh, L.A (1968). Fuzzy algorithms. Information and Control 12 (2): 94-102.
Zadeh, L.A. (1965). Fuzzy sets. Information and Control 8 (3): 338-353.
Biographies
Caner ZCAN was born in Karabk, 1986. He graduated from Yldz Technical University in 2004. He received his
M.Sc. degree in Computer Engineering from Karabuk University.
He is interested in fuzzy logic, artificial intelligence and real time process control systems.
Research Assistant Ozcan still works for Karabuk University, Department of Computer Engineering.
Raif BAYIR was born in Karabk, 1973. He graduated from Gazi University in 1995. He received his M.Sc. degree from
the same university in 1998 and his Ph.D. degree in 2005.
He is interested in robotics, expert systems and fault diagnosis of electric machines.
Assoc. Prof. Dr. Bayir still works for Karabuk University, Technology Faculty, Department of Mechatronics Engineering as
Department Chair of Mechatronics Engineering.
Cemal ZCAN was born in Karabk, 1982. He graduated from Sleyman Demirel University in 2005. He received his
M.Sc. degree from the Karaelmas University in 2007 and his Ph.D. degree from the Bartn University in 2011.
He is interested in combustion properties, thermal conductivity and process control systems.
Research Assistant Dr. Ozcan still works for Karabuk University, Department of Furniture and Decoration Education.
1016
Abstract.We introduce an approach to reconstruct the residual (non-tidal) component of historical water
level data with artificial neural network. Long-term historical water level data is important for coastal studies
such as coastal flood hazard analyses. However, long-term water level observations are limited to specific
sites. The approach presented here uses the ANN methods and the correlation between the residuals of a
subordinate station and a reference station in order to estimate the historical residual of the subordinate
station. It is tested on the data of a reference station and three subordinate stations located in North San
Francisco Bay, California. The artificial neural network-estimated residuals are compared to the observed
residuals for the subordinate stations. Comparison results show that the historical residual of a subordinate
station can be estimated successfully using this approach in the availability of a strong correlation with the
residual of a nearby reference station.
Keywords: tidal residual, artificial neural network, reconstructing historical data.
Introduction
The water levels measured at tide stations include tidal (astronomic), and non-tidal components. The non-tidal
component of water level is caused by some meteorological effects such as high/low atmospheric pressure, high
winds and possibly by other effects such as riverine flow, and El Nino. Tidal elevation is the predicted elevation
of water surface solely due to astronomic forcing on water body. Tidal predictions are commonly calculated
using harmonic constituents at a given station [1]. The harmonic constituents for a station are calculated based
on the measured water levels at that station. Astronomic tide can be predicted accurately due to its periodic
nature. Contrary to the tidal component, it is not easy to predict the non-tidal component of the water level since
the driving forces are non-periodic, non-stationary and highly uncertain. As illustrated in [2], the non-tidal
component of water level can easily dominate the water level during strong meteorological events (e.g. very
high/low pressure, strong wind etc.). So, the non-tidal component of water level is as crucial as the tidal
component of water level for accurate water level estimation.
Artificial neural networks (ANNs) have been widely used in multivariate nonlinear time series
modeling in many research areas. ANN is capable of directly correlating the time series of multiple input
variables to the output variables through the interconnected nodes using trainable weights and bias signal [3].
There are several applications of ANN for water level or tidal level predictions in the literature. These
applications can be placed in two groups based on the aim of the study as: (1) forecasting future water levels;
and (2) estimating the historical water levels. One of the pioneering researchers of the first group, [4] developed
an ANN model to forecast one month ahead sea level using the past 12 months sea level data. [5] developed an
1017
ANN method coupled with auto regressive moving average method (ARMA) to forecast future water levels
based on the current and past water level data from the same station. [6] extended that application to stations
with mixed tides. [7] proposed the application of back-propagation neural network (BPN) combined with the
equation of harmonic analysis for the long-term prediction of the tidal level based on the previous observations
in the same station. [8] used four input factors such as the wind velocity, wind direction, pressure and harmonic
analysis-tidal level to forecast future water levels. [9] developed an ANN model to forecast the non-tidal
component of water level using previous non-tidal water levels measured at the same station, wind
measurements, and harmonic water level forecasts. [10] Chang and Lin (2006) used ANN to establish a
relationship between tidal elevations and tide-generating forces. They propose using their model to forecast the
tidal elevations. [2] forecasted water level anomaly (residual) reasonably well up to nine hours ahead using the
observed wind velocity at the entrance of Galveston Bay, Texas. [11] developed a typhoon-surge forecasting
model with a back-propagation neural network (BPN). They used the typhoons characteristics, local
meteorological conditions and typhoon surges at a considered tidal station at previous times as input data of the
model to forecast typhoon surges 13 hours ahead at the following time.
In the second group of applications, [12] used observed water levels at a reference station to estimate
the water levels at subordinate stations using ANN. [13] developed a regional neural network tool for coastal
water level predictions. They predicted long-term water levels in a coastal inlet based on the input of data in a
remote NOAA (National Oceanic and Atmospheric Administration) station in the region. Similar to [12], they
did not separate the tidal and non-tidal component of water level in their analyses; however, they examine the
ANN predictions of the non-tidal water level variations using 36-h low pass filtering. [14] Chen et al. (2007)
developed a coupled wavelet-ANN model to estimate the tidal elevation at a station using tide elevation data
from a nearby station. They remove the high frequency noise from the observed water elevation data before
using in ANN. [15] Liang et al. (2008) developed a minus-mean-value neural network method along with two
other ANN methods for non-stationary tidal level predictions. With the minus-mean-value neural network model
they predict the tidal level between two or more stations. They define the hourly residual as the observed hourly
water level minus the daily mean of observed water level. They set up two ANN models; one predicts the daily
mean water level, and the other one predicts the hourly residual based on the synchronous data at the reference
station. They add the two predictions from the ANN models to obtain the predicted hourly water elevation at the
given station.
The present study can be included in the second group of ANN applications on water/tidal levels. Here
an ANN method is developed to predict the historic non-tidal component of water level at a given station.
Basically, an ANN structure is optimized for the given data set and the optimized ANN structure is trained to
learn the correlation between the residual of a reference station and the residual of a subordinate station. The
optimized and trained ANN is then used to estimate the missing residual of the subordinate station from the
residual of the reference station. The main requirement in this methodology is the availability of a strong
correlation between the residuals of the reference and subordinate stations. Once the trained and validated ANN
is developed, output variables are directly calculated from the input variables and the vectors of weights and bias
in the network nodes.
The novelty of this study is that the tidal and non-tidal components of the observed water level are
separated and the ANN estimation technique is applied directly on the non-tidal component. This approach has
two significant advantages over the preceding approaches. First, by subtracting the tidal elevation from the
observed water level the periodic and deterministic component is eliminated from the signal (observed water
level). Therefore the nodes in the ANN model are dedicated to learn the correlation in the random non-tidal
component of the signal. Second, in calculating extreme water level with convolution method one needs the
probability density function of both the tidal and non-tidal components separately [16]. The presented approach
allows reconstructing the non-tidal component of water levels which can be used for developing the probability
density function for convolution method-extreme water level computations. The following sections will provide
the details of the selected ANN architecture, structure, and the training-testing methods used in this study;
various example applications; discussion of findings; and conclusion remarks.
1018
1.
In this study, multilayered perceptron neural network (MLPNNs) is preferred as the architecture of the neural
network owing its simplicity [17], [18]. Feed forward backpropagation (FFBP) with single hidden layer is used
as the ANN structure. In the training of neural network, K-fold cross validation is used in order to minimize the
bias associated with the random sampling of the training data [19].
2.
Example Applications
The proposed approach is applied on residuals from one reference station and three subordinate stations in North
San Francisco Bay operated by NOAA. The aim in these applications is reconstructing the residual records for
the subordinate stations. In order to be consistent in the analyses the study period is set as from January 1, 1898
to December 31, 2008 for all the stations.
The water level and tidal level data are retrieved from NOAA web site [1]. The time interval and
duration of the available water level records are listed in Table 1. It should be noted that, Table 1 only provides
the gross time intervals for the available water level records and the records contain many gaps. The tidal level
data is available for the entire study period for all stations.
Table 3. NOAA Tide Stations and Availability of Water Level Data
Station Name
Start
End
Duration (month)
San Francisco
1898
2008
1331
Richmond
1996
2008
149
Port Chicago
1979
2008
349
Mare Island
1985
2008
65
Both the water level and tidal level data are retrieved for hourly intervals, in meter and relative to Mean Lower
Low Water (MLLW) of the particular station. The hourly residuals are obtained by subtracting the tidal levels
from the available water levels. Then the daily peak residuals are extracted from the hourly residuals in order to
reduce the size of the time series. Working with daily data provides significant efficiency in the ANN training.
Besides the computational efficiency, daily peak data substantially covers the possible lag [2], [20] in the
residual data between the reference and subordinate stations. So, by using daily data we avoided (1) the necessity
for estimating the hourly time lag in the residuals between stations, (2) using multiple (lagged) inputs in the
ANN model. Furthermore, peak residuals in daily intervals provide enough precision for extreme water level
analyses as monthly peaks are used commonly in extreme level analyses. The San Francisco Station is used as
the reference station since it has the most complete water level record compared to the other stations. A separate
network is trained and used for residual estimation for each subordinate station, since the correlation structure
between the reference station and the subordinate station is expected to be different for each subordinate station.
The network is trained for the duration when the observed data is available for the subordinate station. The
quality of the training indirectly depends on the duration of the training data. The longer the training data is the
more the ANN exposed to the correlation structure between the stations. Furthermore, longer training data
duration increases the chance to cover a wider range of variation in the residuals.
As noted earlier, accuracy of the estimated residuals depends on the strength of the correlation between
the residuals of the reference and the subordinate station. The correlation between the residuals of the reference
and subordinate stations is measured by calculating the coefficient of determination (R2) for each subordinate
station. The calculated values of R2 are listed in Table 3. As in Table 3, the residuals of the subject subordinate
stations have a very strong to moderate positive correlation with the residual of the reference station.
1019
Some portion of the subordinate station data is excluded from the training. The excluded data is later
used for testing the performance of the ANN estimation. Table 2 lists the total available data, durations of data
included/excluded in the training and the time interval of the excluded data for each subordinate station.
Table 4.Residual Data Included/Excluded in the ANN Training
Station
Name
Total
Available
(day)
Included in
the Training
(day)
Excluded in
the Training
(day)
Richmond
4452
3599
853
20060706
20081231
Port Chicago
10492
9492
1000
19820524
19850216
Mare Island
1853
859
994
20060330
20081231
Excluded
Start
End
(Y/M/D)
(Y/M/D)
The values of R2 for residuals of the subordinate stations compared to the residual of the reference station based
only on the data excluded in the training are also listed in Table 3.
Table 5.Coefficients of Determination (R2) for the Observed Residuals at the Subordinate Stations compared to
the Observed Residual at the Reference Station
For All Available Data
0.90
0.88
Port Chicago
0.74
0.83
Mare Island
0.70
0.73
Station Name
Richmond
For the Richmond and Mare Island Stations; the latest period of the available time series data are excluded from
the training. For the Port Chicago Station; specifically a time interval when a strong meteorological effect on the
measured water level was observed is excluded from the training. Table 3 shows that, unlike for the other two
stations, for Port Chicago station the values of R2 for the excluded data is significantly higher than the values of
R2 for the all available data. Intuitively, a strong meteorological event is expected to affect the water level
elevations at two nearby stations similarly; while a relatively insignificant meteorological forcing may not affect
the water levels at one station as strongly as it affects the water level at a nearby station. The R2 tests for the
included/excluded data of the Port Chicago Station support that supposition. The performance of the ANN
Model depends on the strength of the correlation between the residuals of the subordinate and reference stations.
Therefore, ANN is expected to perform better for the times when there are strong meteorological events. This is
a significant advantage of using ANN for reconstructing the historical residual data for extreme water level
analyses.
In order to find the optimum ANN structure several networks with different structures are trained and
tested. The parameters used in these training-testing processes and the properties of the best network architecture
are given in Table 4. As listed in Table 4, the number of nodes in the hidden layer for an optimum ANN
structure is determined as 22, 14, and 13 for the Richmond, Port Chicago, and Mare Island Stations respectively.
One input node and one output node are used for all stations. Accordingly, in Table 4 the best ANN architecture
is denoted as 1:22:1 (number of nodes in input layer: number of nodes in hidden layer: number of nodes in
output layer) for the Richmond Station. The number of iterations to achieve to the best structure is also listed in
Table 4. The scan interval for the iterations is from 100 to 10000 iterations with 100-iteration increments.
Similarly, the number of nodes in the hidden layer is tested from 2 to 35 nodes with one-node increments. The
starting learning rate for the variable learning rate backpropagation training algorithm is set to 0.7. The learning
rate is updated by the training algorithm at each iteration. In this study the momentum coefficient is kept
constant. Next, the performance tests of the optimized ANN structures which are already trained to learn the
correlation structure between the residuals of the reference and subordinate stations are presented.
1020
Richmond
Port Chicago
Mare Island
Best Network
Architecture
1:22:01
achieved after 5700
iterations
1:14:01
achieved after
2800 iterations
1:13:01
achieved after 4700
iterations
100-10000
step=100
100-10000
step=100
100-10000 step=100
2-35 step=1
2-35 step=1
2-35 step=1
Variable Learning
Rate
Backpropagation
Variable
Learning Rate
Backpropagation
Variable Learning
Rate
Backpropagation
0.7
0.7
0.7
0.9
0.9
0.9
K-Fold CrossValidation
Scanned
Iteration Interval
Scanned Hidden
Nodes
Training
Algorithm
Starting
Learning Rate
Constant
Momentum
Coefficient
3.
Station Name
Excluded Data
Richmond
0.90
0.87
Port Chicago
0.74
0.82
Mare Island
0.69
0.68
It is demonstrated that in the availability of a nearby reference station with long-term historical water level
record, the residual data of the subordinate station can be estimated using the correlation between the residuals of
the reference and subordinate stations. The performance tests on three subordinate stations indicate that the
accuracy of the ANN estimates is limited with the strength of the correlation between the residuals of the
reference and subordinate station.
1021
The water elevation at a tide station near an estuary is influenced by the riverine discharge. That
influence usually reduces the correlation between the residuals of the reference and subordinate stations. In such
a case, since it depends on the correlation between the residuals of the stations, the accuracy of the ANN
estimations will also be low. Future studies can test the ANN models which include the river discharge as an
additional input in order to improve the overall accuracy of the estimated residual at the subordinate station.
With such an ANN structure the model accuracy is expected to be not limited to the strength of the correlation
between the residuals of the reference and subordinate stations.
References
1.
2.
3.
4.
5.
6.
7.
8.
9.
10.
11.
12.
13.
14.
15.
16.
17.
18.
19.
20.
National Oceanic and Atmospheric Administration (NOAA). (2009a). Web Site: http://opendap.coops.nos.noaa.gov/
Cox, D.T., Tissot, P., Michaud, P., (2002). Water level observations and short-term predictions including
meteorological events for entrance of Galveston Bay, Texas. Journal of Waterway, Port, Coastal and Ocean
Eng., 128 (1), 2129.
Hagan, M.T., Demuth, H., Beale, M., (1995). Neural Network Design. PWS Publishing Company, Boston,
MA.
Vaziri, M., (1997). Predicting Caspian Sea surface water level by ANN and ARIMA models. Journal of
Waterway, Port, Coastal and Ocean Eng., 123, 158162.
Tsai, C.-P., Lee, T.-L., (1999). Back-propagation neural network in tidal level forecasting. Journal of
Waterway, Port, Coastal and Ocean Eng., ASCE 12 (4), 195202.
Lee, T.-L., Jeng, D.S., (2002). Application of artificial neural networks in tide forecasting. Ocean Eng., 29 (9),
10031022.
Lee, T.-L., (2004). Back-propagation neural network for long-term tidal prediction. Ocean Eng., 31 (2), 225
238.
Lee, T.-L., (2006). Neural network prediction of a storm surge. Ocean Eng., 33, 483494.
Tissot, P., Cox, D., Sadovski, A. Michaud, P., Duff, S., (2004). Performance and Comparison of Water Level
Forecasting Models for the Texas Ports and Waterways. Proc., Ports 2004: Port Development in the Changing
World, Conf., ASCE. 0-7844-0727-4.
Chang, H.-K., Lin, L.-C., (2006). Multi-point tidal prediction using artificial neural network with tidegenerating forces. Coastal Eng., 53, 857864.
Tseng, C.M., Jan, C.D., Wang, J.S., Wang, C.M., (2007). Application of artificial neural networks in typhoon
surge forecasting. Ocean Eng., 34, 17571768.
Deo, M.C., Chaudhari, G., (1998). Tide prediction using neural networks. Computer-Aided Civil and
Infrastructure Eng., 13 (2), 113120.
Huang, W., Murray, C., Kraus, N., Rosati, J., (2003). Development of a regional neural network for coastal
water level predictions. Ocean Eng., 30 (17), 22752295.
Chen, B.-F., Wang, H.-D., Chu, C.-C., (2007). Wavelet and artificial neural network analyses of tide
forecasting and supplement of tides around Taiwan and South China Sea. Ocean Eng., 34, 21612175.
Diamantidis, N. A., Karlis, D., and Giakoumakis, E. A. (2000). Unsupervised stratification of cross-validation
for accuracy estimation. Artif. Intell., 116(1-2), 1-16.
Walden, A. T., Prescott, P., Webber, N.B., (1982). An Alternative Approach to the Joint Probability Method
for Extreme High Sea Level Computations. Coastal Eng., 6 (1982), 71-82.
Haykin, S. (1994). Neural Networks: A Comprehensive Foundation. New York: Macmillan College
Publishing Company.
Maren, A., Harston, C. & Pap, R. (1990). Handbook of Neural Computing Applications. London: Academic
Press.
Francois, D., Rossi, F., Wertz, V., & Verleysen, M. (2007). Resampling methods for parameter-free and
robust feature selection with mutual information. Neurocomputing, 70(79), 12761288.
doi:10.1016/j.neucom.2006.11.019.
Liang, S.X., Li, M.C., Sun, Z.C., (2008). Prediction models for tidal level including strong meteorologic
effects using a neural network. Ocean Eng., 35, 666675.
1022
Biographies
smail Halta (Ankara, 1980). B.S., Civil Engineering, Middle East Technical University, Ankara, Turkey, 2002. M.S.,
Civil and Environmental Engineering, University of California, Davis, CA, USA, 2004. Ph.D., Civil and Environmental
Engineering, University of California, Davis, CA, USA, 2006.
He has published several journal papers on scale issues in hydrologic processes. His research interest includes scaleinvariance in hydrology, flood modeling, floodplain mapping, statistical methods in hydrology.
Dr. Haltas is a member of American Society of Civil Engineers (ASCE) and Environmental and Water Resources Institute
(EWRI).
Nihat Ylmaz (Konya, 1975). B.S., Electrical-Electronic Engineering, Seluk University , Konya, Turkey, 1996. M.S.,
Electrical-Electronic Engineering, Seluk University , Konya, Turkey, 1998. Ph.D., Electrical-Electronic Engineering, Seluk
University , Konya, Turkey, 2005.
He has published several journal papers on Artificail Neural Network Modeling . His research interest includes robotics,
expert systems, remote sensing.
Dr. Ylmaz is a member of Institute of Electrical & Electronics Engineers (IEEE).
1023
1. Introduction
Many real world problems can be defined using the Linear Programming model.Decision space is defined based
on the objective function and constraints while the type of decision is deciding under certainty. The Linear
Programming assumes that the parameters A,c,b are all positive numbers. Therefore, creating the Linear
Programming problem requires high-level knowledge. The precision, reliability, and certainty of the data in real
life applications usually demonstrate a fuzzy structure. For this reason, the optimal solution of a linear
programming problem is valid only for some part of the constraints. Thus, most of the available data sets have an
interaction which is not very strong within the solution. The process of decision making in a fuzzy environment
defined by Belman and Askerzadeh has yielded the usage of the Fuzzy Set Theory in their studies on a linear
programming.The initial formulation of Fuzzy Linear Programming was created by Zimmermann
(1978).Later,various approaches were developed by other scientists and categorized.
Formulations in fuzzy linear programming problems are created using membership functions. The membership
degree of problem parameters ascend/descend between 0 and 1. On the other hand, some uncertainty or
hesitation can be observed in membership degrees. Gau and Buehrer have suggested the concept of vague sets
for the uncertainty of membership degrees.
2. Vague Sets
x . A vague set V in X is
characterized by a truth-membership function tv and a false-membership function f v . tv is lower bound on the
Defination 1:Let X be a space of points, with a generic elementof X denoted by
1024
x . tv and f v both associate a real number in the interval [0,1], with each
fv 1 .
That is
t v : X [0,1]
f v : X [0,1]
This approach bounds the grade of membership of x to a subinterval [t v (x),1 f v (x)] of [0,1]. Other words, the
exact grade of membership v ( x ) of x may be unknown, but is bounded by t v (x) v (x) 1 f v (x) , where
t v (x) f v (x) 1 .
The precision of our knowledge about x is immediately clear, with ouruncertainty characterized by the difference
1 f v (x) t v (x) . If this is small, our knowledge about xis relatively precise; if it is large, we know
correspondingly little. If 1 f v (x) is equal to t v ( x ) ,our knowledge about x is exact, and the theory reverts back
to that of fuzzy sets. If 1 1 f v (x) and t v ( x ) are both equal to 1 or 0, depending on whether xdoes or does not
belong to V, our knowledge about xis very exact and the theory reverts back to that of ordinary sets (i.e., sets
with two-valued characteristic functions).
When X is continuous a vague set V can be written as
V [t V (x),1 f V (x)] / x, x X
X
V [ t V ( x ),1 f V ( x )] / x, x X
i 1
The vague set theory can be interpreted using a voting model. The number of total voting people is assumed to
be 10. Suppose that A is a vague set in U,
f A (u) 1 0.8 0.2 . Then, we can say that the degree of u A is 0.6, and the degree of u A is 0.2. The
vague value [0.6,0.8] can be interpreted as the vote for a resolution is 6 in favor, 2 against, and 2
abstentions(Lui, Wang, Feng,2008).
1025
Max Z cx
Subject to
(3.1)
~
(Ax) i b i
x 0, i 1,2,..., m
~
Where bi , i , are in bi , bi pi with given p i (yolerance).Werners proposed that the objective function of
Equation (3.1) should be fuzzy because of fuzzy total resources or fuzzy inequality constraints. Then for Eqution
(3.1),Werners first defined Z 0 and Z l as follows:
Max Z 0 cx
Subject to
(3.2)
(Ax) i b i
x 0, i 1,2,..., m
Max Z1 cx
Subject to
(3.3)
(Ax) i b i p i
x 0, i 1,2,..., m
Optimal solution will be in between Z 0 and Z l , the satisfaction of the optimal solution will increase when its
value increases.The membership function o of the objective is:
1
1
Z cx
o ( x ) 1 1
0
Z Z
0
cx Z1
, Z 0 cx Z1
,
(3.4)
cx Z 0
With the above membership function, we can then use the max-min operator to obtain optimal decision. Then
equation (3.1) can be solved by solving max x 0
1026
That is:
Max
Subject to
(3.5)
cx [Z Z ] Z
(Ax) i p i b i p i
0,1 and x 0
1
The vague linear programming proposed herein utilizes the steps of the approach described above. All resources
of the proposal are vague sets and the steps are explained as follows:
Step 1: First, the values of z0 and z1 are computed using equations (3.2) and (3.3).
Step 2: Next, the true and false membership functions for objective function and constraints are determined using
z0 and z1.
o
,
if cx z l
(z l cx )
, if z o cx z l
t ~o ( x ) o 1 l
o
z z
0
,
if cx z o
(3.6)
(Ax) b
i
i
t ~b ( x ) i 1
i
pi
0
(3.7)
if (Ax) i b i
, if b i (Ax) i b i p i
,
if (Ax) i b i p i
vo
,
if cx z l
(z l cx )
, if z o cx z l
1 f ~o ( x ) v o 1 l
o
z z
0
,
if cx z o
vi
(Ax) b
i
i
1 t ~b ( x ) v i 1
i
pi
(3.8)
if (Ax) i b i
, if b i (Ax) i b i p i
,
if (Ax) i b i p i
(3.9)
Step 3: Two different LP problems are defined depending on the true and false membership functions which
were determined for the objective function and constraints.
Max
1027
Subject to
cx [ Z1 Z 0 ].
(3.10)
Z0
bi pi
(Ax) i p i
Max
Subject to
(3.11)
cx [ Z1 Z 0 ]. Z 0
(Ax) i p i b i p i
4.Numerical Example
Let us consider a product-mix selection problem.
Max.Z 3x 1 5x 2 4x 3
Subject to
(4.1)
x 1 x 2 x 3 60
(Material A)
2x 1 5x 2 3x 3 180
(Material B)
x 1 3x 2 4x 3 130
(Material C)
2x 2 5x 3 150
(Material D)
Let us assume that the available total material A,material B, Material C and material D are imprecise and their
maximum tolerances are 10, 20,15 and 5 units respectively. Then, the membership functions of constraints are:
1
,
x 1 x 2 x 3 60
60
15
2
3
1 ( x ) 1 1
, 60 x 1 x 2 x 3 70
70
10
0
,
x 1 x 2 x 3 70
1
,
2x 1 5x 2 3x 3 180
2x 1 5x 2 3x 3 180
2 ( x ) 1
, 180 2x 1 5x 2 3x 3 200
200 180
0
,
2x 1 5x 2 3x 3 200
1028
1
,
x 1 3x 2 4x 3 130
x 3x 2 4x 3 130
3 ( x ) 1 1
, 130 x 1 3x 2 4x 3 145
145 130
0
,
x 1 3x 2 4x 3 145
1
,
2x 2 5x 3 150
2x 5x 3 150
4 ( x ) 1 2
, 150 2x 2 5x 3 155
155 150
0
,
2x 2 5x 3 155
Then, solving Equation (3.2) and (3.3.), we obtain: Z o 224.29 and Z l 255 . The membership function o of
the objective is defined as:
255 3x 1 5x 2 4x 3
O ( x ) 1
255 224.29
3x 1 5x 2 4x 3 255
, 224.29 3x 1 5x 2 4x 3 255
,
3x 1 5x 2 4x 3 224.29
Max
Subject to
3x 1 5x 2 4x 3 30.71 224.29
x 1 x 2 x 3 10 70
2x 1 5x 2 3x 3 20 200
x 1 3x 2 4x 3 15 145
2x 2 5x 3 5 155
x 1 , x 2 , x 3 0 and [0,1]
The solution is:
Step 1: The solution of equations (3.2) and (3.3) gives Z o 224.29 and Z l 255 .
Step 2: The true and false membership functions for the objective function and constraints are determined as
follows:
0.9
,
3x 1 5x 2 4x 3 255
(
255
3x
5
x
4
x
)
1
2
3
t ~o ( x ) 0.91
, 224.29 3x 1 5x 2 4x 3 255
255
2294
.
29
0
,
i3x 1 5x 2 4x 3 224.29
1029
0.93
,
3x 1 5x 2 4x 3 255
(255 3x 1 5x 2 4x 3 )
1 f ~o ( x ) 0.931
, 224.29 3x 1 5x 2 4x 3 255
255 2294.29
0
,
i3x 1 5x 2 4x 3 224.29
0.9
,
x 1 x 2 x 3 60
x x 2 x 3 60
t ~b ( x ) 0.91 1
, 60 x 1 x 2 x 3 70
1
70 60
0
,
ix 1 x 2 x 3 70
0.95
,
x 1 x 2 x 3 60
x x 2 x 3 60
1 f ~b ( x ) 0.951 1
, 60 x 1 x 2 x 3 70
1
70 60
0
,
ix 1 x 2 x 3 70
0.85
,
2x 1 5x 2 3x 3 180
2x 5x 2 3x 3 180
t ~b ( x ) 0.851 1
, i180 x 1 5x 2 3x 3 200
2
200
180
0
,
2x 1 5x 2 3x 3 200
0.9
,
2x 1 5x 2 3x 3 180
2x 5x 2 3x 3 180
1 f ~b ( x ) 0.91 1
, i180 x 1 5x 2 3x 3 200
2
200
180
0
,
2x 1 5x 2 3x 3 200
0.93
,
x 1 3x 2 4x 3 130
(Ax) 3 130
t ~b ( x ) 0.931
, 130 x 1 3x 2 4x 3 145
3
145
130
0
,
x 1 3x 2 4x 3 145
0.96
,
x 1 3x 2 4x 3 130
x 3x 2 4x 3 15
1 f ~b ( x ) 0.961 1
, 130 x 1 3x 2 4x 3 145
3
145
130
0
,
x 1 3x 2 4x 3 145
0.8
,
i 2x 2 5x 3 150
2
x
5
x
150
3
t ~b ( x ) 0.81 2
, 150 2x 2 5x 3 155
4
155
150
0
,
2x 2 5x 3 155
1030
0.85
,
i 2x 2 5x 3 150
2x 5x 3 150
1 f ~b 3 ( x ) 0.851 2
, 150 2x 2 5x 3 155
155 150
0
,
2x 2 5x 3 155
Step 3: The problem,defined using the true membership function and the false membership function, is as shown
below.
Max
Subject to
3x 1 5x 2 4x 3 34.122 224.29
x 1 x 2 x 3 11.11 70
2x 1 5x 2 3x 3 23.529 200
x1 3x 2 4x 3 16.129 145
2x 2 5x 3 6.25 155
x 1 , x 2 , x 3 0 and [0,0.8]
3x 1 5x 2 4x 3 33.02 224.29
x 1 x 2 x 3 10.53 70
2x 1 5x 2 3x 3 22.22 200
x 1 3x 2 4x 3 15.625 145
2x 2 5x 3 5.88 155
x 1 , x 2 , x 3 0 and [0,0.85]
x1
x2
x3
Z
35.714
35.77
35.85
15.36
15.07
15.106
13.9
14.21
14.14
Z 239.5
Z 239.5
Z 239.64
0.499
0.445
0,465
1031
The vague value [0.445;0.535] indicates that the exact grade membersip
O ( x ) is bounded by
1.0
Against Region
0.93
0.9
Hesitation Region
Support Region
0.0
cx
224.29
255
5. Conclusion
The proposed approach shows that the degree of hesitation is also a significant factor in explaining the results of
the assumptions on the membership degree 1.0 in classical fuzzy linear problems.In order to get the lowest
degree of the hesitation region as low as possible, the values of tolerance have to be determined precisely. There
is no certain way to determine the value of tolerance in the problems of fuzzy linear programming. Currently, the
value is only determined by means of the information obtained from the decision maker. At this point, an
important assessment has to be made: If the tolerance value is given for constraints with a slack variable larger
than zero, the solution will result larger slack value than expected. As larger slack value has been yielding a
lower degree of membership,at the other side a hesitation region may remain at higher value.In order to
determine the tolerance value the ranging intervals must be evaluated by using sensitivity analysis . It is also
possible to fix the slack values to zero and rearranging the constants on the right hand side by DeNovo
Programming. Finally, in a fuzzy linear programming approach, more effective solution can be achieved by
arranging correct tolerance values.
1032
References
1.
2.
3.
4.
5.
6.
7.
R.E. Bellman, L.A. Zadeh, Decision-making in a fuzzy fuzzy system reliability, Presented in National
Conf. On environment, Management Sciences 17 (1970) B141-B164
W.L. Gau, D.J. Buehrer, Vague sets, IEEE Trans. On of India, INDIA. Systems, Man, and Cybernetics
23 (1993) 610-614.
Amit Kumar, Kavita Devi, Shiv Prasad Yadav (2009), Method To Solve Linear Programming Problems
Using Vague Sets, 2009 IEEE International Advance Computing Conference (IACC 2009) Patiala,
India, 6-7 March 2009
N. Ramakrishna,(2009) Optimization Through Vague Sets, Internatonal Journal Of Computatonal
Cognton Vol. 7, No. 1, March 2009
B.Werners. An interactive fuzzy programming system, Fuzzy Sets and Systems 23 1987, 131-147.
Liu Yong, Wangl Guoyin and Lin Feng, General Model For Transforming Vague Sets nto Fuzzy Sets,
M.L. Gavrilova Et Al. (Eds.): Trans. On Comput. Sci. II, LNCS 5150, Pp. 133144, 2008
H. J. Zimmermann. Fuzzy Programming And Linear Programming With Several Objective Functions.
Fuzzy Sets And Systems, 1:4555, 1978.
1033
1034
Chapter 6
ARCHITECTURE
1035
Abstract. Today, precision in architectural design has attained a new and a broad meaning with the advent of
computational design and manufacturing/fabrication technologies, i.e with CAAD/CAM. The architectural
design process which requires integration of several knowledge domains, handling diverse and mass amount
of data and their processing has found new expansions with computational tools. Construction process has
also been undergoing a transformation and construction has been turning more into manufacturing and
fabrication process. Terms like customized mass production, concurrent architecture, and integrated design
have become a part of architectural design vocabulary. All these paradigmatic changes force architects and
thus architectural design education to be acquainted with CAAD/CAM and thus computational design
paradigm and a new level of complexity. In this context, design process itself searching for the optima in all
the levels, and diverse models which are any time products of design process, their reproducibility in
different phases of the design process in various scales with different fabrication/manufacturing technologies,
their explorations and experiencing through simulations, bring subject of precision in the realm of
architecture as a mean of controlling the design process in terms of complexity and in terms of intactness. In
this paper, this new role of precision and thus computational design process in the realm of performance
based design has been exemplified by the term project given in spring 2010 Digital Design Studio given by
Assoc. Prof. Dr. ArzuGnenSorgu and Res. Asst. FatmaZehraakc in METU, namely the modulator. In
the design, manufacturing and fabrication processes of the modulator, students have been introduced with
several level of precision and reflect their experience in 1/1 scale model of their design process or the product
itself.
Keywords: Arch 475, CAAD/CAM, complexity, design optimization, modulator, performance based design,
precision
1 Introduction
Architecture as one of the oldest disciplines has always been one of the major actors responsible for shaping the
world we live in. Architects are expected to orchestrate materials, forms, functions together with diverse
knowledge fields to provide quality spaces and thus built environment responding men's needs. From tents to sky
scrapers architecture has been continuously experiencing available technologies of the era they belong to. In this
regard, architecture can be considered as the manifestation of "the present state of technology" with all the pros
and cons.
Therefore, forms, materials and construction techniques used in the buildings reflect the design and its
integration with other domains, involving in the architecture, technology and "the level of precision" in the
whole process. Precision as the term to define the quality of being reproducible in amount or performance or
level of accuracy [1], has been incorporated in the 20 th century architecture mostly as a parameter to define the
intactness of the several forms in their dimensional accuracies as the state of technology allows.
1036
1037
dimensionality of these computational models as the medium of integration of form, functions structures and
materials, under the influence of any external and internal force(s) can be considered as analogous to the
complexity of programming and precision and accuracy as the measure in the architectural design.
Consequently, architectural design vocabulary has been enriched by the terms like emergence, genetic
algorithms, self-organizations, evolutionary algorithms digital morphogenesis, design informatics and etc.
The resulting complex and mostly manifested as non-Euclidian forms, generated by the design processes and
their geometric constructions for the material/(immaterial) world force architects to device the algorithms of
their own. This compels the understanding of geometry and proper application of algorithmic techniques as Berg
and et.al stated [6].
1038
5 The Modulator
Students are assigned to design a space modulator which can exist anywhere and should reflect the
environmental forces (internal and external) where it is situated [7-11]. This very short and rather ill-defined
design problem aimed to provoke discussions on
Parametric design and parameterization of the problem by defining first the problem then the variables
and parameters of the design
Shift from designing the end product to the designing process by discussing complexity, systems and
their architecture as well as precision as a measure of the complexity
Digital morphogenesis, generic/or special algorithms developed for form generation
Different CAAD environments, the out coming models and their potentials, shortcomings regarding
the requirements of the design performance and the process, and precision as a measure to control the
integration of design process with all the domains and actors involving in it
Experiencing the design by simulations and the level of precision required in these simulations
Precision in the materialization from virtual to physical
Precision in fabrication/manufacturing/construction regarding the design process and model generated
in the process
The timeline of the detailed design process is shown in Fig. 1 including different temporal outcomes i.eany
time models of the process.
Fig. 1. Time line of the Modulator Design Process and Any Time models
Instructor ArzuGnenSorgu, Assistant FatmaZehraakc, Students; CemKorkmaz,
Fratzgenel, Mate Csocsics
BelamirKse,
As it is shown in Fig. 1, in the design and fabrication of the modulator, students had chance to experience
performance based design and be acquainted with computational design as well. Students, Korkmaz, Kse,
zgenel and Csocsics, in the search of optimal solution for the modulator, were also accustomed with
optimization process which is the backbone of the performance based design idea. Precision of different phases
of the design process served as the major feedback and feed forward information/knowledge generated in the
design process determines the convergence of the design as well.
Mapping the design process or providing models for it is still a subject of concern since creativity and
intuition is a part of the design. The design process diagram (model) for the modulator is shown in Fig. 2.
1039
6 Conclusion
In this paper, the impact of computational and information technologies in architectural design and as an
outcome computational design paradigm and performance based design approach in architecture have been
briefly discussed. It should be pointed out that methods and models used in computational design, the design
process itself exhibiting complexity and requiring a high level of system integration have resemblance with
problems dealt in software engineering and complexity science.
Regarding this fact, precision gained a new meaning in architectural design. It is not only a measure of
reproducibility but it can be defined as a measure of complexity of the design process controlling the architecture
of the design process as well as providing the feedback in the search of optimum performance. Moreover,
precision controlling the whole design process and integration of the knowledge/information with its hybridized
design model force to develop new instructional models for architecture. Thus understanding the complexity and
precision is essential for architecture transforming by the computational technologies.
1040
References
WordNet, a large lexical database of English [On-line], Retrieved February 24, 2011 from the World Wide Web:
http://wordnetweb.princeton.edu/perl/webwn.
2. Bachman, L. (2003) Integrated Buildings: The System Basis of Architecture John Wiley and Sons.
3. Oxman, R. (2008) Performance-based Design: Current Practices and Research Issues, International Journal of
Architectural Computing, issue 01, vol. 06.
4. GnenSorgu A., Arslan S. (2009) Art and Literature as a Teaching/Learning Interface of Mathematics for
Students of Architecture, In Proceedings of ECAADE 2009 Conference, Istanbul, 16-19 September, pp. 465-472.
5. GnenSorgu A. (2010) "MimarlktaSaysalTeknolojilerinKullanm: YeniTeknolojilerveHibridleenMalzemeler",
MimarlktaMalzemeDergisi, Year 5, No 15, pp. 41-46.
6. Berg M., Cheong O., Kreveld M., Overmars M. (2008) Computational Geometry Algorithms and Applications,
Springer-Verlag Berlin Heidelberg.
7. Arch 475 Advanced Digital Design Studio, Official Course webpage [On-line], last Retrieved May 1, 2011 from
the World Wide Web: http://archweb.metu.edu.tr/dds/index.htm.
8. METU Arch 475 DDS II, Course Blog [On-line], last Retrieved May 1, 2011 from the World Wide
Web:http://metuarch475.blogspot.com/
9. zgenel, F. The Modulator [On-line], last Retrieved May 1, 2011 from the World Wide
Web:http://www.grasshopper3d.com/forum/topics/the-modulator.
10. ODT MimarlikBlmrencilerininGelecekTasarm: 'the MODULATOR' [On-line], last Retrieved May 1,
2011 from the World Wide Web: http://www.yap.com.tr/Haberler/odtu-mimarlik-bolumu-ogrencilerinin-gelecektasarimi-the-modulator_80776.html.
11. Fibercement; The METU students meet the MODULATOR [On-line], last Retrieved May 1, 2011 from the
World Wide Web:http://www.prefabrikyapi.com/english/guncel/the-modulator.html.
1.
Biographies
Arzu Gnen Sorgu born in Ankara, graduated from Middle East Technical University De-partment of
Mechanical Engineering. She completed her Master and Ph.D studies in the same department on Dynamic
Systems and Control. She was also employed as research assistant during her graduate studies. She went also to
Technical University of Delft to conduct experimental part of her PhD studies by NATO. After the completion
of her Ph.D thesis, she went to Tokyo Institute of Technology for post doctoral studies and since then she visits
TIT as visiting professor conducting research studies and projects. Currently she is working as an associate
professor in Department of Architecture in METU in the field of computational design.
She is conducting her studies on computational design and technologies, performance based design, acoustics
and control. Her recent publications include Teaching Mathematics in Archi-tecture (2005), Role of System
Architecture in Developing New Drafting Tools (2005), The Evolving Role of Mathematics in Designing in
Digital Medium (2006), Yapay Zeka Aratrmalar ve Biomimesis Kavramlarnn Gnmzde Mimarlk
Alanndaki Uygulamalar: Akll Mekanlar (2006), Refining technique for multilevel graph k-partitioning and
its appli-cation on domain decomposition non overlapping Schwarz technique for urban acoustic pollu-tion
(2009), Bilgisayarak renmek, Bilgisayarla renmek (2009), Art and Literature as a Teaching/Learning
Interface of Mathematics for Students of Architecture (2009), Origamics in Architecture: A Medium of Inquiry
for Design in Architecture (2009).
Fatma Zehra akc born in orum, graduated from Seluk University Department of Archi-tecture. She
completed her Master studies in Middle East Technical University Department of Architecture. Currently, she
conducts her Ph.D studies in the same department on Building Science while employing as research assistant in
METU.
She studies on performance based design, energy performance in buildings, and digital design and fabrication
technologies in architecture.
1041
Abstract. This paper generally aims to investigate the methodologies to increase the form flexibility of
common deployable scissor structures. For this purpose, the Modified Scissor-like Element (M-SLE) which
has been proposed in the articles of Akgn et al. [1-3], have been applied to the common spatial scissor
systems; and the form flexibility of this hybrid systems have been investigated. In order to investigate the
behavior and feasibility of these hybrid systems, the study has employed architectural computer simulations
as the main research methodology. During the study, SolidWorks 2011 has been used as the computer
simulation media. According to the obtained data from these simulations, the proposed system is evaluated in
an architectural aspect, and the potential applications are discussed.
Keywords: Deployable Scissor Structures, Kinetic Architecture, Architectural Simulations
1 Introduction
Deployable structures are structures capable of large configuration changes in an autonomous way. Most
common configuration changes are those that change from a packaged, compact state to a deployed, large state
[4]. Usually, these structures are used for easy storage and transportation. When required, they are deployed into
their service configuration. Simple examples include umbrellas or tents; or solar arrays and antennas on
spacecraft, which have to be compactly stowed for launch, but then must unfold autonomously to their final
state.
Scissor mechanisms are typical examples of deployable structures and many key researchers have proposed
different systems using scissor mechanisms. To form the primary unit of a scissor mechanism, which is called as
Scissor-like Element (SLE), two bars are connected to each other at an intermediate point through a pivotal
connection which allows them to rotate freely about an axis perpendicular to their common plane but restrains all
other degrees of freedom (DoF) [4] (See Fig.1a). By addition of intermediate elements to the end points of SLEs,
which allow perpendicular connections, a spatial scissor unit can be obtained as well. This spatial primary unit is
called Spatial Scissor-like Element (S-SLE) (See Fig.1b). In addition to these common primary elements, Akgn
et al. [1-3] have proposed a novel primary element called Modified Scissor-like Element (M-SLE). An M-SLE is
obtained by the connection of four struts by three hinges on a common point. Thus, each of the four struts can
rotate freely (See Fig.1c). When an M-SLE is used in a scissor mechanism, it increases DoFs of the whole
system and the form flexibility as well.
1042
a.SLE
b. S-SLE
c. M-SLE
1043
a
a
c
b
As the first application, a spatial scissor system which is composed of the units in Fig. 2b and Fig. 2a have
been assembled in SolidWorks medium. In the final product, there are 7 SLEs and 2 spatial M-SLEs which are
shown in Fig.2a. The assembly in Fig. 2b has four M-SLEs (two at each parallel structure). After constructing
the assembly, the system has been forced to transform. The beginning and the result shapes are presented in
Fig.3.
At first sight, it can be thought that the transformation capability of the first version should be very high.
However, as it is seen in Fig.3, the structure in Version 1 can only meet homogeneous arcs. Thus, the mechanism
meets the shape alternatives which keep 2, 3 and 4 angles equal. Contrary, when 2, 3 and 4 angles are
different, depth of the structure change heterogeneously; and the system will take a shape like as it is in Fig.3b.
The structure in this version has only revolute joints; so the form in Fig.3b is geometrically impossible. To
constitute this shape, universal or spherical joints should be used instead of revolute joints. However, using
spherical joints may result in excess mobility of the system and the necessary number of actuators increases.
This decreases the feasibility of the system. In addition, the structure in Fig.3b does not allow attaching another
unit next to it; so it cannot be used as a member of a roof structure.
1044
a
4b
isometric
d2
3b
2b
d2
2b= 3b = 4b
top view
d2
d2
b
3c
isometric
d4
4c
2c
2c 3c 4c
d3 d4 d5
d3
top view
d3
d5
d4
As the second application, contributions of the spatial M-SLEs to the spatial scissor shell structures have been
investigated. The simplest transformable spatial structure is formed by the connection of two perpendicular
planar scissor structures. In this study, this structure is called as Star-shape System; and an example of this
structure is drawn in Figure 4. The structure is composed of SLEs, M-SLEs and one spatial M-SLE (Fig.2b)
located in the middle of the system. Each planar sub-structure of the Star-shape system in the figure has two MSLEs; but mobility of the whole system is equal to one. This is because of the joint types. At this structure, only
revolute joints are used, and revolute joints can only allow planar rotations. Because of the revolute joints, each
planar structure can only transform on its own plane. Intersection of two perpendicular planes constitutes an
edge. For this structure, this edge is on the z axis; and the intersection element can only move along the z axis.
z
M-SLE
M-SLE
y
x
1045
Star-shape system cannot constitute asymmetrical shapes; and to increase the transformation capability of the
system, some of the revolute joints should be switched to spherical joints. An example of switching can be seen
at the structure in Fig.5. In this figure, revolute joints at the points A and B are switched to spherical joints
(universal joints may be used as well). By this small change, the planar sub-structure on x-z plane can rotate
around x axis. This small modification increases the transformation capability of the whole structure. However,
rotation of the plane of the structure on x-z axis causes extra load over the perpendicular sub-structure; and this
decreases the feasibility of this solution.
Spherical Joint
B
Spherical Joint
References
1.
2.
3.
4.
5.
6.
7.
8.
9.
Akgn, Y.; Gantes, G.; Kalochairetis, K.; Kiper, G. (2010). A novel concept of convertible roofs with high
transformability consisting of planar scissor-hinge structures. Engineering Structures, 32 (9), pp. 2873-2883.
Akgn, Y.; Gantes, G.; Sobek, W., Korkmaz, K.; Kalochairetis, K. (2011). A Novel Adaptive Spatial Scissor-hinge
Structural Mechanism for Convertible Roofs. Engineering Structures, 33 (4), pp. 1365-1376.
Akgn, Y.; Korkmaz, K.; Sobek, W.; Gantes, C.; (2010). A Novel Transformable Roof System with Spatial
Scissor-hinge Structural Mechanisms, In Proceedings of AZcIFToMM 2010, International Symposium of
Mechanism and Machine Science, Akgn, y. and Gezgin E. eds., 5-8 October, pp. 116-122.
Gantes, C. J. (2001). Deployable Structures: Analysis and Design. Boston: WIT Press.
Chilton, J. (2000). Space Grid Structures, Architectural Press, Oxford, 2000.
Escrig, F.; Valcarcel, J.P. (1993). Geometry of Expendable Space Structures, International Journal of Space
Structures, 8: pp. 71-84.
Escrig, F. (1984). Expandable Space Frame Structures. In Proceeding of the 3rd International Conference on Space
Structures, Guildford, UK: Elsevier Applied Science Publishers; pp. 845-850.
Escrig, F. (1985). Expendable Space Structures, Space Structures Journal, 2 (1): pp. 79-91.
Escrig, F.; Valcarcel, J.P. (1987). Curved Expandable Space Grids, Proceedings of the International Conference on
the Design and Construction of Non-Conventional Structures, London, England. Civil-Comp Press, pp. 157-168.
1046
Biographies
Dr. Yenal Akgn was born in Elaz/ Turkey in 1978. He was graduated from Istanbul Technical University Department of
Architecture (Turkey) in 2000. Then, he completed his M.Sc. in 2004 at Izmir Institute of Technology Department of
Architecture (Turkey); and PhD in 2011 at zmir Institute of Technology Department of Architecture (Turkey) and University
of Stuttgart Institute for Lightweight Structures and Conceptual Design (Germany). At the moment, Dr. Akgn works as an
Assistant Professor at Gediz University.
He has four international journal articles and several symposium papers on the field of scissor-hinge structures,
deployable structures and overconstrained mechanisms. His current research field is based on the architectural uses of spatial
linkages and multi-loop mechanisms.
Dr. Akgn is member of Turkish Chamber of Architects (UIA) and eCAADe.
Dr. Koray Korkmaz was born in zmir/Turkey in 1973. He was graduated from Dokuz Eyll University Department of
Architecture (Turkey) in 1995. Then, he completed his M.Sc. in 1998 at Izmir Institute of Technology Department of
Architecture (Turkey); and PhD. in 2004 at Izmir Institute of Technology Department of Architecture (Turkey). One year he
served as a visiting researcher at University of Pennsylvania Department of Architecture. At the moment, Dr. Korkmaz works
as an Assistant Professor at Izmir Institute of Technology.
He has four international journal articles and several symposium papers on the field of deployable structures. His current
research field is based on the architectural uses of planar and spatial linkage mechanisms.
Dr. Korkmaz is member of Turkish Chamber of Architects (UIA).
1047
Abstract. Visualization is inherent to the conduct of urban design as a direct connection between the designer
and three dimensional reality of urban settlements. Visualization of urban environments and urban design
projects is vital, since most designers prefer to understand place and context through visualization. Dynamic
computer models can present an ideal environment to visualize the change in respect to time. Digital tools are
much more efficient than conventional methods in explaining the growth and change of urban environments.
Especially, incremental growth requires features not found in static/analog media. Christopher Alexander
and his colleagues, in their book A New Theory of Urban Design, tried to justify their ideas about
piecemeal growth by an experiment. The analog methods, such as physical models, two-dimensional
diagrams, have been used to conduct the experiment and to convey their ideas about the design process. This
research tries to produce a dynamic/digital model that could be utilized in their experiment instead of
static/analog methods. Results prove that spatial data should be considered as dynamic or changing identities
rather than as simple, static features. Recent technological developments are increasing computer hardware
and software capabilities so that this dynamic aspect of data can be accounted for by todays systems.
Keywords: CAD, urban design, visualization, piecemeal (incremental) growth, 3-D modeling
1 Introduction
Computer use in urban design is distinguished from the applications in other related fields, due to the necessity
of a mixture of geometric, geographical, and annotative information. This difference requires opportunities for
collaboration, needs for data integration, and examples of the increasing importance of rich datasets as a basis
for design work [1]. For instance; planning requires visualization of information on a two-dimensional map; or
typically, in architecture, perspective rendering of a single building is considered.
Visualization can be defined as the ability to create mental pictures which lead to the manifestation of a
design solution [2]. Urban design is very much engaged with the visual form of the city. Therefore,
visualization is inherent to the conduct of urban design, not just as a representative media, but as a direct
connection between the designer and three-dimensional reality of human settlements. The need for threedimensionality suggests effective utilization of digital graphics in the field and increases the importance of
computer-aid in urban design.
The city does not have a static character and urban settings cannot be seen as finished products of a single
design process. Places certainly change; and this change can be explained as a function of time. Urban designers
1048
can utilize time-patterning technique as a learning process, studying historic patterns, and formative processes
that created them, and translating and interpreting these patterns through graphic diagramming techniques as a
key part of the design process [3]. Adding the fourth dimension to three-dimensional models can be a helpful
tool to explain the growth and change in urban design environments.
Dynamic computer models can present an ideal environment to visualize the change in time. The study aims to
prove that digital tools are much more efficient than conventional methods in explaining the growth and change
of urban environments. Especially, incremental growth requires features not found in static/analog media.
Christopher Alexander and his colleagues, in their book A New Theory of Urban Design, tried to justify their
ideas about piecemeal growth by an experiment [4]. The analog methods, such as physical models, twodimensional diagrams, have been used to conduct the experiment and to convey their ideas about the design
process. This study make an effort to produce a dynamic/digital tool that could be utilized in their experiment
instead of static/analog methods.
1049
The practice of urban design is an intentional act of integrating culture, biological environs, artefact and
functional needs; and especially focuses on the visual qualities of urban environments [5]. Visualization assists
the designer in visually perceiving and evaluating the underlying conditions and changes. Visualization can be
used as a mental tool to be able to study the urban issues. Lynch states that: Showing how a
placewilldevelopover time is moreeffectivethandrawingwhatits final form will be. [6].
However, visual thinking is not only applicable to the field of design. In scientific visualization, special
emphasis is given to the relationships between formalization and the aspects of cognition and communication
that makes it possible to be used successfully in a problem solving context [7]. For many professionals from
different fields and disciplines, visualization becomes a mental tool to create and solve problems.
Visualization assists the designer in the interpretation of complex urban patterns. The use of visualization as a
cognitive process can be generative, since spatial relationships are formed and evolved as a result of the process
[8].
The use of three-dimensional drawings can give a better opportunity to integrate urban information into a
meaningful context. Visual presentation of a design problem is closely linked with its solution [9]. Through
artistic inquiry, the designer states a position, describes a philosophy, explains a set of values. The principal
means of this act is visualization - as a language for analyzing and revealing the order of things. Through
visualization, the designer tries to "simplify complexities and, by simplifying and diagramming, to express
essential sensory relationships in the form of design" [10].
In every step of the urban design process, visualization modelling becomes an important tool. Design can be
defined as "the fundamental skill required to structuring urban space-a skill that distinguishes from other
activities of planning and engineering the built environment [11]. And in this distinguished task, visualization
assists the designer, not only with its communication potentials, but also in creating a method to understand,
analyze and implement design decisions.
4 Integration of Systems
Computer applications in urban design present a distinct identity, based on a blend of computer-aided design
(CAD), urban-geographical information systems (Urban-GIS), World Wide Web (WWW) and interactive
multimedia. The different capabilities of computer-aided design and geographical information systems force
these technologies to become a perfect combination for urban design applications. The ideal system is likely to
include 'three-dimensional modelling and visualization tools of CAD' in combination with 'data storagemanipulation and analytical tools of GIS'. Each can do what the other cant; thus together they become ideal
partners. There is an ongoing debate about how this integration should be. The leading software companies in the
market have made attempts to unite these systems in a single environment. Some have produced functional tools
for database management as additions to their traditional CAD packages. On the other hand, GIS systems have
improved their 3-D modelling and visualization capabilities to be able to offer outputs as satisfactory as CAD
packages [12], [13].
Although CAD and GIS packages have started to show similarities in terms of modelling and visualization,
there are enough reasons to hold these systems quite distinct. GIS may include a very important visual
component, but more than that special emphasis is on non-visual components to geo-data. Correspondingly,
when CAD is configured to support non-graphical properties of objects for simulations, the types of data and
functional requirements differ from the ones in GIS. A wide range of data from natural and artificial sources
have to be captured in GIS. In spite of the fact that these systems are of different character, both sides learn much
from the researches of the other and benefit from studying the connections in-between [14].
Virtual Reality seems to be the next logical step in computer-aided design technology. These two systems
have close relations. VR systems benefit from the modelling capabilities of CAD systems. And CAD models
find themselves a true interactive environment to be presented [15]. It is not surprising that VR has become
popular for many applications; such as computer games since it makes the player an integral part of the game.
1050
For urban design applications, it could provide many opportunities when this technology is utilized with real
spatial data and true 3D models. Then it would be possible to create virtual environments, which correspond to
real world locations. This would make great contributions to public participation and decision-making process.
But, today the limitations of computing power, required to handle such voluminous urban data, prevent the use
of such systems.
1051
However, CAD/GIS integration is not an easy task. According to Levy, CAD and GIS come from two
different traditions and their language and structure are different as well [20]. Then, integration can only mean
utilizing CAD tools to create, edit and maintain graphical information and using GIS tools to model and
analyze the relative data. Autodesk has taken a different approach with its AutoCAD Map product: We spent a
lot of engineering resources developing object technology to let us read and write our competitors files in their
original formats. But now, it is possible to use a GIS database file in its original format in conjunction with a
.dwg-based file (AutoCAD drawing file format) without translation [21].
1052
Animations showing the piecemeal growth can also be produced as a series of frames (images) by the addition of
each increment in order of growth.
Figure 4. Initial and a sample screenshot from the models software interface.
For visualization purposes, an html-based interface has been created in the scope of this study. The interface
basically operates on two menu bars: Top menu bar includes the time-layers, and side menu bar includes the
view angles. The user chooses the time interval of piecemeal growth from time-layering menu bar and the
viewing angle from views menu bar to visualize, then the computer generates the image of the urban
environment for that particular period of time from the camera chosen. The number of time-layers depends on
the input system of the increments in CAD software. The number of views can be infinite; but this beta version
of the interface only includes some sample cameras for visualization purposes. The initial and a sample screen of
the interface are shown in Figure 4.
7 Conclusion
In todays digital design world, most CAD systems are configured to model only a static snapshot of the
environment. Yet, it is not possible to experience urban environments as a single snapshot. Instead, phenomena
change dynamically over space and time. In urban design, final design recommendations are dynamic, not static,
and assured of changing even after final construction drawings are completed. Within this framework, spatial
data should be considered as dynamic or changing identities rather than as simple, static features. Viewing and
analyzing change requires additional information, which depends on greater value of data. This study has tried to
produce a dynamic/digital tool that could be utilized in their experiment instead of static/analog methods.
Furthermore, this study has tried to present a way for dynamic modelling of urban environments with the goal
of developing a digital visualization model suitable for incremental growth. The model is based on an
experimental study, but can also be utilized for professional practice. The present and future states of the urban
environment were modelled in a linear timeline model, but in further studies a branched timeline model can also
be considered.
However, it is not always the present and future states that are to be modelled. Visualization of urban
environments requires a distinctly rich hybrid of geometric, geographic, and annotative information. These
requirements suggest new techniques for collaboration, data integration, and rich datasets. Digital systems can
offer useful and practical visualization tools for urban designers.
As a final point, the digital visualization model created within the scope of the research was produced by using
CAD systems that are widely available in the commercial market today. The hardware and software used during
the production of the digital model and interface were not configured for experimental studies, but mostly for
conventional drafting purposes. However, this study could have been undertaken in a much different manner by
utilizing systems with higher capacities and by the help of software experts. Especially studies in systems
integration could be taken a step further.
1053
In conclusion, this study has contributed new insights into dynamic CAD modelling. Additional key areas can
be identified for further studies. Some of the interesting research areas that could follow this work include:
Temporal CAD systems; Integration of spatial information systems for urban design; Visualization tools for
urban environments; Criteria for digital visualization modelling in urban design.
References
1.
2.
3.
4.
5.
6.
7.
8.
9.
10.
11.
12.
13.
14.
15.
16.
17.
18.
19.
20.
McGullough and Hoinkes (1996). Dynamic Data Sets as Collaboration in Urban Design. In Proceedings of the
Sixth International Conference on Computer Aided Architectural Design Futures. Toronto, Canada, 24-26
September
1995.
Retrieved
June
16,
2003
from
the
World
Wide
Web:
http://www.clr.toronto.edu:1080/PAPERS95/ CAADF95.DynamicData/CAADF95.DynamicData.html.
Dong W. and Gibson K. (1998). Computer Visualisation. New York: McGraw-Hill. p.4.
Kasprisin R. and Pettinari J. (1995). Visual Thinking for Architects and Designers. New York: Van Nostrand
Reinhold. p.112.
Alexander, C. et al. (1987). A New Theory of Urban Design. New York: Oxford University Press Inc.
Relph, E. (1982). Rationallandscapesandhumanisticgeography. London: Croom-Helm. p. 229.
Lynch, K. (1974). City Sense and City Design. Cambridge: The MIT Press, Cambridge. pp. 274-275.
Cassettari, S. (1993). Introduction to Integrated Geo-information Management. London: Chapman and Hall. p.183
Arnheim, R. (1969). Visual Thinking. Berkeley and Los Angeles: University of California Press. p.27.
Doubilet, S. (1984). The Big Picture. Progressive Architecture, 65 (May): 140-146.
Trancik, R. (1986). Finding Lost Space: Theories of Urban Design. New York: Van Nostrand Reinhold. p.228.
Ibid., p.225.
Mahoney, D.P. (1998). Merging CAD and GIS. Computer Graphics World, 21(March): 45-48.
Sullivan, A.C. (1998). Urban Simulations. Architecture, 87(5): 212-216.
Worboys, M.F. (1995). GIS: A Computing Perspective. London: Taylor and Francis. p.4.
Bertol, D. (1997). Designing Digital Space. New York: John Wiley & Sons.
[16] Campbell, S. and Fainstein, S.S. (1996). Introduction: The Structure and Debates of Planning Theory. In
Readings in Urban Theory, edited by S. S. Fainstein and S. Campbell. Cambridge, Mass., USA : Blackwell
Publishers.
McLeod, D. (1996). Planning Theory. Retrieved May 13, 2003 from the World Wide Web:
http://3.sympactico.ca/david.macleod/pthyr.htm .(9-1996)
Alexander, C. et al. (1987). A New Theory of Urban Design. New York: Oxford University Press Inc. p.5.
Broadbent, G. (1990). Emerging Concepts in Urban Space Design. New York: Van Nostrand Reinhold. p.323.
Mahoney, D.P. (1998). Merging CAD and GIS. Computer Graphics World, 21: 48.
1054
Biographies
Sabri ALPER Born in Manisa, Turkey in 1975; SabriAlper, studied architecture at the Department of Architecture in
Middle East Technical University, Ankara, Turkey and School of Architecture in University of Newcastle-upon-Tyne, UK.
He received his Masters Degree in Urban Design in 2002 and has carried on PhD research on urban morphology in Izmir
Institute of Technology, Turkey , where he also worked as a research and teaching assistant. He has also been practicing
architecture as a partner of a design office and taught as a part-time lecturer in the Faculty of Fine Arts and Design in Izmir
University of Economics.
Dr. Alper is currently teaching at the Department of Architecture in Gediz University, Izmir. His research interests include
computer-aided design, design technologies, urban design, design theory, and urban morphology.
Erkal SERM Graduated from the Department of City Planning in Middle East Technical University, Ankara, Turkey.
ErkalSerim holds a M.Sc. degree in city planning from University of Strathclyde, Glasgow, UK; and completed his doctoral
studies in EgeUniversity, Izmir, Turkey.
He made post-doctoral research at The Danish Royal Academy of Fine Arts, Copenhagen, Denmark, with the thesis entitled
Slum Clearance and Urban Renewal Process in Denmark. He worked as Asst. Prof. Dr. at the Department of Islamic
Architecture, College of Engineering & Islamic Architecture, in Umm Al-Qura University, Saudi Arabia, and in Dokuz Eylul
University, Izmir, Turkey.
Dr. Serimis currently teaching at the Department of City and Regional Planning in Izmir Institute of Technology. He is also a
member of the Chamber of City Planners, Turkey and Royal Town Planning Institute, UK.
1055
1. Introduction
Architectural programming is the research and decision making process that defines the problem to be solved
by design [1]. Human factors and functions are important design criteria for architectural programming at predesign level of a project. A significant step in planning such human factors is establishing the functional
relationship diagram. Functional relationship diagrams are typical tools that architects use during the pre-design
level of a project. These diagrams indicate the level of connection between different spaces of a building so as to
aid in the functional organization. There are two well-known methods for establishing functional relationships:
bubble diagram and matrix diagram. Either method indicates not only the relationships between activities but
also the nature of that relationship [1].
Recent advances in information technology have offered new horizons in architecture. Such advances are not
only used to make the buildings smarter and interacting, but also enabled opportunities to collect data for the
process of their design. Example applications in this direction include context aware technologies for tracking
elderly/disabled people, mostly for the health support systems [2], and also collecting data for optimizing
functional relationships of living spaces [3]. This study is investigating the opportunity to optimize the functional
relationship diagram of workspaces, besides to test the ergonomics of the office equipments.
1Architecture Department, Faculty of Engineering & Architecture, Gediz University, zmir, 35665, TURKEY
2 Mechanical Engineering Department, Faculty of Engineering & Architecture, Gediz University, zmir, 35665, TURKEY
1056
Figure 2-Gediz University Site Plan, Case Study at Block D: Faculty of Architecture& Engineering
We measure these using both statistical questionnaires and temporal data measurement techniques. The target for
this analysis is the offices located within Engineering & Architecture Faculty of Gediz University.
2. Methods
In this study we measure physical activity levels of employees at work place for understanding the nature of the
functional relationship between various spaces in an office setting. Specifically, what we are interested in is to
find out how much time, and at which locations that an employee spends his/her time during the working hours.
According to this data, we wish make some correlations between the distance between the spaces and how
frequently the employees use the routes determined. The case study for this analysis is the offices located within
Engineering & Architecture Faculty of Gediz University.
1057
Temporal and statistical physical activity levels of the employees a particular section in Gediz University
campus was measured in order analyze the correlation between the architectural design and the actual usage
patterns of common working spaces. The activity relations between the offices and the other service spaces were
measured by both temporally using inertial sensors and statistically (or spatially) by using questionnaires on test
subjects. These measurements were designed so as to sense both the physical distance and the nature of the
activity. The details of these measurements are explained in the following subsections.
2.1 Temporal Data Collection
At the first step, we aim to collect temporal movement data of the employees for aid in the analysis of the
functional relationships as opposed to employee movement statistics between different spaces within the
building. To measure this we used an INS (Inertial Measurement System) sensor known as the SHAKE (Sensing
Hardware Accessory for Kinaesthetic Expression) SK7 seen in Figure 3 below. This sensor can measure 3D
accelerations, 3D rotation rates, 3D magnetic field and the heading data of the test subject and it is specially
designed for human motion tracking purposes. The SHAKE SK7 has embedded Bluetooth radio and can send the
sensor data wirelessly to any Bluetooth enabled system.
To detect the temporal employee activity in the context of functional relationship analysis we prepared a data
logging software in Mobile Python language implemented on Nokia Symbian S60 mobile phone. The mobile
phone served as a data acquisition computer while SHAKE SK7 provided the measurement data. For the purpose
of the data collection both the sensor and the mobile phone were carried on the test subjects. In the interest of
keeping data log size small we chose the sampling interval to be 1Hz.
Figure 3- Data collection system composed of an the INS sensor and a mobile phone
serving as data acquisition computer used during this study
In post processing the data we implemented a discrete time low-pass Butterworth filter to get rid of highfrequency components in the sensor data particularly affecting in gyro and acceleration measurements. The
specific filter we implemented was designed in Matlab and is a 2nd order filter with a sampling interval of 1Hz
and a cut-off frequency of 0.5Hz. The transfer function of the filter is given as
figure. During the data collection the subject moved in an out of office, and spent about half of the time outside
his office and moving between the kitchen facilities, the rest room and nearby offices.
1058
The next plot corresponds to the orientation datain [deg] that is composed of the heading, roll, and pitch of the
INS sensor fixed on the test subject shown in Figure 6. The heading informationisoutput directly by the sensor
that is obtained internally from the magnetometer data. Roll and pitch readings are computed using acceleration
signals based on the following relationships:
1059
ay
roll angle= tan1 and
az
a
pitch angle= tan1 x
az
for t = 1: length(time) :
walk_flag(t) = 1
else
walk_flag(t) = 0
end
end
The resulting walk instance detection plot is seen at the bottom of Figure 7. In the same figure the filtered
heading measurement details are seen in the middle plot. The heading at the sitting posture is also shown on this
figure with a dashed line. The subject travelled to the kitchen and came back to his desk in the first 3 minutes of
this measurement and he interacted with other employees in his office intermittently during the rest of the 10minute period. Using the algorithm it is possible to detect how much of the time that the subject was walking and
sitting during the measurement period.
1060
Figure 7 Looking at an instance of activity data of the test subject based on the heading, and vertical
acceleration measurements
While the measurement was useful in detecting the sitting and walking instances the measurement is not
sufficient to detect the exact location tracking of the test subjects. The heading data is obtained from the
magnetic field measurements, which is prone to errors due to the presence of metal objects. We tried to infer the
position of the subject integrated during the walking instances however measurement errors prevented automatic
categorization of the destination of the test subject. GPS location measurements could potentially be added to the
sensors set but this would not be useful for in-building applications due to the lack of GPS signals in enclosed
locations.
2.2 Spatial Data Collection
The second method used in this study is collecting spatial data from the test subjects, the employees locating at
the 2nd floor of Block D. These spatial data is effective for making the correlations between the frequency of
using the route and the distances. The subjects are given an unorganised bubble diagram that is used to determine
the frequency of visits from their own offices to the spaces indicated. And they were required to mark once every
time they use the route during a 3 day measurement period. The bubble diagram seen in Figure 8 is composed of
the following 7 mostly-used office spaces: (i) Kitchen, (ii) Other offices, (iii) WC, (iv) Administration &
Management Offices, (v) 1st Floor Lecture Halls, (vi) Cafeteria, and (vii) Ground Floor Lecture Halls.
1061
1062
Then,
wemeasuredthedistances
in
metersbetweentheemployeesownoffice
(located
at
2 ndfloor)
andthespacesindicated (located at differentfloors) shown in Figure 10. Thenweusedthemean of theactualdistances
in ordertocorrelatewiththefrequency of use of eachspecificrouteseen in Figure 11.
Figure 10- Actual Office Locations and Their Distances as Obtained from the Building Plan
Figure 11- Correlation Analysis Between the Actual and Suggested Functional Relationships
1063
3. Conclusions
After analysing the collected data, we reached o the conclusion that the distance from the employee offices at 2 nd
floor to the kitchen, other offices, restrooms, and administration & management offices correlate well with the
frequency of use in the sense that the more the route is used the closer is the distance, as needed. On the other
hand, 1st floor lecture halls should be closer to the employee offices because they are used more frequently as
seen at the table (frequency rate is 3, however distance grade is 2). Again the cafeteria should be closer to the
employee offices as it is used more frequently as seen at the table (frequency rate is 2, however distance grade is
1). Moreover, the ground floor lecture halls may be farther away to the offices because they are used less
frequently as seen at the table (frequency rate is 1, however distance grade is 3). Consequently, this analysis
model may be used to optimize and re-organize the functional diagram of workspaces, which already have some
expected usage patterns seen in this case study.
In this paper, we suggested alternative analysis techniques for analysing and improving functional relationship of
spaces, which was demonstrated on an office space setting. By using such approaches it may be possible to
increase the efficiency and effectiveness of the buildings through re-organization.
References
[1] Edith Cherry, Programming for Design: From Theory to Practice, John Wiley and Sons Inc., New York, 1999.
[2] S. S. Intille, K. Larson, J. S. Beaudin, J. Nawyn, E. Munguia Tapia, P. Kaushik, A living laboratory for the design and
evaluation of ubiquitous computing technologies in Extended Abstracts of the 2005 Conference on Human Factors in
Computing Systems, New York, USA, 2004.
[3] K. Larson, S. Intille, T. J. McLeish, J. Beaudin, and R. E. Williams, "Open source building-reinventing places of living"
BT Technology Journal, vol. 22, pp.187-200, 2004.
[4] E. Munguia Tapia, S. S. Intille, L. Lopez, and K. Larson, "The design of a portable kit of wireless sensors for naturalistic
data collection," in Proceedings of PERVASIVE 2006 , vol. LNCS 3968, K. P. Fishkin, B. Schiele, P. Nixon, and A. Quigley,
Eds., Springer-Verlag, pp.117-134, 2006.
Biographies
Dr. Seza Filiz Dr. Filiz was born in Denizli, Turkey in 1974. She received her BSc degree in architecture from Istanbul
Technical University, Istanbul, Turkey in 1999 and MA in Interior Architecture from Marmara University Faculty of Fine
Arts at Istanbul, Turkey in 2002. Dr. Filiz received her PhD in architectural design program from Istanbul Technical
University, Graduate School of Science Engineering and Technology, in 2010. Her master study was researching the effect of
culture-function relation to the design organization of house kitchen. Her PhD research was about the future of housing
within the context of sustainability and technology. Therefore, her research area comprimises both design problems of
interior architecture and also sustainable architecture, specifically in housing design.
Dr. Filiz was employed as instructor in Istanbul Aydn University Faculty of Architecture between 2008-2010 after she
worked several years at field of architectural practice as designer and technician during 1999-2008. During this time she also
attended some design concurs and has awarded for jury special award from Siemens Facade Design Contest in 2002. Dr. Filiz
is currently with Gediz University, Interior Architecture Department, where she is employed as an Assistant Professor. Her
current research interest includes sustainable architecture, interior design and housing.
Dr. Selim Solmaz Dr. Solmaz was born in Izmir, Turkey in 1978. He received his BSc (High Honours) degree in
aeronautical engineering from Middle East Technical University, Ankara, Turkey in 2001 and MSc from Purdue University
at West Lafayette, School of Aeronautics and Astronautics in 2003. Dr. Solmaz received his PhD in 2007 from Hamilton
Institute, National University of Ireland-Maynooth. As part of his PhD project he spent a year as a visiting researcher at
Daimler Chrysler Research AG in Esslingen, Germany in 2003-2004. Upon his return to Ireland he conducted research on
1064
1065
1066
Chapter 7
FUNDAMENTAL
SCIENCES
1067
Abstract. Cubic B-spline differential quadrature methods have been introduced. Nonlinear Burgers' equation
is solved numerically using cubic b-spline differential quadrature methods. Numerical simulations are studied
for two well-known test problems. Stability analysis is studied. Discrete root mean square error norm and
maximum error norm are computed for both test problems and a comparison with some earlier works are
given.
Keywords: Differential quadrature Method, Burgers' equation, B-spline.
1 Introduction
Since Bellman et al. [1] have introduced the differential quadrature method, lot of basis have been used to
improve the differential quadrature method. Lagrange polynomials, Hermite polynomials, spline functions, radial
basis functions and Sinc functions can be counted as some of them [2-6].
Since it has analytical solutions, the Burgers. Equation has been used as test problem to verify the numerical
algorithms. The form of the one-dimensional nonlinear Burgers equation with the initial and boundary
conditions is
U t UU x vU xx 0
( 1)
U ( x,0) f ( x), a x b
U (a, t ) , U (b, t ) , t 0, T
where v 0 is the kinematic viscosity coefficient.
Burgers equation is a quasi-linear parabolic time dependent partial differential equation, whose analytic
solutions can be constructed from a linear partial differential equation by using Hopf-Cole transformation[7-8].
Besides, this equation is very important fluid dynamical model both for the conceptual understanding of a class
of physical .flows and for testing various numerical algorithms. The Burgers. equation has been founded to
model in many fields such as gas dynamics, elasticity, turbulence problems, heat conduction, number theory,
shock waves and continuous stochastic processes, etc. Burgers equation can be solved analytically for arbitrary
initial conditions. But some analytic solutions involve infinite series, which may converge very slowly for small
viscosity coefficient v 0 [9].
1068
2 Numerical Method
U x( r ) ( xi ) wij( r )U ( x j ), i 1,2,..., N
( 2)
j 1
where
(r )
ij
w represents the weighting coefficients, and is the number of grid points in the whole domain. The
discretization of a PDE is incredibly easy by using DQM. The key procedure of the method is the determination
of the weighting coefficients using the basis functions. In this study, we use cubic b-spline functions to
determine the weighting coefficients of the first order derivative approximations.
We consider a uniform partition of the domain axb at the knots x m , m=1,2,...,N, such that a= x1 < x 2 <...<
x N =bwith h= xm xm1 . Then, the cubic b-spline functions Pm (x) , m=0,1,...,N+1 defined by
( x xm2 ) 3
3
h 3h 2 ( x xm1 ) 3h( x xm1 ) 2 3( x xm1 ) 3
1
Pm ( x) 3 h 3 3h 2 ( xm1 x) 3h( xm1 x) 2 3( xm1 x) 3
h
( x x) 3
m 2
0
, xm2 , xm1
, xm1 , xm
, xm , xm1
, xm1 , xm 2
(3)
, others
form a basis for the functions defined over the interval [a,b] [10].
dPm(xi )
dx
where
m1
j m1
(1)
i, j m
P ( x j ),
i 1,2, ,N , m 0,1,..., N 1
(4)
wm(1,)1 , wm(1,)0 , wm(1,)N 1 , wm(1,)N 2 are the support coefficients and x-1, x0, xN+1 ve xN+2 are ghost grids.
In order to determinate the weighting coefficients
1069
dP (x )
wm(1,)1 0 m
dx
(1)
1 4 1
wm , 0 dP1(x m )
1 4 1
w (1)
dx
m
,
1
dP (x )
2
m
dx
1 4 1
1 4 1
dPN (x m )
1 4 1 wm(1,)N 1
dx
PN' 1(xm )
P0''(xm )
(5)
N2
w
j 1
(1)
m, j
P0' ( x j ) and
N2
w
j 1
(1)
m, j
PN ( x j ) are added to the system (5) to obtain a solvable system, that is,
d 2 P0(x m )
wm(1,)1
dx 2
(1)
3 0 3
wm , 0
dP0(x m )
1 4 1
w (1)
dx
m ,1
dP
(x
)
1 m
1 4 1
dx
1 4 1
1 4 1
dPN 1(x m )
3 0 3 wm(1,)N 1
dx
(1) d 2 P (x )
N 1 m
wm , N 2
2
dx
(6)
1070
chosen as different, the weighting coefficients will be determined as the same, since Eq.(2) is unique for each r.
In this study, we use the DQM formulation based on Lagrange interpolation polynomials [11].
The third method to determine the weighting coefficients is derived from definition of derivative and
matrix multiplication approach. Briefly, the weighting coefficients of the second order derivative approximation
can be determined by using the weighting coefficients of the first order derivative approximation.
N
For the determination of the weighting coefficients of higher order derivatives, the weighting
coefficients of lower order approximations are used [11].
where
B (wi(,11) wi(,21) ) 2 (wi(,1N) wi(,2N) ) .is resulted by the boundary conditions. Then, the system (7)
K i U ( xi ) on the right hand side of the Eq.(7) and writing it in the matrix form we obtain
U ( x 2 , t )
U ( x2 , t )
t
U ( x , t )
U (x , t)
2
K
B
U ( x N 1 , t )
U ( x N 1 , t )
( 8)
where
(1)
( 2)
3 w3,1 vw3,1
K
w (1) vw( 2 )
N 1,1
N 1 N 1,1
3w
vw
(1)
( 2)
N 1 wN 1, N 1 vwN 1, N 1
2 w2(1, )2 vw2( 2, 2)
3w
(1)
3, 2
2 w2(1, )N 1 vw2( 2, N) 1
vw
( 2)
3, 2
(1)
3, N 1
i ,
( 2)
3, N 1
2.78 ti 0 as if all the eigenvalues are purely imaginary, then they must be 0 ti 2 2 . If the
eigenvalues are complex with real parts, then all
1071
4 Test Problems
In order to verify the accuracy of the proposed methods, we have chosen shock-like solution and
sinusoidal disturbance solutions of Burgers Equation.
The error between the numerical and exact solutions are computed via discrete maximum norm and
discrete mean square norm, that are,
N
L2 h U j (U N ) j
j 1
L U U N
max U j (U N ) j )
j
U ( x, t )
x
t
t
x2
1
exp( )
t0
4vt
t 1,
0 x 1.2
( 9)
1
) . This solution simulates a deflating shock wave as time goes. The functional value of Eq.(
8v
9) at t 1 is taken as initial condition as the boundary conditions are forced to be U( 0,t) U(1.2,t) 0 .
The routine is run up to the terminating time t=3.6 with the initial data set N 241, t 0.001,v 0.005
-5
-5
for all three methods. The L2 and L are computed as 2.69 10 and 8.30 10 respectively in Method I.
where
The
t0 exp(
L2
and
are computed as L2 7.85 10 and L 3.18 10 . CPU times are recorded as 0.64, 0.37 and 2.87 for
Method I,II,III, respectively.
Stability analysis of the methods are also studied for various grid numbers. In Method I, the eigenvalues
are complex numbers for N=21 and N=31. When the numbers of grids are chosen as N=41 and N=61, the
-5
max 765.536961481966
max 2707.46270556291 for N=61. In Method II, the real parts of the eigenvalues with nonnegative real
parts are determined as
1072
respectively. In Method III, most of the eigenvalues are determined as complex. All the eigenvalues are in a
good agreement with stability conditions.
[0,1] with the boundary conditions U(0,t)=U(1,t)=0. In order to make a comparison with the exact solution in
this case to test the accuracy of the present method, the exact solution is calculated using
(x )2
x
exp
exp
u 0 ( )d d
t
4vt
2v 0
, t 0, 0 x 1
U ( x, t )
( 10)
(x )2
exp 4vt exp 2v 0 u0 ( )d d
where u 0 denotes the initial condition. The procedure is run with the parameters v / 100, t 0.0025
up to the terminating time t=0.50 for various numbers of grid points. Even though less grids are used in the
Method I, the results obtained by Method I is more accurate than Wavelet Galerkin Method (WGM). A
comparison of L is given in Table 1.
Table 1. Comparison of CBSDQ and Wavelet Galerkin Method
Method
Method I
WGM[12]
N
16
32
64
128
16
32
64
128
t
0.0025
0.0025
0.0025
0.0025
0.0001
0.0001
0.0001
0.0001
t=0.14
0.0659
0.0133
0.0074
0.0022
0.1588
0.0523
0.0117
0.0030
t=0.26
0.3559
0.2865
0.1334
0.0155
0.6856
0.4737
0.2127
0.0503
t=0.38
0.3507
0.2853
0.1024
0.0135
0.6955
0.3676
0.1550
0.0422
t=0.50
0.3234
0.2242
0.0544
0.0119
5 Conclusion
Cubic b-spline differential quadrature methods have been developed to solve the nonlinear Burgers
equation. Two classical solutions of the Burgers equation are chosen as test problems. Numerical results show
that all presented methods generate successful numerical solutions for Burgers equation.
References
[1] Bellman, R., Kashef, B., & Casti, J. (1972). Differential Quadrature: A Technique for the Rapid
solution of Nonlinear Differential Equations. Journal of Computational Physics, 40-52.
[2] Bellman, R., Kashef, B., Lee, E., & Vasudevan, R. (1976). Differential quadrature and splines.
Computers and Mathematics with Application, 371-376.
[3] Bonzani, I. (1997). Solution of non-linear evolution problems by parallelized collocation interpolation
methods. Computers and Mathematics with Applications, 34(12), 71-79.
[4] Cheng, S., Wang, B., & Du, S. (2005). A theoretical analysis of piezoelectric/composite laminate with
larger-amplitude defection effect, Part II:Hermite differential quadrature method and application.
International Journal of Solids and Structures, 42(61816201), 6181-6201.
[5] Shu, C., & Richards, B. (1992). Application of generalized differential quadrature to solve two
dimensional incompressible Navier Stokes equations. Inter. Jour. for Num. Meth. in Fluids, 15(791798).
[6] Shu, C., & Wu, Y. (2002). Development of RBF-DQ method for derivative approximationand its
application to simulate natural convection in concentric annuli. Computational Mechanics, 29, 477-485.
1073
[7] Cole, J. D. (1951). On a Quasi-linear Parabolic Equation in Aerodynamics. Quarterly of Applied Math.,
225-236.
[8] Hopf, E. (1950). The Partial Differential Equation Ut+UU_x = vU_xx. Comm. Pure App. Math., 201230.
[9] Miller, E. L. (1966). Predictor-Corrector studies of Burgers model of turbulent flow. Newark,
Delaware: University of Delaware.
[10] Prenter, P. M. , Splines and Variational Methods, J. Wiley, New York, 1975.
[11] Shu, C., 2000, Differential Quadrature and its Application in Engineering, Springer-Verlag London
Limited.
[12] Kumar, B.V.R and Mehra, M.,"Wavelet-Taylor Galerkin method for the Burgers' equations", BIT
Numerical Mathematics, Vol. 45,543-560, (2005).
Biographies
Alper Korkmaz was born in Izmir, Turkey in 1976. Having graduated from zmir Karabalar Cumhuriyet
Lisesi, received his BSc degree on Mathematics from Dokuz Eyll University. Completed his MSc degree on
"Differential Quadrature methods for numerical solutions of nonlinear wave equations" at Eskiehir Osmangazi
University. In the same university, he obtained the PhD on Applied Mathematics with the thesis called
Numerical Solutions of time dependent nonlinear PDEs with b-spline differential quadrature methods.
He has extensively published in various peer-reviewed academic journals. His research interest includes
numerical solutions of dynamical partial differential equations, numerical stability, differential quadrature
methods.
Dr. Korkmaz is currently a faculty member of the Department of Mathematics at the ankr Karatekin
University.
dris Da was born in Kbrsk county of Bolu in 1963. Graduated from Ankara University, Department of
Mathematics in 1985. Between 1986-1988, he worked as research assistant in Erciyes University Department of
Mathematics. Received his MS. in the same university in 1987. He got his PhD in University of Wales in 1994
and he began to work as assistant professor in Eskiehir Osmangazi University in the same year. Between 19992005, he worked as associate professor. In 2005, he became professor in the same university.
Numerical methods for nonlinear wave equations, b-spline functions, numerical stability, dynamical
systems, are the fundamental research areas of him. He has lots of peer reviewed articles on various journals.
Prof. Da is the member of Scientific Research Commission since 2002 and member of Scientific
Research and Ethics Committee since 2005.
1074
PhD. Student, Gazi University, Engineering Faculty, Civil Engineering Dep., Ankara
2
Prof. Dr., Gazi University, Engineering Faculty, Civil Engineering Dep., Ankara
1
kagan.cebe@gmail.com, 2lalebal@gazi.edu.tr
Keywords: hydrodynamics, coastal dynamics, numerical model, turbulence, diffusion, finite difference,
finite element, pollution.
Introduction
HYDROTAM, a three-dimensional, baroclinic, mathematical model is developed to simulate the circulation
and transport mechanism in coastal waters. The model is calibrated and tested by field experiments for over 15
years. HYDROTAM has been applied successfully to Gulf of Fethiye, Gulf of zmir, Gksu Lagoon, Gulf of
Marmaris, Gulf of Bodrum, Gulf of ldeniz, Belceiz, Kemer, amyuva, Gynk, Belek, Gulf of Mersin, Gulf
of skenderun, Gulf of zmit, Gulf of Antalya, Gulf of Cardiff, Karasu and Gulf of Gkova [1], [2], [3], [4], [5],
[6], [7], [8].
There are one, two and three-dimensional numerical models which have been used for the estimation of
circulation patterns and the water quality in coastal water bodies. Most of those models assume laterally uniform
flow. According to zhan and Balas [9], the wind induced currents have a strong three-dimensional character
and two dimensional models are poorly adapted to the simulation of circulation induced by the wind. In many
applications, it is important to predict the vertical structure of the flow due to salinity or temperature variations,
which also require a three-dimensional model [10]. As a three-dimensional numerical model, HYDROTAM
stands out among most of the existing models in literature with its comprehensiveness and numerical scheme.
1075
Model Structure
There are various forces affecting the flows in coastal waters. According to Balas [10], major forces that affect
the flow mechanism are gravity, the tide producing forces due to the Sun and the Moon, the stress of the wind
acting on the sea surface, the Coriolis force, the pressure gradient force and the viscous stresses. Besides the
changes in climatic conditions and the action of wind stress exerted on the water surface, the gravitational
circulation induced by variations in water density resulting from temperature and salinity differences influence
the flow in coastal waters.
Turbulent motions in coastal waters also contribute significantly to the transport of momentum, heat and mass,
and therefore have a determining influence on the distributions of velocity, temperature or species concentration
over the flow field [10]. With the increasing development in computer technology, research on the numerical
modeling has increased over the last two decades and simulations of turbulent transport processes have become
possible.
Governing Equations
The model HYDROTAM has hydrodynamic, transport and turbulence model components. In the
hydrodynamic model component, the Navier-Stokes equations are solved with hydrostatic pressure distribution
assumption and the Bussinesq approximation. In the transport model component, three-dimensional convective
diffusion equations to simulate pollutant, temperature and salinity transport in water are solved. In the turbulence
model component, the k- formulation is solved to calculate the kinetic energy of the turbulence and its rate of
dissipation, which is the measure of the vertical turbulent eddy viscosity. The Smagorinsky algebraic sub grid
scale turbulence model is used to simulate the horizontal eddy viscosity. The governing model equations in the
three-dimensional Cartesian coordinate system are as follows [10], [11]:
Continuity equation:
u v w
0
x y z
(1)
vy
t
x
y
z
0 x
x x y y x z z x
(2)
v
v
v
v
1 p v u
v v w
vz
u v w fu
v x 2 v y
t
x
y
z
0 y x x y
y y z z y
(3)
gz v x
vy
2 vz
t
x
y
z
0 z
x x z y y z
z z
(4)
where, x,y: Horizontal coordinates; z: Vertical coordinate; t:Time; u,v,w: Velocity components in x,y,z
directions at any grid locations in space; x, y, z: Eddy viscosity coefficients in x,y and z directions,
respectively; f: Coriolis coefficient; (x,y,z,t): In situ water density; o: Reference density; g: Gravitational
acceleration; p: Pressure.
The conservation equation for a pollutant constituent is [10], [11]:
C
C
C
C C C C
+ Dz
+u
+v
+w
= Dx
+ Dy
k pC
t
x
y
z x
x x
y z
z
(5)
where, C: Pollutant concentration, kp: Decay rate of the pollutant, Dx, Dy and Dz: Turbulent diffusion
coefficient in x, y and z directions respectively.
1076
Boundary Conditions
There are four different boundary conditions defined in the model. These are free surface, sea bed, open sea
and coastal land boundaries [10].
Free Surface Boundary Condition: The wind acting above the free surface changes water velocities just
under it. The wind-induced shear stress at the surface results in a water velocity gradient below the surface.
Sea Bed Boundary Condition: The bottom shear stress at the sea bed is determined by matching velocities
with the logarithmic law of the wall. The gradients of temperature, salinity and the pollutant are taken as zero at
the sea bed. This model implies that there are no advective and diffusive fluxes from the sea bed to the water
body.
Open Sea Boundary Condition: The open sea boundary is the lateral boundary through which mass fluxes
can be occurring.
Coastal Land Boundary Condition: Shorelines in the model are defined in order to simulate the flooding
and drying processes, which can be also seen in nature depending on seasonal variations, and cause some water
areas to dry out or land areas to be flooded. At the end of each step, once the free surface and the new water
velocities are computed, the total water depth and vertical grid spacing are updated before proceeding to the next
time step. The water surface slope is calculated at each time step with the new values of water depth. If the water
surface slope is positive at coastal land boundary, the water surface is extended using the calculated slope until it
intersects the coastal land boundary. This feature allows the model to redefine the shorelines at each step
according to the calculated water depth values.
The normal gradients of temperature, salinity and the pollutant across the shoreline are taken as zero which
means that there are no advective and diffusive fluxes across the shoreline boundary.
Numerical Scheme
Equations are solved numerically by approximating the horizontal gradient terms using a staggered finite
difference scheme. In the vertical plane, however, the Galerkin Weighted Residual Method with linear shape
functions of finite elements is utilized in order to convert the differential equation to a discrete problem. Water
depths are divided into the same number of layers following the bottom topography. At all nodal points, the ratio
of the length (thickness) of each element (layer) to the total depth is constant. To increase the vertical resolution
wherever necessary, grid clustering can be applied in the vertical plane. The mesh size in the horizontal plane
can be variable.
1.a)
1.b)
Fig. 1.a) Location where u, v and all other variables are defined in the Staggered Finite Difference Scheme,
v locations, u locations, locations of all other variables [10].
Fig.1.b) The discretization in the finite element scheme [10]
1077
R N
dz 0
1, 2
for
(6)
lk
The integrations are performed over the length (lk) of the element k, which varies from point to point over the
grid area and also in time. After the application of the Galerkin Method, any derivative terms with respect to
horizontal coordinates appearing in the equations are replaced by their central finite difference approximations.
The mesh size may also vary in the horizontal plane. Then, one obtains local matrices and vectors describing the
nonlinear equations for each of the element k over the water depth as follows [10]:
A1k,1
A2k,1
k
G1k
k
A1k, 2
a1
t
k
k
k
A2, 2 G2
a 2
(7)
where, G1 and G 2 represent any of the variable at the beginning and at the end of the kth element, respectively;
Ak and ak are the elements of local matrices and vectors respectively.
The local element matrices for all elements on a vertical line are grouped together to form the global matrix
equation for the unknown nodal time derivatives of the variables at a grid point on the horizontal plane. The sea
bed and the sea surface boundary conditions are taken into account while building up the global matrices over
the water depth. Variables at the sea surface and the sea bottom, and the gradient of the variable at the sea
bottom are known. Using these known values a system of nonlinear equations can be obtained.
This system of nonlinear equations is solved by the Crank Nicholson Method which has second order accuracy
in time. To provide this accuracy, difference approximations are developed at the midpoint of the time step. The
temporal first derivative is approximated at (t+1/2) and all other variables and derivatives at this time are
determined by averaging the difference approximations at the beginning (t) and at the end (t+1) of the time
increment. The resultant set of implicit equations is solved by an iterative method, which is controlled by underrelaxation [10]. Under-relaxation is typically employed to make a non-convergent system convergent or to
hasten convergence by dampening out the oscillations.
Continuity Equation
After the estimation of horizontal velocities, the vertical velocities, w, are calculated from the continuity
equation, at each time step. The approximate expressions for horizontal velocities u and v are substituted into the
continuity equation. Then, the residual error is minimized by the Galerkin method. Since the horizontal
velocities are known, the equation to calculate the vertical velocity becomes as such [10], [11]:
w2k w1k
u k u k
1 l k k
u1 u 2k l k 1 2
2 x
x
x
l k u1k u 2k
l k k
v k v k
v1 v 2k l k 1 2
y
y
y
z 2 z1 z1 z 2
k
k
2
l x l 2 x v1 v 2
k
k
z 2 z1 z1 z 2
2
l y l 2 y
k
k
(8)
1078
where
k=1,2....,
m
represents
the
layer
number
and
lk
is
the
length
of
the
element k. At each node, vertical velocities are computed, from the bottom layer to the surface layer, in order to
satisfy the continuity equation.
4078380
4077880
HIZ (cm/s)
4077880
1855
0 20
HIZ (cm/s)
1855
0
1755
4077380
10
1755
4077380
1655
1655
1555
4076880
1555
4076880
1455
1455
1355
4076380
1355
4076380
1255
1255
4075880
4075380
S0
4074880
955
855
755
655
4074380
1155
1055
4075380
TC
1055
y (m)
1155
TC
y (m)
4075880
S0
4074880
455
655
555
455
4073880
355
355
255
4073380
255
4073380
155
155
4072880
4072380
283560
55
284560
285560
286560
287560
288560
289560
855
755
4074380
555
4073880
955
4072880
4072380
283560
55
284560
285560
286560
287560
288560
289560
x (m)
x (m)
2.a
2.b
Fig. 2.a) Current velocity pattern and bacterial concentration distribution at water surface after 2 hours of
wind speed of 5m/s from SSW direction [7]
Fig. 2.b) Current velocity pattern at bottom after 2 hours of wind speed of 5m/s from SSW direction [7]
A very common condition conforming to the wind-induced currents in coastal waters can also be observed in
the figures above. Here, the wind action which creates a shear-induced drift current leads to an upward tilt, or
wind set-up, of the water surface in the downwind direction, and a mass balancing gravity-driven reverse current
near the bottom.
1079
Acknowledgements
This paper is a part of the field study conducted in Antalya Bay which is financially supported by the
TUBITAK project titled Investigation of Uncertainties in Predicting Bacterial Concentrations of Discharged
Sewage from Deep Sea Outfalls When Submergence Occurs in Receiving Media.
References
1.
Balas, L. (2001). Simulation of Pollutant Transport in Marmaris Bay.Chineese Ocean Engineering Journal (15): 565578.
2. Balas, L. and zhan, E.(2001). Applications of a 3-D Numerical Model to Circulations in Coastal Waters.Coastal
Engineering Journal (43): 99-120.
3. Balas, L. and zhan, E. (2003). A Baroclinic Three Dimensional Numerical Model Applied to Coastal Lagoons.Lecture
Notes in Computer Science.Vol. 2658, pp. 205-212.
4. zhan, E. and Balas, L. (2003). Simulation of Water Exchange in Enclosed Water Bodies.Lecture Notes in Computer
Science.Vol. 2658, pp. 195-204.
5. Kkosmanolu, A.; Balas, L.; and zhan, E. (2003). Modelling of Mixing and Flushing in Fethiye Bay.Sixth
International Conference on the Mediterranean Coastal Environment, MEDCOAST 03, 7-11 October, Ravenna, Italy,
pp. 2099-2107.
6. Balas, L. and nan, A. (2007). A Composite Finite Element-Finite Difference Model Applied to Turbulance
Modelling.ICCS 2007: 7th International Conference on Computer Science, Lecture Notes in Computer Science,
2007.Vol. 4487/2007, pp. 1-8.
7. Cebe, K. and Balas, L. (2010). HDROTAM-3 Boyutlu Hidrodinamik Tanm Modeli.Trkiye Kylar 10:
Trkiyenin Ky ve Deniz Alanlar VIII. Ulusal Kongresi Bildiriler Kitab.Vol. 3, pp.1121-1135.
8. Balas, L. and zhan, E. (1994). Modeling of Water Quality in Enclosed Water Bodies.Proceedings of the Second
Workshop on Coastal Engineering.Istanbul: Yldz Technical University, pp. 43-54.
9. Balas, L. and zhan, E. (1997). A Comprehensive Mathematical Model for Coastal Water Pollution
Studies.Proceedings of International Conference on Water Problems in the Mediterranean Countries. Nicosia, North
Cyprus:Near East University, pp. 1049-1156.
10. Balas L. (1998). Three-dimensional numerical modelling of transport processes in coastal water bodies. Unpublished
PhD thesis. The Graduate School of Natural and Applied Sciences, Middle East Technical University, Ankara, Turkey.
11. Balas, L. and zhan, E. (2000). An Implicit Three Dimensional Model Simulate Transport Processes in Coastal Water
Bodies.International Journal for Numerical Methods in Fluids(34): 307-339.
Biographies
Kaan CEBE was born in Ankara in 1975. He received his Bachelors degree from Gazi University,
Department of Civil Engineering, Ankara in 1999 and earned his Master of Science degree from Gazi
University, Institute of Science and Technology in 2002. He has been pursuing his Ph.D. research in the same
institution since 2009. His field of study is coastal hydraulics, coastal engineeringand numerical modeling. Mr.
Cebe is a member of Ky Alanlar Ynetimi (KAY) Trkiye Milli Komitesi (Turkish National Committee of
Coastal Management).
Lale Balas was born in Ankara in 1969. She received her B.Sc. in Civil Engineering from Middle East
Technical University, Ankara in 1990 and earned M.Sc. degree again from Middle East Technical University in
1992. She received her Ph.D. in Coastal Engineering in 1998. Dr. Balas worked as Assistant Professor in Gazi
University, Faculty of Architecture and Engineering, Ankara from 1999 to 2002. In 2002 she was appointed as
Associate Professor in same faculty. She is serving as a Professor in Gazi University, Civil Engineering
department since 2008. Her field of study is hydromechanics, fluid dynamics, coastal engineering and numerical
modeling.
She participated in and directed various scientific and engineering research projects in the field of coastal
engineering and numerical modeling. Some of these projects are; British Council- ALP: Three Dimensional
Hydrodynamic Modelling of Coastal Water Bodies,(1992-1993);NATO -CCMS: Pilot Study on Ecosystem
Modelling of Coastal Lagoons for Sustainable Management, (1995-1997); TBTAK-NTAG 821: Three
Dimensional Mathematical Modelling of Hydrodynamics, Salinity, Temperature and Pollution Distribution of
Coastal Lagoons, (1996-1997); TBITAK-MSAG 85: Nonlinear and Linear Self Control Strategies of Open
Loop Unsteady Reactors, (1996-1998) and TBTAK-NTAG 835: Development of a Reliability Model for
the Preliminary Design of Coastal Structures, (20002002). She is a member of MEDCOAST Institute and
director of Ky Alanlar Ynetimi (KAY) Trkiye Milli Komitesi (Turkish National Committee of Coastal
Management).
1080
Abstract. Besides providing resources like product-servise and human at intended time and demanded
place, logistics also has meaning as operation responsability, geogaphic positioning of raw material, handing
process and meeting the requeriments with the minimum cost. The most expensive part of logistics is
transportation. In recent years combined transformation using maritime lines come into prominence due to its
advantages. In order to minimize the logistics costs, a different perspective is used at this study. Preventing
the loses of container-carriying ships while coming alongs due to the harbors, hydrodinamical, geotechnical
and structural design principles of breakwaters these are important infrastructural harbor member, are studied.
Most economic material which can be used at the armour layer of the rubble mound based composite
breakwaters is determined by calculating thickness of rubble mounds one by one.
Keywords: Rubble- Mound Breakwaters,Composite Breakwaters,Hydrodynamic Analysis, Cost Analysis
1 Introduction
Logistics can be divided into three process as provision logistic, material transport and delivery logistic. Material
transport is the most expensive and requiring the best planning part of the logistic. Rotating, programming,
stocking cost, position of delivery centers, type and quantity of transported material, rivalry strategies and traffic
problems are the issues we have to face during the process. At this study, design of harbor infrastructure that
effects logistic cost indirectly when maritime transportaion is used as transport method, is studied.
Breakwaters which are built for minimizing the wave effects at harbors to maintain comfortable movements of
ships and preventing the scour of solid material at harbor entrance and increasing the capacity of dock with
extending its length, are one of the most important harbor infrastructure. It has types that changes according to
existing conditions as quarry stone monolithic and special type. Quarry stone has the longest service life and
they are most resistant to wave effects. Monolithic breakwaters require small amount of material and they need
small area. During the planning process, the most convenient breakwater type suitable for the environmental
conditions should be chosen by considering these advantages. Then hydro-dynamical, geotechnical stability and
economical analysis have to be made. At this study analysis for rubble mound and composite breakwaters, that
are commonly used breakwaters types, are made.
1081
blocks, having weight that cannot move by the predicted wave effect. Especially in our country rubble mound
breakwaters are the most used type.
sH3
K D ( S r 1) 3 cot (1)
slop with
contains parameters affecting the stability of block such as; armour unit type, block surface roughness, porosity,
and floor geotechnics.
Tablo 1. KD Stability Coefficients
On Core KD
On Armour Layer KD
Revetment
Placement
Slope
NonNonn
Breaking
Breaking
System
Pattern
Breaking
Breaking
Wave
Wave
Wave
Wave
Smooth
1,52 Random
1,2-2,1
2,4
1,1-1,7
1,9
Stone
2,0
Rough
1,51 Random
2,9
2,3
Stone
2,0
Tetrapod
2 Random
7
8
4,5
5,5
2,0
Tribar
1 niform
12
15
7,5
9,5
Some of the values of stability coefficients are given in Table (1). The block weight can be obtained using these
values in the formula developed by Hudson.
gT 2
2 (2)
K s H / H 0' H 0' K r H 0
L0
(3)
b0
b1
Kr
cos 0 cos 1
(4)
H H0 Kr Ks
(5)
b is the spacing between two wave orthogonals and b 0 is in the deep water. Calculations along the surface of
breakwater with varyingdepthandanglevalues are presentedin tables within the scope ofnumericalapplication.
1082
W
B nK
s (6)
The equation (6) is giving the width of the peak. n and K respectively shows number of lines of stones and
coefficient of layer. Layer thickness is calculated in the following relation;
W *
r nK
s
1/ 3
(7)
W* indicates the weight of the rubble mound in the related layer. Number of blocks is found with following
equations;
P a
N AnK 1
100 W
1/ 3
(8)
Aa N.W
Aat Lb Aa
(9)
Subbase armour layer volume (VL) and weight(Awt)can be written as;
E 12 E 12 (r1 / cos )
r1 r2
VL
r1 10
(10)
2 sin
2 sin
sin
Awt VL a Lb (11)
Where E is the breakwater peak elevation, ra is armour layer thickness; rl is subbase armour layer. The core
settlement volume (Vc) and weight (Awc) are expressed by the following relations;
2
Vc
r1
1
12 E
1,5 cot (12)
2
cos
Awc Vc a Lb (13)
Total breakwater weight is calculated by the next equation;
Atotal Awc Awt At (14)
Multiplying of unit prices concerning different filler materials used in construction of breakwater with total
weight calculated by type of material provides the cost of rubble mound breakwater.
1083
Fig.1.Hydrodynamic Pressure Acting on the Caison Breakwater d the water height from armourstone of vertical
breakwater, h is the water depth infront of the structure, h w is the depth of vertical breakwater body from the
static water level, hc is the height of crest from the static water level
Pressure values calculated by GODA method are expressed by;
P2
P1
cosh( 2h / L)
(16)
P3 3 .P1 (17)
P4 4 .P1
(18)
1084
Similarly, the equations for the total lifting force and moment;
U (1 / 2).PU .B (23)
M u (2 / 3).U .B (24)
where B is the width of the caisson structure. The sliding and overturning resistance of caisson breakwater are
defined as follows respectively;
SFs .(W U ) / P
(25)
SFo ( wt M U ) / M P
Where SFs is the safety coefficient aganist sliding and SFo is the the safety coefficient aganist turnovering. These
coefficient values are wanted to be larger than 1,2 and coefficient of friction between the armourstone and the
body of breakwater can be taken as =0,60. W is the body weight, t is the distance from heel to center of
gravity of body.
4. Numerical Application
Design wave parametres are shown in Table (2) for the slope-surfaced rubble mound breakwater and the various
between the wave characteristics parametres acting on the breakwater and depth is given in Table (3).
1085
wave
Deep-water
angle
0)
8 sn
45
6m
Bottom
(m)
Velocity of
wave (c)
65,04 m
27,390
8,11 m/sn
1,67 m
57,69 m
24,21
7,18 m/sn
1,69 m
6,00 m/sn
1,75 m
4,33 m/sn
1,98 m
4m
48,02 m
19,88
2m
34,72 m
14,480
slope
0,02
wave
of
According to the calculations, wave breaking depth was determined as 2,85m. In this case wave is unbroken in
8m depth thus this depth is suitable for construction of the breakwater. The breakwater as shown in Fig (3). has
been designed to absorbe the waves in the A surface of 300m along the line.
Cost (TL)
Hearthstone
Tetrapod
Tribar
Material Type
Dolos
1086
As a result of the economic analysis, using hearthstone instead of specially produced prefabricated blocks gets a
significant advantage in terms of cost. Vertical composite breakwaters analyses have been done on a model
whom plan and cross section are given in Fig.(5).
Fig. 5. Plan and Cross Section of the Vertical Faced Composite Breakwater
Design wave parametres to be used in hydrodynamic analysis are given in Table (4).
Deep-water
height (H0)
3,0 m
600
10 sn
Bottom
slope (m)
0,01
Scope of this study, following hydrodynamic, geotechnical and stability analysis, the cost has been calculated
according to amount of required materials. Variation of the cost has been investigated by changing the thickness
of the rubble mound base layer of the composite breakwater.
Cost (TL)
caisson
rubble
mound
Total
Height of rubble mound layer (m)
1087
design. This study aims to reduce the cost of maritime transport which is the most preffered channel in rapidly
developing combined transport with overseas trade, with economic optimization of port infrastructure facilities.
Breakwaters are one of the most important facilities in the ports thus designing of the breakwaters accurately
contributes to the economic aspects in the long term.
For this reason, the cost of logistics service associated with the port facility cost indirectly. Using rubble mound
breakwaters which are more stable than the others, especially in deep waters is recommended in this study
Although the less costly material is used for composite, facility is more costly. If the design load is exceed,
caisson is damaged and loses its function. These resons feature the rubble mound breakwaters in terms of
strength and economy.
References
Bruun, P., Design and Construction of Mounds for Breakwaters and Coastal Protection, Elsevier
Science Publishing Co, Amsterdam, 1985.
2. Goda, Y., Random Seas and Design of Maritime Structure, University of Tokyo Press, Tokyo, 1985.
3. Ergin, A.,Coastal Engineering ,Department of Civil Engineering Middle East Technical University,
Ankara, 2009.
4. Gokkus, U., Marine Structures, Manisa, 2001.
5. Japan International Agency(JICA,2), Bearing Capacity and Slope Stability, International Course on
Port and Harbor Engineering, Bureau of Ports and Harbors, Ministry of Transport, Japan, 1990.
6. Kapdasl, S., Ky Mhendislii, Istanbul Technical University Civil Engineering Faculty Printing
House, 1968.
7. Hall, K. R., HETTIARACHCHI, S., Mathematical Modelling of Wave nterpaction with Rubblemound
Breakwaters, Coastal Structures and Breakwaters, Conference by the Instution of Civil Engineers,
British Library, 1992.
8. Mc. Carthy, D. F.,Essential of Soil Mechanics and Foundations, Prencite-Hall CO., Virginia, 1982.
9. Tanimoto, K., Takahashi, S., Japanese Experiences on Composite Breakwaters, Proceedings of
International Workshop on Wave Barriers in Deepwater, Port and Harbor Research Institute, Japan,
1994.
10. Ministry of Transport,Railways, Ports andAirportsConstructionGeneralDirectorate,Kiyi Yapilari
ve Limanlar Planlama ve Tasarimi Teknik Esaslari, Ankara, 2007.
11. Yuksel, Y.,Cevik E. ,Ky Mhendislii, Beta Press,Istanbul, 2009.
12. http://www.frmtr.com/kultur/3450864-turkiye-limanlarimiz.html.
1.
1088
Abstract. The aim of this paper is to establish the existence of positive solutions of a third-order
-point
boundary value problem on time scales with an increasing homeomorphism and homomorphism, which
generalizes the usually -Laplacian operator. By using the Krasnosel'skii-Guo fixed point theorem and Four
Functional fixed point theorem, we establish the existence of at least one positive solution and by using a
generalization of Leggett-Williams fixed point theorem, we establish the existence of at least three positive
solutions for the third-order
) is
1 Introduction
The theory of time scales, which has recently received a lot of attention, was introduced by Stefan Hilger in
his PhD thesis in 1988 [1]. Theoretically, this new theory not only unify continuous and discrete equations, but
has also exhibited much more complicated dynamics on time scales. Moreover, the study of dynamic equations
on time scales has lead to several important applications, e.g., insect population models, biology, neural
networks, heat transfer and epidemic models (see [2,3] and references therein). Some preliminary definitions and
theorems on time scales also can be found in books [2,3] which are excellent references for calculus on time
scales.
Multipoint boundary value problem (BVP) arise in a variety of different areas of applied mathematics and
physics [4]. The study of multipoint BVPs for linear second-order ordinary differential equations was initiated
by Ilin and Moiseev [5], since then many authors studied more general nonlinear multipoint boundary value
( )
problems ([6,7] and references therein). Recently, when
is -Laplacian operator, that is
| |
(
) and the nonlinear term does not depend on the first-order derivative, the existence problems of
positive solutions of boundary value problems have attracted much attention. One can notice that the oddness of
the -Laplacian operator is key to the proof. However in the operator is not necessary odd, so it improves and
generalizes the -Laplacian operator.
Avery, Henderson and ORegan [8] present Four Functional Fixed Point Theorem which is a major
generalization of the original Leggett-Williams Fixed Point Theorem [9]. This fixed point theorem was also
1089
studied by Sun et all.[10]. By using the four functional fixed point theorem and five functional fixed point
theorem, they obtain the existence criteria of at least one positive solution and three positive solutions for the
nonlocal BVP with -Laplacian. Liu and Sun [11] discussed the existence of at least three and arbitrary odd
number positive solutions for -point boundary value probles on time scale. They used a generalization of
Leggett-Williams fixed point theorem [12]. Xu and Meng [13] considered the third order -laplacian -point
boundary value problems on time scales. They obtained the existence of positive solutions by using fixed point
theorem in cones. In particular, the nonlinear term ( ) is allowed to change sign.
Motivated by the above results, in this study, we consider the following third-order
problem (BVP) on time scales
( .
( )/)
(
( ) (
(
))
( ))
( )
( )
( )/
(1)
where
is an increasing homeomorphism and homomorphism and ( )
A projection
called an increasing homeomorphism and homomorphism if the following conditions are satisfied:
i.
ii.
iii.
If
is
( ) for all
then ( )
;
is continuous bijection and its inverse mapping is also continuous;
( )
( ) ( ), for all
We will assume that the following conditions are satisfied throughout this paper:
(H1)
,
(H2)
such that ( )
( )
(
(H3)
) is continuous,
(,
- ,
+ exist.
By using the Krasnosel'skii-Guo Fixed Point Theorem [14,15] and Four Functional fixed point theorem [8], we
establish the existence of at least one positive solution and by using a generalization of Leggett-Williams fixed
point theorem [12], we establish the existence of at least three positive solutions for the above boundary value
problem. In particular, the nonlinear term ( ( )) is allowed to change sign. The remainder of this paper is
organized as follows. Section 2 is devoted to some preliminary lemmas. We give and prove our main results in
Section 3.
2 Preliminaries
,
( )
( )
) ( )
( )/)
( )
( )
has the unique solution
) ( )
(3)
( )
( )/
))
1090
(4)
( )
where
( )
( )
(5)
( )
where
- and ( )
| ( )|
,
- be endowed with the ordering
,
- and
( ) for all
if ( )
|
(
)|
(
)
is
defined
as
usual
by
maximum
norm.
Clearly,
it
follows
that
is
a
Banach
space. For
, the convenience, let
Let
( )
( )
( )
where
*
( )
( )
*
( )
+
is defined in Lemma 3. Define the operators
(
(
)( )
.
(
( ) (
( ))
(6)
,
- +
,
and
by setting
4 ( ) (
( ))
)( )
where
{ (
( ) (
( ))
(
( )
) ( )
(9)
( )
( )) } Obviously,
Lemma 4.
) ( )
5
(8)
where
(7)
( ))
( )
( ))
and
( ))
is completely continuous.
by
can
3 Main Results
Theorem 5. [14,15]Let be a Banach space and let
be a cone in . Assume
are open subsets of
(
)
with
and let
be a completely continuous operator such that, either
(i)
and
or
(ii)
and
(
)
Then has at least one fixed point in
Set
* ( )
+
{ ( )
}
(10)
, - ( )
Lemma 6. [15] defined above has the following properties:
(a)
(b)
is open relative to
(c)
if and only if
, - ( )
,
( )
(d) if
then
for
For the convenience, we introduce the following notations:
1091
) ( )
) ( )
)
)
(
(
)(
)(
,
,
)
)
,
,
) ( )
( ))
( )
4 ( )
)( ( )
( ))
and
and
( )
be defined as
Also
( )
Therefore,
(
)( )
Next, for
( (
) ( )
for
) ( )
(11)
( )
( )
( ))
( )
)( ( )
( )
( )
( )
Therefore,
(
This implies that
)( )
) ( )
for
) ( )
(12)
1092
with
) ( )
) ( )
( ( )
) ( )
(B1) (
(B2) (
. / for all (
/ for all (
(B3) ( )
for all ( ) ,
then the BVP (1) has a fixed point
Define maps
( )
( )
(
and let
,
,
( )
) with ( )
(
-,
( )
( ) (15)
)( )
) we have
)( )
and
(
(
)( )
)
(
) ( )
we have
)( )
) ( )
( ( )
) ( )
) with ( )
(
is completely continuous.
and
)
) ( )
) with ( )
(
( (
Clearly,
and
8 is satisfied.
(
For all
,
such that
( )
) is bounded and
(
)
( )
which means that (
( )
(14)
with
( ) ( )
( ) ( )
, ) and (
) be defined by (13).
(
4 (
) ( )
1093
) Clearly,
( )
Thus (iii) and (v) in Theorem 8 are satisfied. has a fixed point
in (
,
- By condition (B3), we have ( ( ))
,
- that is
( ( ))
( ( )) Hence,
This means that is a fixed point of operator
Therefore, the BVP (1) has at least one positive
solution.
Suppose
(C2) (
. / for (
(C3) (
2 .
2 .
)
,
- If there are positive numbers
for
3 such that the following conditions are satisfied
/3 for (
( )
with
( )
( )
/3 for (
satisfying
( )
( )
,
( )
( )
Proof. By the definition of , it suffices to show that the conditions of Theorem 10 hold with respect to the
operator S. Firstly, we show that if the condition (C1) is satisfied, then
(
(
)
)
(20)
(
) then assumption (C1) implies
If
(
(
{ (
) ( )
)}
) ( )
1094
{ (
) ( )
,
)}
)
(
As the argument above, we have (
,
- It is easy to see that
We choose ( )
) ()
( ) (
,
(
) there are
( )
Therefore, for
,
- So, by the definition of
( )
. /,
(
( )
( )
Hence, condition (i) of Theorem 10 holds. We finally prove that (iii) in Theorem 10 holds. In fact, for
(
) with ( )
we have ( )
(
) Clearly,
Thus from Theorem 10, the BVP (1) has at least three fixed points
,
( )
( )
( )
By condition (C1), we have
( ( ))
,
- that is
( ( ))
( ( )) Hence,
This means that
are fixed points of
operator Therefore, the BVP (1) has at least three positive solutions.
References
1.
2.
3.
4.
5.
6.
7.
8.
9.
10.
11.
12.
13.
14.
15.
Hilger, S. (1990). Analysis on measure chains-a unified approach to continuous and discrete calculus.
Results in Mathematics. 18:18-56.
Bohner, M.; and Peterson, A. 2001. Dynamic Equations on time scales, An Introduction with
Applications. Birkhauser
Bohner, M.; and Peterson, A. 2003. Advances in Dynamic Equations on Time Scales. Birkhauser.
Adem, J.; and Moshinsky, M. (1950). Self-adjointness of a certain type of vectorial boundary value
problems. Boletn de la Sociedad Matematica Mexicana. 7:117.
Ilin, V. A.; and Moiseev, E.I. (1987). Nonlocal boundary value problem of the second kind for a
Sturm-CLiouville operator. Difference Equation. 23: 979987.
Bai, Z.; Ge, W.; and Wang, Y. (2005). Multiplicity results for some second-order four-point boundaryvalue problems. Nonlinear Analysis. 60:491500.
Zou, Y.; Hu, Q.; and Zhang, R. (2007). On numerical studies of multi-point boundary value problem
and its fold bifurcation. Applied Mathematics and Computation. 185:527537.
Avery, R.; Henderson, J.; and O'Regan, D. (2008). Four functionals fixed point theorem. Math. Comput.
Modelling. 48:1081-1089.
Leggett, R.W.; and Williams, L.R. (1979). Multiple positive fixed points of nonlinear operators on
ordered Banach spaces, Indiana Uni. Math. Journal. 28:673-688.
Sun, T.-T.; Wang, L.-L.; and Fan, Y.-H. (2010). Existence of positive solutions to a nonlocal boundary
value problem with p-laplacian on time scales. Adv. Difference Equations. Art. ID 809497:1-15.
Liu, J.; and Sun, H.-R. (2011). Multiple positive solutions for m-point boundary value problem on time
scales. Bound. Value Problems. Art. ID 591219:1-11.
Bai, Z.; and Ge, W. (2004). Existence of three positive solutions for some second-order boundary value
problems. Computers & Mathematics with Applications. 48:699707.
Xu, F.; and Meng, Z. (2009). The existence of positive solutions for third-order p-laplacian m-point
boundary value problems with sign changing nonlinearity on time scales. Adv. Difference Equations.
Art. ID 169321:1-14.
Deimling, K. 1985. Nonlinear Functional Analysis. Berlin: Springer.
Guo, D.; and Lakshmikantham, V. 1988. Nonlinear Problems in Abstract Cones. New York: Academic
Press.
1095
Biographies
Fatma Tokmak was born in Marmaris, Turkey on August 2nd, 1987. She received a Bachelors degree in
Mathematics from Ege University and started a master programme in the field of analysis and the theory of
functions analysis at the same university.
She is interested in analysis on time scales.
lkay Karaca was born in zmir, Turkey on May 24 th, 1967. She received a Bachelors degree in Mathematics
from Ege University in Turkey, a Masters and a PhD in Mathematics from Ege University.
She is an Associate Professor of Mathematics at Ege University in Izmir, TURKEY. Dr. Karacas research
interests include spectral theory, and time scales.
1096
Abstract. In this paper, a collocation method based on the Bessel polynomials is introduced to solve system of linear
Fredholm-Volterra integral equations (FVIEs). Numerical examples are given to demonstrate the validity and the
applicability of the method and the comparisons are made with the existing results. The results show the efficiency and
accuracy of the present method. All of the numerical computations have been performed on a computer using a program
written in MATLAB v7.6.0 (R2008a).
Keywords:System of Volterra integral equations; Bessel polynomials and series; collocation method; approximate method;
collocation points.
1. Introduction
Systems of the linear integral equations, integro differential equations have a very important place in science and
engineering. Since the systems of the mentioned equations are usually difficult to solve analytically, it is
required to obtain the approximate solutions. Recently, several authors have solved the mentioned systems using
the numericalmethods given in [1-9]. In addition, the Taylor, Chebyshev and Legendre (matrix and collocation)
methods have been used by Sezer et al. [10-12] to solve linear differential, integral and integro-differential
equations and systems of them.
Yzba et al. [13-15] has been worked the Bessel collocation method to obtain the approximate solutions of
the linear neutral delay differential, the pantograph and Fredholm integro-differential equations.
In this study, we will develope the Bessel collocation method in given [13-15] for the approximate solutions
of a system of linear Fredholm-Volterra integral equations (FVIEs) with variable coefficients in the form
b k
x k
k
f
P
(
x
)
y
(
x
)
g
(
x
)
K
(
x
,
t
)
y
(
t
)
dt
i , j
i, j
K iv, j ( x, t ) y j (t )dt
j
i
j
j 1
a j 1
a j 1
i 1, 2, , k , 0 a x, t b
(1)
, k can be
expandedmaclaurinseries.
Our aim is to find an approximate solution of (1) expressed in the truncated Bessel series form
N
yi ( x )
i ,n
J n ( x) ,
i 1, 2,
(2)
,k
n 0
so that ai ,n , n 0,1, 2,..., N are the unknown Bessel coefficients and J n ( x ) , n 0,1, 2,
polynomials of first kind defined by
1097
x
k !( k n)! 2
( 1)
J n ( x)
k 0
2 k n
, n
, 0 x .
J ( x) DX ( x) J ( x) X( x) D
T
where
J ( x) J 0 ( x)
(3)
J N ( x) , X( x) 1
J 1 ( x)
and the coefficientmatrix D can be seenclearlyfrom[13-15]. We consider the desired solution y j ( x ) of Eq. (1)
defined by the truncated Bessel series (2). Now, we write the matrix form of the function defined in relation (2)
as
[y
( x)] J ( x ) A , A j a j ,0
j
a j ,1
a j ,N
j 1, 2,
,k
,k .
(4)
y ( x) X( x) DA
(5)
where y ( x) y1 ( x)
y2 ( x )
Xr ( x) X( x) , r 1, 2,
A A1
, k , D diag D1 , D2 ,
, Dk , D r D
, X k ( x) ,
and
Ak .
T
A2
(6)
b
so that P( x) pi , j ( x) , y ( x) yi ( x) , g( x) gi ( x) , F ( x ) = K ( x,t )y (t ) dt ,
f
K ( x,t ) K i , j ( x, t ) ,
f
V ( x ) = K ( x,t )y (t ) dt , K ( x,t ) K i , j ( x, t ) ,
v
F( x) = Fi ( x) ,
V( x) = Vi ( x) ,
, k , j 1, 2,
i 1, 2,
b
,k ,
Fi ( x ) K i , j ( x, t ) y j (t ) dt ,
f
(7)
j 1
Vi ( x )
and
K
a
v
i, j
(8)
( x, t ) y j (t ) dt .
j 1
Let us now find the matrix form of F ( x) in Eq. (6). The kernel function K i , j ( x, t ) can be approximated by
the truncated Maclaurin series of degree N in the form
N
K i , j ( x, t )
f
T,f
i, j
m n
k mn x t ;
T,f
k mn
i, j
m n
K (0, 0)
m ! n ! x t
so that m, n 0,1, 2, , N , i 1, 2, , k , j 1, 2, , k .
The expression (9) can be put in the matrix form
m
(9)
m0 n 0
Ki , j ( x, t ) X( x)K T , f X (t ) ;
f
i, j
K T , f
i, j
T,f
kmn .
i, j
(10)
1098
Also, the Bessel matrix form of the kernel function K j ( x, t ) can be written as
Ki , j ( x, t ) J ( x)K B , f J (t ) .
f
i, j
(11)
J ( x)K B , f J (t ) X( x)K T , f X (t ) .
i, j
i, j
(12)
Substituting the matrix J ( x ) in Eq. (3) into the relation (12) and simplifying the result equation, we find the
matrix relation
K B , f (D ) K T , f D
i, j
i, j
kmn , m, n 0,1, 2,
B, f
, N , i, j 1, 2,
i, j
,k .
(13)
f
Substituting the matrix forms (11) and (4) corresponding to the functions K i , j ( x, t ) and y j ( x ) into the Fredholm
integral part Fi ( x ) given by Eq. (7), we get the relation
b
[ Fi ( x)]
J ( x )K
i, j
B, f
J (t ) X(t ) D A j dt
T
a j 0
J ( x )K
i, j
B, f
Q f ( x) A j
(14)
j 0
H f ( x) X (t ) X(t ) dt hr ,s ( x) , hr , s ( x )
r s 1
r s 1
[ Fi ( x)] X( x) D K B , f Q f ( x) A j .
T
i, j
(15)
j 0
Smilarly, we obtain the matrix form of the Volterra integral part Vi ( x ) in Eq. (8) as
k
i, j
(16)
j 0
so that K B ,v (D ) K T ,v D , K T ,v k mn , K i , j ( x, t )
i, j
i, j
i, j
T ,v
i, j
T ,v
i, j
m n
k mn x t ,
m0 n 0
m, n 0,1, 2,
,N ,
i, j 1, 2,
Qv ( x) DH v ( x)D ,
T
,k ,
H v ( x ) X (t ) X(t )dt hr ,s ( x ) ,
T
hr , s
r s 1
r s 1
(17)
PY G + F + V
(18)
where
P = diag P( x0 ), P( x1 ),
G g( x0 )
g( x1 )
V V( x0 )
V( x1 )
, P( xN ) , Y y ( x0 )
y ( xN )
y ( x1 )
g ( xN ) ,
F F( x0 )
F( x1 )
F ( xN )
and
V ( xN ) .
T
By substituting the collocation points (17) into Eq. (5), we obtain the matrix relation
(19)
y ( xs ) X( xs )DA , s 0,1, , N Y XDA
so
that
X X( x0 )
X( x1 )
X( xN ) ,
T
X( xs ) diag X1 ( xs ), X2 ( xs ),
, Xk ( xs ) ,
r 1, 2, , k and s 0,1, , N .
By placing the collocation points (17) into Eqs. (15) and (16), we have matrix relations
1099
Xr ( xs ) X( xs ) ,
[ Fi ( xs )] X( x) D K B , f Q f ( xs ) A j ;
T
i, j
(20)
j 0
i, j
, N , i 1, 2,
(21)
, k.
j 0
Similarly, substituting the collocation points (17) into the matrix F ( x) and V ( x) defined in relation (6) and using
relations (20) and (21), we gain the matrix relations
F( xs ) = F1 ( xs )
F2 ( xs )
Fk ( xs ) X( xs ) DK f Q f ( xs ) A ;
(22)
V( xs ) = V1 ( xs )
V2 ( xs )
Vk ( xs ) X( xs ) DK vQv ( xs ) A
(23)
where X( xs ) diag X1 ( xs ), X2 ( xs ),
D diag D1 , D2 ,
, Xk ( xs ) ,
, Dk ,
Xr ( xs ) X( xs ) ,
r 1, 2,
s 0,1,
,k ,
K f = K B , f ,
Dr D ,
Q f ( xs ) diag Q f ,1 ( xs ), Q f ,2 ( xs ),
, Q f , k ( xs ) ,
, Q v, k (xs ) and Qv ,r ( xs ) Qv ( xs ) .
,N ,
i, j 1, 2,
i, j
,k ,
K v = K B ,v ,
Q f ,r ( xs ) Q f ( xs ) ,
i, j
Thus, by using Eqs. (22) and (23), the matrices F and V in Eq. (18) can be expressed in matrix form in terms of
the matrix of Bessel coefficients A as follows:
F F( x0 )
F( x1 )
V V( x0 )
V( x1 )
F( xN ) XDK f Q f A ;
(24)
V( xN ) XDK vQv A
(25)
so that X diag X( x0 ), X( x1 ),
, X( xN ) , D diag D1 , D2 ,
, N 1 , K f diag K f ,1 , K f ,2 ,
p 1, 2,
K v diag K v ,1 , K v ,2 ,
, DN 1 , D p D ,
, K f , N 1 , K f , p K f , Q f Q f ( x0 )
, K v , N 1 , K v , p K v and Qv Qv ( x0 )
Q f ( x1 )
Q f ( xN )
Q v ( xN ) .
T
Qv ( x1 )
By replacing relations (19), (24) and (25) into Eq. (18), we have the matrix equation
PXD - XDK Q
f
XDK vQ v A G .
(26)
Hence, the matrix equation (26) corresponding to Eq. (1) can be written in the form
(27)
WA G or [ W; G ]
which corresponds to a linear system of k ( N 1) algebraic equations in k ( N 1) the unknown Bessel
coefficients so that
W PXD - XDK f Q f XDK vQv [ wp ,q ] , p, q 1, 2,
j 1, 2,
, k ( N 1) .
A k is determined solving this linear
A2
Since the truncated Bessel series (3) is approximate solution of (1), when the function yi , N ( x) , i 1, 2, , k ,
and its derivatives, are substituted in Eq. (1), the resulting equation must be satisfied approximately; that is,
for x xq [a, b] q 0,1, 2,
m
Ei ( xq )
n 0
Pi , j ( xq ) y j ( xq ) g i ( xq )
j 1
Ei ( xq ) 10
If max 10
a j 1
kq
xq
K i , j ( xq , t ) y j (t )dt
f
v
i, j
( xq , t ) y j (t )dt 0 , (28)
i 1, 2,
,k
and
a j 1
( k q positive integer).
kq
10
( k positive integer) is prescribed, then the truncation limit N is increased until the
k
difference Ei ( xq ) at each of the points becomes smaller than the prescribed 10 , see [7,10-13, 15].
1100
4. Numerical examples
To show the efficiency of the proposed numerical method, we consider the following examples. The
computations asssociated with the examples were performed on the computer using a program written in Matlab
7.6.0 v2008a. In this regard, we report in Tables and Figures, the values of the exact solutions yi ( x ) ,
i 1, 2,...,k ,
the
approximate
solutions
i 1, 2,..., k ,
yi , N ( x ) ,
and
the
absolute
error
functions
ei , N ( x) yi ( x) yi ,N ( x) , i 1, 2,..., k , at the selected points of the given interval. In addition, we computed the
maximum absolute errors for example 2 using
ei , N yi , N ( x) yi ( x) max yi , N ( x) yi ( x) , a xb.
Example 1. Consider the system of the linear FVIEs given by
1
x
2
2
y
(
x
)
2
xy
(
x
)
g
(
x
)
(
x
t
)
y
(
t
)
(
x
t
)
y
(
t
)
dt
2
1
1
2
x ty1 (t ) xt y2 (t ) dt
1
0
0
1
x
y ( x) x 2 y ( x ) g ( x ) ( xt t 2 ) y (t ) xt 2 ) y (t ) dt 2 xty (t ) ( x 2t ) y (t ) dt
2
1
2
1
2
1
2
0
0
so that 0 x, t 1, g1 ( x) x 2 x
2
25
4x
3
x
6
12
20
16
67 4 4 37 3 1 5
2
x
x x.
and g 2 ( x) 6 x x
15
60 3
6
2
5
4
x
5
4
3
(29)
x ,
Example 2. [16] Let us consider the linear Volterra integral equation system
x
y
(
x
)
g
(
x
)
0
, 0 x, t 1
x
y ( x) g ( x) y (t ) ( x t ) y (t ) dt
2
2
1
2
0
with
the
exact
solutions
y1 ( x) cos( x)
and
y2 ( x) sin( x).
(30)
Here,
g 2 ( x) sin( x) x
and
1101
Fig. 1. Comparison of the absolute errors functions e1, N ( x ) for N 5 and N 10 of y1, N ( x) of the system (30).
Fig. 2. Comparison of the absolute errors functions e2, N ( x ) for N 5 and N 10 of y2, N ( x) of the system
(30).
Table 1 Comparison of the absolute errors for N 5, 10 of y1 ( x ) of the system (30)
Taylor method [7]
xi
e1,5 ( x )
Absolute errors
with TAM [16]
0
0.2
0.4
0.6
0.8
1
0
8.89e-008
5.6727e-006
6.43851e-005
3.59957e-004
1.364361e-003
0
9.27188e-004
5.45507e-003
1.38644e-002
2.56349e-002
4.19808e-002
Present method
e1,5 ( x )
e1,10 ( x )
0
2.0571e-007
2.3189e-006
3.8191e-005
2.8744e-004
1.3807e-003
0
1.7097e-014
1.6798e-013
1.2788e-011
2.9213e-010
3.2045e-009
1102
xi
e2,5 ( x )
Absolute errors
with TAM [16]
0
0.2
0.4
0.6
0.8
1
0
2.5e-009
3.244e-007
5.5266e-006
4.124248e-005
1.95682e-004
0
1.14715e-004
8.57201e-003
2.78243e-002
5.91212e-002
8.70896e-002
Present method
e2,5 ( x )
e2,10 ( x )
0
7.2964e-007
6.5961e-007
3.7140e-006
2.9572e-005
1.7560e-004
0
8.7264e-013
1.7413e-012
3.1888e-012
2.3141e-011
2.7802e-010
Table 3 The maximum absolute error e1, N for various N of y1 ( x ) of the system (30)
N
e1,N
5
6.7068
7
2.5150
8
3.6695
10
3.2045
102
105
107
109
12
5.56 10
11
Table 4 The maximum absolute error e2, N for various N of y2 ( x ) of the system (30)
N
e2,N
5
7.7642
7
2.5183
8
3.7146
10
2.7873
102
106
108
1010
12
6.51 10
11
5. Conclusions
In this study, we presented the Bessel collocation method for the approximate solutions of system of (FVIEs). In
addition, an interesting feature of this method is to find the analytical solutions if the system has the exact
solutions that are polynomial functions. This feature can be seen from Example 1. The comparision of our
results with other methods shows that the present method is a powerful mathematical tool for finding the
numerical solutions of system of linear Fredholm-Volterra integral equations. One of the considerable advantage
of the method is that the approximate solutions are obtained very easily by using the computer code written in
MATLAB v7.6.0 (R2008a).
References
[1] Yusufolu, E. (2007). An efficient algorithm for solving integro-differential equations system, Appl. Math.
Comput. 192(1): 5155.
[2]Javidi, M. (2009). Modified homotopy perturbation method for solving system of linear Fredholm integral
equations, Math. Comput. Modell. 50(1-2): 159-165.
[3] Maleknejad, K.; and Mirzaee, F. (2003). Numerical solution of linear Fredholm integral equations system by
rationalized Haar functions method, Int. J. Comput. Math. 80(11): 13971405.
[4] Arikoglu, A.; and Ozkol, I. (2008). Solutions of integral and integro-differential equation systems by using
differential transform method, Comput. Math. Appl 56(9): 24112417.
[5] Pour-Mahmoud, J.; Rahimi-Ardabili, M.Y.; and Shahmorad, S. (2005). Numerical solution of the system of
Fredholm integro-differential equations by the Tau method, Appl. Math. Comput. 168(1): 465478.
[6] Saberi-Nadjafi, J.; and Tamamgar, M. (2008). The variational iteration method: A highly promising method
for solving the system of integro differential equations, Comput. Math. Appl. 56(2): 346351.
[7] Sorkun, H.H.; and Yalinbas, S. (2010). Approximate solutions of linear Volterra integral equation systems
with variable coefficients, Appl. Math. Modell. 34(11): 3451-3464.
[8] Biazar, J.; Babolian, E.; and Islam, R. (2003). Solution of a system of Volterra integral equations of the first
kind by Adomian method, Appl. Math. Comput. 139(2-3): 249258.
[9] Maleknejad, K.; and Kajani, M.T. (2004). Solving linear integro-differential equation system by Galerkin
methods with hybrid functions, Appl. Math. Comput. 159(3): 603612.
[10] Akyz-Dacolu, A.; and Sezer, M. (2005). Chebyshev polynomial solutions of systems of higher-order
linear FredholmVolterra integro-differential equations, J. Franklin Ins. 342(6): 688-701.
1103
[11] Glsu, M.; and Sezer, M. (2006). Taylor collocation method for solution of systems of high-order linear
Fredholm Volterra integro-differential equations, Intern. J. Computer Math. 83(4): 429448.
[12] Yalnba, S.; Sezer, M.; and Sorkun, H.H. (2009). Legendre polynomial solutions of highorder linear
Fredholm integro-differential equations, Appl. Math. Comput. 210(2): 334-349.
[13] Yzba, .; ahin, N.; and Sezer, M. (2011). A Bessel polynomial approach for solving linear neutral delay
differential equations with variable coefficients, Journal Advanced Res. in Differential Equations 3(1): 81- 101.
[14] Yzba, .; ahin, N.; and Sezer, M. (2011). A Bessel collocation method for numerical solution of
generalized pantograph equations, Numer. Methods for Partial Diff. Eq., in press, (2011), doi:
10.1002/num.20660.
[15] Yzba, .; ahin, N.; and Sezer, M. (2011). Bessel matrix method for solving high-order linear Fredholm
integro-differential equations, Journal Advanced Research in Applied Mathematics, in press, (2011), doi:
10.5373/jaram.
[16] Rabbani, M.; Maleknejad, K.; and Aghazadeh, N. (2007). Numerical computational solution of the Volterra
integral equations system of the second kind by using an expansion method, Appl. Math. Comput. 187(2): 1143
1146.
Biographies
uayip YZBAI was born on 20 September 1984 in Karabk, Turkey. He entered the Department of
Mathematics of Faculty ofScience of Mula University in 2002 and graduated from the department with a first
class honours degree in 2005. He was a graduate student in the same department between 2006 and 2009. He has
been Ph.D. studentin the department since 2009 under the supervision of Dr. Niyazi AHN. Since 2007, he has
worked as a research assistant there.
He has published the following research papers related to numericalmethods.
Yzba, .; ahin, N.; and Sezer, M. (2011). A Bessel polynomial approach for solving linear neutral
delay differential equations with variable coefficients, Journal Advanced Res. in Differential Equations
3(1): 81- 101.
Yzba, .; ahin, N.; and Sezer, M. (2011). A Bessel collocation method for numerical solution of
generalized pantograph equations, Numer. Methods for Partial Diff. Eq., in press, (2011), doi:
10.1002/num.20660.
Yzba, .; ahin, N.; and Sezer, M. (2011). Bessel matrix method for solving high-order linear
Fredholm integro-differential equations, Journal Advanced Research in Applied Mathematics, in press,
(2011), doi: 10.5373/jaram.
Yzba, .; ahin, N.; and Sezer, M. (2011). Numerical solutions of systems of linear Fredholm
integro-differential equations with Bessel polynomial bases, Comput. Math. Appl. 61(10): 30793096.
He continues his research on numerical methods for solutions of ordinary and partial differential equations.
Niyazi AHN was born on March 15th 1970 in zmir, Turkey. In 1991, he received his B.Sc. degree from the
Department of Mathematics of Ege University in zmir, and M.A. from the University of Pittsburgh, USA in
1996. He then studied computational fluid dynamics for his Ph.D. degree under the supervision of Prof. William
Layton in the Department of Mathematics at the University of Pittsburgh, and graduated from that school in
2003.
His main research interests are computational fluid mechanics, numerical linear algebra and numerical
solutions of ordinary and partial differential equations. Some of his published articles are:
Yzba, ., ahin, N.; and Sezer, M. Numerical solutions of systems of linear Fredholm integrodifferential equations with Bessel polynomial bases, Comput. Math. Appl. 61(10): 30793096, 2011.
Akgnll, N., ahin, N.; and Sezer, M. A Hermite collocation method for the approximate solutions
of high-order linear Fredholm integro-differential equations,Numer Methods Partial Differential Eq.
2010, DOI: 10.1002/num.20604, 2010.
John V., Layton W. and Sahin N., Derivation and Analysis of Near Wall Models for Channel and
Recirculating Flows, Comput. Math. Appl., 48(2004): 11351151.
Mehmet SEZER was born on March 20th 1954 in Antalya,Turkey. He received his B.Sc. from the Department
of Mathematics of Ege University in 1976 and was research assistant at Balkesir Engineering Academy. His
M.Sc. was from Ege University in 1977. He received his Ph.D. under the supervision of Prof. Dr. Seda
MORALI in the Department of Applied Mathematics at Ege University in 1982 and studied in the field of
Parabolic partial differential equations.
1104
In 1982, he was Assistant Professor at Faculty of Education (Balkesir) of Uluda University, Associate
Professor in 1989, and in 1995 Professor at Faculty of Education of Dokuz Eylul University. Since 2004, he has
been Mathematics Professor at Faculty of Science at Mula University and has been the chair in the same
department.
His main research interests are ordinary and partial differntial equations, integral and difference equations,
and their numerical solutions. Prof. Sezer has been reviewer for numerous influential journals, has published
research articles related to differential equations, linear algebra, analytic geometry and calculus; and has been
authored over 80 papers.
1105
Abstract. The Monte Carlo simulation technique has been used to study thin film growth and crystallization
process. We exclusively investigated evaporative technique since it is the most preferred on the film growth
processes. The fomation of thin film has been simulated on 50x50x50 unit cubic region with 10000 A type,
10000 B type atoms with 10000 Monte Carlo Steps. Temperature effect has been handled with Boltzman
statistics. The role of substrate on crystallization of the thin film has been verified.
1 Introduction
Technological advancements reached a level where nano scale applications have become an important part of
scientific research. Among much different research area, development of thin film of nanometer thickness has
many applications. Experimental methods in nano scale applications require very advanced instrumentation [1].
As an alternative way to study the physical properties of thin films, computational powers of todays desktops
can be used. A physical model for the formation of the crystal structure in a thin film can be studied by
computational models [2]. Experimental studies give the end state of crystal structures without giving any clue to
mid-states while ordered lattice structure evolves. Although physical interactions between atoms are very
complicated, they can be represented by symbolic parameters and then the whole process of formation of the thin
film can be mimiced by Monte Carlo method. The temperature effect can be handled with Boltzmann statistics.
[3-4]. In this study, a thin film growth and its ordered crystal structure of chemical composition of two different
atoms (AB) has been studied.
2 Literature Review
Nowadays, an ordinary desktop has an enormous computational capability that was not available decades ago
on main frame computers which were available only at selective research centers in the world. Advancements on
computing technology makes computational study as a third branch in science besides experimental and
theoretical studies. Computational studies about thin film growth have been done on single atom films [5-7]. In
these studies, mostly Molecular Dynamics (MD) has been used to mimic the process. MD has advantage of using
energy and momentum conservation when inter-atomic interactions are handled. This requires substantially high
calculating power because to conserve energy and momentum of the system, the program must keep track of all
particles energy and momentum all the time throughout the simulation process. As a stochastic approach Monte
Carlo method with Metropolis algorithm [8] can be more useful than a deterministic one (MD) for this study. In
MC method, only the small part (arbitrarily chosen) of the system is taken into account at each decision process.
We use MC method with Metropolis algorithm to deal temperature effect (thermal fluctuations can be dealt
better by using MC method). We also study the process of crystallization by following the film growth through
analyzing successive MC steps (chemical composition is consisted of two different types of atoms). The process
1106
of reaching ordered state could be traced by investigating mid MC steps. This enables us to study the role of
substrate in obtaining good quality thin films. Previous studies have been concentrated films of only one type of
atom [7]. In this study we consider formation of thin film of two different type of atoms (AB type). Thus in this
work, the ordered crystallization can be traced through the process.
3 Methods
3.1 The Monte Carlo Method
In our simulation method, the most important part is to mimic interatomic forces with suitable models to be
able to write a feasible algorithm considering limited calculation capability of desktop computers.
The system is consisted of 50 unit length borders for each coordinate (x,y,z). 10000 A type, 10000 B type,
totally 20000, particles are represented A and B type atoms in the simulation region.
(a)
(b)
Fig. 1. Particle distribution (a) without any external effect after 10000 MC step (b) with %60 downward z
direction effect after 10000 MC steps.
According to the Fig 1a, particles are distributed randomly without any external effect like gravity or any
force while Fig 5b gives the particle distribution with applying %60 downward effect on the system.
Mathematicly, each particle has a label and all particle coordinate numbers are handled with matrices. First
all particles randomly distributed to the system uniquely. Next, one of them is chosen randomly. All possible
reaction steps of the choosen atom are checked to see which one is noteworthy to evaluate. The reactions may, of
course, involve neighboring sites [12]. 3 neighbouring cells for all directions were computed. By utilizing
Metropolis algorithm a choice is made and the chosen atom is moved accordingly. Then another particle is
chosen and the precious process is repeated. After each particle has been processed, a Monte Carlo time step has
been completed. The full simulation consists of a large number of such steps. The total physical time is
expressed in the number of Monte Carlo steps.
The interaction between atoms was mimiced by using representative interaction parameters which would be
determined in order to represent the fundamental physical expectations when two dissimilar (AB) or
similar(either A-A or B-B) type of atoms come close enough to each other [9]. Since we study the film of
chemical composition of AB type, we must somehow distinguish the two cases of having interactions between
similar and dissimilar atoms. AB type of atomic combination means that attractive force between atom A and
atom B must be stronger than the attraction between similar atoms (A-A or B-B).
There are many dosens of deposition technologies about material formation. [13], [14] These varius type of
coating technologies generally used to perform semiconductor materials. The evaporative ones handle thermal
conditions for growth of thin film over the substrate material. Thermal effects causes temperature fluctuations.
Temperature fluctuations were taken into account by using Boltzman statistics with MC method [8] [9], [10].
1107
Ea
k
e B T
(1)
Ea k T
B
e
Ep
Ea
In equation (1), Ea , denotes binding energy between the particles. According to this equation, the connection
between particles scaled to thermal effect. So the binding probabilty, E p , directly proportional to temparute (T).
The temperature effect over the particles indicates the crystallization of the system entirely. We measured the
crystallization that pictures the physical state of the whole system utilizing the z layers one by one.
Fig. 2. (a) Type of perfect crystallization: 6 dissimilar atoms (B type) (b) partial crystallization: 4 dissimilar
atoms (B type) and 2 similar (A type) atoms that neighbouring to S(i,j,k) (A type)
The number of dissimilar atoms that exist at neighbouring cells divided by total neighbour numbers and summed
up with CP. The mean value of CP for a certain level can be obtained as,
CP 100
1 N
CPi ,
N i 1
N=2500
(2)
Equation (2) mention about the crystallization for a z layer in 3D coordinate space.
The most suitable number of MC steps was determined by trial and error method to optimize computational
time with the stability of the end results [10].
1108
Fig. 3. Filling ratio of the first 6 layer according to the temperature scale
We studied temperature effect on crystallization. At low temperatures close to zero Kelvin, the filling factor
was not affected much as seen Fig 3 but crystallization shows a different behaviour than filling factor as seen Fig
4 which gives the impression that the crystallization (long-range order) has a discontinuity at a certain
temperature that may be be interpreted a phase transition on crystallization.
Fig. 4. Crystallization ratio of the first 6 layer according to the temperature scale
We also studied the effect of using suitable substrate on crystallization of thin film. Fig 5 shows the case for
low temperature where Fig 5a gives the distribution of atom A (green) and B (red) without using substrate where
Fig 5b gives the result with substrate.
1109
(a)
(b)
Fig. 5. Particle distribution of first layer at T=0.01 temperature (a) without suitable substrate (b) with suitable
substrate
Its clear that using substrate greatly improves crystallization that can even be detected visually. Fig 6
compares crystallization ratio of two cases analytically which supports the improving role of using proper
substrate.
Fig. 6. Crystallization ratio against the MC step for first layer with substrate and without substrate at T=0.01
temperature.
At higher temperature, Fig 7 gives the atomic distribution with substrate and without substrate. The long-range
order for both with/without substrate has been lost at this range of temperature as seen in Fig 8.
(a)
(b)
Fig. 7. Particle distribution of first layer (z=1) at T=0.15 (a) without suitable substrate (b) with suitable substrate
(WS)
1110
Fig. 8. Crystallization ratio against the MC step for first layer with substrate and without substrate at T=0.15
temperature
References
1.
2.
3.
4.
5.
6.
7.
8.
9.
10.
11.
12.
13.
14.
1111
Department of Mathematics, zmir Institute of Technology,35430, Glbahe Campus, Urla, zmir, Turkey
Department of Civil Engineering, zmir Institute of Technology, 35430, Glbahe Campus, Urla, zmir, Turkey
1
nurcangucuyenen@iyte.edu.tr, 2gamzetanoglu@iyte.edu.tr, 3gokmentayfur@iyte.edu.tr
Abstract. In this paper, iterative operator splitting method is used to solve numerically the transient two
dimensional (2D) solute transport problem in ground water. The method is based on first splitting the
complex problem (full advection-diffusion equation) into simpler sub-problems (advection and diffusion).
Then each sub-equation is combined with iterative schemes. The consistency of the proposed method is
investigated. The stability analysis of the iterative splitting is discussed in the sense of von Neumann for two
dimensional cases. It is seen that the von Neumann analysis achieves the linear stability criteria. The
numerical results show that the iterative splitting method provides more accurate results than other classical
splitting methods.
Keywords: Solute transport model, Iterative operator splitting method, Von Neumann analysis, Consistency
1 Introduction
The iterative splitting is a recent popular technique which is based on first splitting the complex problem into
simpler differential equations. Then each sub-equation is combined with the iterative schemes, each of which is
efficiently solved with suitable integrators [1], [2], [3], [4].
The iterative method is employed by several researches in solving one dimension convection-diffusion equation,
elastic wave equation, heat equation [1], [4] etc. The stability of this scheme is mostly carried out using Astability of ODEs sense [1], [4], [5], [6]. In this study, the iterative splitting method is the first time applied to
solve the transient two dimensional solute transport in ground water flow, [7], [8], [9], [10], [11], [12], [13].
Furthermore, this study explicitly derives the stability criteria using Fourier analysis, see [14], [15], [16].
The paper is organized as follows. In Section 2, the transient two dimensional solute transport model is
described. In section 3, the iterative operator splitting method and its application to the solute transport is given.
The consistency of the proposed method is explained in Section 4. In Section 5, the general linear stability
criteria which based on von Neumann stability analysis is derived and then tested on transient two dimensional
solute transport model. In Section 6, which is followed by the numerical results and conclusion part, numerical
experiments are carried out.
1112
water under the ground. One important segment of such management model is the numerical simulation of solute
transport. Since 1970s many models have been developed, based on analytical solutions such as Laplace
transform [7], finite difference methods [11], [12], [13], finite element method [10], finite volume method [8],
etc. All these models one way or another employ different solution techniques. This study, as an alternative to
the existing methods, employs the iterative splitting method to solve the transient 2D solute transport in ground
water flow, [8].
C
Dh C VC Rd C
(1)
t
where C=solute concentration; V =pore-water velocity vector; Rd =retardation factor; =first-order decay
coefficient; Dh =hydrodynamic dispersion tensor, are generally functions of velocity and the Dxx , Dyy ,
Rd
Dxy ( Dyx ) are generally functions of molecular diffusion. In this study, the case of homogenous and isotropic
medium under 2D ground-water flow with 2D dispersion is considered. Equation (1), using the iterative
splitting method, is solved to simulate unsteady 2D solute transport between two impervious boundaries. A
finite-length strip solute source, whose concentration is a given function of time, is located asymmetrically along
the y-axis at x=0 in a unidirectional seepage velocity field. The rectangular domain is 75m in the x-direction and
50m in the y-direction. The uniform pore velocity v is 0.1 m/day. The longitudinal, transverse, and cross
dispersivities, Dxx 1 , Dyy 0.1 ,and Dxy 0
respectively.The
retardation
factor
Rd 1 , and
decaycoefficient
C ( x, y,0) 0 .
(2)
C (0, y, t ) 0, 0<z<5 m
(3)
C (0, y, t ) 1, 0<z<15 m
(4)
C (0, y, t ) 0, 15<z<50 m.
(5)
Note that for the comparative study analysis we employed the same problem above which was considered by [8].
( A B ) t
c0 , t [0, T ] .
(7)
The method is based on iteration by fixing the splitting discretization step size t on time interval [ t , t
The following algorithms are then solved consecutively for i=1,3,,2m+1.
n
ci (t ) Aci (t ) Bci 1 (t ), ci (t n ) c n
(8)
ci 1 (t ) Aci (t ) Bci 1 (t ), ci 1 (t n ) c n
1113
n 1
].
where
c n is the known split approximation at time level t t n and c0 0 is the initial guess. The split
Ct vCx .
(10)
We then combine these equations by using the iterative splitting algorithms as follows:
(11)
n=1,2,,M+1. Then, after assembling the unknowns in and for each m, n, embedding the boundary conditions,
we have the following systems of equations in matrix form:
(12)
cxx A1 c , cyy A2c and cx B1c
Redefining equations (11) we have
(13)
(14)
where A Dxx A1 Dyy A2 , B vB1 are constant matrixes and f, g are vectors which come from boundary
n
n 1
conditions. After applying the midpoint rule to each iteration on each [ t , t ] interval, then we have:
t
t
t
t
ci n1 ( I A)1 (( I A)c ni B(c ni 1 c n 1i 1 ) ( f n g n f n 1 g n 1 ))
2
2
2
2
t 1
t
t
t n
n 1
n
n
n 1
ci 1 ( I B) (( I B)c i 1 A(c i c i ) ( f g n f n1 g n1 ))
2
2
2
2
where
t is time discretzation step and starting for i=1, initial guess c0 (t ) (0,
c1 (t ) c0 , c2 (t ) c0 .
c(t ) e
( A B ) t
t ] interval is
c0
( I ( A B)t
( A B) 2 2 ( A B )3 3
t
t O(t 4 ))c0 .
2!
3!
(15)
Considering the iterative splitting algorithms (8) and applying them to Cauchy problem (6) and combining with
2nd
order
Mid-point
rule
on
[ 0, t ]
interval,
we
have:
t 1
t
t
t
A) ( I A)c1 (0) ( I A)1 B(c0 (0) c0 (t ))
2
2
2
2
t
t
t
t
c2 () ( I B)1 ( I B)c2 (0) ( I B)1 A(c1 (0) c1 (t ))
2
2
2
2
c1 () ( I
1114
inserting the initial values c1 (0) c2 (0) c0 and initial guess c0 (0) c0 (t ) c0 ,
c1 (t ) into c2 (t )
t
t
t
t
c2 (t ) ( I B)1 ( I ) ( I B) 1 A
2
2
2
2
( I
and substituting
(16)
t 1 t
t
t
t
t
t
B)
A( I A)1 ( I A) ( I B) 1 A( I A) 1 (tB))c0
2
2
2
2
2
2
2
A2 AB B 2 BA
)
(17)
2
2
2
2
A3 A2 B BA2 BAB B3 B 2 A B3
t 3 (
))c0 O(t 4 )
4
4
4
4
8
4
8
c2 (t ) ( I t ( A B) t 2 (
A3 3 A2 B 6 AB 2 3BAB 3B 2 A B3 3
t O(t 4 )
12
and | c2 (t ) c(t ) | 0 as t 0 .
c2 (t ) c(t )
(18)
c ' Lc
(19)
where L is linear operator and suppose that we have a linear map resulting from the application of iterative
splitting schemes (8) with the 2nd order midpoint rule to equation (19) over one time step such that
c1n 1
c1n t
(20)
La n 1 Lb n
c2
c2 2
T
where La , Lb are matrixes of linear operators, ( 1 , 2 ) is vector with known constant components.
c10 c0 and c20 c0 are the initial conditions, and c11 , c12 are the approximations of c1 , c2 in function
space at time
t t 0 t .
In order to apply the stability theory, La , Lb must be manupulated into a matrix of scalars. This is done by
taking Fourier transforms of (20) as would be done in a von Neumann stability analysis. We get constant system
and to find the stability criterion for the linear map, we need to check the eigenvalues of its matrix, they must be
i 1 , where i= 1,2.
Example: Two dimensional Solute Transport equation
We consider the solute transport equation specified in Section 2. We can then write eqn. (1) as
t C vCx D xx Cxx Dyy Cyy 0
(21)
where the uniform pore velocity v is 0.1 m/day. The longitudinal and transverse dispersivities,
Dxx 1, Dyy 0.1. After rearranging equation (21) as in the previous example, we have
t C L1C L2C
(22)
where L1 v x and L2 Dxx xx Dyy yy . Application of iterative splitting schemes (8) with the 2nd order
midpoint rule to equation (22) over single time step, and then taking Fourier transform yields
1115
1
(v1 )i
1
(v1 )i
Ci n 1
2
n 1
t
t
Ci 1
(v1 )(1
(v1 )i )i
2
2
(1 t (v )i )(1 t ( D 2 D 2 ))
1
xx 1
yy 2
2
2
n
Ci
t
C n
1
( Dxx12 Dyy2 2 ) i 1
2
t
1
( Dxx12 Dyy2 2 )
2
t
t
1 ( Dxx12 Dyy2 2 )
(v1 )i
and
.
2
2
2
1
t
t
1 ( Dxx12 Dyy2 2 )
1 (v1 )i
2
2
For stability we know that | 1 | 1 and it is true for any choice of v,
1
2
1
Note
that
c ( )
fourier
1
(2 )
N
2
e
R
i x
(1 , 2 ) . The inequality | 2 | 1 is
Dyy2 0 .
transform
in
dimensions
of
the
function
c( x)
where
x R N is
t 1 day and t 10 . Here, for the purpose of comparative analysis, the same discretization numbers and
numerical parameters used in [8] are employed. Figure 1 represents the longitudinal concentration distributions
for y=10, y=16.25 m and the lateral concentration distributions for x=5 and x=20 m at t=100 days obtained with
different splitting methods. It is seen from the Figure that the iterative splitting gives accurate solutions. We
present the errors of iterative splitting and classical splitting methods in Table 1 and Table 2. We see that
iterative splitting gives better results than other classical splitting methods.
The iterative splitting method is superior to the others, because it includes all operators in each subequation
unlike the traditional operator splitting methods. This is physically the best and so we obtain the consistent
approximations after each inner step.
Table1: The errors of transverse concentration distributions for
At x=20
At x=5
Error L
t 1 at t=100 days.
Iterative splitting
Lie-Trotter
splitting
Strang splitting
6.2517e-006
0.0011
3.3810e-005
0.0022
6.9421e-005
6.3913e-005
At y=10
t 1 at t=100 days.
At y=16.25
Iterative splitting
2.3716e-005
Lie-Trotter splitting 0.0025
Strang splitting
1.6546e-004
4.5885e-005
0.0018
5.4205e-004
1116
(a)
(b)
References
1.
2.
3.
4.
5.
6.
7.
8.
9.
10.
11.
12.
13.
14.
15.
16.
1117
serife.nur.yalcin@ege.edu.tr, 2 ali.caliskan@ege.edu.tr
Abstract. In this paper, we study the relation between line congruence and screw system. To form this
relation; first three linearly independent dual vectors are handled and then it is showed that the external
bisectors of these vectors are belonged to the same recticongruence. Then the line congruence, which is made
the same angle between the recticongruence, is obtained. After that types of line congruences are examined.
Finally, it is obtained that a line congruence become a third order screw system in a special condition.
Keywords: line geometry, line congruence, screw theory
1 Introduction
Line congruences arise in the classical method of transforming one surface to another by lines. These
transformations are particularly interesting if some geometric properties of the original surface is preserved [1].
Line congruences, then, are two parameter families of lines and can be studied as surfaces in the space of lines.
And since a screw is a line with an associated pitch, it follows that a line, in its own right, is a screw with zerovalued pitch. On the other hand, also line congruences and screw systems have their own properties, there is a
clear similarity between them.
In this paper, a new approximation, that in which condition a line congruence become a screw system, is given.
By this way, a relation between congruences and screw systems is illustrated.
X 1
Q1 Q 2
2 1 cos 1
, X 2
Q 2 Q3
2 1 cos 2
, X 3
Q3 Q1
2 1 cos 3
(2)
X 1 , X 2 , X 3 dual external bisectors are belong to the same recticongruence because these bisectors are linearly
dependent [2].
To prove this, let show there are a, b , c dual numbers, which are not pure dual, provide the following equation:
(3)
aX bX cX 0 .
1
1118
1 1
0 1
0
1
(4)
Then because of Q1 , Q 2 , Q 3 lines are not parallel to each other, is equal to zero. And thus a, b , c dual
numbers verified in eq. (3) can be founded.
On the other hand, let show the perpendicular bisectors of X 1 , X 2 , X 3 , two by two, is the same line.
Q Q 2 Q 2 Q3 Q3 Q1
(5)
X 1 X 2 1
Q1 Q 2 Q 2 Q3 Q3 Q1
is the perpendicular bisector of X 1 , X 2 . Then with the aid of eq. (3) we can write
X X 2
X 3 1
0.
(6)
X 1 X 2
Because of this X 1 , X 2 , X 3 external bisectors are belong to the same recticongruence [2]. Such that the axes of
recticongruence is
Z
(7)
Y
Z
where Z Q1 Q 2 Q 2 Q3 Q3 Q1 , Z Z .
Q1 , Q 2 , Q3
Q Z Q1 Q1 Q2 Q2 Q3 Q3 Q1
Q1 Y 1
(8)
Z
Z
Z
Q1 , Q 2 , Q3
Q1 , Q 2 , Q 3
Q3 Y
Q 2 Y
,
(9)
Z
Z
And thus let say
Q1 , Q 2 , Q3
cos
(10)
Z
so
(11)
Q1 Y Q 2 Y Q3 Y cos
is obtained.
2.3 Conclusion
Consequently, any three lines such as Q1 , Q 2 , Q 3 are in the same line congruence [2], which has constant dual
slope, and which recognizes the axis of the recticongruence belonging to the external bisectors of these three
lines as its own axis, and which is formed by the lines generated by the constant dual angle in eq.(11).
1119
Q 2 Q 3
2 1 cos 2
A *
Q 2 Q3
(12)
2 1 cos 2
To verify types of line congruence we need to know w2 and ww equals to zero or not where
W 2 (w w)2 Cu 2 Cv 2 (Cu Cv )2 ( 2 0)
(15)
sin cos
(16)
3
sin 2 sin 2 3 sin 2 cos 2
.
2ww
4
w2
(17)
However it is enough to examine below equations instead of eqs. (16) and (17).
(18)
sin 0, sin 0
Let examine four condition which provide eq. (18)
1.
If 0 ; Q1 line pass through the point which A and A* intersect each other. So we can say
normal congruence to the C and illustrate as below.
2.
1120
3.
4.
1121
c0
.
(21)
c
In kinematics, screws are used to describe both infinitesimal and finite displacements of a rigid body. These
displacement screws are usually referred to as twists. The pitch of a twist is conventionally defined as
tij
(22)
p
ij
p
where
ij and tij
On the hand, Dimentberg [4] introduced a different definiton of pitch in formulation the resultant screw
displacement of two successive screw displacements. He showed that if a finite twist of a rigid body from
position i to position j is defined as
j
(23)
Tij tan E ij
2
where
ij ij tij
is the dual angle [3] of the twist then the resultant finite twist ( T13 ) of two successive
The twist can also be written as the product of an amplitude and a unit screw, of which the pitch is defined by eq.
(25). That is,
Tij tan
ij
Sij .
(26)
2
It is well known that specifying the position of a rigid body requires six independent parameters, but it takes only
four independent parameters to describe the position of a line. It follows that there are two free parameters in
choosing a screw for displacing a line from one position to another. It is of interest to know if these screws can
form a screw system of the third order. That is, we seek a mathematical representation of the following form [5]:
(27)
S f1S1 f 2 S2 f3 S3
where f1, f 2 and f 3 are scalar functions of the free parameters and S1, S2 and S3 are linearly independent
screws.
Tsai and Roth [6] formulated an incompletely specified displacement as finding the resultant of two successive
displacement screws. The twist S for displacing a certain element of a body can be thought of as the resultant of
a twist K , containing free parameters, followed by a twist L that actually displaces the element. In the case of
displacing a line, the axis of K is taken to be the initial position of the line. The translation and rotation
parameters of K , d and are free parameters because they do not alter the position of the line. The axis of L is
selected to be the line perpendicular to both the initial and final positions of the line and the rotation and
translation parameters of L are respectively the angle and distance between the two positions of the line.
1122
Fig. 1.
As shown in Fig. 1., let the Plcker coordinates of two positions of a line be
(28)
Qi qi , q0i qi , bi qi (i 1, 2)
The unit screw of K , whose axis coincides Q1 , is
K k, k0 q1 , q01 pK q1
(29)
(30)
(31)
tan S
2
where
tan
tan
1 tan tan K L
(33)
2
2
is the rotation parameter of S . Since K and L intersect each other perpendicularly eq. (33) becomes
2
Then
2
2
substituting
the
2
equations
k tan k l
(34)
(29)
to
2
tan S tan tan
2
2 l0
2
b1 k tan b1 l k tan k l0
2
2
(32)
into
eq.
(34)
gives
cos
tan
k l
d
(35)
S1 tan S2
S3
2
2
1 cos
Consequently, the above equation is in the form of eq. (27) and thus all possible screws for displacing a line
form a three system.
tan
1123
References
1.
2.
3.
4.
5.
6.
Dimentberg, F. M., 1965, The Screw Calculus and Its Application in Mechanics (in Russian), Moscow.
Huang, C., Wang, J-C., 2003, The Finite Screw System Associated With the Displacement of a Line,
Journal of Mechanical Design, Vol.125, pp.105-109
Kula, M., 1978, ki Dorunun Dual Aortaylar ve Bu Aortaylarla lgili Baz Doru Kongranslar,
E.. Fen Fak. Der; S. A, C.II., S.2, 155.
Shepherd, M.D. , 1997, Line Congruences as Surfaces in the Space of Line, Suny College at Potsdam,
USA
Tsai, L. W. and Roth, B., 1973, Incompletely Specified Displacements: Geometry and Spatial Linkage
Synthesis, ASME.
Yang, A.T., 1963, Applications of Quaternion Algebra and Dual Numbers to the Analysis of Spatial
Mechanisms, Ph. D. Thesis, Columbia University, NY
Biographies
erife Nur Yaln erife Nur Yaln was born in Afyonkarahisar in 1986. She graduated from Sleyman
Demirel University Faculty of Arts and Science, Department of Mathematics in Isparta-Turkey in 2008. She took
her master degree on geometry in 2010 from Sleyman Demirel University. Now she is Phd student and research
assistant in Ege University in zmir.
Her study field is Differential Geometry, Dual and Line Geometry, Kinematics.
Ali alkan Ali alkan was born in Yozgat in 1947. He graduated High Teaching School and Ege
University, Faculty of Science, Department of Mathematics-Astronomy in zmir in 1971. He graduated master
between 1974-1975 and post graduated about differential geometry in 1978, then he became science doctor. He
worked as assistant professor in Ege University, Faculty of Science, Department of Mathematics. He became
associate professor in 22.11.1984. He took professor degree in 27.02.1990.
His study fields are Differential Geometry, Dual and Lorentz Geometry, Kinematics, Computer Aided
Geometric Design, Geometry on Time Scales, and Mathematics Education.
Prof. Dr. alkan continues his job as head of department and geometry president in Ege University, Faculty of
Science, Department of Mathematics.
1124
Abstract. In this paper the method for obtaining an exact solution of the boundary layer problem which is
described by Navier-Stokes system of equations of fluid flow at high Reynolds number in a class of
discontinuous functions is suggested. For this aim, the system of equations of the boundary layer problem
is reduced to the one dimensional nonlinear scalar equation of parabolic type in von-Mises coordinates. Then
the special auxiliary problem having some advantages over the main problem is introduced. Using these
advantages of the auxiliary problem the exact solution of the boundary layer problem is found. The
differentiable properties of the solution of the main problem is also investigated.
Keywords: Boundary layer problem, auxiliary problem, exact solution in a class of discontinuous functions
1 Introduction
Boundary layer problems arise in many scientific and practical applications. Among them there are some
problems which are described by Navier-Stokes system of equations of fluid flow at high Reynolds number
[1],[2], mathematical models of liquid crystal materials and chemical reactions [3], electrical networks [4], and
etc.
One of the most important applications of boundary-layer theory is the calculation of the friction grad of bodies
in a flow, e.g., the grad of a flat plate at zero incidence, the friction grad of a ship, an airfoil, the body of
airplane, or a turbine blade. One particular property of the boundary layer is that, under certain conditions, a
reverse flow can occur directly at the wall.
The following tendency is emerging from the rapid developments in the area of supercomputing. The future will
consist more of direct numerical solutions of the Navier-Stokes equations without any simplifications, and also
of the computation of turbulent flows using direct numerical simulation, i.e without using a turbulence model or
by modeling only high frequency turbulent fluctuations. However, numerical methods in computing flows at
high Reynolds numbers become efficient only if the particular layered structure of the flow, as given by the
asymptotic theory, is taken into account, as occurs if a suitable grid is used for computation.
In this paper, in order to investigate the natural future of the flow of the fluid adjacent to a rigid wall, the new
method for obtaining the exact solution is suggested.
1125
u
u
2u
v
= 2,
x
y
y
(1)
u v
= 0,
x y
(2)
ux,0 = 0, ux, = u x
(3)
where u ( x) = u 0 x n and u 0 is a real constant. At first, we reduce this problem to one-dimensional parabolic
type equation. For this aim, we apply the von-Mises transforms as follows u = ux, y , = x, y , where
u=
.
, v=
y
x
2u
u u
,
and
in new coordinates. We have
x y
y 2
u
u
u
x
y y
u
u
v
x
y y
x
.
x
(4)
u
It should be noted that by calculation of the derivative in the expression, the function u depends of y as
x y
a parameter. Similarly,
u
u
=
y
x
x y
u
= u
x
(5)
and
2u
= u u
y 2
x y
=
x x
u
u
=u
x x y
u
u
x x
(6)
can be obtained. Putting (4), (5) and (6) into (1) we get
u
u
u u v
u
vu
x
= u
u
u
.
x x
Thus, von-Mises transformation reduces the system of equations (1), (2) to the one-dimensional nonlinear heat
equation in scale of ( x, )
u
u
=
u
x
(7)
1126
Equation (7) is a nonlinear parabolic type partial differential equation which degenerates to the first order
equation when u = 0 . The problem of finding the solution of the equation (7) under the conditions (3) is called
as the main problem.
u( x, ) = f ( Dx ) = f ( ) ,
where is a special coordinate traveling with speed
From (8), we have
u
= Df ' ,
x
(8)
u
= f ' ,
2u
= f ' .
(9)
df
d df
f
.
=
d
d d
Df = f
df
c.
d
For the sake of simplicity, we assume that c = 0 . Integrating again the last equation for the solution we find
f ( ) = C
The arbitrary constant C is defined as C = 0 from the condition f = 0 for = 0 . Since the function u( x, ) is
equal to zero on the right hand of the front point = Dx , the solution of the problem (7), (4) can be written as
D
( Dx ),
u ( x, ) =
0,
< Dx,
(10)
Dx.
Hence, supp u( x, ) = 0 Dx . As it is seen from (10), the solution of the problem (7), (4) is continuous,
u ( x, ) is piecewise continuous and u is equal to zero. On the other hand, the equation (7) is degenerated to
the first order nonlinear equation
u
u
=
x
whose solution is u = ax
1127
( Dx ) = ax
we get D = a . Therefore, the speed of traveling wave is proportional to the square root of .
uf x u f dxd 0
D
2
u 2 u 2
=
.
x 2 2
According to [6], we propose auxiliary problem which is equivalent to the problem (7), (4) in known sense. Let
A be the operator A() =
2 ()
. Assuming that A 1 exists and applying it to both sides of the equation (7) we
get
A 1
u 1 2
= A Au
x 2
or
A 1u 2
= u .
x
2
(11)
u = Av( x, ) .
(12)
v 2 v
=
.
x 2 x 2
(13)
v(0, ) = v0 ( ), u( x, ) = u0 x
(14)
d 2 v0 ( )
d 2
= u ( ,0) .
1128
Theorem 1. If v( x, ) is the solution of the problem (13), (14), then the function u( x, ) defined by (12) is the
weak solution of the problem (7), (4).
Now, let us find the solution of the problem (13), (14). For this aim, the solution v( x, ) is expressed in the form
v( x, ) = ( Dx ) .
(15)
Calculating the necessary derivatives and putting the obtained expressions into (13) we have
D =
2 .
1 D
v( x, ) =
3 2
1/2
1/2
D
2
Dx c1 c2 .
(16)
2 v( x, )
2
= u ( x, ) .
(17)
v( x, ) =
1
( )u ( x, )d
(0)
(18)
where (x) is the first kind Euler function. As it is seen from (18), if u( x, ) is twice differentiable function,
equation (17) would become exact. However, the differentiability property of u( x, ) function is less than 1, i.e.
the differentiability property of u( x, ) is in the order of 0 < < 1 .
The derivatives of u( x, ) function with respect to x and variables can be expressed as
Dx u ( x, ) =
u
1
u ( , )
=
( )
d , 0 < < 1
x
( ) 0
and
D u ( x, ) =
1
u ( x, )
( x )
d , 0 < < 1
( )
where Dx u( x, ) and D u( x, ) are the Caputo derivatives with respect to x ve variables, respectively.
1129
3 Conclusion
In order to investigate the nature of the flow of the fluid adjacent to a rigid wall, the new method for finding the
exact solution of the boundary layer problem is suggested.
For this reason, the auxiliary problem having some advantages over the main problem is introduced. The
differentiable properties of the main problem are also investigated.
References
1.
2.
3.
4.
5.
6.
7.
1130
burak.sahiner@bayar.edu.tr, 2 ali.caliskan@ege.edu.tr
Abstract.Ruled surface design and its applications are widely used in CAGD (Computer Aided Geometric
Design), industrial areas and many other areas such as motion analysis and simulation of rigid bodies. Thus,
ruled surface design has become an important issue in engineering and engineering mathematics. In this
paper, the helicoid, which is a particular ruled surface, is drawn by using the dual De Casteljau algorithm. It is
shown that the characteristics of the helicoid provide the requirement of the dual De Casteljau algorithm. A
method for the screen representation of the rulings of the helicoid is propounded. The method present the
rulings of the helicoid as a clipped line segment on the screen. The two new methods represent an approach to
draw helicoid by a combination of the theoretical and practical aspects of dual methods.
Keywords: Computer Aided Geometric Design, dual space, dual De Casteljau algorithm, the helicoid
surface, screw theory.
1 Introduction
Recently, the dual design methods of curves and surfaces have been widely researched. Juttler [1] used
quaternion representation of RMM (Rotating Minimizing Motions) to sweep surface modelling. Pottmann and
Chen [2] presented the approximation algorithm in 4-D space to draw ruled surfaces. Ding [3] applied the dual
De Boor algorithm in 6-D space to draw particular surfaces, which are duplicated surfaces. In the same paper,
the problem of drawing infinitely long lines, on a screen, as represented by Plcker coordinates, was touched
upon. The objective of Leopoldseders paper [4] was to provide algorithms on the approximation of developable
surfaces with pieces of right circular cones. Pottmann and Wallner [5] used a dual approach to solve the drawing
methods of differential geometry of developable surfaces, which include ruled surfaces. Sprott and Ravani [6]
introduced the De Casteljau algorithm in screw space to resolve ruled surface drawing and mesh generation
problems. Ding and Zhang [7] used the dual De Casteljau algorithm to draw ruled surfaces with particular shapes
which are the hyperbolic paraboloid, the cylinder, the cone, and the hyperboloid. They gave the principles of that
method which were the applications, extensions and variations of the dual De Casteljau algorithm. In addition,
they gave a new method for the screen representation of the rulings. This method presented the rulings as clipped
line segments on the computer screen in order to show the proportion of the ruled surface on the computer
screen. But they didnt suggest any method to draw the helicoid which is also a ruled surface with particular
shapes. Our work is inspired by the paper of Ding and Zhang [7].
1131
2 Screw Theory
A general screw S , from which we derive a line by specialisation, a real 3-vector S which indicates the
direction of the screw, and a real 3-vector S p which locates S is the moment of the screw about the origin [3]. A
screw is formed as
S S Sp S ( pS S0 ) ,
where S0 R S V S and is a quasi-scalar satisfying 2 0 . Here p is the pitch of the screw and S 0 ,
which is the moment of the line of the screw about the origin, is orthogonal to S , and is derived as from the
origin radius vector R , or more generally from any point V of the screw. The length S of the real part S
is the real magnitude of the screw. Figure 1 illustrates a geometric interpretation of a screw.
0 X
0.
where cos1 a and a b cos X
0
1
2. Find the screw axis for the motion from
0 X
0.
s X
0
1
1132
S sz
0 sx ,
sy sx
0
which has the property that [S ]S s S for any screw S . See Figure 2 for illustration.
4. If the two control screws are parallel, the displacement matrix should be written as
A I D ,
0 (t ) X
, r 1 is the complex angle between ruled screws X
r 1 and X
r 1 , and s r 1 is the line of the
where X
i
i
i
i 1
i ,i 1
i ,i 1
r 1 and X
r 1 . The ruled screws on the surface is given by X
n (t ) .
screw axis between X
i
i 1
0
1133
where u and range from negative infinity to positive infinity, while and c are constant. If is positive
then the helicoid is right-handed as shown in the Figure 3; if is negative then the helicoid is left-handed. For
simplicity, we consider the right-handed helicoid in this paper, such that 1 and c 1 .
The helicoid will become
u, u cos , u sin ,
We rewrite this helicoid in ruled form
u, 0,0, u cos , sin , 0 .
The striction line is 0, 0, , that is the z-axis and all the rulings are parallel to xy-plane.
rotates
s . When X
an axis s that is the z-axis. The axis s is also the striction line of the helicoid. Here X
, the helicoid is created. To create the helicoid using the dual De
and translates along s to the position X
t
Casteljau Algorithm, we take the following steps:
A I sin t S 1 cos t S
A X
X
i
1134
5 Presenting the Rulings of the Helicoid as Clipped Line Segments on the Screen
It is very important to represent screws as lines on the screen where the ruled surface is described by using the
screw presentation method. Ding [3] used the line segment algorithm [8] to draw the infinitely long screws on
the screen. However, the screws on the screen are still infinite lines. Intuitively, an infinite line must be
converted into line segments to render a ruled surface on the screen. Ravani and Wang [12] proposed a method
of representing line segments based on the centre point of a line. The line coordinates were augmented with
two additional parameters representing the distance along the line on either side of the centre point. Ge and
Ravani [13] propose an intrinsic method based on the striction curve of the ruled surface. Sprott and Ravani [6]
proposed to select a reference point on the first control line and draw a reference line through the point. Then the
reference line is interpolated along with the control screws, in that the control screws and the reference line
construct a rigid body.
In this study, two points must be chosen on the ruling of the helicoid to represent the ruling of the helicoid as
a clipped line segment on the screen. Then, the line segment, which is determined by these two points, is drawn
on the screen. Two points on a ruling of the helicoid can be chosen by the following way. Since the directrix of
the ruling and a point, which the ruling passes through, are known, the equation of the ruling can be written as
r(t ) OP t u ,
where P is a point, which the ruling passes through, u is the directrix of the ruling and t is the parameter
drawn the ruling. Here, we can chose the point P as the intersection of the ruling and z-axis. Note that, P is
also the striction point on the ruling of the helicoid. In the equation of the ruling, when we write d and d
instead of t , we can determine two points on the ruling. The line segment, which lies between the two points, is
drawn on the screen. Thus, the infinity line is presented as a clipped line segment on the screen. When this
procedure is applied to all the rulings, which are created by using dual De Casteljau algorithm, the representing
of the helicoid on the screen is completed.
6 Conclusion
This paper represents the methods for Computer Aided Geometric Design, the dual drawing method of the
helicoid, and the screen representation of the rulings of the helicoid. This paper introduces the drawing method
of the helicoid by using the dual De Casteljau algorithm. The ruling of helicoid is demonstrated as a clipped line
segment in this screen representation method. These methods show that the helicoid, which is drawn by real
space methods, can be drawn by dual space method.
References
1.
2.
3.
4.
5.
6.
7.
8.
Juttler, B., 1998, Rotation minimizing spherical motions, Advances in Robot Kinematics: Analysis and
Control, Kluwer Academic Publishers, 413-422pp.
Pottmann, H., Chen, H.Y., 1999, Approximation by ruled surfaces, Journal of Computational and
Applied Mathematics 108, 143-156.
Ding, R., 2002, Drawing ruled surfaces using the dual De Boor algorithm, Proceeding of Computing:
The Australasian Theory Symposium CATS 2002, Elsevier Science B.V.
Leopoldseder, S., 2001, Algorithms on cone spline surfaces and spatial osculating arc splines,
Computer Aided Geometric Design, 18, 505-530pp.
Pottmann, H., Wallner, J., 2001, Computational Line Geometry, Springer Verlag, Heidelberg, Berlin,
565pp.
Sprott, K., Ravani, B., 1997, Ruled surfaces, Lie groups, and mesh generation, 1997 ASME
Engineering Technical Conferences, Sacramento, California, USA.
Ding, R., Zhang, Y., 2006, Dual space drawing methods for ruled surfaces with particular shapes,
IJCSNS International Journal of Computer Science and Network Security, Vol. 6, No. 1.
Hill, F.S. Jr., 1990, Computer graphics. Macmillan Publishing Company, New York.
1135
9.
10.
11.
12.
13.
Farin, G., 1988, Curves and surfaces for computer aided geometric design: A practical guide, Academic
Pres, Inc., Boston, Mass.
Fomenko, A.T., 1991, Elements of the Geometry and Topology of Minimal Surfaces in Threedimensional Space.
Carmo, D., Mathematical Models from the Collections of Universities and Museums (Ed. G. Fischer).
Braunschweig, Germany: Vieweg, pp. 44-45, 1986.
Ravani, B. and Wang, J.W., 1991, Computer aided geometric design of line constructs, Transactions of
the ASME Journal of Mechanical Design, 113:363-371.
Ge, Q. Jeffrey and Ravani, B., 1994, On representation and interpolation of line-segments for computer
aided geometric design, In ASME Design Automation Conference, volume 69-1, pages 191-198.
Biographies
Burak ahiner - was born in zmir in 1986. He graduated from Ege University, Faculty of Science, Department
of Mathematics in zmir in 2008. He graduated master between 2008-2010 in Geometry/Mathematics in Ege
University.
His study field is Computer Aided Geometric Design.
Mr. ahiner is still working as a research assistant in Celal Bayar University, Department of Mathematics.
Ali alkan - was born in Yozgat in 1947. He graduated High Teaching School and Ege University, Faculty of
Science, Department of Mathematics-Astronomy in zmir in 1971. He graduated master between 1974-1975 and
post graduated about differential geometry in 1978, then he became science doctor. He worked as assistant
professor in Ege University, Faculty of Science, Department of Mathematics. He became associate professor in
22.11.1984. He took professor degree in 27.02.1990.
His study fields are Differential Geometry, Dual and Lorentz Geometry, Kinematics, Computer Aided
Geometric Design, Geometry on Time Scales, and Mathematics Education.
Prof. Dr. alkan continues his job as head of department and geometry president in Ege University, Faculty
of Science, Department of Mathematics.
1136
1,2
Abstract. The purpose of this study is to give a Taylor polynomial approximation for the solution of hyperbolic type partial
differential equations (HPDEs) with variable coefficients. Some numerical examples, which consist of initial and the
boundary conditions, are given to illustrate the reliability and efficiency of the method. Also, the obtained results are
compared by the known results.
Keywords: Taylor polynomial solutions; hyperbolic partial differential equations; Taylor matrix method, double Taylor
series.
1. INTRODUCTION
Hyperbolic partial differential equations play dominant role in many branches of science and engineering.
These equations model the vibrations of structures (e.g., buildings, beams and machines) and are the basis for
fundamental equations of atomic physics. Wave equation and telegraph equation are typical examples of
hyperbolic partial differential equations. Telegraph equation is commonly used in signal analysis for
transmission and propagation of electrical signals and also has applications in other fields (see [2] and the
references therein). On the other hand, much attention has been given in the literature to the development,
analysis and implementation for the numerical solution of second-order hyperbolic equations, specially telegraph
equations, see for example [3-13].
Let
us
consider
the
second-order
linear
hyperbolic
partial
differential
equation
2
A( x, t )
u
x
B ( x, t )
u
t x
C ( x, t )
u
t
D ( x, t )
u
x
E ( x, t )
u
t
F ( x , t ) u G ( x , t ) a x b, 0 t T ,
(1)
B ( x, t ) 4 A( x, t )C ( x, t ) 0
with the initial conditions
x a, b
u ( x, 0) f ( x) ,
2
u ( x, 0)
m( x ),
(2)
x a, b
u ( a, t ) h(t ), 0 t T
u (b, t ) k (t ), 0 t T
(3)
1137
u ( x, t )
( x c0 ) (t c1 ) , ar , s
r ,s
(r ,s )
(4)
2. FUNDAMENTAL RELATIONS
To obtain the numerical solution of the hyperbolic partial differential equation method presented, we first
evaluate the Taylor coefficients of the unknown function. For convenience, the solution function (4) can be
written in the matrix form
(5)
u ( x, t ) X( x, t ) A
where X( x, t ) is the 1x( N 1) matrix defined as
2
A= a 0,0 a 0,1 ... a 0, N a1,0 a1,1 ... a1,n ... a N,0 a N,1 ... a N, N .
T
On the other hand, the relation between the matrix X( x, t ) and its derivatives X
X
( x, t ) X( x, t )(B) (B)
where
( m,n )
0
0
0
0
0
0
, B
0
0
0
1 0
0 2
0 0
0 0
( m,n )
( x, t ) is
(6)
B 0
0
0 B
, B=
NI
0 0
0 2I
0
0
I is the ( N 1) ( N 1) identity matrix. Using the relations (5) and (6) we have
u
( m,n )
( x, t ) X
( m,n )
( x, t ) A X( x, t )(B) (B) A .
m
(7)
3. METHOD OF SOLUTION
Our purpose is to investigate the approximate solution of Eq. (1), under the given conditions, in the series
form
N
u ( x, t )
( x c0 ) (t c1 )
r
r ,s
r 0 s 0
or in the matrix form u ( x, t ) X( x, t ) A . To obtain the solution, we first reduce terms of Eq. (1) to matrix forms
uxx ( x, t ) X( x, t )(B) A
(8)
uxt ( x, t ) X( x, t )BBA
(9)
utt ( x, t ) X( x, t )(B) A
(10)
ux ( x, t ) X( x, t )BA
(11)
ut ( x, t ) X( x, t )BA .
(12)
1138
A( x, t ) aij ( x c0 ) (t c1 ) ,
i
i 0
j 0
C ( x, t ) cij ( x c0 ) (t c1 ) ,
i
i 0
j 0
E ( x, t ) eij ( x c0 ) (t c1 ) ,
i
i 0
j 0
B ( x, t ) bij ( x c0 ) (t c1 ) ,
i
i 0
j 0
D( x, t ) d ij ( x c0 ) (t c1 ) , (13)
i
i 0
j 0
F ( x, t ) fij ( x c0 ) (t c1 )
i
i 0
j 0
which are the Taylor polynomials at a point ( x, t ) (c , c ) . The matrix representation of the Taylor expansion
0
( x c0 ) (t c1 ) u
i
( x c0 ) u
i
(t c1 ) u
j
( x c0 ) u
i
(t c1 ) u
j
( x , t ) X ( x, t )C j B A
r
( x , t ) X( x , t )C i B A
(0, s )
(0, s )
( x , t ) X( x , t )C i B A
( r ,0)
( r ,0)
( r ,s )
(14)
( x , t ) X( x , t )C j B A
s
( x c0 ) (t c1 ) u
i
( r ,s )
( x , t ) X( x , t ) C i C j B B A
s
where
Cp 0
0 C
p
Cp
0 0
1;
Cp cij
0;
Cp ( N 1) ( N 1)
2
i j p
p 0,1,...,
otherwise
i , j 0,1,..., N
and
i j ( N 1) p
1;
Cp cij
0;
p 0,1,... ,
otherwise
i, j 1, 2,..., ( N 1) .
2
G ( x, t )
( x c0 ) (t c1 ) ,
r
r ,s
gr ,s
(r ,s )
r 0 s 0
( x, t )
(15)
r !s!
or in matrix form
G( x, t ) X( x, t ) G
(16)
where
Substituting the expressions (5), (8)-(14) and (16) into Eq. (1) and then simplifying the result, we have the
fundamental matrix equation
N
a C C B
ij
i 0
bij Ci C j BB cij Ci C j B d ij Ci C j B
2
(17)
j 0
1139
Denoting the expression in parenthesis of (17) by W, the fundamental matrix equation (17) is reduced to form
WA G , W wi , j ,
i, j 1, 2,..., ( N 1)
(18)
coefficients ar , s .
We can use the following formulas in order to construct matrix representation of conditions: For the initial
conditions (2),
(19)
u ( x, t ) X( x)Q(t ) A
and
ut ( x, t ) X( x)Q(t )BA ;
(20)
u(x, t) T(t)J(x) A .
(21)
The matrix representations of non-homogeneous terms of Eq. (2) and Eq. (3) can be written in the forms
F f0 f1 ... f N ,
f ( x ) X( x ) F,
fn
(n)
( c0 )
(22)
n!
M m0 m1 ... mN ,
m( x ) X( x ) M ,
mn
(n)
( c0 )
(23)
(24)
n!
(n)
h(t ) =T(t ) H,
H h0 h1 ... hN ,
hn
h (c1 )
n!
(n)
K k 0 k1 ... k N ,
k (t ) T(t ) K ,
kn
k (c1 )
(25)
n!
where
X( x ) 1 ( x c0 ) ( x c0 ) ... ( x c0 )
2
(t ) 1 (t c1 ) (t c1 ) ... (t c1 )
2
J ( x ) I ( x c0 )I ( x c0 ) I ... ( x c0 ) I
2
T (t ) 0
0 T (t )
Q (t )
0
0
0
0
T(t )
Note that the matrices J (x) and Q(t) are dimension of ( N 1) ( N 1) . By substituting relations (17)(25) into (2)-(3) and then simplifying the result, we get the matrix forms of the conditions respectively, as
2
K 1 A Q(0) A F
(26)
K 2 A Q(0)BA=M
(27)
K 3 A J (a) A=H
(28)
K 4 A J(b)A=K .
(29)
To obtain solution of Eq. (1) under the conditions (2) and (3), it is formed the augmented matrix as follows:
K1
K
2
W;G K 3
K 4
W
; F
; M
; H
; K
; G
(30)
1140
A W
where
W; G is generated by using the Gauss elimination method and then removing zero rows of the gauss
eliminated matrix. Here W and G are obtained by throwing away the maximum number of row vectors from
W and G, such the rank of the system defined (30) cannot be smaller than ( N 1) 2 . This process provides
higher accuracy because of the decreasing truncation error.
4. NUMERICAL EXAMPLES
The method of this study is useful in finding the solutions of hyperbolic type partial differential equations in
terms of Taylor polynomials. We illustrate it by the following examples. The numerical computations have been
done using Maple 9.
Example1. Consider the hyperbolic partial differential equation with the initial conditions
u
2
u
2
(1 2 x )
xt
u
2
( x x 2)
2
0,
u ( x, 0) x,
u ( x, 0)
1.
t
The matrix representation of equation is
(B)
A 0
( N 1) 1
By taking
Thus the method gives the solution u( x, t ) x t , which is exact solution [14].
Example 2. Let us now consider the initial-boundary value problem
u( x, t ) u( x, t )
u( x, t )
2
2
2
u( x, t )
(2 2t t )( x x )exp(t ) 2t exp(t ) 0 t 1, 0 x 1,
2
2
t
t
x
2
u ( x, 0)
u ( x, 0) 0
t
and the Dirichlet boundary conditions
u(0, t ) u(1, t ) 0
the exact solution is [12,13]
u( x,t ) t 2 ( x x 2 )exp( t ) .
By taking c0 c1 0 and N=8, we solved the above problem by means of the fundamental matrix equation.
Then exact solution and approximate solution for N=8 are plotted in Fig.1. Also the errors associated with the
present method and Rothe-Wavelet method and Chebyshev-Tau method are compared in Table I. we see that we
can achieve a good approximation with the exact solution by using only a few terms of Taylor polynomials.
1141
xr
Analytical
sln.
RotheWavelet [13]
0
0.125
0.250
0.375
0500
0.625
0.750
0.875
1
0.00
0.04023
0.06897
0.08622
0.09196
0.08622
0.06897
0.04023
0.00
0.00
15E-5
2.2E-5
2.3E-5
0.0E-5
0.0E-5
2.4E-5
3.1E-5
0.00
ChebyshevTau
m=n=8
0.00
2.4E-9
4.2E-9
5.2 E-9
5.6E-9
5.2E-9
4.2E-9
2.4E-9
0.00
Present
Method
for N=8
0
1.55E-12
0.66E-12
1.13E-10
1.43E-11
2.88E-12
4.71E-10
1.41E-10
7.71E-9
FIG. 1. Plot of exact solution (right) and approximation solution with N=8 (left), from Example 2.
5. CONCLUSION
Analytic solutions of second order linear partial differential equation with variable coefficients are usually
difficult. In many cases, it is required to approximate solutions. For this purpose, a new Taylor method based on
the truncated double Taylor series is developed to numerically solve second order HPDEs with the initial and
boundary conditions. Obtained results in this paper better than results obtained by the other methods in the
references.
It is observed that the method has a great advantage when the known functions in the equation can be
expanded in Taylor series. To get the best approximation, we must take more terms from the Taylor expansion of
functions; that is, the truncation limit N must be chosen large enough. Another considerable advantage of the
method is that the N th order approximation gives the exact solution when the solution is polynomial of degree
equal to or less than N. If the solution is not polynomial, Taylor series approximation converges to the exact
solution as N increases. Moreover, this method is applicable for the approximate solution of parabolic and
elliptic type partial differential equations with variable coefficients, and also can be extended to higher-order
partial differential equations.
1142
References
1.
2.
3.
4.
5.
6.
7.
8.
9.
10.
11.
12.
13.
14.
B. Blbl, M. Sezer, (2011). Taylor polynomial solution of hyperbolic type partial differential equations
with constant coefficients, Numer. Methods Partial Diff. Eq. 88 (3), 533-544.
Dehghan, M., Shokri A., (2008). A numerical Method for solving the hyperbolic telegraph equation.
Numer. Methods Partial Diff. Eq., 24, (4): 1080-1093.
M .S. El-Azab and M. El-Gamel, (2007). A numerical algorithm for solution of Telegraph equations, Appl.
Math. Comput. 190 757-764.
F. Gao and C. Chi, (2007). Unconditionally stable difference schemes for a one-space- dimensional linear
hyperbolic equation, Appl. Math. Comput. 187 1272-1276.
R. K. Mohanty and M. K. Jain, (2001). An unconditionally stable alternating direction implicit scheme for
the two space dimensional linear hyperbolic equation, Numer. Methods Partial Diff Eq. 17 684-688.
R. K. Mohanty, M. K. Jain and U. Arora, (2002). An unconditionally stable Adi method for the linear
hyperbolic equation in three space dimensions, Int J Comput Math 79 133-142.
M. Dehghan and M. Lakestani, (2009). The Use of Chebyshev cardinal functions for solution of the secondorder one-dimensional Telegraph equation, Numer Methods Partial Diff Eq. 25 (4) 931-938.
S. A. Yousefi, (2009), Legendre multiwavelet galerkin method for solving the hyperbolic Telegraph
equation, Numer Methods Partial Diff. Eq. 26 (3) 535-543.
J. Biazar and H. Ebrahimi, (2005). An approximation to the solution of hyperbolic equations by Adomian
decomposition method and comparison with characteristic method, Appl Math Comput 163, 633-638.
A. Ashyralyev and M. E. Kksal, (2009). On the numerical solution of hyperbolic PDE with variable space
operator, Numer. Methods Partial Diff. Eq. 25 (5) 1086-1099.
A. Ashyralyev and M. E. Koksal, (2005). On the second order of accuracy difference scheme for hyperbolic
equations in Hilbert space, Numer Funct Anala Optimiz 26 739-772.
A. Saadatmandi, M. Dehghan, (2009). A numerical method for solving the hyperbolic telegraph equation
using the Chebshev tau method, Numer Methods Partial Diff. Eq. 26 (1), 239252.
M. S. El-Azab and M. El-Gamel, (2007). A numerical algorithm for the solution of telegraph equations,
Appl. Math. Comput. 190 757764.
J. Biazar and H. Ebrahimi, (2005). An approximation to the solution of hyperbolic equations by Adomian
decomposition method and comparison with characteristic method, Appl. Math. Comput. 163, 633-638.
Biographies
Mehmet SEZER - Prof. Dr. Mehmet SEZER was born on March 20th 1954 in Antalya, Turkey. He received his
B.Sc. from the Department of Mathematics of Ege University in 1976 and was research assistant at Balkesir
Engineering Academy. His M.Sc. was from Ege University in 1977. He received his Ph.D. under the supervision
of Prof. Dr. Seda MORALI in the Department of Applied Mathematics at Ege University in 1982 and studied
in the field of Parabolic partial differential equations.
In 1982, he was Assistant Professor at Faculty of Education (Balkesir) of Uluda University, Associate
Professor in 1989, and in 1995 Professor at Faculty of Education of Dokuz Eylul University. Since 2004, he has
been Mathematics Professor at Faculty of Science at Mula University and has been the chair in the same
department.
His main research interests are ordinary and partial differential equations, integral and difference equations, and
their numerical solutions. Prof. Sezer has been reviewer for numerous influential journals, has published research
articles related to differential equations, linear algebra, analytic geometry and calculus; and has been authored
over 80 papers.
Berna Blbl: Berna Blbl was born on January 17th 1980 in Izmir, Turkey. he received her B.Sc. from the
Department of Mathematics of Mula University in 2002. he was research assistant at Mula University at
2003. Her M.Sc. was from Mula University in 2005. he has been authored 3 papers about numerical solution
of ordinary and partial differential equations.
1143
Islamic Azad University, South Tehran Branch, Tehran, Iran/2School of Computer Science Australian National
University Canberra, Australia/3Department of Computer Science, University of Tehran, Tehran, Iran/ 4Department of
Mathematics and Computer Science, Shahid Bahonar University of Kerman, Iran
1
sh_pirnia@azad.ac.ir, 2afzaly@ipm.ir, 3a.kashefi@ut.ac.ir, 4sharifizadeh@gmail.com
Abstract.Constraint satisfaction problems are the mathematical problems defined as a set of objects whose
states must satisfy some constraints. In some CSPs, the goal is to find the smallest subset of a set which
gratifies the constraints. A typical approach to solve this class of problems is to apply a generation algorithm
and inspect each subset for satisfaction of constraints; then the smallest feasible subset is selected. The two
most common methods of generating the subsets of a set are lexicographic ordering and Gray code; however,
they are not efficient when the set is large. In this paper, we propose a novel efficient generation algorithm,
Optimal-Banker, which can be used to find the smallest feasible subset. It generates a sequence of all subsets
of a set, in which not only the number of elements of subsets is monotonically increasing, but also the change
between each two consecutive subsets is minimal.
Keywords: rank, unrank, successor, constraint satisfaction problem, subset, generation algorithm.
1 Introduction
A generation algorithm is an algorithm which exhaustively lists all the subsets of a set [1]. A finite set is an
unordered combinatorial structure, i.e. a finite collection of objects. Therefore a finite set has a finite number of
elements[2], i.e. every finite set is an n-set for some specific natural number n [3, 4]. It may be desirable to
predetermine the position of a given object in the generated list without producing the whole list [5]. A subset
describes a selection of objects, where the order among them does not matter. There are 2 n distinct subsets of an
n-element set; including the empty set as well as the set itself. Many important algorithmic problems seek the
best subset of a group of things. For instance, a constraint satisfaction problem is defined as a set of objects
whose states must satisfy some constraints [6]. Some CSPs seek to find the smallest subset of a set which
satisfies some limitations. For example, the vertex cover problem tries to achieve the smallest subset of vertices
to touch each edge of a graph [7]. A solution to these problems is to generate all ofthe 2n possible subsets and
test each one until either we find a desired subset or weexhaust all possible candidates before we declare that
there is no possible subset. Therefore, a generation algorithm is used. The task of constructing the subsets of a
set grows exponentially with the size of the set. In section 2, we will show that applying each of the two common
methods of generating the subsets of a set, the lexicographic ordering and Gray code is not efficient when the set
is large.
In this paper we propose a novel efficient generation algorithm, the Optimal banker, which can be used to
find the smallest feasible subset. It generates a sequence of all subsets of any given set, in which not only the
number of elements of subsets is monotonically increasing, but also the change between each two successive
subsets is minimal. The rest of the paper is organized as follows: Section 2 investigates two known systematic
1144
methods of generating the subsets of a set, the lexicographic ordering and Gray code, as well as the Bankers
algorithm. The Optimal banker has been introduced in Section 3. Then its Rank, Unrank and Next algorithms [5]
are presented in Section 4, and finally the paper will be ended in section 5 by Conclusion.
2 Generating Subsets
The two common methods for generating all subsets of a set, the lexicographic order and Gray code, are
investigated in the following:
2.1
The lexicographic ordering on the set of subsets of S is derived from the lexicographic ordering of the
characteristic vectors [5]. If characteristic vectors are considered as binary representations of the integers 0 to 2n1
, then Lexicographic order corresponds to the usual order of the integers. With respect to this order, the rank of
a subset, denoted by rank(T), is the integer with binary representation (T) [5]. Suppose that n is a positive
n
integer, and S={1, 2,, n}. Suppose P(S) is the set of all 2 subsets of S. Subsets in P(S) are generated in
n 1
lexicographic order, then rank( T ) xi 2 n . Although the Lexicographic order is simple to describe, the
i 0
distance between two successive subsets in the order may be irregular or large. Moreover, it never considers the
last element in the set until halfway through the sequence i.e. 2n-1subsets of the first n-1 elements must be
generated before the nth element of the set is considered. Hence the performance of this generation algorithm for
large values of n is poor.
rank
(T)
0
1
2
3
4
5
6
7
[0, 0, 0]
0
[0, 0, 0]
{3}
[0, 0, 1]
1
{3}
[0, 0, 1]
{2}
[0, 1, 0]
3
{2,3}
[0, 1, 1]
{2,3}
[0, 1, 1]
2
{2}
[0, 1, 0]
{1}
[1, 0, 0]
6
{1,2}
[1, 1, 0]
{1,3}
[1, 0, 1]
7 {1,2,3}
[1, 1, 1]
{1,2}
[1, 1, 0]
5
{1,3}
[1, 0, 1]
{1,2,3}
[1, 1, 1]
4
{1}
[1, 0, 0]
1145
{1}
{2}
{3}
{1,2}
{1, 3}
{2,3}
{1,2,3}
Binary
value
0
4
2
1
6
5
3
7
The Banker sequence is an efficient order for constraint satisfaction problems which seek the smallest subset
Tmin S. However, it is not optimal because the difference between any two successive subsets is not minimal.
The new method presented here is constructed by the revolving door [5] method which is a minimal change
ordering, however, we modified it to make an optimal generation algorithm. For a Positive integer n, let S =
{1,2,... ,n}. We define S to consist of
n
K-element
k
element subsets of S, and T1 is not equal to T2, then dist (T1, T2) 2. Hence, if
element subsets of S, then a minimal change ordering on
Sn
k
Sn
k
n
k
k-
consecutive subsets is two. Now, we study a minimal change ordering called the revolving door. The revolving
door ordering for
Sn
k
n
k
n,k
n,k n,k
n,k
A
[ A0 , A1 ,..., A
]
n 1 (1)
k
The revolving door algorithm is motivated by Pascal's identity:
n n 1 n 1
=
k k 1 + k
(2)
n 1
k 1
n
k-element
k
An1,k 1
n 1
k-element subsets
1146
n ,k
n 1,k 1
n 1,k 1
{ n },...,A0
{ n}]
,A
n 1 1
k 1
(3)
This recursive definition can be applied whenever 1<k<n-1. The lists An ,n and An ,0 are given as initial conditions to
start the recursion. They are as follows:
,
A n ,0 =[]
A n ,n = [{1,,n}](4)
Therefore, it can be verified that for integers k and n, we have [5]:
n ,k
A
{ 1,2 ,...,k 1, n }
n 1
k
(5)
n ,k
A0 { 1,2 ,...,k } (6)
With this background, we want to introduce a new total order called the Optimal banker to generate all subsets
of a set. This order is a Bankers one, i.e. subsets are generated in an increasing order with respect to their size.
On the other hand, the Optimal banker is a minimal change order among all potential Banker orders. We can
prove this claim as follows:
It is obvious that the smallest distance between any two k-subsets is two, while the distance between a ksubset and a (k+1)-subset is at least one. Hence the distance between consecutive subsets is two when they are of
the same size, and it is one otherwise. That is the reason why the order is called the Optimal banker. Thus, it
applies minimum number of changes to generate all subsets among all other Bankers orders. All subsets in an
Optimal banker sequence, use the revolving door method on k-subsets. Let An ,k be the list of all k-subsets in the
revolving door order and An ,k be the list of all k-subsets in the reverse revolving door order. The Optimal banker
order is defined as follows:
A
n ,k
(7)
where s1.s2 shows the concatenation of the two lists, s1 and s2.
Theorem: The Optimal banker is a minimal change Banker order.
Proof: As it is shown in [5], for all integers n,i the lists An ,i and hence An ,i are in a minimal change order, i.e.
the distance between each two successive subsets is exactly two. Then we must show that the distance between
the characteristic vectors of any two successive subsets of size k and k+1 is simply one.
1) If k is even, the two successive subsets are the last element of An , k and the first element of An ,k 1 , i.e. the last
element of An ,k 1 i.e.
n ,k
A n
1
k
n ,k 1
n ,k
A n
A n
{k }.
1
k 1 1
k
n ,k
A
{ 1,2 ,...,k 1, n }
n 1
k
and
n ,k 1
A n
{ 1,2 ,...,k 1, k , n }
k 1 1
, so
Hence the characteristic vectors of these subsets differ just in the position 'n-k'.
2) If k is odd, the two consecutive subsets are the last element of An ,k (which is the first element of An , k ) and
n,k
n,k 1 {1,2,...,k , k 1}
the first element of An ,k 1 , i.e. A0n ,k 1 , A0n ,k . According to (6): A0 {1,2,...,k} , and A0
, so
n ,k 1
A0
contains exactly one element more than A0n , k . Therefore, their characteristic vectors differ
only in one position.
= A0n ,k
{ k 1}
In this section, Rank, Unrank and Next Algorithms are shown. The last one generates an Optimal bankers
sequence for a set of n items in one possible way. Let R be the output Rank and K be the size of the output set.
OptBanker_Rank gets a sequence like T and generates its Rank.
OptBanker_Rank( T[1..N] )
K=0
// the size of the input subset
R=0
// the rank of the output subset
S = -1
// sign
for i=1 to N
if ( T[i] != 0 )
1147
K=K+1
R = R + S * ik
S=-S
if ( K is Even ) R = R + kj 11 Nj
else R = R + kj 1 Nj
return R
end
OptBanker_Next generates this sequence. Let T be the binary representation of the input subset, T be the next
sequence and N is the size of the set.
OptBanker_Next ( T[1..N] )
for i= 1 to N do
T[i] = T[i]
if( T = 0 ) // the given subset is the empty set
T[N]=1
else
T[N +1]=1
T[N +2]=1
j=1;
while ( T[j] = 1 )
j=j+1
i=j
while ( T[i] = 0 )
i=i+1
if ( j is Odd )
if ( j == 1 )
T[i] = 0
T[i-1] = 1
else
T[j] = 1
T[j-2] = 0
else
if ( T[i+1] == 0 )
T[j-1] = 0
T[i+1] = 1
else
T [i+1] = 0
T [j] = 1
return T[1..N]
end
OptBanker_Unrank gets R, the Rank of a sequence and generates the binary representation of R.
OptBanker_Unrank ( R )
K=0;
while ( R > KN -1 )
R = R - KN
K=K+1
if ( K is Odd )
R = KN - R - 1
for ( i=1 to N )
T[i]=0
X=N;
for ( i=k Downto 1 )
1148
while ( iX > R )
X=X1
T[X+1] = 1
R = iX 1 - R - 1
return T
end
Conclusion
One category of constraint satisfaction problems seeks the smallest subset of a given set which satisfies certain
constraints. In this paper, we proposed a novel generation algorithm, the Optimal banker, to help solving the
above mentioned category of CSPs. It generates a sequence of all the subsets of any given set, in which not only
the number of elements of subsets is monotonically increasing, but also the change between each two
consecutive subsets is minimal. Thus, if we start examining from the first subset in the sequence for being a
solution, the first solution found is an optimal one, so there is no need to produce and inspect other subsets.
Eventually, the three important algorithms: Next, Rank, and Unrank were presented.
In comparison with the two popular generation algorithms, i.e. Gray code and the lexicographic order, we
should mention that the proposed algorithm does not have their disadvantages when the set is large. Indeed,
when we use them, not all the elements of the set have the chance to be selected in the possible solution unless
half of all subsets (i.e. 2n 1 subsets) are inspected. In addition, due to the fact that the number of elements in
their produced sequences is not increasing, all 2 n subsets should be produced and investigated to find an optimal
solution. On the contrary, if the Optimal banker is utilized as a generator of subsets, the optimal solution is
obtained in much less time. However, while the change between each two successive subsets in a sequence
produced by Gray code is 1, it is 1 or 2 in a sequence produced by the banker.
No other known generation algorithm produces a monotonic minimal change sequence. In addition, the
Optimal banker is optimal among all monotonically increasing generation algorithms with respect to the distance
between two consecutive subsets.
Acknowledgements
It is pleasure to acknowledge financial support of Islamic Azad University, South Tehran Branch for this
research project.
References
1.
2.
3.
4.
5.
6.
7.
8.
Peter Eades, Brendan D. McKay: An Algorithm for Generating Subsets of Fixed Size With a Strong
Minimal Change Property, Information Processing Letters, Volume 19, Number 3, 19 October 1984.
Patrick Suppes, Axiomatic Set Theory, D. Van Nostrand Company, Inc., 1960.
Gian-Carlo Rota, Studies in Combinatorics. MAA Studies in Mathematics. Mathematical Association
of America. p. 3. ISBN 0-88385-117-2, 1978.
Jech, Thomas, Set Theory. Springer-Verlag. ISBN 3-540-44085-2, 2002.
Donald L. Kreher, Douglas Robert Stinson, Combinatorial algorithms: generation, enumeration, and
search, CRC press LLC 1999.
Tsang, Edward (1993). Foundations of Constraint Satisfaction. Academic Press.
http://www.bracil.net/edward/FCS.html. ISBN 0-12-701610-4.
Steven S. Skiena, The Algorithm Design Manual, Second Edition, Springer-Verlag London Limited
2008.
Hamming, Richard W., "Error detecting anderror correcting codes", Bell System Technical Journal 29
(2): 147160, MR0035935, 1950.
1149
9.
J. Loughry, J.I. van Hemerty, L. Schoofsz, Efficiently Enumerating the Subsets of a Set, December
2000. http://applied-math.org/subset.pdf.
10. Donald E. Knuth. The Art of Computer Programming. Addison-Wesley Publishing Company, Reading,
Massachusetts, 1973.
Biographies
Shahriar Pirnia Naeini was born in Esfahan, Iran in 1971. He received a B.Sc. degree in Electrical
Engineering-Control from Tehran University, Tehran, Iran in 1995, and an M.Sc. degree in Computer
Engineering-Software from Islamic Azad University, South Tehran Branch, in Tehran, Iran in 1999. He has been
a lecturer of Computer Engineering at Islamic Azad University, South Tehran Branch, in Tehran, Iran since
2000. He is the first author of two international or national conference papers. He is interested in pattern
recognition, neural networks, computational and algorithmic problems, and machine learning. Mr. Pirnia is a
peer reviewer of IEEE/IACSIT conferences and a member of IACSIT.
Seyedeh Narjess Afzaly was born in Rasht, Iran in 1985. She received a B.Sc. degree in Computer Science
from the Shahid Beheshti University (SBU) in Iran, and an M.Sc. degree in computer Science, from the
University of Tehran in 2011. She likes Combinatorics and Algorithms in the field of computer Science. She is
currently working on graph theory.
Amir Hosein Kashefi was born in Tehran, Iran in 1984. He received a B.Sc. degree in Computer Engineering
from the Islamic Azad University of South Tehran, in 2008. He received an M.Sc. degree in Computer Science
from the University of Tehran, in 2011. His interest areas are Artificial Intelligence, Machine Learning,
Intelligent Systems, Bioinformatics and computational and Algorithmic problems. He is the author or co-author
of more than 4 international conference papers in his research area.
Fatemeh Sharifizadeh was born in Tehran, Iran in 1985. She received a B.Sc. degree in Computer Science
from the University of Tehran, in 2009. She is now a student of Computer Science in M.Sc. degree. Her interest
areas are Machine Learning, Intelligent Systems, Bioinformatic Algorithms, and Computational Neuroscience.
She is the author or co-author of more than 15 international journal and conference publications in her research
area.
1150
Abstract: In this study, we consider one-dimensional Burgers' equation. The equation is non-linear partial
differential equation which can be solved analytically for the arbitrary initial and boundary conditions. The
equation has a coefficient of the kinematic viscosity, first and second order derivatives.Initial and boundary
conditions are chosen as homogeneous. We consider the spatial domain [0,1] is partitioned into N finite
element as a uniform. Cubic B-spline functions are used as basis function.Cubic B-spline collocation finite
element method for calculating numerical solution of one-dimensional Burgers' equation is used. Spacesplitted Burgers' equation is solved numerically.
Keywords: B-Spline, Burgers' Equation,Collocation Method, Space-Splitted
1. Introduction
Burgers' equation was first introduced by Bateman [1].Burgers' extensive work on Burger equation as a
mathematical model for the turbulence.For this reason, this equation is known Burgers' equation.The equation is
used wide field as heat conduction [2], gas dynamics [3], shock waves [4], number theory [5] and so
forth.Burgers' equation is solved exactly for an arbitrary initial and boundary conditions [6,7,8].Many authors
have used a variety method of numerical techniques for getting numerical solution of Burgers' equation.Great
numbers of those techniques for solving Burgers' equation spline and B-spline functions are used in variable
methods.For instance,cubic B-spline collocation method is suggeted for one-dimensional Burgers' equation
[9,10,11].The cubic spline function technique and quasilinearisation for the numerical solutions of the Burgers'
equation have been used in one space variable at low Reynolds numbers [12].The equation is solved numerically
by the collocation method with cubic B-spline interpolation functions over uniform elements [13] .A finite
element solution of the Burgers' equation based on Galerkin method using B-splines as both element shape and
test functions is developed in the studies [14,15,16]. Least-squares formulation using quadratic B-splines as trial
function is given over the finite intervals [17]. A Petrov- Galerkin method is used by a quadratic B-spline spatial
finite elements in [18], and it is also used a least-squares technique using linear space-time finite elements in
[19].
Spline functions are adapted to numerical methods to get the solution of the differential equations.Numerical
methods with spline functions in getting the numerical solution of the differential equations lead to band
matrices. Solutions of this matrix systems are very easy and useful for computer programming.The term B-spline
was coined by Isaac Jacob Schoenberg (1946) and is short for "basis spline".B-spline functions are piecewise
value and their derivatives of lower degree are continous.
B-spline functions are very popular for the smooth approximations and are being used by various researchers
for different problems ([20],[21],[22],[23],[24]) gave recurrence relations for the purpose of computing
1151
coefficients ([25],[26]). Yang used a technique to deal with the boundary conditions using splines [27]. Many
others have used B-splines for various applications where B-splines has been proved as a useful tool for
approximations [28,29].In addition to this, Thomas presented extensive use of splines for Boeing [30].Numerical
methods using B-spline functions has been successfully applied to solve various linear and nonlinear partial
differential equations.The success of the B-spline collocation method is dependent on the choice of B-spline
basis.In this study we used cubic B-spline bases to build up the approximate solutions for Burger's equation.
The one-dimensional Burgers' equation is given as follows.
(1.1)
is the coefficient of the kinematic viscosity and subscripts and denote differentiation. We consider the
-is partitioned into finite element as a uniform such that
spatial domain ,
and
throughout paper.Burgers's equationis splitted in space then cubic B-spline collocation method is
applied. For this purpose, we set
in the Burgers' equation.
(1.2)
The initial and boundary conditions are chosen as follows.
(
)
(
(
where
( )
( )
(
)
(
(
(1.3)
)
(
(
)
)
)
(
(
)
)
(
(
by the knots
,
at these knots
)
)
(2.1)
where
for all m. The values of cubic B-spline function ( ) and all its first,second
-[31]. One can easily verify that the values of
derivatives vanish outside the interval,
( )
and its derivatives are as shown. The collocation method is applied to find approximate solutions of the
system (1.2). Collocation approximant can be expressed for ( ) and ( )in terms of element
parameters
and , respectively.
(
( )
)
(
( )
(
(
( )
(2.2)
)
)
1152
( )
(2.3)
)
(
(
)(
)
(2.4)
(
where
denotes
differentiation
with
respect
to
time
and
(2.5)
)unknowns.
(2.6)
Where
,
We use boundary conditions to eliminate parameters
,
,
,
from the system (2.6) to have an
) (
) 7-band matrix system.
equal solvable (
Before the solution process begin iteratively, we need to compute the initial parameters
(
) and
(
) by using initial conditions of Burgers's equation.
And thus, we get a matrix system which can be solved iteratively with the help of Thomas algorithm.
(
( )
( )
(2.7)
(
(
)
)
( )
( )
1153
(2.8)
(2.9)
(2.10)
) unknowns.
(2.11)
, ( )
( )
( )
( )
( )
( )
)
(
(
)
)(
)-
)
(2.12)
1154
/
.
(
/
)
(
(
)
)
(
where are the modified Bessel function. For small the convergence of this series is very slow.
The numerical solution of Burgers' equation for the problem reported by Caldwell [32] and by Ali
[33]andthat obtained by the present collocation method are compared with the analytic solution in Table
(3.1). These numerical solutions are for the case
using a time step
.
h
0.25
0.50
[32]
[33]
Present
Exact
0.7071
0.7071
0.7071
0.7071
0.05
0.3734
0.4082
0.4130
0.4131
0.10
0.2202
0.2463
0.2536
0.2536
0.15
0.1334
0.1463
0.1567
0.1566
0.20
0.0814
0.0898
0.0964
0.0964
0.25
0.0510
0.0540
0.0592
0.0592
1.0000
1.0000
1.0000
1.0000
0.05
0.5563
0.5979
0.6091
0.6091
0.10
0.3263
0.3578
0.3715
0.3716
0.15
0.1944
0.2142
0.2269
0.2268
0.20
0.1173
0.1283
0.1385
0.1385
0.25
0.0614
0.0768
0.0845
0.0845
0.75
0.7071
0.7071
0.7071
0.7071
0.05
0.4263
0.4384
0.4503
0.4502
0.10
0.2444
0.2601
0.2727
0.2726
0.15
0.1423
0.1542
0.1644
0.1644
0.20
0.0843
0.0917
0.0995
0.0994
0.25
0.0501
0.0547
0.0603
0.0603
Table (3.1)
4.Conclusion
In this paper a numerical solution for non-linear Burgers equation is proposed using collocation method with the
cubic B-splines. We obtained approximate numerical solutions maintain good accuracy compared with the exact
solutions.
1155
References
1.
2.
3.
4.
5.
6.
7.
8.
9.
10.
11.
12.
13.
14.
15.
16.
17.
18.
19.
20.
21.
22.
23.
24.
25.
26.
27.
28.
Bateman, H. (1915). Some recent researches on the motion of the uids; Monthly Weather Rev.
43:16370.
Cole,J. D. ( 1951). On a quasi-linear parabolic equation occurring in aerodynamics; Q.Appl.Math.9,
225236.
Lighthill,M. J. (1956). Viscosity effects in sound waves of nite amplitude; Surveys in Mechanics (G.
K. Batchlor and R. M. Davies, eds.), Cambridge University Press, Cambridge, , pp. 250351 (2
plates).
Burgers, J.M. (1948). A mathematical model illustrating the theory of turbulence; AdvancesinApplied
Mechanics; Academic Press, New York, pp. 171199.
Van der Pol, B. (1951). On a non-linear partial differential equation satised by the logarithm of the
Jacobian theta-functions, with arithmetical applications, Proc.Acad.Sci.Amsterdam 013.
Burgers ,J.M. (1948). A mathematical model illustrating the theory of turbulence; Advances in applied
mechanics. New York: Academic Press; p. 17199
Hopf, E. (1950). The partial dierential equation Ut + UUx mUxx = 0; Commun Pure Appl
Math.3:20130.
Cole, J.D. (1951). On a quasi-linear parabolic equations occurring in aerodynamics; Quart Appl
Math.9:22536.
Rubin, S.G. and Graves, RA. (1975). Viscous ow solutions with a cubic spline approximation;
Comput Fluids.3:136
Rubin ,S.G.and Khosla, PK. (1976). Higher-order numerical solutions using cubic splines; AIAA
J.14:8518.
Caldwell ,J. Applications of cubic splines to the nonlinear Burgers equation; In: Hinton E et al.,
editors. Numerical Methods for Nonlinear Problems, 3. p. 25361.
Rubin, S. G.; and Graves, R. A. (1975). Cubic spline approximation for problems in uid mechanics,
Nasa TR R-436 .District of Columbia
Ali, A.H.A.; Gardner,G.A.; and Gardner.L.R.T. (1992). A collocation solution for Burgers equation
using cubic B-spline nite elements; Comput. Method Appl. M. 100 no. 3, 325337.
Ali,A.H.,A.; Gardner,L.R.T.; and Gardner,G.A. (1990). A Galerkin Approach to the Solution of
Burgers Equation; University College ofNorthWales, Bangor,Maths Preprint Series, no. 90.04.
Davies, A.M. (1978). Application of the Galerkin method to the solution of the Burgers equation;
Comput. Method Appl. M. 14, 305321.
Gardner,L.R.T.; Gardner,G.A.; and Ali, A. H. A. ( 1991). A method of lines solutions for Burgers
equation. Proceeding of the Asian Pacic Conference on Computational Mechanics; A.A. Balkema/
Rotterdam/ Brookeld, Hong Kong.
Kutluay, S.; Esen, A.; and Da,. (2004). Numerical solutions of the Burgers equation by the leastsquares quadratic B-spline nite element method. J.Comput. Appl. Math. 167, no. 1, 2133.
Gardner, L.R.T.; Gardner,G.A.; and Dogan,A. (1996). A least-squares finite element scheme for
Burgers equation; University of Wales, Bangor, Mathematics, Preprint 96.01
Gardner, L.R.T.; Gardner,G.A.; and Dogan,A. (1997). A PetrovGalerkin finite element scheme for
Burgers equation; Arab.J. Sci.Engrg.22, 99109
Kashiwagi, M.(1998). A B-spline Galerkin scheme for calculating the hydroelastic response of a very
large oating structure in waves .Journal of Marine Science and Technology ,Vol. 3, pp. 37-49.
Prautzsch, H.(1989). A round trip to B-spline via de Casteljau. ACM Transactions on Graphics(TOG).
Vol. 8, No. 3, pp. 243-254.
Bahadir, A.R.(2004). Application of cubic B-spline nite element technique to the thermistor
problem. Applied Mathematics and Computation. Vol. 149, pp. 379-387.
Xuebiao, L.(1988). The B-spline nite elemenet method applied to axisymmetrical and nonlinear eld
problems. IEEE Transactions on Magnetic. Vol. 24, No. 1, pp.27-30.
Guangzheng, N.; Xiaoming, X.; Baidun, J.(1990). B-spline Finite Element Method for Eddy Current
eld analysis . IEEE Transactions on Magnetic .Vol. 26, No. 2, 1990, pp.723-726.
de Boor,C.( 1972). On Calculating with B-splines. Journal of Approximation Theory. Vol. 6,1972,
pp.50 - 62.
de Boor, C. (2001). A Practicle guide to splines. Applied Mathematical Sciences, Springer-Verlag
Yang, L.F. The Spline Matrix Method. J. Guangxi Univ. Vol. 18, No. 2, 1994, pp.61-65.
Zong, Z.; Lam, K.Y.(1998). Estimation of complicated distributions using B-spline functions. Struct.
Safety. , Vol. 20, 1998, pp.341-355.
1156
29. Cheung, Y.K.; Fan, S.C.; Wu, C.Q.(1982). Spline nite strip in structural analysis. Proceedings of
Intrnational Conference on Finite Element Method. , pp.704-709.
30. Grandine, T.A.(2005). The Extensive use of Splines at Boeing. SIAM News. Vol. 38, No. 4
31. Prenter , P.M. (1975). Splines and variational methods. J.Wiley, New York.
32. J. Caldwell,(1987) Application of cubic splines to the nonlinear Burgers' equation, in: E. Hinton
et al., eds.,Numerical Methods for Nonlinear Problems, Vol. 3 (Pineridge, Swansea),253-261.
33. Ali, A.H.A.(1992).A collocation solution for Burgers' equation using
cubic B-spline finite elements.Computer Methods in Applied Mechanics and Engineering, NorthHolland 100, 325-337
Glnur Yel - MERSN 09.08.1987
Undergraduate : Mugla University Department of Mathematics (2005)
Master Student: Mugla University Institute of Science Department of Mathematics, Applied Mathematics (2009)
National Journal Papers
G. Yel, M..NVAR, N.LER, .IIK, Studies on Generalization of Special Equations at the q-Integers, 5.
Ankara Mathematics Days Symposium at TOBB Economics and Technology University,3-4 June 2010
International Journal Papers
G. Yel, M..NVAR, N.LER, .IIK, Studies on Generalization of Special Equations at the q-Integers, The
6th Dynamical Systems and Applications, , Volume 11, Number J11 ,ISSN: 0974-570X, 2011
Zeynep Fidan Koak ANKARA 07.02. 1949
1157
1158
1159
Absract:In this study in order to find the approximate solution of the linear Lane-Emden equation Hermite polynomials are
used. In this method an alternative and approximate solution of this equation is introduced which based on Hermite
polynomials. Each equation in this method we use has matrix form and we use these forms. After applying these procedure.
Finally the exact and the approximate solution is compared.
1.Introduction
In recent years, the studies of singular initial value problems in the second order ordinary differential equations
have attracted the attention of many mathematicians and physicists. One of the equations describing this type is
the LaneEmden-type equations formulated as
(1)
where
and
(2)
are constant,
P2 (t ) 1, P1 (t )
g (t ) C[0,1] . In [1], the authors proved local existence of solutions and under additional assumptions,
uniqueness of the solutions for the initial value problem (1). Since, LaneEmden type equations have significant
applications in many fields of scientific and technical world, a variety of forms of f (t , y ) and g (t ) have been
investigated by many researchers. A discussion of the formulation of these models and the physical structure of
the solutions can be found in the literature. For example, for f (t , y) y , g (t ) 0 , Eq.(1) is the standart
Lane-Emden equation that was used to model the thermal behavior of a spherical cloud of gas acting under the
mutual attraction of its molecules and subject to the classical laws of thermodynamics [2-3] and for
m
f (t , y) e y , g (t ) 0 , Eq.(1) is the isothermal gas sphere equation, where the temperature remains constant
[3]. Several methods for the solution of Lane-Emden equations have been presented, which are the modified
adomian decomposition method [4], integral operators method [5], Hermite collocation method [6] and other
methods [7-13]. The aim of this study is to get solution as truncated Hermite series defined by
N
y (t ) a n H n (t )
(3)
n 0
where
1160
2.Fundamental Relations
In this method our aim is to find the matrix form of each equation in which we use for approximate solution.
Firstly we convert that y (t ) and
y k (t ) such that
k=0,1,2
(4)
H(t ) [ H 0 (t ) H1 (t ) H N (t )]
where
and
A [a0
a1 a N ]T .
(5)
t 2r
H n (t ) is
H 2 n (t )
(2r )!
2r
2 n0 (2n)!(r n)!
(6)
and
t 2 r 1
H 2 n 1 (t )
(2r 1)! r
, r 0,1,2...
2 r 1
2
n 0 ( 2n 1)!( r n)!
(7)
By using (6) and when n=0,1...N the matrix form of the relation is
(8)
here
X(t ) [1 t t N ]
and
for even N
1!
0
0
1!0!0!
1!
0
0
2!1!1!
2!
2!
M
0
2
2!0!1!
2 2!1!
N!
N!
0
N
2 N 0!( N )!
2 N 2!( 1)!
2
2
0
N!
2 N N !0!
(9)
0!
1!0!0!
2!
M 2
2 0!1!
1!
211!1!
2!
2 2!1!
N!
N
2 N 2! 1!
2
0
N!
2 N 0! N !
(10)
1161
H(t ) X(t )M T ) 1
is obtained and also the k-th power of each matrix in the equation is
H ( k ) (t ) X ( k ) (t )(M T ) 1 k=01,2...
(11)
where
Clearly, we know that
X1 (t ) X(t )B1
X 2 (t ) X1 (t )B X(t )B 2
X k (t ) X k 1 (t )B X(t )B k
(12)
(13)
where
0
0
B 0
0
0
1 0
0 2
0 0
0
0 0
0
0
0
N
0
y ( k ) (t ) X(t )B k (M T ) 1 A
(14)
3.Method of solution
t t as
0 lim P1* (t ) P1 (t )
0
and
(15)
P XB (M
2
) 1 P1 XB(M T ) 1 P0 X(M T ) 1 A G
*
(16)
where
0
0
P0 (t 0 )
0
P0 (t1 )
0
P0 0
0
P0 (t 2 )
0
0
0
0
0
P1 (t0 )
P1 (t1 )
0
0
P1 (t2 )
P1 0
0
0
P0 (t N )
0
0
0
0
P1 (t N )
0
0
0
1162
has a
P1 (t ) P1 (t ) . And then
i
by using collocation points t i
in our equation we obtain
N
P0 (t i )X(MT ) 1 )A g (t i )
1 x0
0
0
0
P2 ( x0 )
P2 ( x1 )
0
0
1 x1
P2 0
0
P2 ( x 2 )
0 X 1 x 2
1 x
0
0
0
P2 ( x N )
N
W A G or
W ; G W
wi , j
x0
x1
x2
xN
N
x0
N
x1
N
x2
N
x N
, i,j=0,1,2,...,N
(17)
(18)
w00
w
10
~ ~
W ; F
wN 20
u 00
u10
w01
w11
wN 21
u 01
u11
w0 N
w1N
wN 2 N
u0 N
u1N
; g (t 0 )
; g (t1 )
; g (t N 2 )
;
0
;
1
~
~ ~
rankW rank W; F N 1 then
~
~
A (W ) 1 F can be written. A is uniquely determined. And then the Eq(1) has the unique solution the
solution is by the Hermite polynomials is
of this method can be checked easily. Since the truncated Hermite series is an approximate solution of
Eq.(1),when the solution y N (t ) and its derivatives are substituted in Eq.(1) the result matrix must be satisfied
approximately;that is, for
t t q 0,1 ,q=0,1,2,...
y ''
2 '
y (4t 2 6) y 0
t
June 1 -4, Kusadasi, Aydin, Turkey http://iscse2011.gediz.edu.tr
1163
y(0) 1 ,
y ' (0) 0 .
y e t .In our problems we use Maple 9 program in order to obtain the approximate
solutions. We solve this equation for N 10,15,20 . We get approximate solutions of this equation then the
2
comparison between the exact and approximate solutions is given in Table 1 and plotted errors functions in
Figure 1.
Exact
Solution
1.0000000
1.0100501
1.0408107
1.0941742
1.1735108
1.2840254
1.4333294
1.6323162
1.8964808
2.2479079
2.7182818
0.0
0.1
0.2
0.3
0.4
0.5
0.6
0.7
0.8
0.9
1.0
Ne=10
Ne=15
Ne=20
0.7000E -6
0.3547E-6
0.4546 -6
0.4982E-7
0.5455E-6
0.6029E-6
0.6763E-7
0.7735E-6
0.8963E-5
0.1091E-5
0.4729E-5
0.5300E-27
0.3429E-10
0.3924E-10
0.4260E-10
0.4632E-10
0.5104E-10
0.5719E-10
0.6527E-10
0.7591E-10
0.9018E-10
0.17033E-10
0.2220E-35
0.1279E-14
0.1429E-14
0.1539E-14
0.1667E-14
01834E-14
0.2053E-14
0.2342E-14
0.2724E-14
0.3228E-14
0.1835E-14
-6
x 10
Ne=10
Ne=15
Ne=20
0.1
0.2
0.3
0.4
0.5
0.6
0.7
0.8
0.9
Example 2:
Our last example is
y ''
2 '
y y 6 12 x x 2 x 3
x
1164
Biographies
Hatice YALMAN Hatice Yalman was born on March 6th 1986 in Mersin ,Turkey. She received her B.Sc
from the department of Mathematics of nn University in 2009 and has been a research assistant at Mula
University since 2010.Her M.Sc is at Mula University and now she is studying Numerical Solutions of
Ordinary Differential Equations.
Yaln ZTRK - Yaln ZTRK was born on November 01th 1983 in Antalya,Turkey. He received his
B.Sc. from the Department of Mathematics of Mula University in 2005 and began his M Sc. in 2006 and
finished his M. Sc. in 2009. He has been research assistant at Department of Mathematics of Mula University
1165
since 2006. He began his Ph.D. under the supervision of Associate Professor Mustafa GLSU in Department of
Mathematics of Mula University in 2009.
His main research interests are computational numerical linear algebra and numerical solutions of differential
and integral equations. Some of his published articles are:
- Mustafa Glsu, Yaln ztrk, Mehmet SEZER, A new collocation method for solution
of mixed linear
integro-differential-difference equations, AMC 216:2183-2198(2010)
- Mustafa Glsu, Yaln ztrk, Mehmet SEZER, On the solution of the abel equation of the second kind by the
shifted Chebyshev polynomials, AMC 217:4827-4833(2011)
- Yaln ztrk, Mustafa Glsu, Mehmet SEZER, Chebyshev polynomial bases for solving linear FredholmVolterra delay differential difference equations. The 6th International Conference: dynamical systems and
applications-2010 Antalya-Turkey, July 10-14, 2010.
-Mustafa Glsu, Yaln ztrk, Chebyshev polynomial approximation for solving the second kind linear
Fredholm integral equation, Erciyes University Journal of the Institute of Science and Technology 26(3):203216(2010)
Mustafa GLSU Assoc.Prof.Dr. Mustafa Glsu was born on October 29th 1965 in stanbul,Turkey. He
received his B.Sc. from the Department of Mathematics of Marmara University in 1988 and was research
assistant at Marmara University. His M.Sc. was from Northeastern University in 1997. He received his Ph.D.
under the supervision of Prof.Dr.Turgut zi in the Department of Applied Mathematics at Ege University in
2002.
His main research interests are ordinary and partial differential equations, integral and difference equations and
their numerical solutions. Assoc.Prof.Dr. Glsu has been reviewer for numerous influential journals, has
published research articles related to differential equations, partial differential equations, integral and difference
equations and has been authored over 20 papers.
- Mustafa Glsu, Yaln ztrk, Mehmet SEZER, A new collocation method for solution
integro-differential-difference equations, AMC 216:2183-2198(2010)
of mixed linear
- Mustafa Glsu, Yaln ztrk, Mehmet SEZER, On the solution of the abel equation of the second kind by the
shifted Chebyshev polynomials, AMC 217:4827-4833(2011)
- Yaln ztrk, Mustafa Glsu, Mehmet SEZER, Chebyshev polynomial bases for solving linear FredholmVolterra delay differential difference equations. The 6th International Conference: dynamical systems and
applications-2010 Antalya-Turkey, July 10-14, 2010.
-Mustafa Glsu, Yaln ztrk, Chebyshev polynomial approximation for solving the second kind linear
Fredholm integral equation, Erciyes University Journal of the Institute of Science and Technology 26(3):203216(2010)
Mehmet SEZER - Prof. Dr. Mehmet SEZER was born on March 20th 1954 in Antalya, Turkey. He received
his B.Sc. from the Department of Mathematics of Ege University in 1976 and was research assistant at Balkesir
Engineering Academy. His M.Sc. was from Ege University in 1977. He received his Ph.D. under the supervision
of Prof. Dr. Seda MORALI in the Department of Applied Mathematics at Ege University in 1982
and
studied
in
the
field
of
"Parabolic
partial
differential
equations".
In 1982, he was Assistant Professor at Faculty of Education (Balkesir) of Uluda University, Associate
Professor in 1989, and in 1995 Professor at Faculty of Education of Dokuz Eylul University. Since 2004, he has
been Mathematics Professor at Faculty of Science at Mula University and has been the chair in the same
department.
His main research interests are ordinary and partial differential equations, integral and difference equations, and
their numerical solutions. Prof. Sezer has been reviewer for numerous influential journals, has published research
articles related to differential equations, linear algebra, analytic geometry and calculus; and has been authored
over 80 papers.
1166
Abstract.In this paper a new method is suggested for obtaining the exact and numerical solutions of the
initial-boundary value problem for a nonlinear parabolic type equation in the domain with the free boundary.
With this aim, a special auxiliary problem having some advantages over the main problem and being
equivalent to the main problem in a definite sense is introduced. The auxiliary problem allows us to obtain the
weak solution in a class of discontinuous functions with regular and singular properties. Moreover, on the
basis of the auxiliary problem a higher resolution method is developed so that the solution accurately
describes all physical properties of the problem. In order to extract the significance of the numerical solutions
obtained by using the suggested auxiliary problem, some computer experiments are carried out.
Keywords: free boundary problem, effect localization, main and auxiliary problem, weak solution,
1 Introduction
It is known that many practical problems such as distribution of heat waves, melting glaciers, filtration of a gas
in a porous medium etc. are described by nonlinear equations of the parabolic type.
In [1], at first the effect of localization of the solution of the equation describing the motion of perfect gas in a
porous medium is observed and the solution in the traveling wave form is structured. Then, the mentioned
properties of the solution for the nonlinear parabolic type equation are studied in [2],[3], etc.
These problems also are called free boundary problems. Therefore, it is necessary to obtain the moving unknown
boundary together with the solution of a differential problem. Its nature raises several difficulties for finding
analytical as well as numerical solutions of this problem.
The questions of the existence and uniqueness of the solutions of the free boundary problems are studied in
[4],[5]. In [5], Oleinik introduced the notion of a generalized solution of the Stefan problem whose uniqueness
and existence were guaranteed in the class of measurable bounded functions.
In the literature there are some numerical algorithms (homogeneous schemes) which are approximated by finite
differences of the differential problem without taking into account the properties occurring in the exact solution
[6],[7],[8].
In [4], Kamenomostskaya considered the classical quasilinear heat conduction equation and constructed the
generalized solution by the use of an explicit difference scheme.
1167
u 2 (u)
=
, in R2
t
x 2
(1)
u( x,0) = u0 ( x ) = 0, in I = [0, )
(2)
(3)
1
D 1 1 ( Dt x ) 1 , 0 < x < Dt ,
(4)
u( x , t ) =
0,
x Dt .
u
= Du ( x, t ) are continuous in D(t ) = ( x, t ) | 0 x Dt ,0 t T , but
x
1
1
the solution is equal to zero at x = (t ) = Dt . Here D =
is the speed
u
1 0
1
of the front of the traveling wave and
u((t ), t ) = 0.
(5)
formula (4), when n =
u Du( x , t )
dx
x
Taking into account (5), we get
= t =
= D.
u
u
dt
x
x
It is clear that, when u = 0 the equation (1) degenerates to the first order equation.
The weak solution is defined as follows.
Definition 1. A nonnegative function u( x, t ) which is satisfying the initial condition (2) and boundary
conditions (3),(5) is called a weak solution of the problem (1)-(3) and (5), if the following integral relation
f ( x, t ) u ( x, t ) f ( x, t )
u ( x, t )
dxdt = 0
D (t )
(6)
1168
0=
u( x, t )
0 0
T Dt
0 0
u ( x, t )
f ( x, t )
dxdt D
t
f ( x, t )
f ( x, t )
Du ( x, t )
dxdt
t
x
T Dt
u( x, t )
0 0
f ( x, t )
dxdt .
x
Changing the order of integration in the first integral and then applying the integration by parts to the inner
integrals of the first and second term in the last expression with respect to t and x , respectively, we have
DT T
0 x/D
u ( x, t )
f ( x, t )
dtdx D
t
T Dt
u ( x, t )
0 0
f ( x, t )
dxdt = 0.
x
If we reapply integration by parts to the inner integrals in the first term and second term by t and x ,
respectively we prove that the solution in the form (4) satisfies the integral relation (6).
v
v
=
,
t x x
(7)
v( x,0) = v0 ( x ),
(8)
u(0, t ) =
v(0, t )
= u0 t n
x
(9)
dv0 ( x )
= 0.
dx
(10)
D 1 1 1 ( Dt x ) 1 , x < Dt ,
v( x, t ) =
0,
x Dt
(11)
in the traveling wave form [9]. As is seen from (11) the differentiability property of the function v( x, t ) is more
than the differentiability property of the solution u( x, t ) . In addition to this, from (11) we get
u( x , t ) =
v ( x, t )
.
x
(12)
Theorem 1. If the function v( x, t ) is a soft solution of the problem (7)-(9), then the function u( x, t ) obtained by
(12) is a weak solution of the main problem (1)-(3) in sense of (9).
The auxiliary problem has the following advantages:
1169
Figure 1: a) The exact solution of the main problem, b) The exact solution of the auxiliary problem
v( x, t )
, b) Numerical solution obtained by using the classical algorithm
x
(17)-(19)
= ( xi , t k ) | xi = Dt i , t k = k , i, k = 0,1,2,...,
where is the step of the grid with respect to t variable.
Now, we will develop a numerical algorithm as follows. Since the function
u
is continuous, we can
x
i 1,k
Vi ,k Vi,k Vi 1,k ,
(13)
1170
Vi,0 = v0 ( x i ),
(14)
(15)
(i = 0,1,2,...;k = 0,1,2,...). Here, h D and v0 ( x i ) is any grid function of the equation (10).
It can be easily shown that,
U i ,k =
Vi 1,k Vi ,k
(16)
and the grid function U i ,k 1 defined by (16) is a solution of the nonlinear system of algebraic equations
U i,k 1 = U i,k
i 1,k
2U i,k U i1,k ,
(17)
(i = 1,2,...;k = 0,1,2,...). The initial and boundary conditions for (17) are
U i,0 = 0,
(i = 0,1,2,...)
U 0,k = u0 t kn ,
(18)
(k = 0,1,2,...).
(19)
Here, Vi ,k and U i ,k are the approximation values of v( x, t ) and u( x, t ) at any point ( xi , t k ) of the grid .
5 Numerical Experiments
In order to extract the significance of the suggested method, the numerical solution obtained using the proposed
auxiliary problem is compared with the exact solution of the problem (1)-(3) on an equal footing. With this aim,
firstly, computer experiments are performed on the algorithm (13)-(15) with the values = 3 , T = 2,4 and 6.
The graphs of the obtained numerical solutions are presented in Figure 3 a). The numerical solutions of the main
problem (1)-(3) obtained by using the algorithm (13)-(15) are shown in Figure 3 b). The numerical solution of
the main problem obtained using the algorithm (17)-(19) is demonstrated in Figure 2 b).
Figure 3: a) Numerical solution of the problem (13)-(15), b) Numerical solution of the main problem obtained
by using the solution of the problem (13)-(15)
Comparing the Figures 1 a) and 3 b) shows that the numerical and the exact solutions of the main problem (1)(3) coincide. Moreover, the graphs of the functions v( x, t ) and Vi ,k coincide, too.
1171
Thus, the numerical experiments carried out show that the suggested numerical algorithms are efficient and
economical from a computer point of view. The proposed algorithms permit us to develop the higher resolution
methods where the obtained solution correctly describes all physical features of the problem, even if the
differentiability order of the solution of the problem is less than the order of differentiability which is required by
the equation from the solution.
The finite differences scheme (13)-(15) has the first order approximation with respect to t . But using the
different versions of the Runge-Kutta's methods we can increase the order of the algorithms mentioned above.
6 Conclusions
The new method is suggested for obtaining the regular weak solution for the free boundary problem of the
nonlinear parabolic type equation.
The auxiliary problem which has some advantages over the main problem is introduced and it permits us to find
the exact solution with singular properties.
The auxiliary problem introduced above allows us to develop the higher resolution method where the obtained
solution correctly describes all physical features of the problem, even if the differentiability order of the solution
of the problem is less than the order of differentiability which is required by the equation from the solution.
References
1.
Barenblatt, G.I., Vishik, M.I. (1956). On the Finite Speed of Propagation in the Problems of Unsteady
Filtration of Fluid and Gas in a Porous Medium, Prikladnaya Mat. Mech. (Applied Mathematics and
Mechanics), 20(3): 411-417.
2. Antoncev, S.N. (1981). On the Localization of Solutions of Non-linear Degenerate Elliptic
andParabolic Equations, Soviet Math Dokl., 24: 420-424.
3. Martinson, L.K., Pavlov, K.B. (1972). On the Problem of Spatial Localization of Thermal Perturbations
in the Theory of Non-linear Heat Conduction, USSR Computational Math and MathPhysics, 12(4): 261268.
4. Kamenomostskaya, S.L. (1961). On Stefan's Problem, Math. Sb., 53(95) 4: 489-514.
5. Oleinik O.A. (1960). A Method of Solution of the General Stefan Problem, Soviet Math. Dokl., 1(6):
1350-1354.
6. Abramov, A.A., Gaipova, A.N. (1971). On the Numerical Solution of Some Sets of Differential
Equations for Problems Stefan's Type, USSR Computational Math and Math Physics, 11(1): 162-170.
7. Baklanovskaya, V.F. (1962). The Numerical Solution of a Nonstationary Filtration Problem, USSR
Computational Math and Math Physics, 1(1): 114-122.
8. Budak, B.M., Uspenskii, A.B. (1969). A Difference Method with Front Straightening for Solving Stefan
Type Problems, USSR Computational Math and Math Physics, 9(6): 83-103.
9. Rasulov, M. A. (1992). A Numerical Method of Solving a Parabolic Equation with Degeneration, Dif.
Equations, 18(8): 1418-1427.
10. Sinsoysal, B. (2010). A New Numerical Method for Stefan-Type Problems in a Class of Unsmooth
Functions, Int. J. Contemp. Math. Sciences, 5(27): 1323-1335.
1172
Abstract. This report presents a test of the Babinet's principle in the realm of quantum mechanics. The
principle states that complementary objects in configuration space will end up having the same diffraction
pattern. The same pattern can be obtained by using first principles calculation based on Huygens' principle,
too. Yet there is no obvious reason why these principles must also be valid for material points as well, since
both of the aferomentioned principles were set for optics. In this work we calculate the numerically exact
scattering solutions of the time dependent Schrdinger equation on a two dimensional grid. Although, there
seems to be no enforing condition between Huygens' principle and first principles of quantum mechanics
similar results are obtained.
1 Introduction
To this day in optics the calculation of diffraction patterns of two dimensional objects have been carried out
by means of Rayleigh Sommerfeld formulation of diffraction integrals or Huygens-Fresnel-Kirchoff integrals.
The latter stem from Huygens' principle. This principle is axiomatic in nature in the same sense as the Hamilton's
principle of analytical mechanics. It is only natural to ask whether the implications of the same principle hold for
the wave mechanics of the quantum world where the essential actors are electromagnetic waves but wavicles
(wave particles) that are subject to the time dependent Schrdinger equation. In passing, we also note that
Huygens' principle is older than Maxwell's and Schrdinger's equation.
When Fresnel applied Huygens' principle [1] to the diffraction of light by two dimensional apertures,
Poisson, a strong advocate of the Newtonian corpuscle theory, was opposed to Fresnel's results. He applied the
same method to a circular obstacle in lieu of a circular aperture and found out that there must be a bright spot
right behind the obstacle, a seemingly unreasonable phenomenon. Poisson used this result as a mockery of and
objection to Fresnel's theory. Later, Arago has performed the experiment proposed by Poisson and confirmed the
presence of a bright spot as predicted by Fresnel's theory. Due to this ironic history, Poisson's name is coined
with the counter intuitive bright spot. (Also referred as Poisson Arago spot.) Poisson's spot [2] can also be
regarded as the result of a different principle in physical optics, known as Babinet's principle [3-5]. The
applications [6-10] of these principles in optoelectronics is ubiquitous. However, there seems to be a lack of
quantum mechanical computations for the general theory of a Poisson spot with material points although some
experiments with molecules have been performed to prove the existence of similar structures.
1173
In this paper we demonstrate the presence of a Poisson spot through first principles computations of
quantum dynamics. To this end a Gaussian wave packet is scattered from an almost hard scatterer with finite
width and height. For the scattered waves we have calculated the integrated probability flux density across a line
parallel to and far away from the obstacle. Section (2) describes the numerical methods that are applied. Section
(3) presents our results and concludes.
2 Computational Details
The time dependent Schrdinger equation (TDSE) is the equation of motion of quantum mechanics.
f x, y,t
= Hf x, y,t
t
(1)
Here f is the complex valued wave function and H is the Hamiltonian. It consists of the kinetic and -possibly
time dependent- potential energy operators.
H := T + V =
2 2
+ V x, y
2M
(2)
The Laplacian consists of second order partial derivatives with respect to the x, y dimensions. The potential
energy part of the Hamiltonian for this problem is a hard scatterer. However, we employ a potential energy
function of a short range nature with tunable stiffness and height parameters. It is defined as follows.
V x, y =: g x; a x ,bx g y; a y ,by
(3)
where
b
b
g x; a,b := aexp
+
, x x1 , x2 .
2
2
x x1 x x2
(4)
The function g vanishes outside of the interval in equation (4). It has the merit that not only the function itself,
but also all of its derivatives vanish at the boundaries. Therefore the scatterer is truly a short range potential
without any tails whatsoever. (This aspect is to be contrasted with the reciprocal cosh-square functions as
candidates of scatterers.) By varying the positive parameters a, b one can tune the height and the stiffness of the
potential, respectively.
Although, the unitary time development of the state is taken care of by the propagator
U t := exp itH / or the time evolution operator through f x, y,t = U t f x, y,0 , the simultaneous
presence of the kinetic and potential energy terms in the Hamiltonian makes the evolution step difficult due to
the second order partial derivatives in the kinetic energy term. To this end one splits [11] the propagator as
follows.
(5)
The error term depends on the commutators of the kinetic and potential energy operators. Eq. (5), by splitting the
propagator, offers a straightforward evolution algorithm: (1) Given a state f x, y, t one first applies the half
potential energy operator where it is already diagonal in the coordinate space. (2) Then the resulting function is
Fourier transformed [12] to the momentum space where the kinetic energy is diagonal thereby alleviating the
difficulty of applying an exponential differential operator. (3) Finally, an inverse Fourier transform back to the
coordinate space followed by the remaining half of the potential energy operator yields the state
f x, y,t + dt . This paradigm is unconditionally stable and preserves the unitary nature of the time evolution.
1174
A common difficulty encountered in the scattering computations with free boundaries is the problem of
spurious behavior of the wave packet when it approaches a boundary. For instance, when the scattering
computations are run in a box, after a certain amount of time the wave packet reaches one of the boundaries of
the box and gets reflected. There are two remedies suggested in the literature: One can either (1) map the
infinite interval through a logarithmic coordinate transformation to a finite domain, or (2) add a negative
imaginary absorbing potential around the boundaries. In our calculations we have employed the absorbing
boundary conditions by adding an ad hoc negative imaginary potential close to the proximity of the boundaries.
This method treats the grid spacings homogeneously, a feature that is lacked in the coordinate transformation
alternative. Neuhasuer and Baer [13] have given a semi heuristic choice for the functional form of the absorber.
Per their suggestion we use a ramp like function whose slope is determined to maximize absorption and
minimize reflection. (See also the work by Arnold [14].)
The goal of this paper, computation of the Poisson spot, is calculated by the time integral of the probability
current density j x, y, t across a line far from the scattering potential.
Px y =lim
o
(6)
j xo , y, s xds
where x is the unit vector in the positive x direction and the probability current density is defined as
j x, y,t :=
fgradf fgradf
2Mi
(7)
Fig. 1. Integrated probability current density clearly demonstrates the presence of a Poisson spot right
behind the center of the obstacle. It is located at 10 au
1175
We present the integrated probability current density in Fig. (1). The screen was put at x=14.5 au axis. Right at
the projection of the center of the obstacle (the y=10 au point), one observes the presence of a Poisson spot.
Displayed in Fig. (2) are the snapshots of the simulation. t=0 au corresponds to the beginning of the simulation.
The initial wave packet manifests the spherical symmetry of its Gaussian envelope in probability density,
whereas real and imaginary parts reveal the plane wave nature. t=2000 au is a typical moment of scattering
where the probability density is squeezed and real and imaginary parts turn into spherical waves. A weak
Poisson spot at this moment is apparent only after a close inspection. At t=4980 au scattering is over and plane
waves are now completely turned into spherical ones. A bright Poisson spot in this case is evident. Also note the
ripples due to the absorbing potential.
References
1.
2.
3.
4.
5.
Goodman, J. W. 2005. Introduction to Fourier Optics, Greenwood Village CO: Roberts & Company
Publishers.
Lucke, L. R. (2006). Rayleigh-Sommerfeld diffraction and Poisson's spot, Eur. J. Phys. 27(2): 193-204.
Balanis, C. A. 1989. Advanced Engineering Electromagnetics, New York: John Wiley and Sons.
Nussbaum, A. 1998. Optical System Design, NJ: Prentice Hall PTR.
Lipson, A.; Lipson, S. G.; and Lipson, H. 2011. Optical Physics, Cambridge: Cambridge University
Press.
1176
6.
7.
8.
9.
10.
11.
12.
13.
14.
15.
Juffmann, T.; Nimmrichter, S.; Arndt, M.; Gleiter, H.; and Hornberger, K. (2010). New prospects for de
Broglie interferometry Grating diffraction in the far field and Poissons spot in the near field. Found.
Phys. 41(1): 1-13.
Wang, P.; Xu, Y.; Wang, W.; and Wang, Z. (1998). Analytic expression for Fresnel diffraction,
J.Opt.Soc.Am. A 15(3): 684-688.
Frances, J.; Neipp, C.; Perez-Molina, M.; and Belendez, A. (2010). Rigorous interference and
diffraction analysis of diffractive optical elements using finite-difference time-domain method, Comp.
Phys. Commun. 181(2): 1963-1973.
Born, M.; and Wolf, E. 1999. Principles of Optics, Cambridge: Cambridge University Press.
Martinez-Anton, J. C.; Serroukh, I.; and Bernabeu, E. (2001). On Babinets principle and a diffractioninterferometric technique to determine the diameter of cylindrical wires, Metrologia 38(2): 125-134.
Herman, M. R.; and Fleck, J. A., Jr. (1988). Split-operator spectral methods for solving the timedependent Schrdinger equation in spherical coordinates, Phys. Rev. A 38(12): 6000-6012.
Gottlieb, D.; and Orszag, S.A. 1993. Numerical Analysis of Spectral Methods: Theory and Applications,
Pennsylvania: SIAM.
Neuhasuer, D.; and Baer, M. (1988). The time-dependent Schrdinger equation: Application of
absorbing boundary conditions, J. Chem. Phys. 90(8): 4351-4356.
Arnold, A. (1998). Numerically absorbing boundary conditions for quantum evolution equations, VLSI
Design, 6(1-4): 313-319.
Reisinger, T.; Patel, A. A.; Reingruber, H.; Fladischer, K,; Ernst, W. E.; Bracco, G.; Smith, H. I.; and
Holst, B. (2009). Poissons spot with molecules, Phys. Rev. A, 79(5): 053823(1-4).
Biographies
Mustafa Demirplak A native of mid Anatolian town Konya, Mustafa Demirplak was born 1978. He received
his BS in chemistry from Bilkent University in Ankara in 2000. He received his MS (2001) and PhD (2005)
degrees from The University of Chicago. His thesis was on the aspects of adiabatic population transfer and
control. Currently, he is a faculty member of the Department of Chemistry at Fatih University.
Previously, he has published on the dephasing statistics of diatomic systems in dense media and adiabatic
quantum control of simple systems where he introduced the counter diabatic field paradigm. His current research
interests include the rigorous examination of chemical equilibrium via the tools of computational algebraic
geometry.
Osman alar Akn Born in 1971 in Mersin Turkey, Osman alar Akn received his B.Sc. in Physics from
Middle East Technical University in 1994. He received his M.Sc. from METU Physics Department in 1997 wit a
thesis on the design considerations of an IFEL accelerator. He received his Ph.D. degree from University of
North Texas in 2008 with his work on perturbation of renewal processes. Currently he is a faculty member in
Department of Physics at Fatih University.
1177
Abstract: In this study, we investigated the linear and nonlinear optical properties of 22 heterocycyclic aromatic amines
including aminopyridines, aminopyrimidines and aminopyrazines. The theoretical calculations were performed with the
different three hybrid density functional theories (DFT), BPV86, B3PW91 and B3LYP, and used the 6-311++G(3d, 3p) basis
set. The results show that 2-amino-5-nitropyridine and 2-amino-4methyl-5-nitropyridine compounds exhibit large nonlinear
optical properties.We also consider semi-empirical polarizability and molecular volume calculations at the AM1 level of
theory together with QSAR-quality empirical polarizability calculations. Least-squares correlations between the various sets
of results show that these less costly procedures are reliable predictors of <> for these molecules, but less reliable for the
larger molecules.
Keywords: Nonlinear optical properties, aminopyridines, hyperpolarizability, B3LYP
1 Introduction
The search for nonlinear optical (NLO) materials has been of great interest in recent years because of their
potential applications in laser spectroscopy and laser processing [1], [2], optical communications, data storage
and image processing [3], [4], [5], [6] and terahertz (THz) wave generation technology[7] which use in the fields
of semiconductor, tomographic imaging, label free genetic analysis, cellular level imaging, biological sensing
and so on [8]. Organic materials are optically more nonlinear than inorganic materials due to weak Van der walls
and hydrogen bonds which posses a high degree of delocalization. Amino acids and their complexes belong to a
family of organic materials that have applications in NLO [9], [10], [11] Aminopyridines are bioactive Nheterocyclic tertiary amines, which increase the strength of the nerve signal by blocking of the voltagedependent K+ channel [12]. Also, aminopyridines have been proposed as drugs for the treatment of many
diseases such as antithrombus drugs and antimicrobial agents [13]. In particular, 2-aminopyridine is one of the
potential impurities in piroxicam and teroxicam which are non-steroided antiflammatory drugs that used in
musculo-skeletal and joint disorders [14]. Moreover, aminopyridines are commonly present in synthetic and
natural products [15]. They form repeated moiety in many large molecules with interesting photophysical,
electrochemical and catalytic applications [16].
Amino acids are interesting materials for NLO applications as they contain a proton donor carboxyl acid (COO -)
group and the proton acceptor amine (NH2) group with them [8], [17]. 2-Aminopyridines are promising
substituted pyridines which have been shown to be biologically active molecules [18]. Additionally, because of
their chelating abilities, 2-aminopyridines are commonly used as ligands in inorganic and organometallic
chemistry [19]. If substituted with optically active groups, they could potentially serve as chiral auxiliaries or
1178
chiral ligands in asymmetric reactions. For these reasons, 2-aminopyridines are valuable synthetic targets. 4aminopyridine is a voltage-gated K+ (Kv) channel blocker that is useful clinically in the treatment of spinal cord
injuries [20] and multiple sclerosis [21]. In addition to its therapeutic applications, 4-aminopyridine is routinely
used to isolate different types of K v channels expressed in native tissues based on their affinities for the drug
[22].
2 Computational details
The geometries were fully optimized without any constraint with the help of analytical gradient procedure
implemented within Gaussian 03W program [27]. All the parameters were allowed to relax and all the
calculations converged to an optimized geometry which corresponds to a true energy minimum revealed by the
lack of imaginary values in the frequency calculations.
1179
The performance of various DFT functionals and of basis sets in hyperpolarizability calculations have been
extensively studied for organic NLO materials [28], [29]. Accurate calculation of nonlinear optical properties
requires the use of extended basis sets and high level of theory. In particular, these basis sets have to include d
and p polarization functions together with s and p diffuse functions. In the present study, we utilized three hybrid
density functionals (B3LYP, BPV86, B3PW91) methods [30], [31], [32], [33] with 6-311++G(3d,3p) basis set
[34], [35] to estimate molecular static polarizability, anisotropy of polarizability and first static
hyperpolarizability () of 22 heterocyclic aromatic amines including aminopyridines, amiminopyrimidines
aminopyrazines. The first static polarizability () calculated analytically, whereas first static hyperpolarizability
() has been computed by using a finite-field procedure based on the numerical differentiation twice with respect
to electric field.
4.
5.
6.
When the H atom on the 2-aminopyridine is replaced by methyl group (at the R1, R2, R3 and R4 position)
hyperpolarizability of 2, 3, and 4 increased approximately four fold.
When the H atom on the 2-aminopyridine is replaced by NO2 group at the R1 position, hyperpolarizability
of 5 increased approximately 13, 16, 19 fold with B3PW91, BPV86, and B3LYP methods, respectively.
When the H atom on the 2-aminopyridine is replaced by NO 2 group at the R3 position, hyperpolarizability of
6 increased dramatically. Hyperpolarizability value of 2-aminopyridinechanged from 20.16, 28.18, and
22.08 a.u for 6 to 1313.65, 1242.88 and 1345.29 a.u. with BPV86, B3PW91 and B3LYP methods,
respectively.
It can be seen from Table 1, the largest hyperpolarizability are obtained for 6, 8, 13 and 22 among the title
molecules.
A large difference among the hyperpolarizability of aminopyrimidines has not been seen.
As can be seen from Table 1, hyperpolarizability values of aminopyrimidines and aminopyrazines are very
close to each other, but for 22 the hyperpolarizability increased dramatically due to the H atom on the
aminopyrazine is replaced by phenyl group at the R2 position.
1180
Table 1. Static polarizability, first static hyperpolarizability and molecular volume of studied molecules.
Mol.
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
<>/a.u.
AM1
51.74
61.15
63.15
64.43
72.23
74.04
70.92
72.32
81.58
59.71
61.83
72.58
82.87
50.90
59.62
69.07
64.32
74.58
51.70
68.58
108.83
114.35
BPV86
75.55
89.46
92.76
93.05
99.74
103.30
98.17
104.21
118.19
84.91
93.86
107.60
116.71
72.77
85.70
102.34
93.97
108.27
73.70
97.11
151.90
159.28
B3PW91
73.17
86.45
89.37
89.67
96.41
98.94
97.30
99.66
112.26
81.58
90.41
103.45
111.80
70.22
81.67
98.16
89.86
103.52
71.14
93.34
146.45
152.83
B3LYP
73.90
87.20
90.16
90.41
97.24
99.75
98.17
100.51
113.11
82.37
91.24
104.29
112.60
70.92
82.40
98.96
90.55
104.31
71.83
94.15
147.68
154.00
tot/ a.u.
BPV86
20.16
94.01
242.06
229.69
325.68
1313.65
242.80
844.87
528.13
363.47
283.38
184.09
1128.71
230.91
224.96
190.00
252.41
307.72
219.04
184.81
254.80
845.97
B3PW91
28.18
102.04
256.24
244.65
379.50
1242.88
239.66
780.59
673.70
386.59
297.84
206.89
1088.72
247.93
177.94
202.11
234.37
285.05
241.09
216.48
264.55
769.56
B3LYP
22.08
98.00
265.35
244.73
420.40
1345.29
258.61
840.50
705.62
370.82
284.39
209.90
1183.02
238.66
183.78
213.24
238.50
295.11
230.14
230.48
271.32
734.17
Vol.(3)
354.8
404.82
405.08
405.26
403.83
411.38
420.26
426.50
468.12
371.26
392.18
460.07
456.38
337.74
380.33.
446.62
413.70
471.51
338.24
407.86
545.55
551.15
1181
4. Conclusion
In the present study, linear and nonlinear optical properties of aminopyridines, aminopyrimidines and
aminopyrazines calculated with different DFT methods. Comparison of the calculated and previous semiemprical and HF datas for 4-aminopyridine indicates that the semiempirical PM3 values are overestimated, and
B3LYP hyperpolarizability result is very compatible with the ab initio STO-3G level value for 4-aminopyridine.
The largest hyperpolarizability values obtained for 6, 8, 13 and 22 among the studied molecules. This study
reveals that these molecular systems(6, 8, 13 and 22) have large first static hyperpolarizabilities and may have
potential applications in the development of NLO materials.
We consider the likely reliability of various easily computed indices such as the molecular volume and AM1
polarizabilities discussed above. AM1 polarizability calculations are in poor quantitative agreement with the
B3LYP results. Nevertheless, AM1 average polarizability gives a correlation coefficent of 0.99 when compared
to the B3LYP results.
References
1. D.S. Chemla, J. Zyss (Eds.), Nonlinear Optical Properties of Organic Molecules and Crystals, Academic Press,
Orlando, (1987).
2. Y. Shen, The Principles of Nonlinear Optics, J. Wiley, New York,(1984).
3. P.N. Parasad, D.J. Williams, Introduction to Nonlinear Optical Effects in Molecules and Polymers, JohnWiley &
Sons, New York, (1991).
4. H.S. Nalwa, S. Miyata (Eds.), Nonlinear Optics of Organic Molecules and Polymers, CRC Press, Boca Raton, FL,
(1997).
5. S.R. Marder, B. Kippelen, A.K.-Y. Jen, N. Peyghambarian, Design and synthesis of chromophores for electro-optic
and photorefractive applications, Nature 388, 845851, (1997).
6. Y. Shi, C. Zhang, J.H. Bechtel, L.R. Dalton, B.H. Robinson, W.H. Steier, Low (Sub-1-Volt) Halfwave Voltage
Polymeric Electro-optic Modulators Achieved by Controlling Chromophore Shape, Science 288, 119122, (2000).
7. V. Krishnakumar, R. Nagalakshmi, Studies on the first-order hyperpolarizability and terahertz generation in 3nitroaniline, Physica B 403, 18631869, (2008).
8. G.Ramesh Kumar, S.Gokul Raj, R. Mohan, R. Jayavel, Influence of Isoelectric pH on the Growth Linear and
Nonlinear Optical and Dielectric Properties of L-tThreonine Single Crystals, Crystal Growth Design, 6, 1308-1310,
(2006).
9. M. Kitazawa, R. Higuchi, M. Takahashi, Ultraviolet generation at 266 nm in a novel organic nonlinear optical
crystal: l-pyrrolidone-2-carboxylic acid, App. Phys. Lett. 64, 2477-2480, (1994).
10.L. Misogati, A.T.Varela, F.D. Nunes,V.S. Bagnato, F.E.A. Melo, J. Mendes Filho, C.
Zilio, Optical properties of L-alanine crystals, Optical Materials, 6 ,147-152, (1996).
11.W.S. Wang, M.D. Aggarwal, J. Choi, T. Gebre, A.D. Shields, B.G. Penn, D.O. Frazier, Solvent effects and
polymorphic transformation of organic nonlinear optical crystal-pyroglutamic acid in solution growth processes,
J.Cryst. Growth 198, 578 -582, (1999).
12. C. Carlsoom, I. Rosen, E. Nilsson, Can 4-aminopyridine be used tor everse anaesthesia and muscule relaxation?,
Acta Anaesthesiologica Scandinavica. 27, 87-90, (1993).
13.J.I. Segal, B.S. Brunnemann, 4-aminopyridine improves plumonary function in quadriplegic humans with
longstanding spinal cord injury, Pharmacotherapy 17, 415-423, (1997).
14.K. Parfitt (Ed.), Martindale, The Extra Pharmacopoeia, 32nd ed., Pharmaceutical
Press, London, , 48-49, (1999).
15.A. Goel, V.J. Ram, Natural and synthetic 2H-pyran-2-ones and their versatility in organic synthesis, Tetrahydron
65, 7865-7913, (2009).
16.N.A. Al-Hashimy, Y.A. Hussein, Ab initio study on the formation of triiodide CT complex from the reaction of
iodine with 2,3-diaminopyridine, Spectrochimica Acta, Part A 75, 198-202, (2010).
17.M.B. Johnston, L.M. Herz, et al., Low-energy vibrational modes in phenylene oligomers studied by THz timedomain spectroscopy, Chemical Physics Letters. 377, 256-262, (2003).
18.F. Manna, F. Chimenti, A. Bolasco, B. Bizzarri, W. Filippelli, A. Filippelli, L. Gagliardi, Anti-inflammatory,
analgesic and antipyretic 4,6-disubstituted 3-cyano-2-aminopyridines, European J. Med.Chem. 34, 245 -254,
(1999).
1182
Hamit Alyar-Hatay, 1971. Hereceived the B.S. degree in physics from the Atatrk University, Erzurum,
Turkey, 1992, and the M.S. and Ph.D. degrees in physics from the Gazi University, Ankara, Turkey, in 1999 and
2005.
1183
Abstract.In this study, reflection from through single-negative metamaterial slab waveguides is presented.
Reflections versus wavelength are plotted for different number of bilayer. It is shown that reflections
approximately are reached hundred percent with increasing bilayer number.
Keywords: reflection, slab waveguides, metamaterials
1 Introduction
Metamaterials are engineered materials that are not found naturally. Due to unique properties of these
materials, those materials are attracted for scientists especially physicists and electronic engineers. In the last
decade, single-negative metamaterials with negative electrical permittivity produced artificially are important in
optical frequencies. Recently, a considerable amount of work has been reported on metamaterials [1-7].
In 1968, the Russian physicist Victor Veselago is proposed the existence of electromagnetic materials
called metamaterials with negative permittivity and permeability [8]. While single-negative metamaterials are
designed with a negative permittivity or permeability, which are called SNG, double-negative metamaterials are
designed to have both negative permittivity and permeability, which are called DNG.
Metamaterials are produced periodically with periods much less than a wavelength. Negative
permittivity or/and permeability could cause electromagnetic waves travelling in these media to exhibit unusual
properties which are interested by scientists. The waves travelling in these media are refracted at negative angles
according to conventional media. However, due to negative refraction index, the direction of phase and group
velocity is opposite. These unusual characteristics are used for designing new devices such as optical filter, antireflection and high reflection coatings.
Metamaterials are periodic structures which designed to control of propagation of electromagnetic
waves. Therefore, metamaterials have the potential to manipulate the wave propagation in a manner that eludes
the conventional materials due to periodic nature that can either be small-scale or resonant [9-17]. The practical
applications of metamaterials are currently limited to their operational bandwidths.
2 Theory
Metamaterials have several unusual properties with compared to the regular materials. These unusual
properties can be expressed as follow. Phase velocity of electromagnetic waves is anti parallel to the group
velocity and flow of energy while Poynting vector and group velocity are parallel for conventional materials.
Thus, propagation is described by a left hand rule. Waves are refracted at a negative angle relative to the
conventional materials. Evanescent waves grow with distance into the medium.
1184
For a DNG material index of refraction for real permittivity and permeability is n r r and the
intrinsic impedance is n r r .
These periodic structures are produced with two different refractive indexes. One of dielectric material
has low refractive index and the other has high refractive index. This structure is the simplest periodic structure
provided that [18]
n a x a 2
n x 2 1
(1)
n1 a 2 x d
where a1 and a2 are the thicknesses of layer one and two, respectively and d is the sum of a 1 and a2. Schematic
representation of the slab is shown in Figure 1. The refractive index of 1D medium can be written as [19]
n x n x d
(2)
where is an integer number as known period number and d is periodic length in x-direction and called the
lattice constant of the structure.
Er
(3)
2
Ei
where Er is the amplitude of reflected wave and Ei is the incident wave that is thorough interfaces of media.
1185
1186
From Figure 3, the reflection for dielectric/dielectric bilayer is not changed too much. But for
dielectric/metamaterial bilayer, the reflection value of 100 % is reached all the wavelength range of between
300 and 800 nm.
4 Concluding Remarks
In this paper, the reflection coefficients of multilayered structures including both metamaterial and
dielectric layers are studied. We study first a structure consisting of a metamaterial and dielectric slab, second a
structure consisting of dielectric and dielectric slab. It has been found that the metamaterial dielectric structures
have very interesting features. In conclusion, dielectric/metamaterial bilayers can be used for antireflecting
coating in different areas of scientific applications.
References
1. V. Veselago, L. Braginsky, V. Shklover, and C. Hafner, Negative Refractive Index Materials, Journal of
Computational and Theoretical Nanoscience, Vol. 3, No. 2, 189-218, 2006.
2. W. J. Padilla, D. N. Basov, D. R. Smith, Negative Refractive Index Metamaterials, Materials Today, Vol. 9,
No. 7-8, 2006.
3. A. Boltasseva, V. M. Shalaev, Fabrication of Optical Negative-Index Metamaterials: Recent Advances and
Outlook, Metamaterials, Vol.2, No. 1, 1-17, 2008.
4. R. B. Greegor, C. G. Parazzoli, K. Li, B. E .C. Koltenbah and M. Tanielian, Experimental Determination and
Numerical Simulation of the Properties of Negative Index of Refraction Materials, Optics Express, Vol. 11, No.
7, 688-695, 2003.
5. A. Diaz, J. H. Park, and I. C. Khoo, Design and Transmission-Reflection Properties of Liquid Crystalline
Optical Metamaterials with Large Birefringence and Sub-Unity or Negative-Refractive Index, Journal of
Nonlinear Optical Physics & Materials Vol. 16, No. 4, 533549, 2007.
1187
6. G. X. Yu, Y. T. Fang, T. J. Cui, Goos-Hanchen Shift from and Anisotropic Metamaterial Slab, Central
European Journal of Physics, Vol. 8, No. 3, 415-421, 2010.
7. N. C. Panoiu, R. M. Osgood Jr., Numerical Investigation of Negative Refractive Index Metamaterials at
Infrared and Optical Frequencies, Optics Communications, Vol. 223, No. 4-6, 331-337, 2003.
Soviet
Physics Uspekhi, Vol. 10, No. 4, 509-514, 1968.
9. P. Kolinko and D. R. Smith, Numerical Study of Electromagnetic Waves Interacting with Negative Index
Materials, Optics Express, Vol. 11, No. 7, 640-648, 2003.
10. J. Li, Y. Chen and V. Elander, Mathematical and Numerical Study of Wave Propagation in Negative-Index
Materials, Computer Methods in Applied Mechanics and Engineering, Vol.197, No. 45-48, 3976-3987, 2008.
11. S.M. Vukovic, N. B. Aleksic, D. V. Timotijevic, Guided Modes in Left-Handed Waveguides, Optics
Communications, Vol. 281, No. 6, 1500-1509, 2008.
12. Y. He, Z. Cao and Q. Shen, Guided Optical Modes in Asymmetric Left-Handed Waveguides, Optics
Communications, Vol. 245, No. 1-6, 125-135, 2005.
13. M. Cheng, Y. Zhou, S. Feng, J. Lin and R. Chen, Lowest Oscillating Mode in a Nanoscale Planar Waveguide
with Double-Negative Material, Journal of Nanophotonics, Vol. 3,
No. 039504, 1-5, 2009.
14. Z. H. Wang, Z. Y. Xiao and S. P. Li, Guided Modes in Slab Waveguides with a Left Handed Material Cover
or Substrate, Optics Communications, Vol. 281, No. 4, 607-613, 2008.
15. P. Dong and H. W. YANG, Guided Modes in Slab Waveguides with Both Double-Negative and SingleNegative Materials, Optica Applicata, Vol. 40, No. 4, 873-882, 2010.
16. R. W. Ziolkowski and E. Heyman, Wave Propagation in Media having Negative Permittivity and
Permeability, Physical Review E, Vol. 64, No. 5, 1-15, 2001.
17. H. Coryand C. Zach, Wave Propagation in Metamaterial Multi-Layered Structures, Microwave and Optical
Technology Letters, Vol. 40, No. 6, 460-465, 2004.
18. A. Yariv, P. Yeh, Optical Waves in Crystals, John Wiley & Sons, 2003. 19. J. Zheng, Z. Ye, X. Wang, X., D.
Liu, Analytical Solution for Band-Gap Structures in Photonic Crystal with Sinusoidal Period, Physics Letters A,
Vol. 321, No. 2, 120126, 2004.
1188
Abstract. In quantum information processing, spin-3/2 electron or nuclear spin states are known as two-qubit
states. For SI (S = 3/2, I = 1/2) spin system, there are 8 three-qubit states. In this study, first, matrix
representation of three-qubit Hadamard gate applied to 8 three-qubit states in order to obtain the three-qubit
superpositions. Then it is shown that three-qubit superposition states can be also found by using the magnetic
resonance pulse sequence of Hadamard gate.
1 Introduction
In classical computers, the information is stored as bits. In quantum computers a unit of information is called
quantum bits (qubits)[1]. Qubits can be represented by the states of any quantum system such as spin states of
nuclei or electrons. In magnetic resonance quantum computers, Zeeman levels of a spin1/2 nucleus or electron,
| and | are called a single qubit[2]. Usually, these two states are shown as | and | , respectively. Some
linear superpositions of them are also called a single qubit. One of the aims of quantum information theory is to
develop universal computer. To perform this, it uses quantum mechanical principles of physics. Unlike classical
computer, quantum computer performs simulation of physics very well [3]. Quantum information theory can be
implemented in practice by using spectroscopic methods such as NMR and ENDOR. Quantum logic gates can be
implemented with NMR(Nuclear Magnetic Resonance) or ENDOR(Electron-Nuclear Double Resonance)
spectroscopy by using selective and non-selective pulse techniques. These pulses provide transition between
energy levels. Two qubit(S=1/2 I=1/2 spin system) Hadamard gate implemented with pulse NMR techniques
using . / ( ) pulse sequences for each spin [4]. In this study, by using NMR or ENDOR, we aim to find three
qubit Hadamard gate pulse sequence which is not found in the literature. The Hadamard gate is very important
gate for quantum information theory. It makes superpositios states which have equal probability of all quantum
states. Because of its properties, it is used by quantum circuits and algorithms such as Entanglement states,
Grovers search algorithm, Shors factoring algorithm and Deutsch-Jozsa algorithm.
2 Theory
In quantum computers, quantum logic gates are represented by the unitary matrices. Matrix representation of
single qubit Hadamard (H) gate is
1189
(1)
When this Hadamard gate is applied to onequbit states, it creates the superposition states:
|
(| | ) |
(| | )
(2)
|
(3)
possible states:
(
) .
Where
Endohedral fullerene molecule, 31P@C60, can be considered as SI (S=3/2, I=1/2) spin system[5]. In this
molecule electron spin of 31P in the ground state is 3/2 and nuclear spin of 31P is 1/2 with the abundance of
| ,
| ,
|
100%. For
electron spin quantum number, twoqubit states are
| as there are four magnetic quantum numbers.
and
(
) spin system can be
|
|
|
|
considered as three qubit states. For this spin system eight three qubit states |
|
|
|
are found by direct products of two-qubit and one-qubit states. For example |
state
is obtained as following:
|
( )
( )
(4)
( )
31
)(
Table 1. Energies of the spin Hamiltonian and three qubit states for SI(S=3/2, I=1/2) spin
system
Three Qubit State
Energy
|
Time dependence of the density matrix is governed by Liouvillevon Neumann equation [2]
,
(6)
-is the commutator of the density matrix, , and the Hamiltonian,
Where ,
independent, the solution of the Eq.(6) is
( )
- ( )
-(7)
1190
Where H is the total Hamiltonian, which consists of radio frequency and microwave pulse, chemical shift and
spinspin coupling Hamiltonians. For electron S=3/2 and nucleus I=1/2 spins, some quantum logic gates should
be presented by selective magnetic resonance pulses. Modified matrix representation of a selective pulse
operator for S=3/2 is as follows [6];
. /
. /
( )
. /
. /
(8)
Where =0, 1, 2 and 3 for x, y, x and y, respectively. Single, double and triple quantum transition selective
( ),
( ), ( ), and
( ) can be also obtained. They will be used in threequbit
pulse operators of
Hadamard gate in Section 3.
(9)
,|
| -
(10)
Is found.
| |
| |
| |
| |
| |
| |
| |
and
Where
| |
.
By applying the same Hadamard gate, the superpositions of the other 7 threequbit states can be obtained. All
the superpositions of threequbit states are presented in Table 2.
Pseudo-Pure
State
|
),|
),|
),|
),|
),|
),|
),|
),|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1191
In quantum computing systems, the initial state must be pseudo-pure state. There are several methods for
pseudopreparing the pseudo-pure state such as thermal equilibrium density matrix [7]. Density matrix of |
pure state can be writen as
|
(11)
Off-diagonal elements of density matrix are equal to zero. The pulse sequence of there qubit Hadamard gate can
be represented as
. /
| |
. /
| |
. /
| |
. /
| |
. /
| |
. /
| |
( )|
. /
| |
(12)
Where | | represent the transitions from n to m energy levels [8 ].All pulses are applied along x direction
and each pulse can be obtained using its mathematical equations such as
( )
( )| |
( )
(
)
(13)
When Hadamard gate pulse sequence applied to pseudo-pure state of |
,
(14)
(15)
(16)
It is shown that superposition states of three-qubit states can be obtained by using suitable Hadamard pulse
sequences and the same superposition states can be ob-tained with those found by matrix representation of
Hadamard logic gate as given in Table 1.This result can be used in the first step of three qubit Grovers search
algorithm[9].
References
1.
2.
3.
4.
5.
6.
7.
Nielsen, M.A. , Chuang, I.L. (2001). Quantum Computation and Quantum Information, United
Kingdom, Cambridge University Press.
Oliveira, I.S. , Bonagamba, T.J. , Sarthour, R.S. , Freitas, J.C.C. and deAzevede, E.R. (2007). NMR
Quantum InformationProcessing, Netherlands, Elsevier.
Feynman ,Richard P. (1982). Simulating Physics with Computers.,International Journal of Theoretical
Physics 21 467.
Das, R. , Mahesh, T.S and Kumar, A. (2002). Experimental implementation of Grovers search
algorithm using efficient quantum state tomography, Chem. Phys. Lett. 369 8.
Harneit, W. (2002). Fullerene-based electron-spin quantum computer,Phys. Rev. B 65 032322.
Bonk, F.A. , deAzevedo, E.R. , Sarthour, R.S. , Bulnes, J.D. , Freitas, J.C.C. , Guimaraes, A.P. ,
Oliveira, I.S. and Bonagamba, T.J. (2005). Journal of Magnetic Resonance 175 226.
Naydenov, B. , Mende, J. , Harneit, W. , and Mehring, M. (2008). Entanglement in P@C60
encapsulated in a solid state matrix. Phys. Stat. Sol.(b) 245 No.10
1192
8.
9.
Scherer, W. And Mehring, M. (2008).Entangled electron and nuclear spin states in 15N@C60: Density
matrix tomography. Journal of Chemical Physics 128 052305.
Grover , L.K. (1997). Quantum Mechanics Helps in Searching for a Needle in a Haystack. Phys. Rev.
Lett. 79 325.
Biographies
Seluk akmak He was born in Ordu (Turkey) in 1986. He received his B.Sc. in Physics at Ondokuz Mays
University (Turkey) in 2009. He is an M.Sc. student in Physics at Ondokuz May University (Turkey). His
research interests are quantum information and computation theory.
Sevcan orbac She was born in Hatay (Turkey) in 1984. He received her B.Sc. in Physics at Ondokuz Mays
University (Turkey) in 2008. She is an M.Sc. student in Physics at Ondokuz May University (Turkey). Her
research interests are quantum information and computation theory.
Azmi Genten He was born in Samsun (Turkey) in 1965. He received a B.Sc. in Physics at Ondokuz Mays
University (Turkey) in 1986 and a Ph.D. in Physics with Ray Dupree at Warwick University (UK) in 1993. His
major field of study is magnetic resonance. His previous publications are related to product operator theory in
NMR and magnetic resonance quantum information theory. Current research interest involves magnetic
resonance quantum information theory. Previous research interests are NMR of High TC superconductors and
product operator theory in NMR.
1193
Abstract. High Dimensional Model Representation (HDMR) is a divide-and-conquer method and is used to
represent a multivariate function in terms of less-variate functions in order to reduce the complexity of the
scientific computations in computer based applications. Multivariate data modelling problem is one of the
research areas in which it becomes so complicated to determine analytical models through the standard
interpolation methods while the multivariance of the problem increases. The data modelling problems of real
life usually have a multivariate training data set of which the nodes are randomly distributed in the problem
domain. Generalized HDMR method is one that can partition such training data set and allow us to construct
an analytical structure for the given data modelling problem. However, the method includes a linear equation
system to be solved for this purpose and this work aims to bypass this undesired structure and offers a
piecewise algorithm to determine the sought analytical structure of the given problem.
Keywords: HDMR, Generalized HDMR, data partitioning, interpolation, approximation
1 Introduction
Modelling multivariate data and estimating the unknown function values or unknown class values at the given
nodes of the problem domain are the common research areas of numerical methods, data mining and the other
related topics. However, dealing with multivariate data increases the mathematical and computational
complexity of the modelling problem. Hence, to partition the given multivariate data and to take for instance
only univariate or bivariate data structures into consideration in the modelling process reduce the complexity of
the problem and bring many advantages in model construction purposes. Divide-and-conquer algorithms can be
used in this manner and one of the most important methods in data partitioning concept is the High Dimensional
Model Representation (HDMR) method which was first proposed by Sobol[1]. This method expresses a
multivariate function in terms of less-variate functions through a finite sum and offers an algorithm to uniquely
determine the structures of these less-variate functions appearing in the finite expansion composed of
summations.
After Sobol, H. Rabitz and his group have developed several different HDMR based methods for different
areas of real life problems[2-4]. M. Demiralp and his group have developed many HDMR based methods for
different research areas[5-12].
The Generalized HDMR method was developed to model the multivariate data sets in which the function
values are known at arbitrarily distributed nodes of the problem domain[10]. The Generalized HDMR method
uses a general weight function instead of a product type as used in HDMR method to have the opportunity of
1194
partitioning that type of multivariate data. General weight usage in data partitioning process through an HDMR
based method brings a linear equation system solution need. However, depending on the structure of the given
data, there sometimes exists no solution for this linear equation system. To bypass this disadvantage, this work
offers to split the given domain into subdomains and then to use at most the constant component in representing
each subdomain.
In this work, first the details of the HDMR method are given very briefly in the next section while the
Generalized HDMR method is described in the third section. The piecewise based Generalized HDMR method is
given in the fourth section. The fifth section covers the numerical implementations prepared to examine the
performance of our new method. Finally, the concluding remarks are given in the sixth section.
f ( x1 ,, x N ) f 0 f i1 ( xi1 ) f1N ( x1 ,, x N ) .
(2)
i1 1
The right hand side components of this expansion are mutually orthogonal in the Hilbert space of functions
and the following vanishing conditions are used to determine the structures of these components.
bN
b1
dx dx
1
a1
W ( x1 , x2 ,, x N ) f i ( xi ) 0 .
(2)
aN
The weight function appearing in these vanishing conditions is a product type function and a normalization
criterion is defined as
W ( x1 , x2 ,, x N ) W j ( x j ), x j a j , b j ,
j 1
bj
aj
dx j W j ( x j ) 1 . (3)
The main task in the HDMR algorithm is to determine the general structures of the HDMR components by
using these vanishing conditions. This work aims to use only the constant component. The following operator is
defined for the constant component
b1
bN
a1
aN
I 0 F ( x1 ,, x N ) dx1W1 ( x1 ) dx N WN ( x N ) F ( x1 ,, x N ) .
(4)
The higher variate terms can also be determined by defining similar operators.
d j (1( j ) ,, N( j ) , j ),
j f (1( j ) ,, N( j ) ),
1 j m.
(5)
The Generalized HDMR method uses a general weight function and this general weight function is selected as
the linear combination of Dirac delta functions as follows
1195
W ( x1 , x2 ,, x N ) j ( x1 1( j ) ) ( x N N( j ) )
(6)
j 1
where parameters are used for giving different importance to each individual datum[10]. To use the HDMR
philosophy to partition the given multivariate data under a general weight, we need a product type weight
function similar with the weight given in (3). For this purpose, the following auxiliary product type weight
function with a normalization criterion is defined
N
( x1 , x2 ,, x N )
( x j ), x j a j , b j ,
j 1
bj
aj
dx j
( x j ) 1 . (7)
The first step of the Generalized HDMR method is to determine the general structures of the HDMR
components of the general weight function under the auxiliary weight given in (7). The normalization criteria
given in (3) urges us that the general weight function should be chosen as normalized under the auxiliary weight
function, which implies
m
W0 I 0 W ( x1 ,, x N ) j
1,
j 1
( k( j ) )
(8)
k 1
according to the orthogonality conditions given in (2)[10]. The above relation gives a constraint on parameters
and shows that the constant HDMR component of the general weight is 1. To determine the general structure of
the constant Generalized HDMR component, the I 0 operator given in (4) is applied to both sides of the HDMR
expansion by using the HDMR components of the general weight. To this end, the constant component is
obtained as follows[10]
m
f 0 j
j 1
j .
(9)
(10)
where t is the total number of subintervals and c values with superscripts are the lower and upper bounds of
the subintervals. In this work, the interval splitting process is executed to obtain equivalent subintervals and it is
assumed that the number of subintervals is specified by the user. Then, a constant value through the Generalized
HDMR method is obtained for each subinterval as,
construct an analytical structure through the constant values obtained for each subinterval, standard interpolation
techniques such as Lagrange or Cubic Spline interpolations can be used. A possible way of obtaining the
analytical model of the given problem is to evaluate the midpoint of each interval and to apply the interpolation
formula to the set including the following ordered pairs
(1)
, f 0(1) , ( 2) , f 0( 2) , , (t ) , f 0(t ) .
where
(1)
(11)
1196
5 Numerical Implementations
In this section, a number of numerical implementations as the data modelling problems are given through a
number of testing functions. The scripts to obtain the numerical results are run in MuPAD. The plots are also
prepared by using MuPAD. The blue line in the plots presents the original function while the green one stands
for the constant approximation through our new method and is named as s(x) . The testing functions are
selected as follows
f 1 ( x) x 1 ,
2
f 2 ( x) e x ,
f 3 ( x) sinx .
(12)
where the functions have polynomial, exponential and trigonometric natures respectively. The performance of
the method increases as the number of subintervals increase. Here, we use 30 subintervals for each testing
function and Fig 1-3 are given for the testing functions resprectively. It is seen that even the constant component
of our new method can approximately represent the original function.
Fig. 1. Constant approximation for the first testing function with 30 subintervals
1197
Fig. 2. Constant approximation for the fourth testing function with 30 subintervals
Fig. 3. Constant approximation for the fifth testing function with 30 subintervals
5 Concluding Remarks
The Generalized HDMR method is used for multivariate data partitioning to model multivariate data of real life
applications. However, at most the univariate Generalized HDMR components can be taken into consideration
for the modelling process because of mathematical complexities arised in univariate component determination.
Solving a linear equation system causes these mentioned complexities since there sometimes exist no solutions.
In this work, to bypass the disadvantage of solving a linear equation system, we developed a piecewise based
Generalized HDMR algorithm in which we deal with only the constant component by splitting the problem
domain into subintervals for the univariate case. The determination of the constant term in each subinterval and
the interpolation process of these values give us an approximation for the given data modelling problem. The
numerical implementations show us that our new method works well for the univariate case. The performance
of these approximations obtained for various testing functions are given via a number of plots. The results are
very promising to get qualified approximations through constancy for the multivariate case as the future work.
1198
References
1.
2.
3.
4.
5.
6.
7.
8.
9.
10.
11.
12.
13.
14.
15.
16.
17.
18.
Sobol, I.M. (1993). Sensitivity estimates for nonlinear mathematical models. Mathematical Modelling
and Computational Experiments (MMCE) 1(4): 407-414.
Rabitz, H., Al, .F., Shorter, J., & Shim, K., (1999). Efficient input-output model representations.
Computer Phys. Comm. 117: 11-20.
Li, G., Rosenthal, C., & Rabitz, H., (2001). High Dimensional Model Representations. J. Phys. Chem. A
105: 7765-7777.
Li, G., Wang, S.-W., & Rabitz, H., (2002) Practical Approaches To Construct RS-HDMR Component
Functions. J. Phys. Chem. A 106: 8721-8733.
Demiralp, M. (2003). High dimensional model representation and its application varieties.
Mathematical Research 9: 146-159.
Tunga, B. & Demiralp, M. (2003). Hybrid high dimensional model representation approximats and their
utilization in applications. Mathematical Research 9: 438-446.
Tunga, B. & Demiralp, M. (2009). Constancy maximization based weight optimization in high
dimensional model representation. Numerical Algorithms 52: 435-459.
. Yaman, . & Demiralp, M. (2003). High dimensional model representation applications to exponential
matrix evaluation. Mathematical Research 9: 463-474.
Baykara, N.A., & Demiralp, M., (2003). Hyperspherical or Hyperellipsoidal Coordinates in the
Evaluation of High Dimensional Model Representation Approximants. Mathematical Research 9: 4862.
Tunga, M. A. & Demiralp, M. (2003). Data partitioning via generalized high dimensional model
representation (ghdmr) and multivariate interpolative applications. Mathematical Research 9: 447-462.
Tunga, M. A. & Demiralp, M. (2008). A new approach for data partitioning through high dimensional
model representation. Int. Journal of Computer Mathematics 85: 1779-1792.
Tunga, M. A. (2011). An Approximation Method to Model Multivariate Interpolation Problems:
Indexing HDMR. Mathematical and Computer Modelling 53: 1970-1982.
Rao, B. N. & Chowdhury, R. (2008). Probabilistic analysis using high dimensional model
representation and fast fourier transform. International Journal for Computational Methods in
Engineering Science & Mechanics 9: 342-357.
Chowdhury, R. & Rao, B. N. (2009). Hybrid high dimensional model representation for reliability
analysis. Computer Methods in Applied Mechanics & Engineering 198: 753-765.
Sridharan, J. & Chen, T. (2006).Modeling multiple input switching of CMOS gates in DSM technology
using HDMR. Proceedings of Design Automation and Test in Europe 1-3: 624-629.
Ziehn, T. & Tomlin, A. S. (2008). A global sensitivity study of sulfur chemistry in a premixed methane
flame model using HDMR. International Journal of Chemical Kinetics 40: 742-753.
Banerjee, I. & Ierapetritou, M. G. (2004). Model independent parametric decision making. Annals of
Operations Research 132: 135-155.
Banerjee, I. & Ierapetritou, M. G. (2002). Design optimization under parameter uncertainty for general
black-box models. Ind. Eng. Chem. Res. 41: 6687-6697.
Biographies
M. Alper TUNGA He was born in stanbul, Turkey on June 11, 1975. He earned his Bachelors, M.Sc. and
Ph.D degrees from stanbul Technical University, stanbul, Turkey. He is currently interested in multivariate
data modelling and classification problems. He has 6 journal papers in SCI, 4 journal papers not in SCI and 13
international conference proceedings. His research interests are numerical analysis, data mining and data
modelling. Dr. Tunga is the member of Group for Science and Methods of Computing in Institute of Informatics,
stanbul Technical University.
Metin DEMRALP Metin Demiralp was born in Turkey on 4 May 1948. He got his BS, MS, and PhD from
the same institution, Istanbul Technical University. He is working on methodology for computational sciences
through theoretical chemistry and applied mathematics. He has a group (Group for Science and Methods of
Computing) in Informatics Institute of IstanbulTechnical University. Metin Demiralp has roughly 80 papers in
well known scientific journals and is the full member of Turkish Academy of Sciences since 1994. He has also
two important awards of Turkish scientific establishments.
1199
Abstract. This paper deals with the solution of one dimensional heat equation by using collocation method with
cubic B-splines finite elements. The scheme of the method is presented and the stability analysis is investigated by
considering Fourier stability method. On the other hand, a comperative study between the numerical and the analytic
solution is illustrated by the figure and the tables. The results demonstrate the reliability and the efficiency of the
method.
Keywords: Cubic B-splines finite element method, Collocation method, Cubic B-splines, Finite Element Method.
1 Introduction
Consider the one dimensional initial-boundary value problem
(1)
uxx ut 0 , 0 x L , t 0
with initial condition
u( x,0) f ( x)
(2)
and boundary conditions
u(0, t ) u( L, t ) 0 for t 0 .
(3)
This problem is one of the well-known linear partial differential equation [1-3]. It can be expressed as the heat
flow in the rod with diffusion 2uxx along the rod where the coefficient is the thermal diffusivity of the rod
2
and L is the length of the rod [4]. In this model, the flow of the heat in one-dimension that is insulated
everywhere except at the two end points. Solutions of this equation are functions of the state along the rod and
the time t . In the past, this problem has been widely worked over a number of years by numerous authors. But it
is still an interesting problem since many physical phenomena can be formulated into PDEs with boundary
conditions.
The cubic B-splines collocation method was developed for Burgers equation and used for the numerical
solution of the differential equations in [5,6,7,8]. Recently, spline function theory has been extended and
developed to solve the differential equations numerically by various papers [9,10,11,12,13,14,15]. Furthermore
some extraordinary problems has been numerically investigated by finite element methods such as Galerkin
method, least square method and collocation method with quadratic, cubic, quintic and septic B-splines
[16,17,18].
In this study the cubic B-splines collocation method is used for solving the heat equation (1) subject to (2) and
(3) and the solutions are compared with the exact solution.For constructing the cubic B-splines finite element
method, we use collocation techniques as it was extensively used in [9,16,19,20,21-25]. In the section two,
proposed method is presented and it is also given how to apply the collocation method with cubic B-splines
finite element technique. In the section three, the stability analysis is investigated considering Fourier stability
method. Finally, the numerical results and the related tables are given in the next section.
1200
2 Collocation Method
Let us consider the domain
0 x0 x1
[0, L]
xj
such that
xN L , i.e., finite elements of length h x j 1 x j for j 0,..., N 1 , and also suppose that
m ( x ) to be cubic B-splines at the nodal points xm for m 1, , N 1 . Using cubic B-splines m ( x ) , the exact
solution U ( x, t ) is approached by an approximation U N ( x, t ) such that
U N ( x, t )
N 1
m 1
(t ) m ( x )
(4)
( x xm 2 ) 3
,[ xm 2 , xm 1 ]
3
2
2
3
h 3h ( x xm 1 ) 3h( x xm 1 ) 3( x xm 1 ) ,[ xm 1 , xm ]
1 3
m ( x ) 3 h 3h 2 ( xm 1 x ) 3h( xm 1 x ) 2 3( xm 1 x ) 3 ,[ xm , xm 1 ]
(5)
h
3
,[ xm 1 , xm 2 ]
( xm 2 x )
0
, otherwise
in [27].
Considering the approximation function (4) and the cubic B-splines m ( x ) defined in (5), the required values
of U m and its first and the second derivatives with respect to x at the nodal points xm are identified in terms of
m as
U m U ( xm ) m 1 4m m 1
3
(m 1 m 1 )
(6)
h
6
U m U ( xm ) 2 (m 1 2m m 1 ).
h
For instance, let us take 1 , L 1 and f ( x) sin( x) in (1) and (2), respectively. Therefore the problem (1)
subject to (2) and (3) is become
(7)
uxx ut 0, 0 x 1 , t 0
with initial condition
u( x,0) sin(x)
(8)
and boundary conditions
u(0, t ) u(1, t ) 0 for t 0 .
(9)
U m U ( xm )
One can obtain the approximate solution for the heat equation namely
(10)
U xx Ut 0
by considering the solution of
(11)
(Ut )mn (1 ) f mn f mn 1 0
where
(12)
f mn 2 (U xx )mn
Here, is a parameter that when it takes the value 0 , the scheme is so called forward Euler and also if
1/ 2 , then the scheme is called Crank-Nicholson, and if 1 , the scheme is so called backward Euler. Then
we discretize the time derivative by means of finite difference so we have
(13)
Umn Umn 1 t (1 ) 2 (U xx )mn t 2 (U xx )mn 1 0
Substituting (6) into (13), we obtain the following difference equation system for the variables , which has
n 1 difference equations with n 3 unknown values as
1201
6
12
6
t 2 ) mn 1 (4 2 t 2 ) mn 11 (1 2 t 2 )
(14)
h2
h
h
2
L1 t (1 ) L3
where
3
6
L1 mn 1 4 mn mn 1 , L2 ( mn 1 mn 1 ) , L3 2 ( mn 1 2 mn mn 1 ) .
h
h
Then this set of equations is a recurrence relationship of element parameters vector
d n ( n1 , 0n , 1n , , Nn , Nn 1 ) . Using the boundary conditions (9) and eliminating the parameters 1 , N 1 in
(14), then the system may be rewritten as
U ( x0 ) n11 4 0n 1 1n 1 0 n11 (4 0n 1 1n 1 )
(15)
U ( xN ) Nn 11 4 Nn 1 Nn 11 0 Nn 11 ( Nn 11 4 Nn 1 )
By these substitutions, the equation (14) is turned out to be N 1 unknown at each level of the time n in
order to solve it using by Thomas algorithm.
mn 11 (1
1202
Table 1: Comparison of the solutions for some when t 0.0001 and h 0.0125
x
0.25
0.50
0.75
0.4
0.6
0.8
1.0
3.0
1
Numerical
0.01365
0.00190
0.00026
0.00004
0.00000
1
Analytical
0.01364
0.00189
0.00026
0.00003
0.00000
0.1
Numerical
0.67971
0.66645
0.67974
0.64065
0.52585
0.1
Analytical
0.67973
0.66645
0.65342
0.64065
0.52589
0.01
Numerical
0.70683
0.70669
0.70655
0.70641
0.70501
0.01
Analytical
0.70683
0.70669
0.70655
0.70641
0.70501
0.4
0.6
0.8
1.0
3.0
0.01931
0.00268
0.00037
0.00005
0.00000
0.01929
0.00268
0.00037
0.00005
0.00000
0.96126
0.94250
0.92408
0.90602
0.74309
0.96129
0.94250
0.92407
0.90601
0.74372
0.99961
0.99941
0.99921
0.99901
0.99704
0.99961
0.99941
0.99921
0.99901
0.99704
0.4
0.6
0.8
1.0
3.0
0.01365
0.00190
0.00026
0.00004
0.00000
0.01364
0.00189
0.00026
0.00003
0.00000
0.67970
0.66645
0.65342
0.64065
0.52585
0.67973
0.66645
0.65342
0.64065
0.52589
0.70682
0.70669
0.70655
0.70641
0.70501
0.70683
0.70669
0.70655
0.70641
0.70501
When fixed t 0.1 is taken in calculations with t 0.00001 , the results are shown and compared for
different values of h in Table 2. It is also seen from the table that the results calculated for h 0.025 ,
h 0.0125 and h 0.00625 coincide with the exact solution, while those for h 0.1 , h 0.05 are approached
rapidly to the exact solutions.
Table 2: Comparison of the results for t 0.1 , t 0.00001 , 1
x
h 0.05
h 0.025
h 0.0125
h 0.00625 Analytical
h 0.1
0.1 0.11425 0.11495 0.11513
0.11517
0.11518
0.11517
0.2 0.21732 0.21865 0.21898
0.21907
0.21909
0.21907
0.3 0.29911 0.30094 0.30140
0.30152
0.30155
0.30152
0.4 0.35163 0.35378 0.35432
0.35446
0.35449
0.35446
0.5 0.36972 0.37199 0.37256
0.37270
0.37273
0.37270
0.6 0.35163 0.35378 0.35432
0.35446
0.35449
0.35446
0.7 0.29911 0.30094 0.30140
0.30152
0.30155
0.30152
0.8 0.21732 0.21865 0.21898
0.21907
0.21909
0.21907
0.9 0.11425 0.11495 0.11513
0.11517
0.11518
0.11517
It is observed by two tables that the approximate solutions obtained for different step sizes are closed to the
exact solutions. To further corroborate the applicability of the presented method, numerical solution profile has
been plotted in Figs. 1 for h 0.1 , t 0.01 .
1203
1,0
t=0
0,8
0,6
t=0.05
U
0,4
t=0.1
0,2
0,0
0,0
0,2
0,4
0,6
0,8
1,0
5 Conclusion
In this paper, we apply the collocation method with cubic B-splines finite elements to the heat equation
successfully. In section 3, the stability of this method is analyzed considering Fourier stability method and it is
found that the method is stable for [1/ 2,1] . The results of the numerical solutions in Section 2 confirm that
the accuracy, reliability and efficiency of the presented method which is applied in order to solve this type of the
problem.
References
1. Carslaw, S. and Jaeger , J.C. (1959). Conduction of Heat in Solids,
Oxford University Press.
2. Widder , D.V. (1976). The Heat Equation, Academic Press.
3. Cannon , J.R. (1984). The One-Dimensional Heat Equation, Cambridge
University Press.
4. Wazwaz, A.M. (2002). Partial Differential Equations Methods and Applications, Saint Xavier
University.
5. Kutluay, S., Bahadr, A.R. and zde, A. (1999). Numerical Solution of One-dimensional Burger
Equation: Explicit and Exact-explicit Finite Difference Methods, J. Comp. App. Math., 103, 251-261.
6. Caldwell, J. (1987). Application of Cubic Splines to Nonlinear Burgers Equation, in: E. Hinton et al.
(Eds.), Numerical Methods for Nonlinear Problems, Pineridge Pres., 3, 253-266.
7. Rubin, S.G. and Khosla, P.K. (1976). Higher-Order Numerical Solutions Using Cubic Splines, AIAA
J., 14, 851-858.
8. Calar, H., zer, M. and Calar, N. (2008).The Numerical Solution of the One-dimensional Heat
Equation by Using Third Degree B-spline Functions, Chaos, Solitons & Fractals, 38, 1197-1201.
9. Dag, I., Saka, B. and Irk, D. (2004). Application Cubic B-splines for Numerical Solution of the RLW
Equation, Appl. Maths. and Comp., 159, 373389.
10. Goodman, T. and Hardin, D. (2005). Refinable Multivariate Spline Functions, Topics in Multivariate
Approximation and Interpolation, K.Jetter et al. (eds.), Elsevier, 5583.
11. Dag, I., Irk, D. and Saka B. (2005). A Numerical Solution of the Burgers Equation Using Cubic Bsplines, Appl. Maths. and Comp., 163, 199211.
12. Dag, I. (2001). Approximation of the RLW Equation by the Least Square Cubic B-spline Finite Element
Method, Appl. Math. Modelling, 25, 221231.
13. Zaki, S.I. (2000). A Quintic B-spline Finite Elements Scheme for the KdVB Equation, Comp. Met. in
Appl. Mech. and Eng., 188, 121-134.
14. Bahadr, A.R. (2004). Application of Cubic B-spline Finite Element Technique to the Thermistor
Problem, Applied Mathematics and Computation, 149, 379387.
1204
15. Raslan, K.R. and Hassan, S.M. (2009). Solitary Waves for the MRLW Equation, Applied Mathematics
Letters 22, 984-989.
16. Saka, B. and Dag, I. (2007). Quartic B-spline Collocation Method to the Numerical Solutions of the
Burgers' Equation, Chaos, Solitons & Fractals, 32, 1125-1137.
17. Saka, B., Dag, I. and Boz, A. (2005). B-spline Galerkin Methods for Numerical Solutions of the
Burgers Equation, Appl. Maths. and Comp., 166, 506-522.
18. Ramadan, M.A., El-Danaf, T.S. and Abd Alaal, F.E.I. (2005). A Numerical Solution of the Burgers
Equation Using Septic B-splines, Chaos, Solitons & Fractals, 26, 795-804.
19. Khalifa, A.K., Raslana, K.R. and Alzubaidi, H.M. (2008). A Collocation Method with Cubic B-splines
for Solving the MRLW Equation, Journal of Computational and Applied Mathematics 212, 406 418.
20. Ali, A.H.A., Gardner L.R.T. and Gardner G.A. (1992). A Collocation Method for Burgers Equation
Using Cubic Splines, Comp. Math. Appl. Mech. Eng., 100, 325-337.
21. Ding, J.N., Kan, B. and Cheng, G.G. (2008). Numerical Approach to Torsion Deformation of Armchair
Single Walled Carbon Nanotubes, International Journal of Nonlinear Sciences and Numerical
Simulation, 9, 309-314.
22. Uddin, M., Haq, S. and Siraj-ul-Islam (2009). Numerical Solution of Complex Modified Korteweg-de
Vries Equation by Mesh-free Collocation Method, Computers & Mathematics with Applications, 58,
566-578.
23. Caglar, N. and Caglar, H. (2009). B-spline Method for Solving Linear System of Second-order
Boundary Value Problems, Computers & Mathematics with Applications, 57, 757-762.
24. Kadalbajoo, M.K. and Arora, P. (2009). B-spline Collocation Method for the Singular-perturbation
Problem Using Artificial Viscosity, Computers & Mathematics with Applications, 57, 650-663.
25. Revell , R. and Ridolfi, L. (2008). Generalized Collocation Method for Two-dimensional ReactionDiffusion Problems with Homogeneous Neumann Boundary Conditions, Computers & Mathematics
with Applications, 56, 2360-2370.
26. Sewell, G. (2005). The Numerical Solution of Ordinary and Partial Differential Equations, John Wiley
and Sons.
27. Prenter, P.M. (1975). Splines and Variational Methods, John Wiley, New York.
Biographies
Duygu Dnmez Demir Duygu Dnmez Demir was born in Urla in 1983. She graduated from Celal Bayar
University in 2004. She completed her graduate thesis in 2008. She is still a doctoral student in Celal Bayar
University. She has been an research assistant in Celal Bayar University for 4 years. She is studying numeric
analysis and partial differential equations.
She has four published articles in the journals of national and international. Her one paper has been presented
in the international conference.
Necdet Bildik Necdet Bildik was born in Sivas in 1951. He graduated from Ankara University in 1974. He
completed his graduate thesis in University of Louisville in 1978. He completed his doctorate in Oklahoma State
University in 1982. He was professor assistant in 1995. He was professor in 2003. He has been the head of
department of mathematics in Celal Bayar University since 1997. He is studying numeric analysis, ode
differential equations and partial differential equations.
He has thirty four published articles in the journals of national and international. His twenty six papers have
been presented in national and international conferences. He completed a scientific research projects in Celal
Bayar University.
Professor Necdet BLDK is a member of department of Mathematics.
Simge ztun Simge ztun was born in zmir in 1982. She graduated from Celal Bayar University in 2005.
She completed her graduate thesis in 2009. She is still a doctoral student in Celal Bayar University. She has been
an research assistant in Celal Bayar University for 4 years. She is studying algebraic topology and digital
topology.
She has one published article in the journals of international. Her two paper has been presented in the national
conference.
1205
Abstract.High Dimensional Model Representation (HDMR) based methods which are used to represent a
multivariate function have been developed in last two decays. These methods intend to express a multivariate
function in terms of less-variate functions such as univariate or bivariate or higher variate ones. In literature,
at most bivariate HDMR terms are used which stands for an approximation. In fact, if only the first term,
constant term of HDMR, is used for the representation, although the method has lower complexity, it will be
inadequate to make a good approximation. So the optimization is required for making a better representation
through constant approximation. For this purpose, a new method has been developed to achieve both the
approximation quality and low complexity. In this method, the interval of the given independent variable
which defines the problem is divided into subintervals and a constant term is calculated for each subinterval
to obtain an approximation for the given problem. Hence, as well as computational complexity is reduced, the
representation quality of the method through the constant HDMR component is increased. This work also
includes various numerical implementations to show the mentioned efficiency of the method.
1 Introduction
In scientific problems, dealing with the less-variate functions are better than multivariate ones. Therefore divideand-conquer based algorithms are used. One of the important methods have entered the literature is High
Dimensional Model Representation (HDMR). This method was first proposed by Sobol in 1993[1] and also
called Plain HDMR whose explicit formula is given as follows
N
f ( x1 ,, x N ) f 0 f i1 ( xi1 ) f1N ( x1 ,, x N ) .
(3)
i1 1
fi ( xi ) and the increasing number of higher variate terms. The total number of HDMR expansions components
N
is 2 and all of these components must be used in the expansion to have the ability of exactly representing the
given multivariate function. These HDMR components are determined uniquely by using vanishing conditions
first defined by Sobol for unit constant weight function
bN
b1
dx dx
1
a1
W ( x1 , x2 ,, x N ) f i ( xi ) 0
(2)
aN
1206
under a product type weight function over a given interval, [0,1], with normalization criteria which was first
introduced by Rabitz
N
W ( x1 , x2 ,, x N ) W j ( x j ),
bj
aj
j 1
dx j W j ( x j ) 1, x j a j , b j .
(3)
bN
a1
aN
I 0 F ( x1 ,, x N ) dx1W1 ( x1 ) dx N WN ( x N ) F ( x1 ,, x N ) .
(4)
The projection operator for determination of the general structure of the univariate HDMR components is
defined as follows
b1
b11
a1
a11
dx
bN
b11
(5)
Wi 1 ( xi 1 ) dx N W N ( x N ) F ( x1 ,, x N )
i 1
a11
aN
If these operators are applied to the both sides of the HDMR expansion given in (1), the following constant
and univariate HDMR components are obtained.
f 0 I 0 f ( x1 ,, x N ),
f i ( xi ) I i f ( x1 ,, x N ) f 0
(6)
s 0 ( x1 ,, x N ) f 0 ,
N
s1 ( x1 ,, x N ) s 0 ( x1 ,, x N ) f i1 ( xi1 ) .
(7)
i1 1
Here the important point we need to emphasize on, is how many terms are taken from the expansion for the
qualified representation. To measure the quality of the approximation following measures are defined.
1
f
f0
, 1
1
f
i1 1
f i1
0 , .
(8)
In literature, generally at most bivariate terms are used for the approximate representation. Because if the
higher terms are taken into consideration undesired computational complexities arise. However, when we use
only the constant term in the representation process the presentation quality becomes poor while the
computational complexity decreases sharply. In this work, we split the problem domain into small pieces and a
constant term is obtained for each small piece. Finally, when we interpolate these constant values, the
1207
performance of the constant HDMR component in representing the given univariate function increases
noticeably.
dxw( x) f ( x) f 0( k )
m
ck
k 1
ck 1
ck
ck 1
dxw( x) f ( x) f 0( k )
ck
ck 1
f 0( k )
ck
ck 1
dx
w( x)
ck
ck 1
(9)
dxw( x) .
(10)
f ( x) .
(11)
dxw( x)
To find out the most appropriate coordinates for the randomly selected points, the optimization process is done
by evaluating the following partial derivatives under the given equality
0,
ck
1 k m.
(12)
f 0( k 1) f 0( k )
.
f (c k )
2
To determine the ck ,
(13)
f ( k 1) f 0( k )
c k f 1 0
2
(14)
f ( k 1, j ) f 0( k , j )
c k( j 1) f 1 0
2
where
(15)
1208
Each step of the iteration process is executed through eq. (16). The stoppage criterion for the mentioned
iteration is
ck( j 1) ck( j ) .
(16)
where is a user defined tolerance value. Finally, we have the optimized ck , 1 k m values for creating
piecewise structure to be used in representing the given function through HDMR. To this end, the constant
values for each subinterval constructed through the optimized ck values can be evaluated by using eq. (15).
5 Numerical Implementations
In this section, we arrange some numerical examples to show the efficiency of our new method. The evaluations
are done by using MuPAD within 20 decimal digits precision. To conduct numerical implementations, the
following testing functions are selected
f 1 ( x) e 3 x , f 2 ( x) x 8 4 x 4 5, f 3 ( x) sinx .
(17)
Where they have exponential, logarithmic, polynomial and trigonometric structures respectively. If we take the
expressions of the previous section into consideration and rewrite all these functions in case of constancy
approximation, we obtain the results given in Figures 1-4. In figures, the blue line presents the given original
function and the red line shows the piecewise function which denotes the constant HDMR term between the
points determined in the last step of the iteration process. The points in figures show the middle point of each
step of the piecewise constant function and if we make interpolation by using these points we obtain the
approximation function which is plotted through the line in green. As it is seen in Fig. 3 and Fig. 4, the
approximation quality of our new method becomes better when we increase the number of subintervals during
the node optimization in piecewise HDMR. In addition, we observe that the increase in the number of
subintervals occur especially on the locations where the curvature of the structure increases. This happens
because of the nature of our new method. The approximations obtained through our new method in our
numerical implementations express the given analytical structures almost exactly.
1209
1210
5 Concluding Remarks
In this work, we try to reduce the complexity coming from higher variate terms of the HDMR expansion by
taking only the constant term into consideration as well as to increase the approximation quality of the method
by splitting the whole interval of the problem into a number of subintervals by using node optimization method
and evaluating a constant value for each subinterval to approximate the given function.
In this study, either in the explanation of the new algorithm or in the preparation of the numerical examples,
univariate functions are used. The functions having more than one independent variable are out of the scope of
this article. However, the algorithm can be generalized for these types of functions by making certain changes.
Similar to the other numerical methods, when the number of nodes taken into consideration increases, the
quality of the approximation obtained through piecewise HDMR increases. That is, the more number of
subintervals taken into consideration, the more qualified approximation is obtained.
References
1.
Sobol, I.M. (1993). Sensitivity estimates for nonlinear mathematical models. Mathematical Modelling and
Computational Experiments (MMCE) 1(4): 407-414.
2. Rabitz, H., & Al, . 1999. General foundations of high dimensional model representations. J. Math. Chem.
25:197-233.
3. Al, .F., & Rabitz, H. (2001). Efficient Implementation of High Dimensional Model Representations. J.
Math. Chem. 29: 127-142.
4. Li, G., Rosenthal, C., & Rabitz, H. (2001). High Dimensional Model Representations. J. Phys. Chem. A 105:
7765-7777.
5. Li, G., Wang, S.-W., & Rabitz, H. (2002). Practical Approaches To Construct RS-HDMR Component
Functions. J. Phys. Chem. A 106: 8721-8733.
6. Demiralp, M. (2003). High Dimensional Model Representation and its Application Varieties. Mathematical
Research 9: 146-159.
7. Baykara, N.A., & Demiralp, M., (2003). Hyperspherical or Hyperellipsoidal Coordinates in the Evaluation of
High Dimensional Model Representation Approximants. Mathematical Research 9: 48-62.
8. Tunga, M.A., & Demiralp, M. (2005). A Factorized High Dimensional Model Representation on the Nodes
of a Finite Hyperprismatic Regular Grid. Applied Mathematics and Computation 164: 865-883.
9. Tunga, B., & Demiralp, M. (2003). Hybrid High Dimensional Model Representation Approximants and their
Utilization in Applications. Mathematical Research 9: 438-446.
10. Demiralp, M. (2006). Illustrative Implementations to Show How Logarithm Based High Dimensional Model
Representation Works for Various Function Structures. WSEAS Transaction on Computers 5: 13331338.
11. Tunga, B., & Demiralp, M. (2007). A Novel Hybrid High Dimensional Model Representation (HHDMR)
Based on the Combination of Plain and Logarithmic High Dimensional Model Representations. WSEAS-2007
Proceedings, WSEAS 12-th International Conference on Applied Mathematics for Science and Engineering,
1: 157161.
12. Oevel, W., Postel, F., Wehmeier, S., & Gerhard, J. (2000). The Mupad Tutorial, Springer, New York.
Biographies
Burcu TUNGA She was born in stanbul, Turkey on September 27, 1978. She earned his Bachelors M.Sc.
and Ph.D degrees from stanbul Technical University. She has 3 journal papers in SCI, 5 journal papers not in
SCI, 9 international conference proceedings and a book chapter. Her research interests are numerical analysis,
data modeling and programming languages. Dr. Tunga is the member of Group for Science and Methods of
Computing of Institute of Informatics, stanbul Technical University, stanbul, Turkey since 2000.
Metin DEMRALP He was born in Metin Demiralp was born in Turkey on 4 May 1948. He got his BS, MS,
and PhD from the same institution, Istanbul Technical University. He is working on methodology for
computational sciences through theoretical chemistry and applied mathematics. He has a group (Group for
Science and Methods of Computing) in Informatics Institute of Istanbul Technical University. Metin Demiralp
has roughly 80 papers in well known scientific journals and is the full member of Turkish Academy of Sciences
since 1994. He has also two important awards of Turkish scientific establishments.
1211
Abstract.The disease diagnosis is considered among the most important parts of the treatment process.
The aim of this study is to demonstrate how basic concepts in information theory apply to the problem of
quantifying diagnostic test performance. In this study, the performances of the Dexamethasone Suppression
Test-DST and the Thyroid-Stimulating Hormone Test-TSH, two of the diagnosis tests of Major Depressive
Disorder, are evaluated with the method of Information Theory. The amount of information gained by
performing a diagnostic test can be quantified by calculating the relative entropy between the posttest and
pretest probability distributions. And also demonstrates that diagnostic test performance can be quantified as
the average amount of information the test result provides about the disease state.
Keywords: diagnostic tests, entropy, relative entropy
1 Introduction
Diagnostic tests are widely used in many areas. In particular, these tests have a huge importance in medicine
sector. By courtesy of early and accurate diagnosis can decrease the morbidity and mortality of disease. It is
often important to compare various diagnostics test with each other for specific clinical conditions in order to
determine which one is the best to use. To evaluate the performance of a diagnostic test within a given
population, we ideally determine both the test result and the disease state for every individual in that population.
One of the approaches used to analyze the performance of diagnostic tests in recent years, information theory.
This theory of knowledge and uncertainty for the measurement is based on a mathematical basis. In 1973, Metz,
Goodenough and.Rossmann have developed a formula used in assessing the performance of diagnostic tests
using information theory. (See [1]) After this work, Somoza and Mosmann developed new mathematical and
graphical method to evaluate and compare the performance of diagnostic tests for the value of any prevalence (P)
using the properties of the ROC analysis and information theory approach. (See [2]) In 1999, Lee was obtained
the distance between patients and healthy distributions using the concept of relative entropy. In 2002, Benish
was investigated the concept of relative entropy with a different perspective. The aim of this study is to
demonstrate how basic concepts in information theory apply to the problem of quantifying diagnostic test
performance
1212
tested. On the condition that there is a bilateral class result as disease positive / negative and the test result
positive / negative, a crosstab is created. The crosstab is as below.
Table 1. Relations between prevalence and the level of a test.
Test Result
Diagnosis
Positive Negative
Positive
TP
P
FN
Negative
FP
Q
TN
Q'
P'
1
Let pi be the probability that patient i will get a positive diagnosis and qi be patient i ' s probability of a
positive test.
P is called as the prevalence and it is equal to mean( pi ) value. It is also called as pretest probability of
disease. Q is called as the level of the test and it is equal to mean(qi ) value. We also define
P' 1 P and
Q' 1 Q
TP : If both diagnosis and test are positive, it is called a true positive.
FP : If diagnosis is negative and test is positive, it is called a false positive.
FN : If diagnosis is positive and test is negative, it is called a false negative.
TN : If both diagnosis and test are negative, it is called a true negative.
We are going to mention about 5 criteria for the calculation of the test quality. These criteria will also be used in
the Information Theory.
Sensitivity:The possibility of finding sick people among actually sick people.
SE TP / (TP FN ) TP / P
(4)
SP TN / ( FP TN ) TP / P'
(2)
Efficiency is defined as
EFF TP TN
(3)
The Predictive Value of A Positive Test:Positive test result gives the possibility of being ill.
(4)
The Predictive Value of A Negative Test:Negative test result gives the possibility of not being ill .
(5)
We learn whether the diagnosis test has legitimacy or not with one of the statistical test of Chi-Square. The ChiSquare test depends on sample size ( N 0 ), (1, 0) and (0, 0) values. (1, 0) and (0, 0) are quality indices
and calculated with the formulas below.
(6)
(0,0) ( SP - Q' ) / Q
(7)
1213
H ( X ) is defined by
H ( x) p( x) log 2 p( x)
(8)
in the works Cover and Thomas [4]. While evaluating the performance of the diagnosis test using the
information theory, we need to explain the concepts of test results and disease statement. Disease Statement is
shown with D . On the condition that there are two statements such as the existence or the non-existence of a
disease, we can specify the Disease Statement as follows.
D {Di } i :{, } D =Get ill before diagnosis test D =Get not ill before diagnosis test
The probability distribution of the disease statement before the test is defined with Prevalence ( P( D) ) and 1Prevalence ( P( D) ) values. In this case, the entropy before the test is calculated with this formula:
H ( D) P( D) log 2 P( D) P( D) log 2 P( D)
(9)
The mutual information of two random variables is a quantity that measures the mutual dependence of the two
variables. The interpretation is that when mutual information is absent, marginal distributions are independent
and their entropies add up to total entropy. Mutual information I ( X ; Y ) for the random variables X and Y is
evaluated by the formula
I ( X ;Y ) H ( X ) H ( X \ Y )
(10)
(11)
If H ( D) is defined as pretest entropy, we need to define H ( D \ T ) as the expected value of posttest entropy.
(See Benish [5]) Besides, the difference between H ( D) and H ( D \ T ) ) is called as Mutual Information.
Mutual Information is shown as I ( D; T ) . I ( D; T ) is the reduction in the uncertainty of
knowledge of
due to the
T . Mutual Information is the general criterion of what the diagnosis test will tell us.
The relative entropy (Kullback-Leibler divergence) D( p \ \ q) is an appropriate measure of the similarity of the
underlying distribution. It may be calculated from
1214
D( p \ \ q) p( x) log
xX
(12)
p ( x)
q ( x)
The properties of the relative entropy equation (12) make it non-negative and it is zero if both distributions are
equivalent namely, p q . The smaller the relative entropy is the more similar the distribution of the two
variables and vice versa, see Cover and Thomas [4].
In diagnosis tests, the distance between positive posttest entropy and pretest entropy is relative entropy. This
distance is called as the information content which is provided with positive result and it is shown with I pos.
Similarly, the distance between negative posttest entropy and pretest entropy is called information content which
is provided with negative result and it is shown with I neg. The concepts of I pos. and I neg. are calculated
according to the formulas below.
(13)
(14)
is directly equal to
(15)
The value of D(d \ \ d ) which is called rule in potential is the criterion showing the existence of the
disease. Similarly, the value of D(d \ \ d ) which is called rule out potential is the criterion showing the
non-existence of the disease.
The formulas for the concepts of D(d \ \ d ) and D(d \ \ d ) are as below.
1 SE
SE
( SE ) log
SP
1 SP
(16)
SP
1 SP
(1 SP) log
1 SE
SE
(17)
D(d \ \ d ) (1 SE ) log
1215
4 Application
Major Depressive Disorder is to feel so sad, bored, distressed and depressed that it ruins daily routine and social
functions of the person suffering. Major Depressive Disorder patients can not even manage their daily routines.
In this study, the performances of the Dexamethasone Suppression Test-DST and the Thyroid-Stimulating
Hormone Test-TSH, two of the diagnosis tests of Major Depressive Disorder, are evaluated with the method of
Information Theory. In this application, data of 200 patients who applied Uluda University Medical Faculty
Department of Psychiatry are studied; the data are selected with random sample method. The results of DST and
TSH diagnosis tests are evaluated below.
Table 2. Relations between the measurement probabilities of the outcome, prevalence and the level of diagnostic
tests
DST
TSH
DST
TSH
TP
FN
FP
TN
0.420
0.325
0.180
0.275
0.060
0.020
P
Q
0.340
0.380
Q'
P
'
0.600
0.600
0.400
0.400
0.480
0.345
0.520
0.655
SE
SP
PVP
PVN
EFF
0.700
0.542
0.850
0.950
0.875
0.942
0.654
0.580
0.760
0.705
(1, 0)
(0, 0)
(0.5, 0)
DST
TSH
0.423
0.300
0.688
0.855
0.524
0.444
58.173
51.348
Table 5. Probabilities and conditional probabilities of disease statement and test results for diagnostic tests.
DST
TSH
DST
TSH
P( D)
0.600
0.600
P( D \ T ) 0.875
0.942
P( D)
0.400
0.400
P( D \ T -)
0.346
0.420
P(T )
0.480
0.345
P( D - \ T )
0.125
0.058
P(T )
0.520
0.655
P( D - \ T -)
0.654
0.580
According to the results deduced from the first 4 tables for application, the DST and the TSH have the same
that DST is more efficient and has less error than TSH.
1216
When we have a look at the legitimacy of the test, we can see that both tests are legitimate. Since both Chisquare values are high and p values of both tests are smaller than alpha ( ), we call both tests legitimate.
Table 5 respectively. On the condition that the test is positive, the possibility to diagnose the person as ill is
higher for the TSH test; however, if the test is negative, the possibility to diagnose the person as healthy is higher
for the DST test.
Table 6. Relations among entropy, conditional entropy and mutual information.
TESTS
H ( D)
H (D \ T )
I ( D; T )
DST
TSH
0.970951
0.970951
0.744816
0.752977
0.226135
0.217974
If we take the disease statement as the random variable, the random variable is indicated either as disease exists
before diagnosis test or disease non-exists before diagnosis test. According to this situation the entropy of the
disease is only affected with the possibility of disease existence or disease non-existence. Since these
possibilities are equal in both tests, the entropy of the disease is the same. When we have a look at Table 6, the
entropy value is the same but I ( D; T ) values are different. According to the result of I ( D; T ) , DST provides
more diagnostic information than TSH for value of
I pos.
I neg.
D(d \ \ d )
D(d \ \ d )
LR
LR
DST
TSH
0.266
0.451
0.188
0.094
0.332
0.451
0.284
0.248
4.666
10.833
0.352
0.482
When we have a look at the results of Table 7, we firstly notice that the values of Negative Likelihood Ratio and
the values of Positive Likelihood Ratios are different. If the value of the Negative Likelihood Ratio or the value
of the Positive Likelihood Ratios were the same, two tests could not dominate each other. See Benish [7]
According to the results of Table 9, a positive test result provides more information than a negative test result
for both tests when P is 0.6 (also P is equal for both tests). The positive test result of TSH provides more
information than the positive test result of DST. On the contrary, the negative result of DST provides more
information than the negative result of TSH. Therefore, we can conclude that I pos. is related to PVP and
5 Conclusion
In recent years, for the evaluation of the diagnosis tests, in addition to the other methods, measures calculated
according to the information theory have become important.
In this study, it has been explained that how these measures are calculated for the diagnosis tests used in physic
and the performances of two tests used for the diagnosis of the major depressive disorder are interpreted by
calculating with these methods.
It is aimed that this study will hopefully give various points of view to the researchers who want to make
research on this subject by explaining how the tests used for the diagnosis of various diseases are evaluated with
this way.
1217
References
1.
2.
3.
4.
5.
6.
7.
8.
Metz, C.E, Goodenough D.J, and Rossmann, K. (1973) Evaluation of receiver operating characteristic
curve data in terms of information theory, with applications in radiography. Radiology; 109: 297-303
Mossman, D. and Somoza, E (1989). Maximizing diagnostic information from the dexamethasone
suppression test: An approach to criterion selection using receiver operating characteristic analysis.
Archives of General Psychiatry; 46: 653-60
Westin, L.K. (2001) Receiver operating characteristic (ROC) analysis. Evaluating discriminance
effects among descision support systems. Sweden: Umea universitet - Department of Computing
Science.
Cover, T.M. and Thomas, J.A. 2006. Elements of Information Theory 2nd Edition. New Jersey: John
Wiley & Sons, Inc.
Benish, W.A. (2009) Intuitive and axiomatic arguments for quantifying diagnostic test performance in
units of information. Methods Inf Med 48: 552-557
Lee, W.C. (1999) Selecting Diagnostic Tests for Ruling Out or Ruling In Disease: The Use of the
Kullback-Leibler distance. International Journal of Epidemiology 28: 521-525
Benish, W.A. (2002) The use of information graphs to evaluate and compare diagnostic
tests.
Methods Inf Med 41: 114-118
Boyko, E.J. (1994) Ruling Out or Ruling In Disease with the Most Sensitive or Specific Diagnostic
Test Short Cut or Wrong Turn? Med Decis Making 14: 175-179
Biographies
Armaan KANCAMr. Kanca was born in 1987 in Aydn.He completed his primary and secondary education in Aydn. He
graduated from Dokuz Eyll University Science Faculty Statistics Department in 2009.He is a Master Student in
Dokuz Eyll University Statistics Department.
zlem EGE ORUMrs. Ege Oruc was born in 1972 in Izmir. She completed her primary and secondary education in Izmir. She
graduated from Ege University Science Faculty Statistics Department in 1993. She obtained her MS degree in
statistics in 1997 in Dokuz Eylul University Faculty of Science and Letters Statistics Department and her PhD
degree in 2001 in the same department. She started to serve as research assistant in 1995 in Dokuz Eylul
University Faculty of Sciences and Letters Statistics Department and she is still serving as assistant professor in
the same department. She is married with one child.
1218
Department of Mathematics, Izmir Institute of Technology, Glbahe Campus, Urla, Izmir,35430, Turkey
1
gamzetanoglu@iyte.edu.tr, 2silakorkut@iyte.edu.tr
Abstract.The iterative splitting methods have been extensively applied to solve complicated systems of
differential equations. In this process we split the complex problem into several sub-problems, each of them
can be solved sequentially. In this paper, we develop alternative method based the iterative splitting for
solving non-autonomous problems. We also determine the error bound of the proposed method. Several
numerical examples are illustrated to confirm the theoretical results by comparing frequently used splitting
methods.
Keywords: Iterative splitting,Magnus expansion, non-autonomous system, Schrdinger equation
1 Introduction
In this paper we concentrate on solving linear evolution equations, such as time-dependent differential equation,
Y
A(t )Y (t ), Y (0) Y0
t
where Y0 R , and
d
0 t T0
(1)
branches of mathematics and physics such as quantum mechanics, Hamiltonian dynamics, dynamical systems,
etc. For solving Hamiltonian problems, it is often the case that A(t) = T + V (t), where only the potential operator
V(t) is time-dependent and T is linear operator, see [1, 2, 3, 4, 5]. Operator splitting is a widely used procedure in
the numerical solution of large systems of partial differential equations. It allows us to replace an initial value
problem with a sequence of simpler problems, solved successively in each time step. The general idea behind
splitting is breaking down a complicated problem into smaller parts for the sake of time stepping, so that the
different parts can be solved efficiently with suitable integration formulas. Some splitting methods have been
already used to find numerical solution of the different special non-autonomous system, particularly Hamiltonian
ones [6, 10]. It is important to construct such numerical schemes for Hamiltonian dynamics or Schrdinger
equations that preserve some important qualitative properties and geometric structure of that solution. In this
study, we focus on developing the iterative scheme which conserve the symmetric properties of the Schrdinger
equation. The Magnus expansion [6, 7] is a popular geometric, an attractive and widely applied method of
solving explicitly time-dependent problems. We use this method in the solutions of the split subsystems in the
estimations of the splitting error. Our main focus will be two fold: First, we develop the iterative splitting for
non-autonomous problem. Second, we compare its splitting error with those committed by frequently used
methods such as Lie Trotter, Strang Splitting, symmetrically weighted splitting .
The paper is outlined as follows: In Section 2, the basic idea behind the Magnus method is summarized. In
Section 3, the splitting error is calculated for iterative splitting method, theoretically. In the last section, several
numerical examples are illustrated to confirm the theoretical results.
1219
2 Magnus expansion
The Magnus integrator was introduced as a tool to solve non-autonomous linear differential equations for linear
operators of the form
dY
A(t )Y (t ),
dt
(2)
(t ) ,
(t )
n 1
(4)
(t ),
1 (t )
dt A
1
3 (t )
0
t
2 (t )
1
1
dt
1 dt 2 [ A1 , A2 ]
2 0
0
1
6
t1
t2
where An
A(tn ) . In practice, it is more useful to define the nth order Magnus operator
n
(t )
(t ) O(t n1 )
(6)
such that
Y (t ) exp[
(7)
[2]
(t ) dt1 A(t1 )
0
1
(8)
tA( t ) O(t 3 )
2
We will use the second order Magnus expansion the following sections. In general case, because Magnus
expansion generates more terms in the exponential, more complex splittings are necessary.
Fourth order Magnus operator [9,10],
[ 4]
where A1 A( 1t ) , A2 A( 2t ) and
1
1
3
,
2
6
(t )
1
t ( A1 A2 ) c3t 2 [ A1 , A2 ]
2
1
3
,
2
6
c3
(9)
3 .
12
1220
ui (t )
n
Tui (t ) V (t )ui 1 (t ) t [t n , t n 1 ] ui (t n ) usp
,
t
ui 1 (t )
n
Tui (t ) V (t )ui 1 (t ) t [t n , t n 1 ] ui (t n ) usp
t
forn=0,1,,N-1 whereby
(10)
usp0 Y0 is given from (1) and u0=0. The split approximation at the time level
t t n1 is defined as uspn1 u2 p 1 (t n )
u1 (t h) (t h, t )u0
(t h, )F ( )d
(11)
where F ( ) V ( )u0 and is the fundamental set of solution which ,associated to the homogeneous problem
in (10) is (t , ) e(t )T .Note that (t h, t h) I n and (t h, t ) eTh . If we use the trapezoidal rule to
approximate the integral,
t h
(t h, )F ( )d
h
[ F (t h) F (t )(t h, t )] O(h3 )
2
(12)
h
h
u1n 1 eTh [u1n V (tn )u0 ] V (tn h)u0
2
2
(13)
u2 (t h) (t h, t )u0
(t h, )F ( )d
(14)
( )
exp(
j 1
where
(t ))
(15)
[2]
V ( s )ds
0
tV (t / 2) O(t 3 )
tV (t ) O(t 2 )
And by using the relation in the equation (14) , we have
h
u2n 1 e 2
V ( tn h ) V ( tn )
[u2n
h
h
Tu1 (tn )] Tu1 (tn h)
2
2
(16)
If i is odd
uin 1 eTh [uin
h
h
V (tn )uin1 ] eThV (tn h)uin11
2
2
If i is even
1221
(17)
uin 1 e 2
[V ( tn h ) V ( tn )]
[uin
h
h
Tuin1 ] Tuin11
2
2
(18)
Therefore the following statements show that Iterative splitting have first and second order schemes depend on
the number of iterations considered.
Proposition 3. 1The iterative splitting is at first order for i=1 and u0= 0 the system of equation given as (1) with
the error bound
1
1
(19)
Y ( ) Yitsp ( [T ,V 2 ( )] V ( )T ) 2 O( 3 )
2
2
4
4
Proposition 3.2The iterative splitting is at second order for i=1 and u0= u0 the system of equation given as (1)
with the error bound
Y ( ) Yitsp (
TV ( )(2T V1 ( ) V2 ( ))
TV ( )(2T V1 ( ) V2 ( ))
576
4
4
4
96
2
4
4
TV ( )(2T V1 ( ) V2 ( ))) 3 O( 4 )
72
4
4
(20)
(21)
h
h
uin 1 eTh [uin V (tn )uin1 ] eThV (tn h)uin11
2
2
and by using same procedure for even values of i ,we get the equation (18).
4. Numerical Examples
In order to illustrate the proven orders in our convergence bounds, we consider the Mathieu equation
q ( 2 cos(t ))q 0
with the initial conditions q(0) 1.75, q(0) 0 .
Now the time dependent operator can be splitted as,
0
1 0
A(t )
2
2
( cos t ) 0
h
0.1
0.01
0.001
1 0
0
T V (t )
0 cos t 0
0.0610
0.0066 (0.9658)
6.6819e-004 (0.9946)
0.1015
0.0105 (0.9853)
0.0011 (0.9798)
1222
(22)
0.1
0.01
0.001
2nd order
Iterative Splitting
/Order
Strang
Splitting /Order
Symmetrically
weighted Splitting
/Order
9.8067e-004
8.3542e-006
(2.0696)
8.2197e-008
(2.0070)
0.0011
1.0839e-005
(2.0064)
1.0801e-007
(2.7672)
0.0062
6.3187e-005
(1.9918)
6.3309e-007
(1.9992)
The numerically observed order in the discrete L norm is approximately 1, which is supported by the
proposition 3. 1. In addition, proposition 3.2 predicts order 2. This number is in perfect agreement with table 2.
We also observed in table 2 that second order proposed iterative splitting scheme is more efficient than not only
Strang splitting but also symmetrically weighted splitting.
To illustrate interest of the alternative method proposed in this paper, we consider Schrdinger equation
which is one dimensional harmonic oscillator with a time dependent potential as a second example,
with
( x, t ) 1 2 2 (t )( x 2 1)
( x, t ),
2
t
2
2 x
1
x2
2
(23)
2 (t ) 4 3et .
We solve this problem by using the algorithms in (17) and (18) since the solution is not known. We rewrite the
equation (23) as ODE system with the help of the method of lines as follows
A(t ) u 0 T u 0
V (t , x) u
u 0
0 v
v A(t ) 0 v T 0 v V (t , x)
(24)
where ( x, t ) u( x, t ) iv( x, t ) , then consider the splitting methods with ODE system split in the form, T
corresponds to spatial derivative 2 x2 , we use the second order center difference scheme in order to
approximate it. The system (24) and initial condition in (23) now constitute the complete method of lines
approximate of equation (23), we get 2 N 2 N system.
We suppose that the system is defined in the interval x [10,10] , which is split into M 100 parts of length
x 0.2 . We integrate the system using proposed method with the time-step size t 0.03 up to final time t 3 .
Note that the periodicity of the solution is 3. Figure 1 is illustrated the probability of density, ( x, t ) 2 , for
the equation in (23).
1223
References
1.
D. Baye, G. Goldstein and P. Capel. Fourth-order factorization of the evolution operator for timedependent
potentials. Phys. Letts, A 317, 337 (2003).
2. G. Goldstein and D. Baye. Sixth-order factorization of the evolution operator for time-dependent potentials. Phys.
Rev, E 70, 056703 (2004).
3. S.A. Chin and C.R. Chen. Gradient symplectic algorithms for solving the Schrodinger equation with timedependent potentials. Journal of Chemical Physics, 117(4), 1409-1415 (2002).
4. V.C. Aguilera-Navarro, G.A. Estevez and R. Guardiola. Variational and perturbative schemes for a spiked
harmonic oscillator. J. Math. Phys. 31, 99 (1990).
5. S.A. Chin and P. Anisimov. Gradient Symplectic Algorithms for Solving the Radial
Schrdinger Equation. J. Chem. Phys. 124, 054106, 2006.
6. S. Blanes, F. Casas, J.A. Oteo and J. Ros. The Magnus expansion and some of its applications.
arXiv.org:0810.5488 (2008).
7. S. Blanes and P.C. Moan. Fourth- and sixth-order commutator free Magnus integrators for ,linear and nonlinear
dynamical systems. Applied Numerical Mathematics, 56, 1519-1537 (2006).
8. F.J. Dyson. The radiation theorem of Tomonaga. Swinger and Feynman, Phys. Rev., 75, 486-502 (1976).
9. I.V. Puzynin, A.V. Selin and S.I. Vinisity Comput.Phys. Commun. 123 1-6 , (1999).
10. S. Blanes, F.Casas,A. Murua,Splitting methods for non-autonomous linear systems,. Int. J.comput. Math. 84
(2007), 713-727.
11. W. Magnus,On the exponential solution of differential equations for linear operator,.Commun. Pure and Applied
Math. Vol.7 (1954), 649-673.
1224
Abstract.In this study, we present a practical matrix method to find an approximate solution of higher order
linear difference equation with constant coefficients under the initial-boundary conditions in terms of Taylor
polynomials. To obtain this goal, we first present time scale extension of previous polynomial approach, then
restrict the formula to the Integers with h step. This method converts the difference equation to a matrix
equation, which may be considered as a system of linear algebraic equations.
Keywords: dynamic equations, time scale calculus, matrix method, difference equations
1 Introduction
The theory of time scales, which has recently received a lot of attention, was introduced by Hilger [14] in
order to unify continuous and discrete analysis. This theory is appealing because it provides a useful tool for
modeling dynamical processes. Time scale calculus has been created in order to unify the study of differential
and difference equations, so called dynamic equations [11,12]. The concept of dynamic equations has motivated
a huge size of research work in recent years [1113,15]. Since time scale calculus has main purposes as
unification, extension and discretization [10], it allows us to use the theory to solve a difference equation by the
methods are used to solve differential equations. Several numerical methods were used such as the successive
approximations, Adomian decomposition, Chebyshev and Taylor collocation, Haar Wavelet, Tau and Walsh
series etc [69]. Since the beggining of the 1994, Taylor and Chebyshev matrix methods have also been used
by Sezer er al. to solve linear differential, Fredholm integral and Fredholm integro differential-difference
equations.
Since time scale calculus has unification purpose, it allows us to consider so called dynamic equations as the
unification of differential and difference equations. Also, the Taylor matrix method has been used to find the
approximate solutions of differential, integral and integro-differential equations in recent years. In this paper we
present the time scale analogue of this matrix method by the terms of delta derivatives and restrict this method to
Integers with h step to solve higher order difference equations with constant coefficients.
This approach is based on the matrix relations between Taylor polynomials and their derivatives which are
modified to solve the nth order linear DE with constant coefficients
( )
( )
( )
1225
( )
( )
where
,
and
are constants. In the method of Taylor polynomial solution to differential equations; its
assumed that the solution can be constructed by the Taylor polynomial form:
( )
( )
( )
( )
( )
where
denotes th order delta derivative of and is in an arbitrary time scale.
In [11], Agarwal et. al. presented Taylor polynomial for arbitrary time scales as follows:
( )
where
(
)
(
( )
( )
If we put Eq. (3) in the matrix form, we may obtain the matrix relation
, ( )( )
where
( )
( ) , ( )
( )and
0 ( )
( )
1 0
0 1
0 0
0 0
( )
( )1
( )
On the other hand, its obvious to see the relation between the matrix
()
()
where
0
0
0
0
( )
0
0
.
1
0
()
()
( ) , we obtain the recurrence relations
( )
( )
( )
Substituting Equation (8) to (2) leads us to the matrix relation
1226
( )
( ) is
( )
( )
( )
( )
0 ( )
( )1
( )
We are now able to construct the fundamental matrix equation corresponding to Equation (2). Substituting the
matrix relation (9) into Equation (2) and simplifying lead us to fundamental matrix equation:
{
Now let us consider matrix representation of the initial boundary conditions. We can obtain the corresponding
matrix form for the conditions
( )
( )
( )
( ))
, -
where
(
( )
( ))
To obtain the solution of Equation (2) under the initial-boundary conditions, by replacing the row matrices
(13) by the last rows of the matrix (12) we have the new augmented matrix
w00
w
10
wN m,0
[W : G ]
U 00
U10
U m 1,0
If
[ ]
w01
w11
w0 N
w1N
g N m
.
0
: g0
: g1
wN m,1
U 01
U11
wN m, N
U0N
U1 N
:
:
:
U m 1,1
U m 1, N
1227
( )
( )
is uniquely determined. This solution is given by the Taylor polynomial solution.
When
( )
( )
( )
( )
is
( )
( )
It is possible to construct integer analogues of the previous method. The matrix relation of the ( )stills same
with
( )
)]
and
, ( )
( )
( )( )
Recurrence relations of delta derivative of ( ) presented previous section also becomes
( )
( )
( )
( )
By the equations (1518) and the discrete Taylor expansion of the function ( ), i.e.;
( )
( )
we obtain that the fundamental matrix equation is equal to the Equation (11).
References
1.
2.
3.
S. Yalinba and M. Sezer (2000). The approximate solution of high-order linear Volterra-Fredholm integrodifferential equations in terms of Taylor polynomials, Appl Math Comput 112, 291-308.
M. Glsu and M. Sezer (2005). A method for the approximate solution of the high-order linear difference equations
in terms of Taylor polynomials, Int J Comput Math 82 (5), 629-642.
A. Karamete and M. Sezer (2002). A Taylor collocation method for the solution of linear integro-differential
equations, Int J Comput Math 79 (9) 987-1000., 2002.
1228
Blbl, B., Glsu M., Sezer M. (2010). A new Taylor collocation method for nonlinear Fredholm-Volterra integrodifferential equations. Numer. Methods Partial Diff. Eq., 26,(5) : 1006-1020.
N. Kurt and M. Sezer (2008). Polynomial solution of high-order linear Fredholm inteqro-differential equations
with constant coefficients, J Franklin Ins 345, 839-850.
R.P. Kanwal, K.C. Liu (1989). A Taylor expansion approach for solving integral equations, Int. J. Math. Educ. Sci.
Technol., 20(3),411-414.
B. Assady, M.T. Kajani (2005). Direct method for solving integro differential equations using Hybrid Fourier and
block-pulse functions, Intern. J. Computer Math. Vol. 82, No. 7, 889-895.
J. Biazar and H. Ebrahimi (2005). An approximation to the solution of hyperbolic equations by Adomian
decomposition method and comparison with characteristic method, Appl Math Comput 163, 633-638.
S.M. Hosseini, S. Shahmorad (2002). A matrix formulation of the tau method and Volterra linear integrodifferential equations, Korean J. Comput. Appl. Math., 9(2) 497-507.
R. Agarwal, M. Bohner, D. Oregan, A. Peterson (2002). Dynamic equations on time scales: a survey, Journal of
Computational and Applied Mathematics, 141, 1-26.
M. Bohner, A. Peterson (2001). Dynamic Equations on Time Scales: An Introduction with Applications,
Birkhauser, Boston, Mass, USA.
M. Bohner, A. Peterson (2003). Advances in Dynamic Equations on Time Scales, Birkhauser, Boston, Mass, USA.
F.M. Atici, G. Sh. Guseinov (2002). On Greens functions and positive solutions for boundary value problems on
time scales, Journal of Computational and Applied Mathematics, vol 141, no. 1-2, 7599.
S.Hilger (1990), Analysis on measure chain a unified approach to continous and discrete calculus, Results in
Mathematics, vol. 18, 1-1,pp 1856.
S.P. Atmaca (2010). Normal and osculating planes of delta-regular curves, Abstract and Applied Analysis Volume
2010, Article ID 923916, pp. 8. doi:10.1155/2010/923916.
mer AKGLLER mer AKGLLER was born on July 5th 1982 in zmir, Turkey. He received his B.Sc.
from the Department of Mathematics of Ege University in 2007. His M.Sc. was from Mugla University in 2010.
He is a Ph.D. candidate under the supervision of Prof.Dr. Mehmet SEZER and Assist. Prof. Dr. Sibel P.
ATMACA in the Department of Mathematics at Mugla University.
In 2007, he has been research assistant in the Department of Mathematics at Mugla University and stil at the
same position.
His main research interests are time scale calculus, differential geometry, differential and difference equations,
and their numerical solutions.
Sibel PAALI ATMACA Assist. Prof. Dr. Sibel P. ATMACA was born on November 24th 1971 in Turkey.
She received his B.Sc. from the Department of Mathematics of Ege University in 1991. Her M.Sc. was from
University of Missouri-Rolla in 1996. She received her Ph.D. under the supervision of Prof.Dr. Leon M. Hall in
the Department of Mathematics & Statics at University of Missouri-Rolla in 2002 and studied in the field of
Map Equations of Trochoids.
In 2002, she has been Assistant Professor assistant in the Department of Mathematics at Mugla University and
stil at the same position..
Her main research interests are curves on time scales, time scale calculus, algebraic curves. Assist. Prof. Atmaca
has published research articles related to differential geometry on time scales.
Mehmet SEZER - Prof. Dr. Mehmet SEZER was born on March 20th 1954 in Antalya, Turkey. He received his
B.Sc. from the Department of Mathematics of Ege University in 1976 and was research assistant at Balkesir
Engineering Academy. His M.Sc. was from Ege University in 1977. He received his Ph.D. under the supervision
of Prof.Dr.Seda MORALI in the Department of Applied Mathematics at Ege University in 1982 and studied in
the field of Parabolic partial differential equations.
In 1982, he was Assistant Professor at Faculty of Education (Balkesir) of Uluda University, Associate
Professor in 1989, and in 1995 Professor at Faculty of Education of Dokuz Eylul University. Since 2004, he has
been Mathematics Professor at Faculty of Science at Mula University and has been the chair in the same
department.
His main research interests are ordinary and partial differntial equations, integral and difference equations, and
their numerical solutions. Prof. Sezer has been reviewer for numerous influential journals, has published research
articles related to differential equations, linear algebra, analytic geometry and calculus; and has been authored
over 80 papers.
1229
Abstract.The overall purpose of the research presented is to apply the Bees Algorithm (BA) based method in
order to solve the well-known logic puzzle Sudoku efficiently. The BA is one of the recent meta-heuristic
algorithms proposed to solve global optimization problems. The BA is a population-based search algorithm
which mimics the food foraging behaviour of swarms of honey bees. Sudoku (also known as singular
number in Japanese) has gathered great popularity all around the world and recently, there have been several
studies on solving the Sudoku puzzles with modern methods. In this study, BA is used to tackle the board
game Sudoku due to its success in various other applications. Sudoku puzzles in this study can be formulated
as an optimization problem with number-uniqueness penalties and the optimum solutions are successfully
obtained. A number of design examples is presented to demonstrate the efficiency and robustness of the
algorithm presented.
Keywords: bees algorithm, sudoku puzzle,stochastic games, combinatorial optimization, meta-heuristic
algorithms
1 Introduction
Sudoku (also known as singular number in Japanese) has gathered great popularity in the world. The Sudoku
puzzle consists of 9x9 grid and 3x3blocks for all the 81 cells and one player number puzzle which is based on a
9x9 matrix in its traditional form. The general problem of solving Sudoku puzzles is known to be NP-Complete,
and many methods have been developed to obtain optimal solutions. As the popularity of Sudoku has grown, so
has the number of playing variations. There have been several attempts to solve Sudoku puzzle. In recent years
researchers have started to apply various modern methods such as graph theory [1], artificial intelligence [2], and
genetic algorithm to solve the Sudoku puzzle. Sudoku puzzle solving problem can be formulated as an
optimization problem with minimal violations of the three rules of this puzzle game.
Many methods have been developed to obtain optimal solutions while solving Sudoku puzzles. As the popularity
of Sudoku has grown, so has the number of Sudoku puzzle variations. There have been several studies on
solving the Sudoku puzzles with modern methods. The general problem of solving Sudoku puzzles is given by
Santos-Garcia et al. and several methods have been developed to obtain optimal solutions for this problem [1]. In
some studies, the Sudoku puzzle problem is considered as a combinatoric problem and the solutions of this
problem are found by guessing a random solution or by applying brute force iteration through all possible
combinations [2, 3]. In recent years, various modern methods such as the graph theory, artificial intelligence,
genetic algorithm, and harmony search algorithm have been applied to solve the Sudoku puzzle problems by
researchers. The transformations from a directed or undirected graph to an unlabeled digraph are used to obtain
solutions of the puzzle by Eppstein [4]. An artificial intelligence model named MITS (Mixed Initiative Tutoring
System for Sudoku), in which the tutor takes the initiative to interact when the student lacks knowledge and
makes moves that have low utility, is proposed by Caine and Cohen to solve this puzzle problem [5]. Nicolau
1230
and Ryan [6] developed a system named GAuGE (Genetic Algorithm using Grammatical Evolution) for solving
Sudoku, which uses a position independent representation. Harmony search (HS) algorithm
which mimics musicians behaviours is applied to solving Sudoku puzzle by Geem [7]. In his study, Sudoku
puzzles are formulated as an optimization problem with number-uniqueness penalties.
The bees algorithm (BA) used in this study, which is a meta-heuristic method, is a population-based optimization
algorithm developed by Pham et al. in 2005 [8]. The BA is inspired by natural foraging behavior of honey bees
to use in optimization problems. A colony of honey bees can extend itself over a large area and in multiple
directions to exploit a large number of food sources at the same time [9]. A colony feeds by spreading its
foragers to good fields. Practically, a large number of bees visit and collect nectar or pollen from flower patches
with plentiful amounts of food, instead of visiting flower patches with less nectar or pollen. This behavior
provides feeding of the colony with less effort. Recently, the honey bees behaviour-inspired evolutionary
algorithm has been developed and applied to a wide variety of engineering optimization problems by researchers
[10-15].
i 1
xik 1 x jk 1
j 1
l 1 ( m , n )Bl
mn
45
(5)
where, xik ; the repeated number of integer k which is contained in row i, xjk ; the repeated number of integer k
which is contained in row j, Bl ; set of coordinates for block l.
The first and second term in Equation 1 represents the penalty function for each horizontal row and vertical
column. If all possible values of integer k is repeated only once in one row or column, the penalty value for this
row or column is equal to zero. The third term represents the penalty function for each block in Sudoku matrix. It
should be noted that, although the sum of each block equals 45, it doesnt guarantee that the numbers 1 through 9
are used exactly once in one block, but any violation of the uniqueness affects other row, column, or block which
contains the wrong value jointly.
In this study, BA, a new meta-heuristic method is used to tackle the board game Sudoku due to its success in
various other applications. The BA is an optimization algorithm biologically inspired by the natural behavior
above and the algorithm performs a kind of neighborhood search combined with a random search.
1231
3 Bees Algorithm
The bees algorithm developed by Pham et al. [8] is inspired by natural foraging behavior of honey bees to use in
optimization problems.
At the beginning of algorithm, n scout bees are placed randomly in the search space. And the fitnesses of the
visited points by scout bees are evaluated in step 2. The Bees that have high fitnesses are chosen as selected and
1232
best bees in step 3. Then, the algorithm governs the searches in the neighborhood of the selected and best bees
and of the other selected bees in step 4-5. The more promising solutions can be obtained by searching in the
neighborhood of the best sides. The bees can be selected according to fitnesses associated with the sites they are
visiting. Alternatively, the fitness values can be used to decide the probability of the bees being selected. This
different recruitment by scouting is a key process of Bees algorithm. In nature, there is no such restriction. This
restriction is introduced here to reduce the number of points to be explored. In step 6, the remaining bees in the
scout population are assigned randomly to obtain new potential solutions in search space. These steps are
repeated until the stopping criterion is met. At the end of each iteration, the new population of bee colony
consists of representatives from each selected patch and other scout bees assigned to conduct random searches.
4 Numerical Applications
The Bees algorithm based method was applied to the Sudoku puzzle example proposed by Nicolau and Ryan [6]
as shown in Figure 2(a) and this example solved by Geem using harmony search algorithm [7].
(a)
(b)
(c)
(d)
Fig. 2. Intermediate and Final Solutions of Test Sudoku Puzzle
The BA based method found the optimal solution without any violation after 273 iterations for the same problem
that was solved by Geem after 285 function evaluations [7]. Figure 2 shows the history of reaching global
optimum for this problem. In Figure 2, the black cells with white-colored numbers are constant numbers given at
the beginning of the puzzle, the cells with red-colored numbers means there is at least one violation horizontally,
vertically, or block-wise. In Figure 2, (b) is the solution at 63 iteration, which has a penalty of 43; (c) is the
solution at 174 iteration, which has a penalty of 26; (d) is the solution after 273 iteration, which has a penalty of
0.
1233
The HS model is applied by Geem to another Sudoku puzzle problem classified as hard as shown in Figure
3(a) [7]. Geem obtained the local optima with a penalty of 14 after 1,064 function evaluation as given in Figure
3(b). Same problem handled with BA based model developed in this study and after 1,784 iterations the
optimum solution with a penalty of 0 is found (Figure 3(c)). Also, this second problem is analyzed with different
algorithm parameters to show efficiency of BA based method. Analysis results are given in Table 1.
(a)
(b)
(c)
Fig. 3. Final Solutions of Test Sudoku Puzzle
nsel
35
35
40
30
35
40
40
40
30
30
40
30
nbest
10
15
20
10
15
20
20
30
15
15
25
15
nrb
10
10
10
5
5
5
5
5
5
10
10
5
nrs
5
5
10
5
5
5
5
5
5
5
10
5
prad
3
4
5
4
3
4
3
5
5
5
5
2
itermax
3
3
3
3
3
5
3
3
5
5
5
3
F(penalty)
0
0
1
2
2
4
5
5
7
7
9
14
1234
5 Conclusions
This paper has described a Bees Algorithm and its application to solving a Sudoku puzzle problem. After the
formulation of the optimization problem is constructed, the solution of the design problem is obtained by using
the BA which is one of the recent additions to meta-heuristic techniques. The BA based method could
successfully solve the optimization problem and the algorithm was not trapped at local optima. Also, sensitivity
analysis of the BA parameters is performed to obtain a better idea of the algorithm parameter values. In addition,
the BA based method uses a few parameters which are initially specified and consists of simple steps which
make it easy to implement. As shown in Table 1, the algorithm produced better results but the only disadvantage
is the adjustment of parameters. A number of design examples is presented to demonstrate the efficiency and
robustness of the algorithm presented. The computational results show that the Bees Algorithm performed more
strongly than the existing technique. For studies in the future, the BA model should consider additional problemspecific heuristics in order to efficiently solve a harder puzzle.
References
1.
2.
3.
4.
5.
6.
7.
8.
9.
10.
11.
12.
13.
14.
15.
Caine, A. and Cohen, R. (2006). A Mixed-Initiative Intelligent Tutoring System for Sudoku. In
Canadian AI 2006. Qubec city, Qubec, Canada, 7-9 June. Springer, Heidelberg.
Eppstein, D. (2005). Nonrepetitive Paths and Cycles in Graphs with Application to Sudoku. ACM
Computing Research Repository (cs.DS/0507053).
Felgenhauer, B. and Jarvis, F. (2006). Mathematics of Sudoku I. Mathematical Spectrum 39: 15-22.
Geem, Z.W. (2007). Harmony Search Algorithm for Solving Sudoku. Lecture Notes in Computer
Science 4692: 371-378.
Nicolau, M. and Ryan, C. (2006). Solving Sudoku with the GAuGE System. In EuroGP 2006.
Budapest, Hungary, 10-12 April. Springer, Heidelberg.
Pham D.T., Ghanbarzadeh A. (2007). Multi-Objective Optimisation using the Bees Algorithm. In
Proceedings of Innovative Production Machines and Systems Virtual Conference. Cardiff, United
Kingdom, 2-13 July.
Pham D.T.; Muhamad, Z.; Massudi Mahmuddin; Ghanbarzadeh, A.; Koc, E. and Otri, S. (2007). Using
the bees algorithm to optimise a support vector machine for wood defect classification. In Proceedings
of Innovative Production Machines and Systems Virtual Conference. Cardiff, United Kingdom, 2-13
July.
Pham, D.T., et al. (2005). Technical Note: Bees Algorithm. Manufacturing Engineering Centre. Cardiff
University, Cardiff.
Pham, D.T.; Darwish, A.H.; Eldukhri, E.E.; and Otri, S. (2007). Using the Bees Algorithm to tune a
fuzzy logic controller for a robot gymnast. In Proceedings of Innovative Production Machines and
Systems Virtual Conference. Cardiff, United Kingdom, 2-13 July.
Pham, D.T.; Soroka, A.J.; Ko, E.; Ghanbarzadeh, A. and Otri, S. (2007). Some applications of the
Bees Algorithm in engineering design and manufacture. In Proceedings International Conference on
Manufacturing Automation. Singapore, 28-30 May.
Pham, D.T.; Soroka, A.J.; Ghanbarzadeh, A.; Ko, E.; Otri, S. and Packianather, M. (2006).
Optimising neural networks for identification of wood defects using the Bees Algorithm. In
Proceedings IEEE International Conference on Industrial Informatics . Singapore, 16-18 August.
Pham, D.T., et al. (2006). The Bees Algorithm, A Novel Tool for Complex Optimisation
Problems. In Proceedings
2nd International
Virtual Conference
on
Intelligent Production Machines and Systems. Cardiff, United Kingdom, 3-14 July. Oxford: Elsevier.
Russell, E. and Jarvis, F. (2006). Mathematics of Sudoku II. Mathematical Spectrum 39: 54-58.
Santos-Garcia, G. and Palomino, M. (2007). Solving Sudoku Puzzles with Rewriting Rules. Electronic
Notes in Theoretical Computer Science 176: 79-93.
Yang X.S. (2005). Engineering Optimizations Via Nature-Inspired Virtual Bee Algorithms. Artificial
Intelligence and Knowledge Engineering Applications: A Bioinspired Approach, Lecture Notes in
Computer Science 3562: 317-323. Springer Berlin, Heidelberg.
1235
Biographies
Alper AKIN Eskiehir, 1977. He earned his PhD degree on structural engineering in 2010 at Middle East
Technical University, Ankara, Turkey. His major field of study is structural optimization.
Alper has several papers on the optimum design of concrete structures and he specializes in the engineering
applications of new meta-heuristic optimization algorithms.
brahim AYDODU Antalya, 1980. He earned his PhD degree on structural engineering in 2010 at Middle
East Technical University, Ankara, Turkey. His major field of study is structural optimization.
brahim has several papers on the optimum design of steel structures and he specializes in the engineering
applications of new meta-heuristric optimization algorithms.
1236
Abstract.Cost and duration estimation of construction of a sewage canal, pergola and, germination of an
hospital is performed by a GIS based planning system. To measure the effects of uncertainty of input data,
Monte-Carlo Simulation is implemented. The construction is simulated 1000 times by considering the
variations of meteorological data, productivity of workers and management conditions. A software is
developed in order to perform the cost estimation, scheduling and Monte-Carlo Simulation. Developed
system performs scheduling and quantity take-off if network diagram of the project is provided.
Consequently, the system can assist planning department by shortening the computation duration and
minimizing the mistakes.
Keywords: monte carlo simulation, geographic information systems, project planning
1237
based cost estimation system. Zhong et al. [3] (2004) had simulated a dam construction in order to perform
visualization of construction steps. The system presents optimum equipment set scheme by simulating a variety
of scenarios taking place under different construction conditions. Studies on cost estimation from CAD drawing
can be listed as Bansal and Pal [4] (2006), Bettemir and Snmez [5,6] (2008, 2009). Estimation of uncertainty of
construction cost and duration is studied by Arizaga [7] (2007) and Bettemir [8,9] (2009, 2010).
This study presents a primitive tool for cost and duration estimation of construction activities and the uncertainty
of cost and duration estimation. Developed system, which can reduce the time and cost of preparation of tender
documents, can be beneficial. In addition to this, it is expected to reduce mistakes in calculations and provide
more accurate results by implementing more detailed formulation and making fewer assumptions. Besides the
contractors, the developed system can also be beneficial for governmental institutes, since it is mandatory for
governmental institutes to estimate the cost of the construction before tender.
2 Methodology
Developed GIS embedded system aims to make quantity take-off easier and flawless. Therefore, the system
includes project planning, GIS and construction simulation tools. The system takes construction drawings, soil
data, DEM, weather data, logical relations between the activities and productivity data of workers and
construction equipment as input data. The quantity take-off of earthworks and concreting are performed by
overlay analysis. Obtained quantity take-off data is merged in productivity and unit price data. By using these
data the system executes MCS and predicts cost and duration of construction. Developed system is presented in
Figure 1.
Fig. 1.GIS based system. Square boxes represent data, ellipses output and hexagon tools.
1238
Caterpillar [12] give dependable results for the hourly output of the construction machines. Since productivity
predictions are not exact, they should cover a range which is equal to their significance interval. Uncertainty
analyses involve excavation and compaction activities. Prediction of output of excavator and compactor are
briefly explained. Nunnally (2001) provides the formula to predict the output of an excavator as;
Output = C S V B E
(1)
where, C is the number cycles of the bucket of an excavator made in an hour, S is the depth of excavation and
swing angle of bucket, V is the heaped capacity of the bucket, B is a correction factor for heaped capacity related
with soil type and E is the productivity. Compaction of soil is predicted by the following formula;
Output (m3 /h)=
10 W S L E
P
(2)
where, W is the width compacted per pass in m, S is the compactor speed in km/h, L is compacted lift thickness
in cm, E is job efficiency and P is the number of passes required. Compaction of soils mainly depends on the
moisture content of the soil. For this reason, in case of a rainy weather the number of required passes
significantly changes. The relationship between the moisture content and number of passes can be predicted by
compaction charts.
DEM, project drawings, geological data are overlaid for the determination of amount of earthwork. Overlay
analyzes gives not only the volume of excavation but also volumes of soil classes to be excavated.
Activity durations are estimated by dividing the man-hour and machined-hour requirements of work items to
assigned resources. If more than one resource is assigned to an activity, duration of the activity will be the
maximum of the completion days.
When an activity starts is determined by CPM based on the predetermined logical relationships between the
activities. Crew sizes, number of crews, type and number of construction machines assigned for the activity are
kept constant and they are not optimized. The CPM gives the duration of construction and the direct cost of the
construction is obtained by man-hours, machine-hours and material costs.
(3)
where EWC is efficiency of working conditions, and rand() is a normally distributed random number between 0
and 1. Site conditions mainly depend on weather conditions. Because of this, to determine suitability of site
conditions, it is necessary to simulate weather conditions. According to weather statistics the probability of rain
and heavy rain is determined. Then a random number between 0 and 1 is generated. If the random number is
smaller than probability of heavy rain, heavy rain is assigned for that day. If the random number is bigger than
probability of heavy rain but smaller than probability of rain, that day is assigned as rainy day. If the random
number is bigger than probability of rainy day no rain is assigned for the corresponding day. Site condition is
determined based on the weather conditions. Random number correlated with weather is generated by the
following formula;
1239
(4)
where rand()W is a random number generated for the weather conditions. The second random number is
generated for other weather effects such as wind and temperature. If there is heavy rain ESITE is kept constant, if
there is a mild rain ESITE is increased by 0.25 and if there is no rain ESITE is increased by 0.50. The overall
efficiency of site conditions is determined by;
E= E SITE E WC
(5)
It will not be proper to assign the same efficiency to all type of workers. There must be slight deviations between
the efficiencies of workers. This is realized by a random variable between 0.95 and 1.05 which is generated for
each type of workers. Efficiency of ith worker type is determined by;
Ei = E * [0.95 + 0.1*randi()]
(6)
Random numbers generated for the efficiencies of the worker types should be uncorrelated. Control of
correlation within a sample in MCS is very important. Although sets are paired randomly, there can be
unenviable correlation in the sample set, S. Correlation coefficient matrixes of the input parameters, D, and the
generated sample set, E, should be identical (Helton et al. 2005) [13]. Expected value and the variance of y are
estimated by (7) and (8);
m
E y
i 1
yi
m
V y
y
i 1
(7)
E y
m 1
(8)
3 Case Study
Utility construction of an ongoing hospital construction is used as case study. Small portion of the project, which
consists of 215 meters of sewage water canal, two pergolas, and 5000 m2 of grading and germination, is
analyzed. Some project data are mutilated since the construction is ongoing. The project contains 31 activities
which include 20 resources and materials. Network diagram is shown in Figure 2.
Meteorological data of previous years, DEM, digital drawings of the project, geological data of the region
including the standard penetration tests are loaded. DEM of the region, geological data and drawings of
excavation are overlaid and quantity take-off of excavation is performed. Cost and productivity data from
Ministry of Public Works and Settlement, meteorological data from Turkish State Meteorological Service, and
geological data are obtained from Yuzuncu Yil University [14,15].
Figure 2 illustrates logicalrelationships between activities. Relationships are Finish-to-Start without lag. It is
assumed that in a day 8 hours is worked. Scheduling is performed by considering stochastic variables. Random
parameters are used to model weather, management conditions and productivity of workers. Plain worker,
operator, foreman, steel workers, carpenter, concrete worker, oiler and machinist are the worker types employed
for the construction.
1240
Frequency
300
200
100
0
26
27
28
29
30
31
32
33
34
35
36
37
38
Duration (Day)
Fig. 3.Histogram of construction duration. Construction is simulated 1000 times by generating random numbers
for each parameter. Base value of the construction cost and duration is estimated as 48080.58 TL and 31.7 workdays or 246.13 work-hour. Uncertainty of the base value in %68 confidence (1-sigma) is computed as 972.35 TL
and 1.6 days or 11.7 work-hours.
It is seen that the variation of machine-hour is significantly less than variation in man-hour. In this project,
excavation process is critical on the capacity of excavator only, since there is not hauling. Fluctuations of the
machine-hour of the excavator are only related with itself. However, besides his efficiency, the man-hour of
plain worker is affected from the efficiency of gardeners, timber workers and steel worker. As a result of this,
largest variance is observed in plain workers. Construction simulation is a proper tool to reveal such
relationships and uncertainties. The risks of the schedule are more properly determined and optimum crew size
can be employed.
1241
12
Frequency
10
8
6
4
2
0
1900
2000
2100
2200
2300
2400
2500
2600
2700
2800
300
Frequency
250
200
150
100
50
0
28
30
32
34
36
38
40
42
Excavator machine-hour
4 Conclusion
The study is an endeavor for a possible implementation of information technology in construction. The whole
simulation took about 3 seconds on a computer with i7 CPU. However, quantity estimation could not be
performed automatically since the drawings were not prepared accordingly. Annotation was not accurate and
well prepared enough for an automatic detection by computer. As a result of this, the sewage water canals,
pergola and grading and germination drawings are manually extracted. However, cost and duration estimation
are determined by the system without human endeavor. The system can be a good candidate for an auxiliary
module for the planning department. The most important benefit of the system is the reduction of computations
for the preparation of the tender documents. As a result of this, possibility of making mistakes will be
significantly reduced. In addition to this, uncertainty of the proposed cost and duration of the tender document,
thus the risk would be more or less estimated.
This study is supported by the Scientific Research Project Fund 2010 MIM B023.
References
1.
2.
3.
4.
M. Y. Cheng, and J. T. O' Connor, (1996). ArcSite: Enhanced GIS for construction site layout. J of Const
Eng. Man., vol. 122(4), pp. 329 336.
Min-Yuan Cheng, and Shin-Ching Yang, (2001). GIS-Based cost estimates integrating with material layout
planning. J of Const Eng. Man. vol 127(4), pp. 291 299.
Denghua Zhong, Jingru Li, Huirong Zhu, and Lingguang Song, (2004). Geographic information systembased Visual simulation methodology and its application in concrete dam construction processes. J of Const
Eng. Man, Vol. 130(5), pp. 742 750.
V.K. Bansal and Mahesh Pal, 2007, Potential of geographic information systems in building cost
estimation and visualization, Aut in Const Vol. 12 pp. 311 322.
1242
5.
6.
7.
8.
9.
10.
11.
12.
13.
14.
15.
. H. Bettemir, and R. Sonmez, (2008). Automated cost estimation of dam projects with geographic
information systems. 8th International Congress on ACE, 15 17 September, EMU, Famagusta, North
Cyprus, Volume 2, pp. 557 561.
. H. Bettemir, and R. Sonmez (2009), Estimation of building costs using geographic information systems.
5th international conference on construction in the 21st century (CITC-V) May 20 22, Istanbul, Turkey.
O. J. F. Arizaga, (2007). A methodology for Project risk analysis using bayesian belief networks within a
Monte Carlo simulation environment. PhD, Univ. of Maryland.
. H. Bettemir, (2009), Kaz ve Hafriyat Srelerindeki Belirsizliin Monte Carlo Analizi ile Tahmini, YYU
FBE Dergisi, Cilt: 14(2) pp. 165 173.
. H. Bettemir (2010). Corafi Bilgi Sistemleri ve Yapm Ynetimi Dalnda Uygulama Alanlar. 1. Proje ve
Yapm Ynetimi Kongresi 29 Eyll 1 Ekim, Ankara.
G. Suhanic, (2001). Computer-Aided Project Management, 1st ed, Oxford University Press, Inc. New York.
S. W. Nunnally, (2001). Construction Methods and Management. 5 th Edition, Prentice Hall Ohio.
Caterpillar, (1979). Fundamentals of Earthmoving. Caterpillar Tractor CO.
J.C. Helton, F. J. Davis, J. D. Johnson (2005). A Comparison of uncertainty and sensitivity analysis results
obtained with random and Latin hypercube sampling, Reliability Engineering and System Safety, vol 89,
305 330.
Web reference, Turkish State Meteorological Bureu. Retrieved May 1, 2011 from the World Wide Web,
http://www.dmi.gov.tr/veridegerlendirme/il-ve-ilceler-istatistik.aspx?m=VAN.
2010 unit price indexes of ministry of public works and settlement
Biographies
nder Halis Bettemir Bursa 1979. The author graduated from Middle East Technical University as Civil
Engineer in 2003. He obtained his Master of Science degree from geodesy and photogrammetry division in 2006
in METU. In 2009 he obtained his PhD degree in METU from construction management division. The author
works as Assistant Professor at Yznc Yl University at the department of Civil Engineering.
Photogrammetry, GIS, remote sensing, inertial navigation systems, machine control, construction simulation and
optimization are the study subjects of the author.
Dr. Bettemir is a member of chamber of civil engineers
1243
1,2,3
zgl LHAN,
Niyazi AHN, Mehmet SEZER
Department of Mathematics, Faculty of Science, Mugla University, Mugla, Turkey
oilhan@mu.edu.trnisa70@mu.edu.trmsezer@mu.edu.tr
Abstract. A matrix method for approximately solving certain linear Fredholm integro-differential equations is presented in
this paper. This method converts a linear Fredholm integral equations into a matrix equation. Then, the equation reduces to a
matrix equation corresponding to a system of linear algebraic equations with unknown Morgan-Voyce coefficients. This
method reduces to the analytic solution if the exact solution is polynomial. Numerical examples are included to demonstrate
the applicability of the technique.
Keywords: Morgan-voyce polynomials, linear fredholm integro-differential equations, collocation method
1 Introduction
The linear integral equations and their solutions play a substantial role in science and engineering problems. A
physical event can be modeled by a differential equation, an integral equation or an integro-differential
equation. Since some of these equations can not be solved explicitly, it is often necessary to apply some
numerical techniques which are appropriate combinations of numerical integration and interpolation [1,2]. The
solutions of the integral equations occuring in physics, biology and engineering are based on numerical methods
such as Euler-Chebyshev and Runge-Kutta methods and also in a recent research, the systems of integral and
integro differential equations were solved using an efficient algorithm [3], The Haar functions method [4,5], the
differential transformation method [6], the Variational iteration method [7], the Legendre matrix method [8]. In
addition, the solutions of systems of linear differential and integro differential equations have been found using
the Chebyshev polynomial method and the Taylor collocation method by Sezer (et al) [9,10,11]. Also, the
Morgan-Voyce matrix method has been used to find the approximate solutions of differential, integral and
integro-differential equations.
In this paper, we consider the Fredholm integro-differential equations of the form
b
f
k 0
( x) y ( x) g ( x) K f ( x, t ) y (t )dt ,
a x, t b
(k )
(1)
under the mixed conditions
m 1
[a
k 0
jk
y ( k ) (a) b jk y ( k ) (b)] j ,
j 0,1, 2,
, m 1
(2)
j are stable constants. The constant , the kernel function K f ( x, t ) and the
are defined in the interval given in (1). As herein given, we assume that K f ( x, t ) can be
expandable to Maclaurin series. We want to find an approximate solution of (1) expressed in the truncated
Morgan-Voyce series of the form
1244
y ( x) an Bn ( x)
n 0
(3)
an , n 0,1,..., N are the unknown Morgan-Voyce coefficients and Bn ( x), n 0,1,..., N are the MorganVoyce polynomials formed by [12]
n k 1 k
Bn ( x)
x , n .
k 0 n k
n
D( x) g ( x) I f ( x),
a x, t b
(4)
I f ( x) K f ( x, t ) y(t )dt
a
D( x) f k ( x) y ( k ) ( x).
k 0
Then, we write the solution expressed by (3) and its derivative in the matrix forms as,
(5)
where
B( x) [ B0 ( x) B1 ( x) B2 ( x) ... Bn ( x)]
A a0 a1 a2 aN .
T
On the other hand, we can write a matrix equation for the derivative relation,
B(1) ( x) B( x)T T
where
0
1
0
T
(6)
0 0
0 0
2 0
0
0
0
0
0
0
( N 1) 0
0
N
0 0
0 0
0
0
0
Further, we see that the relation between the matrix B(x) and its derivative
B(1) ( x) B( x)T T
B(2) ( x) B(1) ( x)T T = B( x)(T T )2
(7)
As a result, by substituting the matrix relations (7) into Eq.(5), the matrix relations for y ( x) and
follows
[ y ( k ) ( x)] B(x)(T T )k A
y ( k ) ( x) is as
(8)
1245
To generate a solution in the form (3) of the problem (1), we will use a matrix method based on the MorganVoyce collocation points defined by
xi a (
ba
)i, i 0,1, 2,, N (9)
N
which is called Morgan-Voyce collocation method. When the collocation points (9) are substituted into Eq.(4),
we obtain the system
D( xi ) g ( xi ) I f ( xi ),
D G I f ,
(10)
where
I f ( x0 )
D( x0 )
g ( x0 )
I (x )
D( x )
g(x )
f 1
1
1
D
, G
, I f I f ( x2 ) .
D ( xN )
g ( xN )
I f ( xN )
Let us reduce the part D( x) to the matrix form via the collocation points (8). The matrix D defined in Eq.
(10) is written as:
m
D M kY ( k ) ,
(11)
k 0
where
T
(k )
(k )
(k )
y ( k ) ( xN ) .
Mk diag ( f k ( xi )), i 0,1,..., N Y y ( x0 ) y ( x1 )
When we plug in the collocation points xi , i 0,1, 2,...N in the relation (8), we have the system of equations
Y ( k ) B( x0 ) B( x1 )
B( xN ) (T T )k A B(T T )k A ,
T
(12)
where,
B0 ( x0 ) B1 ( x0 )
B (x ) B (x )
1 1
B= 0 1
B0 ( xN ) B1 ( xN )
BN ( x0 )
BN ( x1 )
.
BN ( xN )
Consequently, from the matrix relations (11) and (12), the fundamental matrix relation for the differential part
D( x) is
m
D Fk B(T T )k A .
(13)
k 0
Now, we obtain the matrix relation for the Fredholm integral part
b
K ( x, t ) kmn
Bm ( x) Bn (t )
(14)
m0 n 0
t
K ( x, t ) kmn
x mt n
(15)
m0 n 0
where
1246
t
kmn
1 m n K (0, 0)
;
m!n ! x m x n
m, n 0,1,...N
Firstly, we convert the expressions (14) and (15) to matrix forms and then equate them as
[ K ( x, t )] B( x) K b BT (t ) X ( x) K t X T (t )
(16)
where
B( x) [ B0 ( x) B1 ( x) B2 ( x) ... Bn ( x)]
X ( x) [ x0 x1 x 2
xN ]
b
t
, Kt kmn
;
Kb kmn
m, n 0,1,..., N
On the other hand, using the Morgan-Voyce formula (2) for n 0,1,..., N we get the matrix relation
BT ( x) RX T ( x) or B( x) X ( x) RT ,
(17)
where
1
0
0
0
2
3
0
1
0
R 3
4
5
2
1
0
n 1
n 2 n 3
n n 1 n 2
2n 1
0
0
Further, substituting B( x) in (17) into the equation (16) and simplifying the result give us the matrix relation
Kb ( R1 )T K t R1
(18)
that is the relation between the Morgan-Voyce and Taylor coefficients of the Kernel function K ( x, t ) .
By substituting the matrix forms (16) and (15) corresponding to the functions K ( x, t ) and y (t ) into the
Fredholm integral part
I f ( x)
(19)
where
b
I f ( xi ) B( xi ) K bQA; i = 0,1,...,N
or shortly the matrix relation
I f BK bQA,
(20)
that is the fundamental matrix relation for the Fredholm integral part
[a
k 0
jk
3. Method of solution
1247
(21)
To construct the fundamental matrix equation defined in (1), we substitute the matrix formulas (13) and (20) into
(10) and simplify this formula. Thus, we get the matrix equation,
T k
(22)
Fk B(T ) BK bQ A G
k 0
which equals a system of ( N 1) algebric equations for the ( N 1) unknown Morgan-Voyce coefficients
a0 , a1 ,..., aN . Eq. (22) can be written in the form
WA = G
or
W ; G ,
(23)
where
m
G g ( x0 ) g ( x1 )
g ( xN )
On the other hand, we can write for the conditions (2) the matrix form (21) as
U j A = j or U j ; j j 0,1,..., m 1
(24)
where
m 1
Under the conditions (2) to obtain the solution of Eq. (1), we replace the last m rows of the matrix (23) by the
rows matrices (24) and we get the new augmented matrix,
w00
w
10
w( N 1 m )0
[W ; G ] w( N m )0
u00
u
10
u
( m 1)1
If
w01
w0 N
w11
w1N
;
;
w( N 1 m )1
w( N 1 m ) N
w( N m )1
u01
w( N m ) N
u0 N
;
;
u11
u1N
;
;
u( m 1)2
u( m 1) N
g ( x0 )
g ( x1 )
g ( xN 1 m )
g ( xN m )
0
1
m 1
(25)
A (W )1G .
an (n 0,1,..., N ) by means of Eq.(25). Besides, via system
(23), some particular solutions may be obtained. If 0 in Eq.(2), the equation becomes the higher-order
linear differential equation; if Bk ( x) 0 for k 0 , the equation becomes the Fredholm integral equation.
Thereby, we uniquely determine the coefficients
4. Accuracy of solution
We now check the accuracy of the method. The truncated Morgan-Voyce polynomial series in (3) must be
approximately satisfying Eq. (1). For each xi [a, b]
E ( xi ) D( xi ) g ( xi ) I f ( xi ) 0
or
E ( xi ) 10 ki
1248
If
max(10 ki ) 10 k
EN ( x) M kYN( k ) ( x) g ( x) I f ( x) .
k 0
5. Numerical Examples
In this section, we give several numerical examples for showing the accuracy and the efficiency of the presented
method. We performed all of these examples on the computer, for that we used a program written in Matlab
v7.11.0584 (R2010b).
0 x, t 1
under the initial conditions y(0) 1 and y(0) 1 and approximate the solution y ( x) by the truncated
Morgan-Voyce series
3
y ( x) an Bn ( x) so that f0 ( x) x, f1 ( x) x, f 2 ( x) 1, g ( x) e x -sinx
n 0
x0 0, x1 , x2 , x3 1 .
3
3
and for
T k
M k B(T ) BK bQ A = G ,
k 0
where
1 2
M 0 diag (0, , , 1)
3 3
1
1
B
1
7
3
8
3
3
40
9
55
9
8
1749
G 1
1637
1 2
M1 diag (0, , ,1) , M 2 I 44 ,
3 3
4
0
217
27 T 0
,
T
0
368
27
0
21
1
0
0
0
0 0
1
2 0
, K b
0 3
0
0 0
1
6
1
2
0
1
12
0
1
6
0
1
6
0
1
36
335 259
252 138
For the conditions y(0) 1 and y(0) 1 the augmented matrix is
U 0 ; 0 1 2 3 4 ; 1
U ; 0 1 4 10 ; 1 .
1
1
So the augmented matrix with the conditions is
1249
0
697
W ; G 1296
0
997
1063
2
1
2
583
544
3
4
1
1749
;
1637
;
1
;
1
12
2833
201
4
10
ytam ( x) e x . Similarly taking N 7 and N 10 , we get the approximate solution. Results are shown in
Figure 1.
Results for N=3,N=7,N=10 and exact solution
2.8
2.6
2.4
y-axis
2.2
2
1.8
1.6
1.4
1.2
1
0.1
0.2
0.3
0.4
0.5
x-axis
0.6
0.7
0.8
0.9
(26)
Taking N 4, N 8 and N 11and proceeding as the previous example, the approximate solution can be
obtained. Results are shown in Figure 2 and Table 1.
Results for N=4,N=8,N=11 and exact solution
25
20
y-axis
15
10
0.5
1.5
x-axis
2.5
1250
xi
0
0.2
0.4
0.6
0.8
1.0
1.2
1.4
1.6
1.8
2.0
2.2
2.4
2.6
2.8
3.0
y ( x) e i 2
3
3.22140275816
3.49182469764
3.82211880039
4.22554092849
4.71828182846
5.32011692274
6.05519996685
6.95303242440
8.04964746441
9.38905609890
13.0231763806
13.0231763806
15.4637380350
18.4446467711
22.0855369232
x
N 4, E ( xi )
N 8, E ( xi )
N 11, E ( xi )
0
0.34894356753
0.61089716802
0.79467731353
0.90778695237
0.95694747345
0.94874849774
0.89044153544
0.79090936098
0.66185001038
0.51922291877
0.38501523692
0.28939921495
0.27336723687
0.39195025898
0.71814881938
0
0.00035112808
0.00058914310
0.00073669150
0.00081353871
0.00083606103
0.00081754399
0.00076914088
0.00070045936
0.00061913747
0.00053072451
0.00044404353
0.00039340338
0.00049806317
0.00109186371
0.00297269410
0.023e-006
0.119e-006
0.191e-006
0.245e-006
0.278e-006
0.296e-006
0.305e-006
0.309e-006
0.305e-006
0.294e-006
0.284e-006
0.274e-006
0.262e-006
0.263e-006
0.097e-006
1.953e-006
References
1 C.T.H. Baker, A perspective on the numerical treatment of Volterra equations J. Comput
Appl. Math. 125 (2000) 217- 249
2. P. Linz, Analytical and Numerical Methods for Volterra Equations SIAM, Philadelphia,
PA, 1985.
3 E. Yusufolu, An efficient algorithm for solving integro-differential equations system ,
Appl. Math. Comput. 192 (2007) 51-55.
4 K. Maleknejad, F. Mirzaee, Numerical solution of linear Fredholm integral equations system
by rationalized Haar functions method , Int. J. Comput. Math. 80 (2003) 1397-1405.
5. K. Maleknejad, F. Mirzaee, S. Abbasbandy, Solving linear integro-differential equations
system by using rationalized Haar functions method , Appl. Math. Comput.
155 (2004) 317-328.
6. A. Arikolu, I. Ozkol, Solutions of integral and integro-differential equation systems by
using differential transform method , Comput. Math. Appl. 56 (2008) 2411-2417.
7. J. Saberi-Nadjafi, M. Tamamgar, The variational iteration method: A highly promising
method for solving the system of integro differential equations ,, Comput. Math. Appl. 56
(2008) 346-351
8. H. H Sorkun, S. Yalnbas, Approximate solutions of linear Volterra integral equation systems
with variable coeffients ,Appl. Math. Modell. (2010), doi: 10. 1016/j.apm.2010.02.034.
9. A. Akyz-Daolu, M. Sezer, Chebyshev polynomial solutions of systems of higher-order linear
Fredholm-Volterra integro-differential equations J. Franklin Ins. 342 (2005) 688-701.
10. M. Glsu, M. Sezer, Taylor collocation method and for solutions of systems of linear differential
equations with variable Coefficients,Intern. J. Computer Math. 83 (2006) 429-448.
11. . Yzba, N. ahin, M. Sezer, Numerical solutions of systems of linear Fredholm integro-differential
equations with Bessel Polynomial bases, Computers&Mathematics with Applications, papers 3079-3096,
22 April 2011
12. M.N.S. Swamy,Further properties of Morgan-Voyce Polynomials Fibonacci Quarterly, Vol. 6, No. 2,
Apr. 1968, pp. 167-175
1251
Biographies
zgl LHAN - Research Assistant zgl LHAN was born on January 01th 1985 in Tokat, Turkey. In 2008,
she received her B. Sc. degree from the Department of Mathematics of Mula University in Mula and she is
currently studying her M.Sc. in the same department under the supervision of Dr. Niyazi AHN. She has
written one paper so far, and it is currently under review.
Niyazi AHN - Dr. Niyazi AHN was born on March 15th 1970 in zmir, Turkey. In 1991, he received his
B.Sc. degree from the Department of Mathematics of Ege University in zmir, and M.A. from the University of
Pittsburgh, USA in 1996. He then studied computational fluid dynamics for his Ph.D. degree under the
supervision of Prof. William Layton in the Department of Mathematics at the University of Pittsburgh, and
graduated from that school in 2003.
His main research interests are computational fluid mechanics, numerical linear algebra and numerical
solutions of ordinary and partial differential equations. Some of his published articles are:
Yzba, ., ahin, N.; and Sezer, M. Numerical solutions of systems of linear Fredholm integrodifferential equations with Bessel polynomial bases, Comput. Math. Appl. 61(10): 30793096, 2011.
Akgnll, N., ahin, N.; and Sezer, M. A Hermite collocation method for the approximate solutions
of high-order linear Fredholm integro-differential equations,Numer Methods Partial Differential Eq.
2010, DOI: 10.1002/num.20604, 2010.
John V., Layton W. and Sahin N., Derivation and Analysis of Near Wall Models for Channel and
Recirculating Flows, Comput. Math. Appl., 48(2004): 11351151.
Mehmet SEZER - Prof. Dr. Mehmet SEZER was born on March 20th 1954 in Antalya,Turkey. He received his
B.Sc. from the Department of Mathematics of Ege University in 1976 and was research assistant at Balkesir
Engineering Academy. His M.Sc. was from Ege University in 1977. He received his Ph.D. under the supervision
of Prof.Dr.Seda MORALI in the Department of Applied Mathematics at Ege University in 1982 and studied
in the field of Parabolic partial differential equations.
In 1982, he was Assistant Professor at Faculty of Education (Balkesir) of Uluda University, Associate
Professor in 1989, and in 1995 Professor at Faculty of Education of Dokuz Eylul University. Since 2004 , he has
been Mathematics Professor at Faculty of Science at Mula University and has been the chair in the same
department.
His main research interests are ordinary and partial differntial equations, integral and difference equations , and
their numerical solutions. Prof. Sezer has been reviewer for numerous influential journals, has published
research articles related to differential equations, linear algebra, analytic geometry and calculus; and has been
authored over 80 papers.
1252
Abstract. A numerical method for solving the generalized (retarded or advanced) pantograph equation with constant and
variable coefficients under the mixed conditions is presented. The method is based on the truncated Boubaker series. The
solution is obtained in terms of Boubaker polynomials. The method is illustrated by studying an initial value problem.
Illustrative examples are included to demonstrate the validity and applicability of the technique. The results obtained are
compared by the known results.
Keywords: Pantograph equation, Boubaker series and polynomials, Boubaker matrix method.
1. Introduction
This study is concerned with a generalization of a functional differential equation known as the
pantograph equation (with constant and variable coefficients) which contains a linear functional argument (with
retarded and advanced cases or with proportional delays). Functional differential equations with proportional
delays are usually referred to as pantograph equations or generalized equations. The name pantograph
originated from the work of Ockendon and Tayler on the collection of current by the pantograph head of an
electric locomotive [1, 2].
In recent years, pantograph equations have been studied by many authors, who have investigated both
their analytical and numerical aspects [8-10]. These equations are characterized by the presence of a linear
functional argument and play an important role in explaining many different phenomena. In particular, they arise
in industrial applications [11] and in studies based on biology, economy, control theory, astro-physics, nonlinear
dynamic systems, cell growth and electro-dynamic, among others [1, 2, 11, 12]. In addition, properties of the
analytical and numerical solutions of pantograph equations have been investigated by several authors [1-5, 1317].
The basic motivation of this work, by considering the mentioned studies, is to develop a new numerical
method, which is called as a matrix method, based on Boubaker polynomials [3-7] and collocation points for the
approximate solution of pantograph equation. The mentioned Boubaker polynomials were firstly by Boubaker et
al. (2006) as guide for solving a one-dimensional formulation at heat transfer equation. In this study, we will
consider the generalized pantograph equation with variable coefficients
J m 1
y(m) ( x ) Pjk ( x ) y( k ) ( jx j ) g( x )
(1)
j 0 k 0
1253
c
k 0
ik
where
y( k ) (0) i ,
i 0,1,..., m 1.
(2)
Pjk (x) and g (x) are analytic functions; cik , i , j and j are real or complex constants. The aim of
this study is to get solution of the problem (1)-(2) as the truncated Boubaker series defined by
N
y( x ) n n ( x ) , 0 x b
(3)
n 0
are defined by
0 ( x ) 1, 1 ( x ) x, 2 ( x ) x 2 2,
for : m 2
m ( x ) xB m1 ( x ) Bm2 ( x )
Later, a monomial definition of these polynomials was established by Labiadh et al. [7]:
(n)
( n 4p) p
n ( x )
Cn p .(1) p .x n 2 p
p 0 ( n p)
(4)
2. Fundemantel Relations
Let us consider the pantograph equation (1) and find the matrix forms of each term in the equation. First we
can convert the solution y( x ) defined by the truncated series (3) and its derivative
y(x) (x)
y( k ) ( x ) ( k ) ( x ) ,
and
k 0,1,2,...
y( k ) ( x ) to matrix forms
(5)
where
X(x) [1 x x N ]
and
0
0
0
0, 0
0
1,0
0
0
2,1
0
2, 0
0
Z
2 N , N
0
2 N , N 1
0
2 N 1, N
0
2 N 1, N 1
0
2 N,0
where
(n 4 p)
q , p
Cnp p (1) p , q 0,1,, N .
(n p)
( k ) ( x ) X( k ) ( x ) Z T .
To obtain the matrix
(7)
(k)
1254
(k)
(k 1)
X (x) X (x)M X(x)Mk (8)
where
0 1 0 0
0 0 2 0
M .
0 0 0 N
0 0 0 0
Similarly, we have the matrix relations between
X( k ) ( jx j ) and X( x ) as follows:
X( jx j ) [1 ( jx j ) ( jx j ) N ] X(x)A( j , j )
(9)
X( k ) ( jx j ) X( jx j )Mk (10)
and from (9) and (10)
0 0 0
( j ) ( j )
0
A( j, j)
0
1 0 1
( j ) ( j )
0
1 1 0
( j ) ( j )
1
2 0 2
( j ) ( j )
0
2 1 1
( j ) ( j )
1
2 2 0
( j ) ( j )
2
N
( j ) 0 ( j ) N
0
N 1 N1
( j ) ( j )
1
N 2 N2 .
( j ) ( j )
2
N N 0
( j ) ( j )
N
Consequently, by substituting the matrix forms (7) and (8) into (5), we have matrix relations
y( k ) (x) X(x)Mk ZT
(12)
y( k ) ( jx j ) B( k ) ( jx j ) X( k ) ( jx j )ZT
X(x)A( j , j )Mk ZT , k 0,1,2,... (13)
3. Method of Solution
We now ready to construct the fundamental matrix equation corresponding to Eq.(1). For this purpose,
substituting the matrix relations (12) and (13) into Eq.(1) and then simplifying, we obtain the matrix equation
J m 1
X( x )M m ZT Pjk ( x ) X( x ) A( j , j )M k ZT g( x ). (14)
j 0 k 0
x i defined by x i
b
i, i 0,1,..., N , we get the system of matrix
N
equations
1255
X( x i )M m ZT Pjk ( x i ) X( x i ) A( j , j )M k ZT g( x i ) i 0,1,..., N
j 0 k 0
m T
XM
Z
Pjk XA ( j , j )M k ZT G (15)
j 0 k 0
where
0
0
Pjk ( x 0 )
0
Pjk ( x1 )
0
0
Pjk ( x 2 )
Pjk 0
0
0
0
X( x 0 ) 1 x 0
X( x )
1
1 x1
X X( x 2 ) 1 x 2
X( x N ) 1 x
N
x0
x1
x2
xN
g( x 0 )
g( x )
1
, G g( x 2 )
g( x N )
Pjk ( x N )
N
x0
N
x1
N
x 2 .
N
x N
0
0
0
W A G or [ W; G] , W [w pq ], p, q 0,1,..., n
(16)
where
J m 1
W [ w pq ] XM m ZT Pjk XA ( j , j )M k ZT
j 0 k 0
Here Eq.(16) corresponds to system of (N+1) linear algebraic equations with the unknown Boubaker coefficients
0 , 1 ,..., N . On the other hand, we can obtain the matrix forms for the conditions (2), as follows:
m1
c
k 0
ik
where
Finaly, to obtain the solution of Eq.(1) under the conditions (2), by replacing the row matrices (17) by the last m
rows of the matrix (16), we have the new augmented matrix [15-17]
w 00
w
10
...
w N m , 0
[ W; G ]
u 00
u10
...
u m1, 0
If
w 01
w 0N
w11
...
w Nm ,1
u 01
.
.
.
.
.
.
w 1N
...
w N m , N
u0N
;
;
;
;
u11
...
u m1,1
u1N
...
u m1,N
;
;
;
g( x 0 )
g( x1 )
...
g ( x N m )
0
1
...
m1
(18)
(W) 1 G
1256
Thus the matrix A (thereby the coefficients 0 , 1 ,..., N ) is uniquely determined. Also the Eq.(1) under the
conditions (2) has a unique solution. This solution is given by truncated Boubaker series (3). On the other hand,
when
W 0,
if
x x q [0,1], q 0,1,2,...
J m 1
E( x q ) y(m) ( x ) Pjk ( x ) y( k ) ( jx j ) g( x ) 0
j 0 k 0
and E( x q ) 10
k q
truncation limit
N is increased until the difference E( x q ) at each of the points becomes smaller than the
kq
prescribed 10 k . On the other hand, the error can be estimated by the function
J m 1
E N ( x ) y(m) ( x ) Pjk ( x ) y( k ) ( jx j ) g( x )
j 0 k 0
If E N ( x ) 0 , when
5. Illustrative Examples
In this section, several numerical examples are given to illustrate the accuracy and effectiveness
properties of the method and all of them were performed on the computer using a program written in Maple12.
The absolute errors in Tables are the values of
yx
3
x
yx y x 2 2 , y0 0 , y0 0 , 0 x 1
4
2
After the ordinary operations and following the method in Section 3, we obtain yx x which is the exact
solution.
2
1
x 1
yx e x / 2 y yx , y(0) 1 , 0 x 1
2
2 2
x
which has the exact solution yx e .
Hence, the computed results are compared with other methods [19, 20, 21] in Table 1.
Table 1. Error analysis of Example 2 for the x value
x
0.2
0.4
0.6
0.8
1
SplineFnk.
Approx[21]
0.198 E-7
0.473 E-7
0.847 E-7
0.135 E-6
0.201 E-6
Taylor
Method[20]
0.100E-8
0.000000
0.100000
0.000000
0.500E-8
ADM
13terms[19]
0.000000
2.22 E-16
2.22 E-16
1.33 E-15
4.88 E-15
Present
Method (N=20)
1.00 E-14
2.00E-14
5.00E-14
2.00 E-14
7.80E-13
1257
Here 1 (x) e 0.5x sin(0.5x) , 2 (x) 2e 0.75x cos(0.5x) sin(0.25x) . It can be seen that the exact solution of
this problem is y(x) e
Exact
Solution
0.0
0.2
0.4
0.6
0.8
1.0
1.000000
0.802410
0.617405
0.452953
0.313050
0.198766
Sezer-Yalnbaahins Method
[15] (N=9)
0.000000
1.3 E-9
1.434 E-7
2.058 E-6
1.212 E-5
4.003 E-5
Present Method
N=9
N=12
0.000000
0.000000
2.00E-10
0.000000
1.00E-10
2.70E-9
0.000000
1.00E-10
4.00E-10
0.000000
2.00E-10
2.45E-8
5. Conclusions
A new matrix method based on the Boubaker polynomials is developed to numerically solve higher
order pantograph equations and multi-pantograph equations with the initials conditions. Comparison of the
results obtained by the present method and those other methods reveals that the present method is vary effective
and convenient. Numerical results show that the accuracy improves when the truncation limit of Boubaker series
is increased and the errors decrease more rapidly.
It is observed that the method has the best advantage when the known functions in equation can be
expanded to Boubaker series. Another considerable advantage of the method is that Boubaker coefficients of the
solution function are found very easily by using the computer programs. Moreover, this method is applicable for
the approximate solution of the pantograph-type Volterra and Fredholm functional integro-differential equation
with variable delays, higher-order differential-difference and integro-differential-difference equations.
References
[1] J.R.Ockendon and A.B.Tayler, (1971). The dynamics of a current collection system for an electric locomotive,
Proc.Roy.Soc.London Ser., A322 447-468.
[2] A.B. Taylor, (1986) Mathematical Models in Applied Mathematics, Clarendon Press, Oxford, pp. 40-53.
[3] K. Boubaker, (2007). Trends Appl. Sci. Res. 2, 540544.
[4] J. Ghanouchi, H. Labiadh, K. Boubaker, (2008). Int. J. Heat Technol. 26 (1), 4953.
[5] O.B. Awojoyogbe, K. Boubaker, (2009). Curr. Appl. Phys. 9, 278283.
[6] K. Boubaker, (March2007), Les Polynmes de Boubaker,Une classe polynomiale qui mane d'un essai de rsolution de
l'quation de la chaleur,Deuximes Journes. Mditerranennes de Mathmatiques Appliques JMMA02, Monastir,
TUNISIE.
[7] H. Labiadh and K. Boubaker, (2007) , A Sturm-Liouville shaped characteristic differential equation as a guide to establish
a quasi-polynomial expression to the Boubaker polynomials, Diff. Eq. and Cont. Proc., 2 pp.117-133.
[8] S.Slama, J. Bessrour, K. Boubaker, M. Bouhafs, (2008). COTUME 2008, pp. 7980.
[9] M. Z. Liu and D. Li, (2004). Properties of analytic solution and numerical solution of multi-pantograph equation, Applied
Math. and Computation 155, 853-871.
[10] G. Derfel and A. Iserles, (1997). The pantograph equation in the complex plane, J. Math. Anal. Appl. 213, 117-132.
[11] Y. Muroya, E. Ishiwata and H. Brunner, (2003). On the attainable order of collocation methods for pantograph integrodifferential equations, J. Comput. Appl. Math. 152, 347-366.
[12] W. G. Ajello, H. I. Freedman and J. Wu, (1992). A model of stage structured population growth with density depended
time delay, SIAM J. Appl. Math. 52, 855-869.
[13] M. D. Buhmann and A. Iserles, (1993). Stability of the discretized pantograph differential equation, Math. Comput. 60,
575-589.
[14] D.Li, M.Z.Liu, (2000). Runge-Kutta methods for multi-pantograph delay equation, App. Math. and Comput.163 (1),
383-395.
1258
Biographies
Mehmet SEZER - Prof. Dr. Mehmet SEZER was born on March 20th 1954 in Antalya,Turkey. He received
his B.Sc. from the Department of Mathematics of Ege University in 1976 and was research assistant at Balkesir
Engineering Academy. His M.Sc. was from Ege University in 1977. He received his Ph.D. under the supervision
of Prof.Dr.Seda MORALI in the Department of Applied Mathematics at Ege University in 1982 and studied
in the field of Parabolic partial differential equations.
In 1982, he was Assistant Professor at Faculty of Education (Balkesir) of Uluda University, Associate
Professor in 1989, and in 1995 Professor at Faculty of Education of Dokuz Eylul University. Since 2004 , he has
been Mathematics Professor at Faculty of Science at Mula University and has been the chair in the same
department.
His main research interests are ordinary and partial differntial equations, integral and difference equations,
and their numerical solutions. Prof. Sezer has been reviewer for numerous influential journals, has published
research articles related to differential equations, linear algebra, analytic geometry and calculus; and has been
authored over 80 papers.
Salih YALINBA -Assc. Prof. Dr. Salih YALINBA was born on July 1st 1969 in Izmir, Turkey. He
received his B.Sc. from Department of Mathematics at Dokuz Eyll University, Buca Faculty of Education in
1990. His M.Sc. was from Dokuz Eyll University in 1994. He received his Ph.D. under the supervision of Prof.
Dr. Mehmet Sezer in the Department of Applied Mathematics at Dokuz Eyll University in 1998 and studied in
the field of Taylor Polynomial Solutions of Volterra-Fredholm Integral and Integro-Differential Equations.
Between 1998-2005, he was Assistant Professor at Faculty of Science (Isparta) of Sleyman Demirel
University. Since 2005, he has been continuing at Faculty of Science at Celal Bayar University. This year, He
became Associate Professor.
His main research interests are ordinary and partial differential equations, integro- differential equations, delay
and neutral differential equations, integral and difference equations, and their numerical solutions. Assc. Prof.
Dr. Yalnba has been reviewer for numerous influential journals, has published research articles related to
differential equations, integral equations and delay and neutral differential equations; and has been authored over
30 papers.
Tue AKKAYA Shewas born on July 9th 1987 in Izmir, Turkey.She graduated from Mathematics
Department at Celal Bayar University in Manisa, Turkey in 2008. She is currently MSc student under the
supervision of Assc. Prof. Dr. Salih YALINBA in Applied Mathematics at the same university and has been
studying in the field of Approximate Calculation of Sturm-Liouville Problems Using Orthogonal Polynomials.
Her main research interests are nonlinear differential equations, integral and integro - differential equations,
difference equations and their numerical solutions. She has published articles in national and international
journals.
1259
Abstract. In this paper, we introduce a digital bundle, a digital fiber bundle, a digital trivial fiber bundle and
define a digital bundle map between digital fiber bundles and an equivalence relation on digital fiber bundles
with the assistance of digital fiber bundle maps. Furthermore, we give a definition of a digital pullback bundle
and study the homotopy invariance of digital pullback bundles. As a result we obtained some conclusions.
Keywords: digital bundles, digital homotopy groups, fiber bundles
1 Introduction
Digital Image Processing is a growing discipline with applications in industry, medicine, meteorology,
geology, among other fields. Digital Topology deals with properties of two and three dimensional digital images
that correspond to topological properties of objects. The main purpose of digital topology is to study topological
properties of discrete objects and it plays an important role in computer vision, image processing and computer
graphics. Researchers (Rosenfeld, Kong, Kopperman, Kovalevsky, Bexer, Chen, Rong, Han, Karaca and others)
in this area have studied to determine the properties of digital images with tools from Topology (especially
Algeraic Topology). Homotopy Theory deals with properties of spaces which are invariant under continuous
deformations. Translating homotopy from general topology to discrete structures raises a number of questions
which are not solved in satisfactory ways. Topological invariants are very useful in digital images and a
geometric modeling.
Fiber bundles play an important role in geometry quite apart from their nice homotopy properties. It is well
known that knowledge of the digital homotopy group is a very important tool for Image Analysis. A general
algorithm to decide whether two digital images have isomorphic homotopy groups would be a very powerful
tool for Image Analysis. Therefore, fiber bundles in digital images will be a very useful tool in the determinating
of digital homotopy groups.
This paper is organized as follows. In the preliminary part we give some basic definitons such as digital adjacencies, a digital (
)-continuous function, a digital
(
)-isomorphism, digitally (
)-homotopic functions, a -contractible digital image. Then we introduce
a digital fiber bundle and a digital trivial fiber bundle. Also we define a digital bundle map between digital fiber
bundles and an equivalence relation on digital fiber bundles with the assistance of digital fiber bundle maps.
Furthermore, we give a definition of a digital pullback bundle and study the homotopy invariance of digital
pullback bundles.
2 Preliminaries
Let be the set of integer. A (binary) digital image is a pair (
), where
for some positive integer
and indicates some adjacency relation for the members of . A variety of adjacency relations are used in the
study of digital images. The following terminology is used. Two points and in
are 8-adjacent if they are
1260
with
. A digital interval is a set of the form
+, in which 2-adjacency is assumed.
is (
Let
be a function. Let
be an adjacency relation
)-continuous if the image under of every -connected
and
be digital images with -adjacency and -adjacency respectively.
+ of
is (
)-continuous if and only if for every pair of -adjacent points *
( ) and ( ) are -adjacent in .
Let
and
be digital images with -adjacency and -adjacency respectively. A function
is a (
)-isomorphism [6]if is (
)-continuous and bijective and further
is
(
)-continuous and it is denoted by
. This suggests using the terminology isomorphic digital images
rather than that homeomorphic digital images. The term homeomorphic may be disleading in that models of
homeomorphic Eucledian sets need not be digital homeomorphic. For example all Euclidean simple closed
curves are homeomorphic but digital simple closed curves of differing cardinalities are not even of the same
homotopy types.
Definition 3. [4] Let
and
be digital images with -adjacency and -adjacency respectively.
Two (
)-continuous functions
are said to be digitally(
)-homotopic in Y if there is a
positive integer and a function
,
such that;
( )
)
( )
i)
( ) and (
( )
ii)
the induced function
,
defined by
( ) for
all
,
- is
(
)-continuous and,
iii)
,
- , the induced function
, defined by ( )
( ) for all
is (
)continuous.
We say that the function
is adigital(
1261
)-homotopic to the
) is adigital subbundle of (
) provided that
|
is a continuous surjection.
is a subset of ,
is
)(
) and (
)be digital images where is -connected space. Let
Definition 7. Let (
be
(
) is adigital fiber bundle with
(
)-continuos surjection, then we say that the quadruple
adigital total set , adigital base set and adigital fiber set if satisfies the following properties:
1. For all
2. For every point
( )
( )
is a (
)-isomorphism.
, there exists a -connected subset of such that
is (
(
))-isomorphism making the following diagram commutes:
(1)
Fig. 1
Example 8. Let
*
, by (
+
)
*
. Then
s the simplest digital fiber bundle, it is called atrivial digital fiber bundle.
(
)and
(
) be two digital fiber bundles. Then a digital fiber bundle
Definition 9. Let
)
(
)consists of a pair of maps
(
)
(
)-continuous,
map (
(
)-continuous subject to the condition that the diagram commutes:
(2)
Fig. 2
(
),
Definition 10. Two digital fiber bundles
(
) (
if there exist digital bundle maps,
(
)
| , denoted by
such that
and
Notice that the relation ' ' is an equivalence relation. If
(
)
) over
)
.
, then
and
and define
*( )
and
( )+
)
Then
1262
(
) is a digital fiber bundle and ( ) is a digital fiber bundle map. We say that (
) and denoted by ( )
digital pullback bundle of (
) is the
(
) be a digital fiber bundle, (
)be a digital image and
Example 11. Let
be a constant
( ) Since is
map. Then there exists a point
such that for all
, ( )
and we get
( )
( )
a digital fiber bundle,
Hence ( ) (
)is a digital trivial bundle.
(
) be a digital fiber bundle and let
Definition 12. Let
-connected subset. Then the restriction of
to , denoted by |
| .
where
( ) and
Proof: |
( ) |
be a
,
- be an digital interval. For
) be a digital fiber bundle and
( )
( ( ) ) Then
is also a digital fiber bundle.
(
Definition 13. Let
(
), define
Proposition 14.. Let
-connected subset. If
( )
*(
), where
be a
( )+.
(3)
Fig. 3
(
)
Define
( )by( )
( ( ) )
( )
( )
( ( ) ). Then is continuous bijection. If we define
we see that
is also continuous so
is an isomorphism. Also,
|
( ), hence the diagram commutes. ( )
( ) (
*( )
*(
)
)
{((
( )|
Claim I:
)
)
* +
Proof of Claim I:
Let
( (
are (
be (
)-homotopic maps. Then, there exists a map
)
)
( ) and
such that (
( ) and (
(
)-continuous for all
,
- and
) (
( )
) ( (
( )|
( ) (
( )
( )+,
( )
( )+
(
)
( )and
( )|
.(
) (
* +
) and
(
* +
* +)
( )
)
(
) },
((
is (
( )
) where
and
)
( ).
|(
* +)
* +
* +) *(
)
(
)
( )
( )+ Define
(
) Then is an
) ))-isomorphism and the diagram commutes:
(4)
1263
* +
Fig. 4
Hence
( )|
( )|
* +
* +
( )| * +
( )| * +
Claim II:
( )
If we prove this claim, then we show that ( )
*
+)
(
)|
)
(
|(
Proof of Claim II:
.(
* +
( )|
* +
) (
.(
) (
* +)
|(
* +
* +)
* +)
* +
/ and
) (
* +) and
* +).
(5)
Fig. 5
* +
*
Here
an isomorphism between
all
, we have ( )
of isomorphism relation,
( ) (
)
( )
(
Corollary 15. Let
-contractible space. Then
+ (
* +
* +. So that we have to find
)
(
) is an isomorphism.
(
)
(
)|
and . Since
and the restriction
* + are digital fiber bundles, then for
(
)
)
. Similarly for all
, we have ( ) (
. From the reflexivity
) ( ) (
). Therefore there exists an isomorphism
we get ( ) (
(
) for all
. So its easy to see that and are isomorphic digital images.
) be a digital fiber bundle where
is a trivial digital fiber bundle.
Proof: Since
is contractible, then the identity map over
( ) Moreover ( )
previous theorem ( )
and
( )
then is a digital trivial fiber bundle.
is a digital
References
[1] Boxer, L. (2006). Homotopy Properties of Sphere-Like Digital Images. J.Math. Imaging Vis. 24: 167-175.
[2] Herman, G.T. (1993). Oriented Surfaces in Digital Spaces. CVGIP: Graphical Models and Image processing
55: 381-396.
[3] Boxer, L. (1994). Digitally Continuous Functions. Pattern Recognit. Lett. 15: 833-839.
[4] Boxer, L. (1999). A Classical Construction for The Digital Fundamental Group. J.Math. Imaging Vis. 10: 5162.
[5] Rosenfeld, A. (1986). Continuous functions on digital pictures. Pattern Recognition Letters 4: 177-184.
[6] Boxer, L. (2006). Digital Products, Wedges, and Covering Spaces Journal of Mathematical Imaging and
Vision 25: 159-171.
[7] Boxer, L.; and Karaca, I. (2010). Some Properties of Digital Covering Spaces. Journal of Mathematical
Image and Vision 37: 17-26.
[8] Boxer, L.; and Karaca, I. (2008). The Classification of Digital Covering Spaces, Journal of Mathematical
Imaging and Vision 32: 23-29.
[9] Boxer, L.; and Karaca, I. (Preprint 2011). Properly Discontinuous Actions in a Digital Covering Space.
[10] Chen, L. (2004). Discerete Surfaces and Manifolds: A Theory of Digital-Discrete Geometry and Topology.
Rockville: Scientific and Practicle Computing.
[11] Chen, L.; Rong. (2008). Linear Time Recognition Algortihms for Topological Invariants in 3D. In the
Proceedings of International Conference on Pattern Recognition.
[12] Cohen, L.R. The Topology of Fiber Bundles Lecture Notes. Department of Mathematics Stanford
University.
1264
[13] Han, E. (2005). Nonproduct Property of The Digital Fundamental Group. Information Sciences. 171: 73-91.
[14] Hatcher, A. 2009. Vector Bundles and K Theory. Version 2.1.
[15] Husemoller, D. 1966. Fiber Bundles. New York: Springer-Verlag, Third Edition.
[16] Kong, T.Y. (1989). A Digital Fundamental Group. Comput. Graph. 13: 159-166.
[17] Kong, T.Y.; Kopperman, R. (2005). On Storage of Topological Information. Discrete Applied Mathematics
147 (2-3): 287-300.
[18] Kovalevsky, A. (1989). Finite Topology as Applied To Image Analysis. Computer Vision. Gaphics and
Image Processing 45: 141-161.
[19] Steenrod, N. 1960. The Topology of Fiber Bundles. USA: Princeton University Press.
Biographies
Ismet Karaca was born in Afyon, Turkey on January 5 th, 1969. He received a Bachelors degree in
Mathematics from Anadolu University in Turkey, a Masters in Mathematics from the university of Miami, and
a PhD in Mathematics from Lehigh University.
He is a Professor of Mathematics at Ege University in Izmir, TURKEY. Dr. Karacas research interests
include homotopy theory, steenrod algebra, and digital topology.
Prof. Karaca is a member of American Mathematical Society.
Tane Vergili was born in Izmir, Turkey on March 21 st, 1986. She received a Bachelors degree in
Mathematics from Ege University and started an integrated PhD programme in the field of algebraic topology at
the same university.
She is interested in algebraic topology and digital topology.
1265
intermediate points of its de Casteljau type algorithm. We also give explicit form of intermediate points in
terms of
Bernstein polynomials.
Keywords:
Bernstein polynomials
1 Introduction
One of the most important representation of curves and surfaces used in computer graphics and computer aided
geometric design (CAGD) is Bzier representation. This curves are first used to design automobile bodies.
A parametric Bzier curve of degree
is defined by
( )
. / (
( )
where
denotes
dimensional Euclidean space. The points are called the control points and the polygon
obtained by joining the control point
with the control point
for
is called the control
polygon. The reason of the popularity of Bzier curves in CAGD is that the control polygon gives information
about the shape of the polynomial curve ( )
The basis functions
( )
. / (
are called Bernstein bases of degree . This polynomials have shape preserving properties and they are not
useful only for CAGD but also for approximation of functions. A constructive proof of Weierstrass
approximation theorem which states that every continuous function can be approximated uniformly, as close as
we wish, by polynomials was given in 1912 via Bernstein polynomials
Although Bzier curves are first publicized in 1962, Paul de Casteljau is the first one who developed them in
1959 by using and algorithm that gives a point on the curve. The most powerful advantage of this method is
1266
subdividing any Bzier curve to make it more flexible. For the given points
algorithm is
( )
for
( )
and
this
( )
and
As a consequence of importance of Bernstein polynomials, several generalization of them have been proposed.
This paper is concerned about two of these generalizations and their de Casteljau type algoithms,
Bernstein
polynomials introduced in [3] and
Bernstein polynomials introduced in [11] which is also a generalization
of
Bernstein polynomials. First we give a new identity on
Bernstein polynomials, then use it to represent
Bernstein polynomials in terms of intermediate points of its de Casteljau type algorithm. In section 3 we give
explicit form of intermediate points of de Casteljau type algorithm related with
Bernstein polynomials.
2
In [3]
Bernstein Polynomials
Bernstein bases defined by
( )
0 1
0 1
for
, -,
,
-
, integer defined by
( )
( )( )
and
1267
)
0
)
)( )
and write
1(
) (
(
(
)
)
)
1(
Then we have
) (
){
)0
)0
)
and
) (
)
1
20
(
) (
)0
( )
where
13
Theorem 1:
( )
( )
)( )
( )
We know that ( )
( )
( )
1268
we have
( )
( )
*(
( )
( )
( )
( )
( )
( )
( )
), that is
( )
( )(
{(
( )
( )
Bernstein Polynomials
We begin by giving some notations and definitions. A generalization of Bernstein polynomials degree
depending on two parameters and (we call
Bernstein polynomials) defined in [11] by
(
| )
0 1
)(
( )
and
Here used
0 1
)(
Pochammer symbol
| )
) (
1269
( )
for
( )
and
with
( )
( )( )
see [11].
The following new result gives explicit form of the intermediate points of the de Casteljau type algorithm in
terms of
Bernstein polynomials.
Theorem 2: The intermediate points of the algorithm (6) are
( )
| )
( )
( )
( )
( )
we may write
( )
| )
| )
and
4
| )
Then
[ ]
) (
1270
Now multiplying
by
) (
(
On the other hand, rearrange the term
[ ]
gives
]
) (
Therefore
[ ]
]}
(
) (
)
)
| )
References
1.
2.
Farin, G. (2002). Curves and surfaces for CAGD, A practical guide(5th ed.). USA: Academic Press.
Phillips, G.M. (1996). A de Casteljau algorithm for generalized Bernstein polynomials. BIT, 36(1),
232-236.
3. Phillips, G.M. (1997). Bernstein polynomials based on the q-integers. Annals of Numerical
Mathematics, 4, 511-518.
4. Diibyk, . & Oru, H. (2007). A generalization of rational Bernstein-Bzier curves. BIT, Numerical
Mathematics, 47(2), 313-323.
5. Diibyk, . & Oru, H. (2008). Tensor product q-Bernstein polynomials. BIT, Numerical
Mathematics, 48(4), 689-700.
6. Goodman, T.N.T., Oru, H. & Phillips, G.M. (1999). Convexity and generalized Bernstein
polynomials. In Proceedings of the Edinburgh Mathematical Society, 42, 179-190.
7. Oru, H. & Phillips, G.M. (1999). A generalization of the Bernstein polynomials, In Proceedings of the
Edinburgh Mathematical Society, 42, 403-413.
8. Oru, H. & Phillips, G.M. (2003). q-Bernstein polynomials and Bzier curves. Journal of
Computational and Applied Mathematics, 151, 1-12.
9. Oru, H. & Akmaz, H.K. (2004). Symmetric functions and the Vandermonde matrix. Journal of
Computational and Applied Mathematics, 172, 49-64.
10. Oru, H. & Tuncer, N. (2002). On the convergence and iterates of q-Bernstein polynomials. Journal of
Approximation Theory, 117(2):301313.
11. Lewanowicz, S. & Wozny, P. (2004). Generalized Bernstein polynomials. BIT, Numerical
Mathematics, 44(1), 63-78.
12. Phillips, G.M. (2008). Survey of results on the q-Bernstein polynomials. IMA J. Numer. Anal., 30(1),
277-288.
1271
Abstract.The numerical and analytic solutions of mixed problem for multidimensional fractional hyperbolic partial
differential equations with the Neumann condition are presented. The stable difference scheme for the numerical solution of
the mixed problem for the multidimensional fractional hyperbolic equation with the Neumann condition is presented.
Stability estimates for the solution of this difference scheme and for the first and second order difference derivatives are
obtained. A procedure of modified Gauss elimination method is used for solving this difference scheme in the case of onedimensional fractional hyperbolic partial differential equations. He's variational iteration method is applied. The comparison
of these methods is presented.
Keywords : Fractional hyperbolic equation; initial boundary value problems, difference schemes ,stability, iteration
method
1 Introduction
It is known that various problems in fluid mechanics (dynamics, elasticity) and other areas of physics lead
to fractional partial differential equations. Methods of solutions of problem for fractional differential
equations have been studied extensively by many researchers (see, e.g., [1]-[14] and the references given
therein)
The role played by stability inequalities (well-posedness) in the study of boundary-value problems for
hyperbolic partial differential equations is well known (see, e.g., [18]-[31]).
In the present paper, finite difference and Hes iteration methods for the approximate solutions of the
mixed boundary-value problem for the multidimensional fractional hyperbolic equation
1
2u (t , x) m
2
a
(
x
)
u
r
xr x
t u (t , x ) u (t , x ) f (t , x ),
2
r
t
r
x ( x1 ,..., xm ) , 0 t 1,
u (0, x) 0, ut (0, x) 0, x ,
u (t , x)
0, x S
n
1272
(1.1)
space.
: x ( x1 ,..., xm ) : 0 x j 1,1 j m
with
boundary
S , S , ar ( x),( x ) and
are
smooth functions
and
f (t , x),(t (0,1), x )
ar ( x) a 0 . The first order of accuracy in t and the second order of accuracy in space variables for
the approximate solution of problem (1.1) is presented. The stability estimates for the solution of this
difference scheme and its first and second order difference derivatives are established. A procedure of
modified Gauss elimination method is used for solving this difference scheme in the case of onedimensional fractional hyperbolic partial differential equations.
0 rj N j , h j N j 1, j 1,..., m
We introduce the Banach space
defined on
L2h L2 (
, Sh
S.
2
2
h L ( ) h ( x) h1...hm .
2
x h
x
x
To the differential operator A generated by problem (1.1), we assign the difference operator Ah by
the formula
m
Ah xu h ar ( x)uxr
r 1
uxh
(2.1)
xr j
1273
2
dt
h
v h (0, x) 0, dv (0, x) 0, x
dt
,
.
(2.2)
In the second step, we replace problem (2.2) by the first order of accuracy difference scheme
(k m )(umh umh 1 )
u h ( x) 2u h ( x) u h ( x)
k
2
1
k
k 1
k 1
1
2
(
k
m
)!
2 m 1 x h
Ah uk 1 f kh ( x)
h
f k ( x) f (tk , x), tk k ,1 k N 1, N 1, x h
h
h
u1 ( x) u0 ( x) 0, u0h ( x) 0, x h .
1
k m
1
2 t
e dt .
Here (k m ) t
2 0
Theorem 1.
Let
and
max
1 k N
ukh
max
1 k N 1
C2 f1h
L2 h
max
ukh ukh1
1 k N
L2 h
C1 max f kh
L2 h
max (ukh ) xr xr
1 k N
max 1 ( f kh f kh1 )
2 k N 1
1 k N 1
L2 h
L2 h
L2 h
L2 h
Here
f kh ,1 k N 1 .
Ahxu h ( x) h ( x), x
(2.3)
Dhu ( x) 0, x S h
h
1274
the
u
r 1
C h
xr xr
L2 h
L2 h
Remark 1 : The stability estimates of Theorem 1 permit us to obtain the estimation of accuracy of
difference schemes. Of course , this approach can be applicable for study of the estimation of accuracy
of
differences
schemes
for
numerical solutions of the initial-boundary value
problem
1
n
2u (t , x) n
2
a
(
x
)
u
b
(
x
)
u
r
xr x
r
xr
t u (t , x ) f (t , x, u (t , x )),
2
r
t
r 1
r 1
x ( x1 ,..., xm ) , 0 t 1,
u (0, x) 0, ut (0, x) 0, x ,
u (t , x)
0, x S
n
D
t
t u (t , x ) u xx (t , x ) u (t , x ) f (t , x ),
8t 2
2
( t ) 2 ) cos( x), 0 t , x 1,
f (t , x) (2 t
3
u (0, x) 0, ut (0, x) 0, 0 x 1,
u x (t , 0) u x (t ,1) 0, 0 t 1
(2.4)
for solving the one-dimensional fractional hyperbolic partial differential equation is considered. Applying
1 m
m 1
k 1
k
k 1
k ( k m )(un un )
2
un 2un un 1
(
k
m
)!
2
2 m 1 k 1
unk nk
2
h
k f (t , x ),1 k N 1,1 n M 1,
k
n
n
un0 0, 1 (u1n un0 ) 0, 0 n M ,
k
u1 u0k uMk uMk 1 0, 0 k N .
(2.5)
We can rearrange the system of equations in the matrix form
1275
AU n 1 BU n CU n 1 Dn ,1 n M 1
U1 U 0 ,U M U M 1 ,
where
0
0
A 0
...
0
0
0 ... 0
0 ... 0
0
0
a
0
0 ... 0
a ... 0
0 ... a
0 ... 0
0
b1,1
b
b2,2
2,1
b3,1
b3,2
b4,2
B b4,1
...
...
bN ,2
bN ,1
b
N 1,1 bN 1,2
and
0
0
0
,
0
...
0
0 ( N 1)( N 1)
0
0
0
0
...
...
0
0
b3,3
...
b4,3
b4,4
...
...
...
...
...
bN ,3
bN ,4
...
bN , N
bN 1,3
bN 1,4 ... bN 1, N
0
0
0
,
0
...
0
1 ( N 1)( N 1)
0
,
...
0
bN 1, N 1 ( N 1)( N 1)
0
0
C A.
0
0
D 0
...
0
0
0 ... 0
0 ... 0
0
0
0
1
0 ... 0
0 ... 0
0 ... 1
0 ... 0
1276
U s0
1
Us
U s2
U s U s3
s n 1, n, n 1 .
...
N 1
U s
UN
s ( N 1)(1)
Where,
1
h2
n0
1
n
n2
.
n ...
N
2
nN 1
n
N
n ( N 1)1
a
So, we have the second order difference equation with respect to n matrix coefficients. To solve this
difference equation, we have applied a procedure of modified Gauss elimination method for difference
equation with respect to k matrix coefficients. Hence, we seek a solution of the matrix equation in the
following form
U j j 1U j 1 j 1 ,
n M 1,...,1 , j ( j 1,..., M ) are ( N 1) ( N 1) square matrices and j ( j 1,..., M ) are
( N 1) 1 column matrices defined by
n1 ( B C n )1 ( A),
n1 ( B C n )1 ( Dn C n ), n 2,3,..., M ,
where
1
0
1 ...
0
0
0
0 1 0
... ... ...
0
0 ... 0
0 ... 0
0 ... 0
,
... ... ...
1 0 0
0 1 0
0 0 1 ( N 1)( N 1)
1277
0
0
0
1 ...
.
0
0
0
( N 1)1
Now, we will give the results of the numerical analysis. First, we give an estimate for the constants
and
C1
C1 max (Ct1 ),
f ,u
C2 max (Ct 2 ),
f ,u
Ct1
max
1 k N
ukh
L2 h
max
h
f1
max ( f f
1 k N 1
(max
1 k N
Ct 2
ukh ukh1
max
1 k N 1
L2 h
L2 h
max (ukh ) xr xr
1 k N
f kh
L2 h
L2 h
) 1 ,
L2 h
2 k N 1
h
k
h
k 1
) .
L2 h
Ct1 and Ct 2 in the case of numerical solution of initial-boundary value problem (2.4) are
computed. The constants Ct1 and Ct 2 are given in Table 1 for N 20, 40,80 and M 80
The constants
respectively.
M=80
N=40
0.2073
M=80
N=80
0.2061
Ct 2 0.2075
0.1223
0.0670
t1
The values of
Second, for the accurate comparison of the difference scheme considered, the errors computed by
2
M 1
E0 max ( u (tk , xn ) u
1 k N 1
n 1
M 1
E1 max (
1 k N 1
n 1
M 1
n 1
1
2
h) ,
2
1
(u k 1 unk 1 )
ut (tk , xn ) n
h) 2 ,
2
k
n
h)
1
2
1278
unk represents the numerical solution at (tk , xn ) . The errors E0 , E1 and E2 results
are shown in the Tables 2 for N 20, 40,60 and M 60 , respectively.
exact solution and
M=80
N=40
0.0037
M=60
N=60
0.0008
Comparison
error E1
of 0.1030
0.0521
0.0491
Comparison
error E2
of 0.1224
0.0806
0.0882
Comparison
error E0
un 1 (t , x) un (t , x)
t
1
2u ( s, x ) m
(
a
(
x
)
u
)
r
xr xr
s u ( s, x ) u ( s, x ) f ( s, x ) ds,
2
r 1
s
(3.1)
where
u 0 , u0 (t , x) is its initial
noticing that u 0 , we obtain :
2 un ( s , x )
ds
2
s
un1 (t , x) un (t , x)
0
(3.2)
un ( s, x ) s t
s
t
2 (t , s)
un ( s, x)ds 0.
2
s
0
un1 (t , x) un (t , x)
( un ( s, x))
s
s t
un ,
(3.3)
2 (t , s)
0
s 2
(3.4)
1279
(t , s)
s
(t , s)
s t
s t
0,
(3.5)
0.
(3.6)
(t , s) s t .
Substituting the identified Lagrange multiplier into Eq. (3.1), following variational iteration formula can
be obtained :
un 1 (t , x) un (t , x)
1
2
m
2
(3.7)
u
(
s
,
x
)
(
a
(
x
)
u
)
D
u
(
s
,
x
)
u
(
s
,
x
)
n
ds.
r
n
x
s
n
n
x
r
(
s
t
)
0
s 2
r 1
f ( s, x )
In this case, let an initial approximation u0 (t , x) u (0, x) tut (0, x) .Then, approximate solution takes
t
the form
u (t , x) limun (t , x) .
n
3
2
8
t
2
( t ) 2 ) cos( x), 0 t , x 1,
f (t , x) (2 t
3
u (0, x) 0, ut (0, x) 0, 0 x 1,
u x (t , 0) u x (t ,1) 0, 0 t 1
(3.8)
for solving the one-dimensional fractional hyperbolic partial differential equation is considered. According
to the formula (3.7), the iteration formula for Eq.(3.8) is given by
un 1 (t , x) un (t , x)
1
un2 ( s, x)
un2 ( s, x)
2
(
s
t
)
D
u
(
s
,
x
)
un ( s, x) f ( s, x) .
s n
2
2
0
x
s
(3.9)
1280
u0 (t , x) 0,
7
5
4
4 2
2
128
t
35
t
35
t
7
5
1
4
4 2
2
u2 (t , x)
128
t
35
t
35
t
420t 2 cos( x)
420
u1 (t , x)
1
420
11
7
4
2
2
0.9058003666
t
0.1510268880
t
0.1719434921
t
cos( x)
0.3281897218t 6 0.01666666667t 5
and so on. In the same manner, the rest of the components of the iteration formula Eq.(3.9) can be
obtained by using the Maple package.
Figures
exact solution
0.5
-0.5
-1
10
8
10
8
4
2
4
2
Fig.1. The surface shows the exact solution u(t,x) for eq. (2.4)
1281
0.5
-0.5
10
8
10
8
4
2
4
2
0.5
-0.5
10
8
10
8
4
2
4
2
1282
References
1. I. Podlubny, Fractional Differential Equations, Academic Press, New York, (1999).
2. S. G. Samko, A. A. Kilbas and O. I. Marichev, Fractional Integrals and Derivatives,
Gordon and Breach Science Publishers, London,(1993).
3. J. L Lavoie, T. J. Osler, R. Tremblay, Fractional derivatives and special functions,SIAM
Review18(2), 240-268,(1976).
4. V. E. Tarasov, Fractional derivative as fractional power of derivative, International
Journal of Mathematics, 18, 281-299,(2007).
5. E.M. El-Mesiry, A. M. A. El-Sayed, H. A. A. El-Saka, Numerical methods for multi-term
fractional (arbitrary) orders differential equations, Appl. Math. Comput. 160(3), 683699,(2005).
6. A.M.A. El-Sayed, F. M. Gaafar, Fractional order differential equations with memory and
fractional-order relaxation-oscillation model, Pure Math. Appl. 12 (2001).
7. A.M.A. El-Sayed, E. M. El-Mesiry, H. A. A. El-Saka, Numerical solution for multi-term
fractional (arbitrary) orders differential equations,Comput. Appl. Math. 23(1), 3354,(2004).
8. R. Gorenflo, F. Mainardi, Fractional calculus: Integral and differential equations of
fractional order,in: A. Carpinteri, F. Mainardi (Eds.), Fractals and Fractional Calculus
in ContinuumMechanics, Springer, Wien, (1997), pp. 223-276.
9. D. Matignon, Stability results for fractional differential equations with applications to
control processing, in: Computational Engineering in System Application 2, Lille,
France, (1996).
10. A. Ashyralyev, A note on fractional derivatives and fractional powers of operators,
Journal of Mathematical Analysis and Applications, 357(1), 232-236, 2009
11. A. Ashyralyev, F. Dal, Z. Pinar, A note on the fractional hyperbolic differential and
difference equations, Appl. Math. Comput. 217(9), 4654-4664,(2011).
12. A. Ashyralyev, F. Dal, Z. Pinar, On the numerical solution of fractional hyperbolic
partial differential equations, Mathematical Problems in Engineering, vol.2009, Article
ID 730465, 2009.
13. F. Dal, Application of Variational Iteration Method to Fractional Hyperbolic Partial
Differential Equations, Mathematical Problems in Engineering, vol.2009, Article ID ,
824385, 2009.
14. I. Podlubny, A. M. A. El-Sayed, On Two Definitions of Fractional Calculus, Solvak
Academy of Science-Institute of Experimental Phys, (1996).
15. H. O. Fattorini, Second Order Linear Differential Equations in Banach Space, Notas de
Matematica, North-Holland, (1985).
16. S. Piskarev, Y. Shaw, On certain operator families related to cosine operator function,
Taiwanese Journal of Mathematics 1(4), 3585-3592, (1997).
17. P. E. Sobolevskii, Difference Methods for the Approximate Solution of Differential
Equations.Izdat. Voronezh. Gosud. Univ., Voronezh (1975), (Russian).
18. S. G. Krein, Linear Differential Equations in a Banach Space, Moscow: Nauka (1966).
19. P.E. Sobolevskii, L. M. Chebotaryeva, Approximate solution by method of lines of the Cauchy problem
for an abstract hyperbolic equations, Izv. Vyssh. Uchebn. Zav.,
Matematika,5,103-116,(1977), (Russian).
20. A. Ashyralyev, M. Martinez, J. Paster, S. Piskarev, Weak maximal regularity for
abstract hyperbolic problems in function spaces, Abstracts of 6-th International ISAAC
Congress, Ankara,Turkey, (2007), p. 90.
21. A. Ashyralyev, N. Aggez, A note on the difference schemes of the nonlocal boundary
value problems for hyperbolic equations, Numerical Functional Analysis and
Optimization 25(5-6), 1-24, (2004).
22. A. Ashyralyev, I. Muradov, On difference schemes a second order of accuracy for
hyperbolic equations. in: Modelling Processes of Explotation of Gas Places and
Applied Problems of TheoreticalGasohydrodynamics, Ashgabat, Ilim, pp.127-138,
(1998), (Russian).
1283
1284
Chapter 8
STATISTICS
AND
DATA ANALYSIS
1285
Abstract. This article investigates the applicability of Zipf Laws on Turkish. These empirical laws
formulated using mathematical statistics, are based on the principle of least effort and they offer a means of
formulization for natural languages by providing relations for four different properties of language. The
formulization of first two properties, word frequencies and vocabulary size, enable researchers to expect the
frequency of each word and total number of different words in a corpus. The relations defined for remaining
properties, number of word meanings and distribution of content words, may be used to create a measure for
the information content of a given corpus. The fitness of a language to Zipf Laws may enrich language
models defined for several Natural Language Processing applications especially as corpus analysis and
evaluation. Experimental results and derived parameters of Zipf relations show that Turkish has concordance
with Zipf Laws.
Keywords: Zipfs Laws, least effort principle
1 Introduction
Zipfs laws which are based on the principal of least effort have found applicability in various physical and
social environments. The basic philosophy is that there is a balance between the parameters that govern a
relation such that as one increases the other decreases in some respect so that the product of these parameters is a
constant constituting the balance in least effort battle. Zipf has stated four laws two of which establish a relation
between the frequency of words and their rank when ordered from the most frequent to the least frequent. The
other two laws are more related with the meaning of the text. His law of meanings relates the frequency of a
word with the number of different meanings that this word is expected to take on and his law of burstiness
claims that content bearing words are likely to appear in clusters. We investigated the applicability of all four
laws of Zipf on Turkish to derive the parameters that govern the distribution of words in a text both syntactically
and semantically.
In this article we first give a brief overview of Zipfs laws and experimental results of applying Zipfs laws on
Turkish, the last part is the conclusion
1286
way. From the speakers viewpoint a single word that would mean whatever he wanted to mean will be
economical (the force of unification). From the auditors viewpoint, a single word meaning everything will face
him by an impossible task of determining particular meaning from almost an infinite space. So auditor will
prefer a distinct meaning for each word to pay least effort in deciphering (the force of diversification). The
opposite forces of unification and diversification will result with a balance in language. Depending on this
principle Zipf has formulated 4 different aspects of language.
2.1 Zipfs Law Zipfian Distribution of Words
In Zipfs first study, each appearance of distinct words (token) in a large corpus is counted up and the words are
ranked in decreasing order of their frequency. The mathematical relation between the frequency and rank of a
word in Zipfs first law is defined as
log( f r ) H N BN log(r ) .
(1)
where f r is the frequency of the word with rank r , B N and H N are numerical are inserted to the formulation
(2)
(3)
where V (i, N ) is the number of words seen i times in an N word corpus, K N and D N are numerical constants
dependent on N . Actually this law of vocabulary has a strong relation with the first law and it is formulated in
[8] as given in equation 4. It depicts that if any distribution is consistent with the first law, it must also be
consistent with the second.
DN BN /(1 BN )
(4)
1287
the other hand there will be fewer than m different words. As a consequence, we may expect that at least
some words must have multiple meanings.
The problem to be solved above is which words will have multiple meanings and how many meanings they
will convey. Zipf has stated that if the frequency of the most frequent word is F1 and it has m1 meanings, the
equation below can be written
F1 m1 f1
(5)
where f 1 represents the average frequency of the occurrence of the m1 meanings. The forces of unification will
tend to increase the size of m1 in the direction of putting all meanings behind a single word. In contrast, the
forces of diversification will tend to increase f 1 in the direction of reducing number of different meanings per
word. The empirical study on frequency-rank curves shows that the effect of conflicting forces arises itself as a
hyperbolic relationship between two variables. As a result m1 and f 1 will also stand in a hyperbolic relationship
with one another with a further result that m1 will tend to equal f 1 generating the equation 6.
(6)
F1 m1 m1 ( F1 m1 )
The equation for the most frequent word can be generalized for all frequencies and
follows:
m
or
m 1
can be formulized as
(7)
(9)
1288
BirTD corpus is compiled by merging METU [12] and Bilkent corpus. Stemmed form of BirTD corpus is
composed of OSTAD and Bilkent corpus.
The language dependent parameters of Zipfs linear function (equation 2) and Mandelbrots equation
(equation 3) are derived on Bilkent corpus. Table 1 summarized the results for both surface and stemmed
forms of corpora. R 2 gives the coefficient of determination and originally is the square of multiple correlation
coefficients in linear regression. As the coefficient grows, the strength of relation increases. N is the number
of words, H N and B N are derived parameters in the model. Although the experiments are repeated with
different values of W ( W 10, 100, 1000, 10000), table 1 presents results of W value which attains the best
curve fit for each corpus.
Table 1. Empirical results of curve fitting experiments on linear function of Zipf and Mandelbrots equations of
Bilkent corpus
Model
Stemmed
form
Surface
form
Zipf
Mandelbrot
Zipf
Mandelbrot
R2
1 000
100
0.974
0.994
0.995
0.998
20 266
20 266
94 228
94 228
HN
6.88
9.58
5.31
5.45
BN
1.62
2.26
1.09
1.12
The results for surface forms are in parallel with the results of Mandelbrot ( W 100 and B N 1.15 ) for
English. So it can be stated that Turkish and English have similar properties in the scope of surface forms. When
the estimated frequencies with Zipfs derived parameters are examined it is seen that Zipfs equation fails to
estimate high frequency items.
The vocabulary balance experiments are utilized on BirTD corpus. In figure 1, horizontal axis gives the
occurrence number ( i =1,2,31004) and vertical axis is the percentage weight of total number of distinct words
that occur i times on the total vocabulary of the corpus.
Fig. 1. Vocabulary balance results of Bilkent corpus (both stemmed and surface form)
Recent researches in different languages showed that number of low frequency words, especially hapax words
(words seen just once in corpus), is an indicative of languages productivity in natural language processing
literature [8]. Baayen in [7] observed that %50 of corpus is composed of hapax words. Figure 1 presents that
hapax words compose about %49.8 of surface form corpus and %36.5 of stem corpus. So we can say that
Turkish is also consistent also with Baayens claim.
1289
The curve fitting experiments of BirTD corpus to Zipfs second law and Zipfs first law generated the values
in Table 2.
To investigate Kornais claim in equation 4 in Turkish, we have used the value of B N 1.2786 at W 1000 to
calculate DN (1.2786) /(1 12786) 0.5611 which approximates to theoretical result, 0.5611. The same
similarity is also observed for stems ( DN (2.2515) /(1 2.2515) 0.6924 0.6994) .
Table 2. Empirical results of Zipfs first and second law experiments on BirTD corpus
First Law
Surface form
Stemmed form
R
0.998
0.995
W
1000
1000
HN
6.709
9.571
BN
1.2786
2.2515
Second Law
BN
R
0.998
2.846
1.000
2.695
2
DN
0.5516
0.6994
Zipf claims that number of meanings that a word conveys is proportional to some power of its frequency and
theoretically this value is -0.5 (m f p ) . The law is empirically supported with his studies on Thorndike
Century dictionary and calculated that power of f is -0.4605. Omitting the most frequent 500 words he
recalculated the power of f as -0.4656. The law is formulized as
mk f
(10)
where m is the number of meanings and , k and p are constants to be estimated. In this study, accepting
p 0.5 we estimated k on stemmed Bilkent corpus. TDK (Institute of Turkish Language) dictionary is used to
tag meanings of the words. In experimentation, the words are grouped in bins of 500 words (e.g first bin includes
most frequent 500 words in the corpus, second bin includes next most frequent 500 words).The experiment gives
k 98.88 (sum of square error SSE=1.89, R 2 0.928 )
Figure 2 presents actual and estimated number of meanings when k is 98.88. Horizontal axis is the rank of the
bins. It is seen that for most frequent 500 words, meaning is over estimated, if first bin is omitted as Zipf has
experimented for English, the estimated values will be closer to the actual values.
Fig. 2. Actual and estimated number of meanings in Bilkent corpus for bin size=500
To validate Zipfs burstiness law on Turkish, we counted the intervals of the same length between the
appearances of word with the same frequency. Figure 3 gives the experimentation results, horizontal axis is the
interval (frequency-1) classes and vertical axis gives the percentage ratio of the number of intervals with given
page length to the total number of intervals. To give an example, actually our corpus (OSTAD) involves 236
distinct words that occur 5 times. The number of intervals for each word in this class is 4 and total number of
1290
intervals for this group is 236*4=944. The %38.6 of these intervals has 1 page length, %8,69of them has 2 page
length, %6.67 of them has 3 page length and %46.08 of them is longer than 3 pages.
Fig. 3. The ratio of intervals with several page lengths to the total number of intervals
Figure 3 shows that as frequency of words increases, the number of one and two page intervals increase. These
results support Zipfs theorem that interval lengths tend to shorten as the frequency of the word increases. The p
parameter in equation 9 for Turkish is found to be in the range 0.52- 0.78 in our observations (For English p is
between 0.96 and 1.29 for English.
4 Conclusions
In this article after giving a brief overview of Zipfs laws based on principle of least effort their applicability on
Turkish is experimentally investigated. Four different properties of language have been investigated: frequency
distribution of words, vocabulary balance, distribution of meanings and burstiness.
The experiments on the frequency-rank distribution in Turkish corpus show that Mandelbrots modified
equation resembles the actual distribution better as it does for English. The curve fitting experiments showed us
that Turkish is convenient to Zipfs principle on vocabulary balance. As a result, parameters can be used in
modeling vocabulary richness of language. Another deliverable of this stage is the evidence on relation between
second and first law as in English [8]. The deliverable convinces that Zipfs laws are modeling different aspects
of Turkish without violating the integrity of language. Experimentation on Bilkent corpus showed us that there is
a correlation between number of meanings and their frequencies. The burstiness experiments showed that
content words are distributed in scattered groups in Turkish corpus as in English. This verifies that Zipfs law of
burstiness does not depend on the language.
Acknowledgements. This work was TBTAK withinthe scope of Project No. 107E192. The authors thank
TBTAK for supporting this project. For further details reader may refer to [12], [13], [14]
References
1.
2.
3.
4.
5.
Zipf, G. K. (1949). Human Behavior and the Principle of Least-Effort. Cambridge, MA: Addison-Wesley
Poole, H. (1985). Theories of the middle range. Norwood: New Jersey: Ablex.
Mandelbrot, B. (1959). A note on a class of skew distribution function analysis and critique of a paper by H.A.
Simon. Inform. and Control 2, 90-99.
Mandelbrot, B. (1952). An information theory of the structure of the language. Symposium on Applications of
Communication Theory, 486-500. London.
Simon, H. A. (1955). On a Class of Skew Distribution Functions. Biometrika 42
1291
Herdan, G. (1960). Type-Token Mathematics: A textbook of mathematical linguistics. The Netherlands: The
Hague: Mouton and Co.
Baayen, R. H. (1996). The Effects of Lexical Specialization on the Growth Curve of the Vocabulary.
Computational Linguistics 22, 455-480.
Kornai, A. (2002). How many words are there?. Glottometrics 4, 61-86.
Tr, G. (2000). A Statistical Information Extraction System for Turkish. Ph.D. thesis, Bilkent University,
Department of Computer Engineering, Ankara, Turkey.
Oflazer, K., Say, B., Hakkani-Tr, D. Z., Tr,G. (2003) Building a Turkish Treebank, Invited chapter in Building
and Exploiting Syntactically-annotated Corpora, Anne Abeille Editor, Kluwer Academic Publishers
Atalay, N.B., Oflazer, K. Say, B. (2003). The Annotation Process in the Turkish Treebank, in Proceedings of the
EACL Workshop on Linguistically Interpreted Corpora - LINC, Budapest, Hungary.
Diner, B. T. (2004). Trke iin statistiksel Bir Bilgi Geri Getirim Sistemi. Ph.D. Thesis, UBE, Ege University,
Turkey.
lgen, B. and Karaolan, B. (2007). Investigation of Zipf's 'Law-of-Meaning' on Turkish Corpora. ISCIS2007.
Istanbul.
Kocaba, . and Kla, T. & Karaolan, B. (2007). Zipf's Law of Burstiness in Turkish. ISCIS2007. Istanbul.
Biographies
Bahar Karaolan has taken her B.S. degree from Electrical and Electronics Engineering, Bogazici University and M.S. and
Phd. Degree from Computer Engineering, Ege University (1991). She is a Professor and Vice Director of International
Computer Institute of Ege University. She is giving several courses such as information retrieval, multimedia systems and
computer architecture.
Senem Kumova Metin Senem Kumova Metin received her BSc. degree in Electrical and Electronics Engineering from Ege
University and MSc. and PhD degrees from International Computer Institute at Ege University. She is currently working as a
lecturer in Izmir University of Economics. She is mainly interested in natural language processing applications and
computational linguistics.
Bekir Taner Diner Bekir Taner Diner received his BSc. degree in Statistics from Middle East Technical University,
MSc. and PhD degrees from International Computer Institute at Ege University. He is currently working as an Assistant
Professor in Mula University. He is mainly interested in information retrieval.
1292
Abstract.It is well known that adaptive IIR (infinite impulse response) filters are useful in many fields such
as echo cancelations, noise reductions, bio systems, speech recognitions communications and control
applications. This systems are recursive in nature
filters. In order to use adaptive IIR filtering, a practical, efficient and robust global optimization algorithm is
necessary to minimize the multimodal error function. On the other hand, Genetic algorithm is a powerful
optimization technique for minimizing multimodal functions. This paper proposes a novel objective
function to enhance the performance of estimated model in both time and frequency responses by genetic
algorithm that shows more its capability to achieve better performance compared with other conventional
fitness functions based on MSE method.
1 Introduction
It is well known that adaptive IIR (infinite impulse response) filters are useful in many fields such as echo
cancelations, noise reductions, bio systems, speech recognitions communications and control applications.
Adaptive IIR filters more accurately model physical plants than equivalent adaptive FIR filters. In addition, they
are typically capable of meeting performance specifications using fewer filter parameters. Despite that, IIR
structures tend to produce multimodal error surfaces for which its cost function is significantly difficult to
minimize. Therefore many conventional methods especially stochastic gradient optimization strategies may
become entrapped to local minima.
There are principally two different important set of applications in IIR filter design: adaptive signal processing
and adaptive system identification. The design of IIR filter for adaptive signal processing is mostly based on the
desired frequency response (passband) whereas in adaptive system identification, IIR filter is employed for
modeling, Therefore it should behave as similar as possible to the real system in both time and frequency
domains.
In order to use adaptive IIR filtering, a practical, efficient and robust global optimization algorithm is necessary
to minimize the multimodal error function. On the other hand, Genetic Algorithm is a robust search and
optimization technique, which finds applications in numerous practical problems. The robustness of GA is due
1293
to its capacity to locate the global optimum in a multimodal landscape[4]. Thus several researchers have
proposed various methods in order to use GA in adaptive IIR filtering applications, for instance, in [3-5] GA and
input-output data in time domain are utilized to estimate the parameters of an IIR model in order to use in
system identification where fitness function based on mean square error (MSE) between the unknown plant and
the estimated modelis used. On the other hand GA is extensively employed in adaptive signal processing
applications. The primary application is reported in [2]. In recent years, several successful techniques have been
introduced to improve GA capability for signal processing applications. [6,8,10,11]. In [8] a stability criterion
embedded in the GA is applied to design robust d-stable IIR filter, but IIR filters, produced by this algorithm, do
not describe the dynamic plant necessarily. Therefore it is not useful for system identification.
In the most applications of optimization methods in adaptive IIR filtering, the objective function is designed
based on the MSE criteria. In [7] another cost function based on Least Mean Squared error (LMS) and mean
absolute error (MAE) is considered alongside MSE. Another example is [6] in that the phase response is
considered besides the magnitude response and it has been tried to design a linear phase response filter via a
fitness function based on variance of the phase different sequence of the designed IIR filter.
In this paper the designing methods of IIR filter in signal processing by using GA based on frequency response
are employed for adaptive system identification and for this purpose, a novel objective function is proposed to
enhance the performance of estimated model in both time and frequency responses.
( )
( )
)( )
( )
where ak's and bk'sare filter coefficients that define its poles and zeroes respectively. In addition, this parameters
are estimated by genetic algorithm so that the error based on fitness function between the frequency response of
designed IIR filter and the real frequency response (of the plant) is minimal.
Genetic algorithm begin with the random set of possible solutions that each one is embedded in one
chromosome. Each chromosome has (M+N+1) genes ([a1aN b0 b1bM]). At every generation, the fitness of
each individual (chromosome) is evaluated by a predetermined fitness function. An individual with lower fitness
value is considered. The population is then evolved based on the circled process of natural selection, survival of
the fittest, and mutation. This cycle is continued in order to find the optimal solution.
2.2 Fitness Function
The typical cost function, in adaptive filtering is the mean squared error (MSE) between the frequency response
of the unknown system (
in which
*, ( )
( )- +
( )):
( )
( )
is the sampling rate and Ns is the number of sampling points in the domain
1294
in [7] another cost function based on Least Mean Squared error (LMS) is considered instead of MSE (4) in
which LMS error function is employed in passband region of a low-pass filter.
( )
* ,
( )- +
( )
Magnitude frequency response would be guaranteed with this fitness functions, but phase response may not be
similar enough respect to the real system. the quality of magnitude response and the desired passband and
stopband frequency response might be insure the desired performance in adaptive signal processing applications,
whereas in adaptive system identification, high quality and similarity in a whole time responses and frequency
responses (magnitude and phase) of the estimated model are considered. As a result of this, the following error
function in order to control the phase response is proposed:
, ( )-
, ( )-+
( )
where
( )
However it will be shown, the other proposed objective function (7) is more better than the first one (6) in some
cases.
( )
3 Simulation Results
In this section, Four typical IIR filters taken from [6] - lowpass, highpass, bandpass and bandstop - are
considered as an unknown plants, in order to examine the suggested objective functions that are employed by
GA for system identification. and their performances are compared with each other and also with the
conventional objective function based on MSE in both time and frequency domain. Figure 1 to 4 depicted
magnitude and phase responses and also their step responses for four plants respectively. Figure 1 show that the
estimated model based on the first proposed objective function (equation 6) has the best performance in both
time and frequency responses whereas figure 2-4 indicate that the estimated model based on the second
proposed objective function (equation 7) is the best for this cases. Thus, this paper proposes equation 6 as a
fitness function when the unknown plant has low-pass frequency response otherwise equation 7 is useful as a
fitness function. This numerical results illustrate that this objective functions could be effective for adaptive
system identification, when there are frequency data from the unknown plant.
4 Conclusion
Magnitude frequency response could be guaranteed with the conventional fitness function , but phase response
may not be similar enough respect to the real system. the quality of magnitude response and the desired
passband and stopband frequency response might be insure the desired performance in adaptive signal
processing applications, whereas in adaptive system identification, high quality and similarity in a whole time
responses and frequency responses (magnitude and phase) are considered, Therefore , the error function based on
variance in order to control the phase response respect to the real system is suggested. Finally two objective
functions based on MSE, LMS and variance are proposed in order to attain acceptable performance for estimated
model in both time and frequency responses.
1295
Four typical IIR filters - lowpass, highpass, bandpass and bandstop - are considered as an unknown plants, in
order to examine the suggested objective functions that are employed by GA for system identification.
Experimental results, including step responses, magnitude and phase responses, illustrate the capability of the
proposed objective functions in compared to conventional one. Numerical results indicate that the suggested
fitness functions are effective in building an acceptable model for adaptive identification.
20
1.4
-20
1.2
-40
-60
0.1
0.2
0.3 0.4
0.5 0.6 0.7
0.8
Normalized Frequency ( rad/sample)
sys-real
sysh-objfunc1
0.9
1
sysh-objfunc2
sysh-objfunc3
100
Amplitude
Magnitude (dB)
Step Response
0.8
sys-real
sysh-objfunc1
sysh-objfunc2
Phase (degrees)
0.6
sysh-objfunc3
0
0.4
-100
-200
0.2
0.1
0.2
0.3 0.4
0.5 0.6 0.7
0.8
Normalized Frequency ( rad/sample)
0.9
10
15
Time (sec) 20
25
30
35
Fig1: The comparative performance of three estimated model (sysh-objfunc(i) based on objfunc(i), i=1,2,3) and
the real plant with low-pass frequency response (sys-real) objfunc1=equation(3), objecfunc2=equation(6),
objfunc3=equation(7) and Ns=100.
Step Response
Step Response
0.3
sysh-objfunc2
sysh-objfunc3
0.2
-40
0.1
0.2
0.9
100
syshq
0.15
sys-real
sysh-objfunc2
sysh-objfunc3
150
Phase (degrees)
sys-real
0.25
-20
Amplitude
Amplitude
Magnitude (dB)
20
0.1
0.05
50
0
0
-50
-0.05
0.1
0.2
0.9
10Time
Time(sec)
(sec)
15
20
25
Fig2: The comparative performance of three estimated model (sysh-objfunc(i) based on objfunc(i), i=1,2,3) and
the real plant with high-pass frequency response (sys-real) objfunc1=equation(3), objecfunc2=equation(6),
objfunc3=equation(7) and Ns=100.
Step Response
sys-real
sysh-objfunc1
sysh-objfunc2
sysh-objfunc3
sysh-objfunc1
sysh-objfunc2
sysh-objfunc3
-20
0.3
-40
-60
0.1
0.2
0.3
0.4
0.5
0.6
0.7
0.8
Normalized Frequency ( rad/sample)
0.9
0.2
0.1
200
Phase (degrees)
sys-real
0.4
Amplitude
Magnitude (dB)
20
0.5
-200
-0.1
-400
-0.2
0.1
0.2
0.3
0.4
0.5
0.6
0.7
0.8
Normalized Frequency ( rad/sample)
0.9
10
15
Time (sec) 20
25
1296
30
35
Fig3: The comparative performance of three estimated model (sysh-objfunc(i) based on objfunc(i), i=1,2,3) and
the real plant with band-pass frequency response (sys-real) objfunc1=equation(3), objecfunc2=equation(6),
objfunc3=equation(7) and Ns=100.
1.3
1.2
sysh-objfunc2
sysh-objfunc3
1.1
-20
1
-40
0.1
0.2
0.3
0.4
0.5
0.6
0.7
0.8
Normalized Frequency ( rad/sample)
0.9
200
Phase (degrees)
sys-real
sysh-objfunc1
1
sys-real
sysh-objfunc1
sysh-objfunc2
sysh-objfunc3
Amplitude
Magnitude (dB)
Step Response
20
0.9
0.8
0.7
0
0.6
-200
-400
0.5
0.1
0.2
0.3
0.4
0.5
0.6
0.7
0.8
Normalized Frequency ( rad/sample)
0.9
0.4
10 Time (sec)
15
20
25
Fig4: The comparative performance of three estimated model (sysh-objfunc(i) based on objfunc(i), i=1,2,3) and
the real plant with band-stop frequency response (sys-real) objfunc1=equation(3), objecfunc2=equation(6),
objfunc3=equation(7) and Ns=100.
References
1.
J.J Shynk, "Adaptive IIR Filtering," IEEE ASSP Magazine april 1989.
2.
D. Etter, M. Hicks, and K. Cho, Recursive adaptive filter design using an adaptive genetic algorithm, proc.IEEE
Int. Conf. on ASSP, vol. 7, May 1982, pp. 635638.
3.
4.
5.
S. Pai., W. K. Jenkins., and D. J. Krusienski, Adaptive IIR Phase Equalizers Based on Stochastic Search
Algorithms, Proc. of the 37th Asilomar Conf. on Signals, Systems, and Computers, November 2003.
6.
Yang Yu and Yu Xinjie, " Cooperative Coevolutionary Genetic Algorithm for Digital IIR Filter Design, " IEEE
Trans. Industrial Electronics ,vol. 54, no.3 , june 2007.
7.
N.Karabogal and B.Cetinkaya, " Performance Comparison of Genetic Algorithm Based Design Methods of Digital
Filters with Optimal Magnitude Response and Minimum Phase, " proc. IEEE Int Symp on Micro-Nano
mechatronics and Human Science 2003.
8.
S.Tai Pan, " Design of robust D-stable IIR filters using genetic algorithms
criterion," IEEE Trans. Signal Processing, vol. 57, no. 8, august 2009.
9.
D. J. Krusienski
and
W. K. Jenkins,
Design and performance of adaptive systems based on
structured stochastic optimization strategies, IEEE Circuits And Systems Magazine, FirstQuarter 2005.
10. J.T.Tsai, J.H. Chou and T.K.Liu, " Optimal design of digital IIR filters by using
genetic algorithm," IEEE Trans. Industrial Electronics ,vol. 53, no.3 , june 2006.
hybrid
taguchi
11. M.Haseyama and D.Matsuura, "A filter coefficient quantization method with genetic algorithm, including
simulated annealing," IEEE Signal Processing Letters, Vol. 13, No. 4, April 2006.
1297
Appendix
Low-pass IIR filter
(
(
(
)
(
)(
)(
)
)
)(
)(
)
)
(
(
)
(
(
(
(
)(
)(
)
)
)(
)(
)
)
(
(
(
Biographies
Tayebeh Mostajabi received her B.Sc. degree in Bioelectric Engineering in 2006 from Shahed University,
Tehran, Iran. Currently, she is pursuing her M.Sc. degree at Iran University of Science and Technology
(IUST) in electrical engineering, Tehran, Iran. Her research interests include swarm intelligence,
computational intelligence and its applications to control system theory, system identification and biomedical
engieering.
Javad Poshtanreceived his B.SC., M.SC and PHD degrees in Electrical Engineering from Tehran
University, Tehran, Iran, in 1987, Sharif University of Technology, Tehran, Iran, 1991, and University of
New Brunswick, Canada, in 1997, respectively. Since 1997 he has been with the Department of Electrical
Engineering at Iran University of Science and Technology. He is involved in academic and research activity
in areas such as control system theory, system identification, and estimation theory.
1298
Abstract. In this study, an application on distribution-free control chart takes place. This order statistic based
distribution-free control chart is proposed by [1]. In this paper, FAR and ARL values for different design
parameters are calculated. Also, this distribution-free control chart is applied to the data obtained from a
production plant. At the end, it is concluded that the process is statistically in control.
Keywords: distribution-free control charts, order statistics, false alarm rate, average run length
1 Introduction
Statistical process control can be defined as the application of statistical methods for monitoring and
controlling a process to ensure that it operates at its full potential for producing conforming products. In
statistical process control, causes of variation in the processes can be detected by using control charts. The
control limits of traditional Shewhart control charts are based on the assumption that the continuous quality
characteristic is distributed normally. But frequently, this normality assumption is not satisfied in real life data
and this affects the statistical properties of standard control charts.
When this is the situation, that is when normality assumption is not satisfied, distribution-free or nonparametric control charts can be used. Distribution-free control charts find wide applications in statistical process
control and there are several papers published in literature. Using distribution-free control charts provide many
advantages to the researchers. The first of these advantages is that, they do not need to assume a particular
probability distribution. Secondly, they do not require known process variance. Besides, they have the robustness
feature of standard nonparametric procedures and therefore, they are less likely to be affected by outliers, or the
shape of the distribution. Also, in-control run length distribution is the same for all continuous distributions.
They are more robust and more efficient in detecting changes when the underlying distribution is non-normal.
Among various distribution-free control chart types, the median based ones are quite advantageous. One of
the most important reasons of this is that, median is less sensitive than other central tendency measures to
outliers, measurement or recording errors, etc. in the data set.
This study consists of 6 sections. In Section 2, literature review on distribution-free median charts is given.
The method proposed by [1] which is studied in this paper is given in Section 3 together with some information
on average run length in Section 4. This method is applied with data obtained from a production plant in textile
sector and this application is given in Section 5. Finally, Section 6 gives the results of this study.
1299
3 The Method
It is assumed that there is a reference sample of size ,
and a test sample of size ,
is taken from it. The reference sample is assumed to have an in-control distribution ( ) and the purpose of this
method is to detect whether there is a change in ( ) towards an out-of-control distribution ( ) or not. Two
specific order statistics are chosen from this reference sample which are used as control limits where
.
,
(1)
|*
+|.
(3)
and are the design parameters of this control chart which can be determined according to a
specified false alarm rate; see [3], [4] and [1].
[1] improved the design parameters
and by [3] and [4] which can be determined according to a
specified false alarm rate given in (4)
(
(4)
where c shows that the process is in-control. FAR is defined as the probability of getting an out-of-control signal
while the process is actually in-control. It can be calculated as in (5) under the null hypothesis
,
(
1300
(5)
where is the probability that the chart does not give a signal. It is clear that
shows the probability that the
chart gives an out-of-control signal. [1] gives the calculation of false alarm rate as follows
) (
(6)
where
(
(
(
.
(
/
)(
)(
(7)
and
.
(8)
(9)
(10)
50
100
0.005
50
100
0.0027
50
100
9
17
27
9
17
27
9
17
27
9
17
27
9
17
27
9
17
27
2
4
6
13
19
24
4
7
9
11
17
21
4
7
9
3
5
17
49
47
45
88
82
77
47
44
42
90
84
80
47
44
42
98
96
84
5
9
14
5
9
14
5
9
14
5
9
14
5
9
14
5
9
14
and specific
6
10
14
1
5
4
2
1
1
1
5
8
1
1
1
6
11
10
FAR
0.00770
0.00827
0.00832
0.00888
0.00896
0.00895
0.00233
0.00249
0.00250
0.00462
0.00431
0.00467
0.00232
0.00249
0.00250
0.00234
0.00208
0.00208
1301
ARLin
20622600
216.491
349.753
481.484
8942.900
29430.200
432549
506.986
949.598
2120.590
8944.190
29430.200
432549
30787.500
19686.600
4 Application
The aim of this study is to use the distribution-free control charts in real life data. As mentioned before,
traditional Shewhart control charts have the assumption of normality. But in real life, data are not always
distributed normally. Therefore, it is better to use distribution-free control charts instead of traditional control
charts when the normality assumption is not satisfied.
For this purpose, the data obtained from a production plant in textile sector are used. The plant which is
located in Bursa has more than 200 employees. This plant produces fancy shirting fabric for women woven
outerwear. The data are the quantities of produced woven fabric by the weaving loom at once. The data set
consists of 50 data. In order to show why traditional control charts cannot be used for these data, AndersonDarling (AD) normality test is realised under the null hypothesis H 0: The data is distributed normally. The result
is given in Figure 1. Figure 1 shows that p value is less than significance level =0.05. Thus, the null hypothesis
H0 is rejected and this means that the data set is not distributed normally at =0.05.
As the data set is not normally distributed, it is convenient to use distribution-free control charts. For the
design parameters given in Table 1, random samples of sizes 9, 17 and 27 are taken from the data set. Here, the
data set of size 50 is the reference sample and random samples of sizes 9, 17 and 27 are test samples. Using this
reference sample and test samples, control limits (
and
), median values ( ) and L values are
determined as shown in Table 2. According to these values, two conditions of the control chart given by [1] are
checked and it is decided whether the process is in control or not. In Table 2, + means that aforementioned
condition is satisfied.
1302
0.005
0.0027
9
17
27
9
17
27
9
17
27
24
26.5
28
26.5
30
31.8
26.5
30
31.8
97
89
86.8
89
84.8
80
89
84.8
80
63.6
47.7
48
47.7
42.4
38
53
42.4
35
8
15
22
7
14
15
7
11
15
+
+
+
+
+
+
+
+
+
Condition 2
(
)
+
+
+
+
+
+
+
+
+
State of the
process
In Control
In Control
In Control
In Control
In Control
In Control
In Control
In Control
In Control
5 Conclusions
The distribution-free control charts are used when the underlying distribution is unknown or normality
assumption is not satisfied. This study shows FAR and ARL values for different design parameters for the
distribution-free control chart proposed by [1]. This chart is applied to a data set obtained from a textile
production plant. First, normality of the data set is tested with Anderson-Darling normality test and it is seen that
the data set is distributed nonnormally. Therefore, as traditional control charts would not be suitable for it,
distribution-free control chart is used with the aim of analyzing whether the process is in control or not. For all
test samples of different sizes, the process is concluded to be statistically in control.
References
[1] Balakrishnan, N.; Triantafyllou, I.S.; Koutras, M.V. (2010). A distribution-free control chart based on order
statistics. Communications in Statistics Theory and Methods 39: 3652-3677.
[2] Montgomery, D.C. 2001. Introduction to Statistical Quality Control, 4th ed. New York: John Wiley & Sons.
[3] Janacek, G.J.; Meikle, S.E. (1997). Control charts based on medians. The Statistician 46(1): 19-31.
[4] Chakraborti, S.; Van der Laan, P.; Van de Wiel, M.A. (2004). A class of distribution-free control charts. Journal of
the Royal Statistical Society: Series C (Applied Statistics) 53(3): 443-462.
1303
Abstract. In this study we made a comparison between thin plate splines, penalized generalized additive
models, and their Bayesian versions: Bayesian adaptive regression spline and Bayesian penalized splines
through the simulation study. We made a simulation study that was conducted to evaluate the performances
of the above nonparametric techniques. The functions for the simulation were taken from the different papers
[1], [2], [3]. For each function we took 100 values with 100 replications. The results of simulation are
compared using log 10 (MSE) with box-plots. The results of performance criteria are compared with each other
to realize the best estimator.
Keywords. Thin plate splines, penalized GAM, Generalized additive model, Bayesian Penalized spline,
simulation study
1 Introduction
Multivariate regression approach such as generalized additive models, thin plate splines, penalized
generalized additive models have a wide implementation in nonparametric problems. Their Bayesian versions,
Bayesian adaptive regression spline, Bayesian penalized splines used in many studies to improve prediction and
assessment of data set.
In this study we made a comparison between thin plate splines, penalized generalized additive models, and
their Bayesian versions: Bayesian adaptive regression spline and Bayesian penalized splines through the
simulation study. The dataset was sampled from five different functions. For this purpose we used one function
from book of [3], two functions from paper of [1] and two functions from paper of [2]. All the used functions
were used in papers previously. [2] used them for make a comparison study of the thin plate regression spline
with other methods. [3] used there function to made a comparison of local polynomial regression models.
The theoretical backgrounds of used techniques are in Section 2. The results of simulation study are
summarized in Section 3. The conclusion section concludes the paper.
1304
(1)
x i . First, we have to introduce additive models. In additive model each smooth function in (1) can be
variables,
introduced using a penalized regression spline basis. The spline basis for smooth functions can be represented in
this form
q 2
f1 ( x1 ) 1 x1 2 R( x1 , x1 j ) j 2
*
j 1
(2)
q 2
f 2 ( x2 ) 1 x2 2 R( x2 , x2 j ) j 2
*
j 1
where j are the unknown parameters for functions f , q is the number of unknown parameters for f ,
while
x *j are the knot location for the function. The wiggleness of the function can be measured as following
f ( x)
dx T S1
dx S 2
f ( x)
(3)
(y
i 1
The parameters
(4)
of the model are obtained by minimization of the penalized least squares objective
y X
Smoothing parameters
and
1 T S1 2 T S 2
(5)
smooth, relative to the objective of closely fitting the response data. The problem of choosing the smoothing
parameter is one of the main problems in curve estimation. We used generalized cross validation method for
choosing smoothing parameter.
y f
J md ( f )
(6)
1305
min y E T
T E subject to T
(7)
Thin plate regression splines [2] are based on idea of truncating the space of the wiggly components of the
thin plate spline. Let E UDU T be the eigen-decomposition of E, where D is a diagonal matrix of eigenvalues
of E arranged so that Di ,i Di 1,i 1 and the columns of U are the corresponding eigenvectors. Now let
Uk
denote the matrix consisting of the first kcolumns of U and D k denote the top right k k submatrix of D .
Restricting
min y U k Dk k T
kT Dk k subject to TT U k k 0
(8)
n k matrix of B-
a is the vector of regression coefficients; secondly, they suppose that the coefficients of adjacent Bsplines satisfy certain smoothness conditions that can be expressed in terms of finite differences of the a i s.
splines and
n
n
S yi a j B j ( xi ) (k a j ) 2
i 1
j 1
j k 1
(9)
For least squares smoothing we have to minimize S in (9). Penalized generalized additive models [7] consider
in the form of g ( ) Ba , where B (1 B1 .... B p ) is the N (1 n ) regressor matrix,
j
p
and
j 1
a ( , a1 ,..., a p ) . P-splines directly fit GAMs through a slightly modified method of scoring algorithm and
avoid the call to backfitting. P-GAM technique essentially eliminates the local scoring algorithm.
1306
3 Simulation Study
For simulation study we used functions from the book of [3], and papers of [1] and [2]. These are following
five functions used in our simulation:
(10)
1
3
f 4 ( x1 , x2 ) exp( 8x12 ) exp( 8 x22 )
5
5
(11)
(12)
(13)
x1 ~ N (0.5,1) , x2 ~ N (0.5,1)
f 5 ( x1 , x2 ) x1 sin(4x2 )
(14)
x1 ~ Uniform(0,1) , x2 ~ Uniform(0,1)
For each function we took 100 values with 100 replications. The results of simulation are compared using
log 10 (MSE) with box-plots. We used R software to make analysis of dataset and simulation study. We create
some helpful subprograms and functions to make our calculation easier. The empirical analysis are made by
mgcv package [12] of R software, the functions for calculating penalized generalized additive models, BayesX
software [AB], http://www.stat.uni-muenchen.de/~lang/bayesx and for the adaptive Bayesian regression spline
we used the functions from the http://www.stat.uni-muenchen.de/sfb386, which was created using C++.
f 1 ( x1 , x 2 )
f 2 ( x1 , x 2 )
1307
f 3 ( x1 , x 2 )
f 4 ( x1 , x 2 )
f 5 ( x1 , x 2 )
Fig. 1. Results of estimation of the function of simulation study
There are results of estimation of simulation data set in figure 1. In each figure you can see boxplots of the
log10 (MSE) . In each boxplot there are the techniques that we have used from left to right(1) P-GAM with 20
knots, (2) P-GAM with 30 knots, (3) P-GAM with 40 knots, (4) GAM with 20 knots, (5) GAM with 30 knots,
(6) GAM with 40 knots, (7) Thin plate splines with 20 knots, (8) Thin plate splines with 30 knots, (9) Thin plate
splines with 40 knots, (10) Bayesian P-spline with b=0.00005, (11) Bayesian 2 dimensional P-spline with
b=0.00005, (12) Adaptive Bayesian regression spline with 30 knots.
4 Conclusion
In this paper we made a comparison study between multivariate regression splines and Bayesian multivariate
splines. We used penalized generalized additive models, generalized additive models, thin plate splines, adaptive
Bayesian regression spline and Bayesian penalized spline. For this purpose we made a simulation study with five
different functions. We can make the following conclusions from the results that we get:
For Bayesian penalized spline was made the different hyperparameter selection. And the best result for
Bayesian penalized spline we could find when hyperparameters are a=1, b=0.00005. The graphs show only
selected hyperparameters results.
The adaptive Bayesian regression spline show the worst result among all techniques used in analysis.
The penalized spline showed best results between generalized additive models with regression spline
techniques. Thin plate spline and Bayesian penalized splines shows best result between all spline approaches that
we used. But in one function penalized spline outperforms another techniques.
The result of regression spline shows that this technique is preferable to use in analysis with not so
scattered dataset.
1308
References
1.
Smith, M. and Kohn, R. (1997). A Bayesian approach to nonparametric bivariate regression, Journal of the
American Statistical Association, 92, pp.1522-1535.
2. Wood, S.N. (2003). Thin plate regression splines, Journal of the Royal Statistical Society B, 65, pp.95-114.
3. Fan, J. and Gijbels, I. (1996). Local Polynomial Modelling and Its Applications, London : Chapman and Hall.
4. Hastie, T. and Tibshirani, R. (1990). Generalized additive models, London: Chapman and Hall.
5. Duchon, J. (1977). Splines minimizing rotation-invariant semi-norms in Sobolev spaces. In: Construction Theory of
Functions of Several Variables. Berlin: Springer.
6. Eilers P.H.C., and Marx B.D., (1996). Flexible smoothing using B-splines and penalized likelihood (with
comments and rejoinders), Statistical Science, 11(2), pp. 89-121.
7. Eilers P.H.C., and Marx B.D., (1998). Direct generalized additive modeling with penalized likelihood,
Computational Statistics and Data Analysis, 28, pp.193-209.
8. Lang S. and Brezger A., (2001). Bayesian P-splines, Journal of Computational and Graphical Statistics, 13,
pp.183-212.
9. Fahrmeir, L. and Lang, S., (2001). Bayesian inference for generalized additive mixed models based on Markov
Random Field Priors, Journal of the Royal Statistical Society C (Applied Statistics), 50, pp.201-220.
10. Carter, C. and Kohn, R. (1994). On Gibbs sampling for state space models, Biometrika, 81, pp.541-553.
11. Biller, C. (2000). Adaptive Bayesian Regression splines in semi parametric generalized linear models, Journal of
Computational and Graphical Statistics, 9 (1), pp. 122-140.
12. Wood, S.N. (2001). mgcv: GAMs and generalized ridge regression in R, R News, 1, pp.20-25.
Biographies
Akhlitdin Nizamitdinov - was born in Kanibadam, Tajikistan, in 1983. He received a diploma at Department of Computer
Science in Khujand branch of Technoligical University of Tajikistan, Khujand, Tajikistan. He is now a PhD student in the
Department of Statistics, at Anadolu University, Eskisehir, Turkey.
His main research interests are smoothing splines, regression splines, generalized additive models, penalized splines,
Bayesian penalized splines.
Memmedaga Memmedli - was born in Azerbaijan in 1947, He received a diploma at Baku State University in 1971, He
received a PhD diploma from Russian Academy of Science in 1977, From 1985 till 2006 he has worked as a Associate
Professor. From 2007 he is working as a Professor at the Department of Statistics, at Anadolu University, Eskisehir, Turkey.
His research interests are Game Theory, Control Theory, Optimization, Artificial Neural Networks, Fuzzy Modeling and
Nonparametric regression analysis. He has published many journal and conference papers in his research areas.
1309
1 Introduction
Forecasting is essential for planning and operation control in a variety of areas such as production
management, inventory systems, quality control, financial planning, and investment analysis. There are many
forecast methods in the literature [1-3]. In this study, taken discrete time series can be viewed as single trial
from a stochastic process since a continuous function can be transformed into a set of discrete observations. To
apply the maximum entropy principle, there are some assumptions about the time series. Firstly, the interval
between successive observations will be assumed to be equal in order to simplify the mathematical treatment.
Another assumption is that the mean of the observations is zero. This is not restrictive at all, because the series
can be easily transformed to another series with zero mean. Furthermore, the stochastic process generating the
observed time series will be assumed to be stationary.
While the raw time series is represented by [y1, y2,,yN], where yj is the j th ordered variate, an important
feature of the time series is given by the covariance between yj and a value yj+k which is k time lags away.
When the information about the time series is given as autocovariances up to a lag m, MaxEnt distribution of the
real time series can be determined as a multivariate normal distribution [1].
p(y )
N
2
1
1 2
exp y y
2
(1)
1310
H p y U ln
2 e 2
1
where is a toeplitz matrix. If we take into account the definition of entropy optimization function U , then (2)
is considered as entropy optimization function which will be required to define MinMaxEnt distribution [4-6].
, n and
define MinMaxEnt. When the future values are more than one, is an unknown vector. Therefore, by virtue of
MinMaxEnt distribution it is possible to estimate several future values in time series.
Fig. 1. Time series plot, ACF and PACF of U.S. unemployment rate
1311
The last five observations are excepted from the analysis in order to compare with predicted values. As a
tentative order, an ARMA(2,0) model is specified. This model is estimated with R 2.12.2. The values of the
estimated AR coefficients are 1 0.9229 and 2 0.2349 . Their estimated standard errors are 0.1113 and
0.1109. Both AR coefficients are significantly different from zero, and the estimated values satisfy stability
condition. The value of the log-likelihood is -47.72 and the AIC takes a value of 103.44. As an alternative, an
ARMA(1,1) model is specified. For this model, the values of the log-likelihood (-45.77) and the AIC (99.54) are
more favorable compared with the ARMA(2,0) specification. In R code 2.1.2, the estimated ARMA(1,1) model
is used to obtain forecasts five years ahead for the logarithm of the U.S. unemployment rate [3]. The obtained
forecasts are 1.603236, 1.660387, 1.689978, 1.705301 and 1.713234.
In the next step, the proposed forecasting process based on MinMaxEnt model for the stationary time series is
implemented by the program written in MATLAB. Once MinMaxEnt Model has been estimated, it can be used
to predict future values of yt . These forecasts can be computed recursively and are presented in Table 1.
Table 2. Actual and forecasted values for 1984-1988 years
Actual Values
1.5040
1.3350
1.3350
1.2809
1.2527
ARMA(1,1)
1.6032
1.6603
1.6899
1.7053
1.7132
MinMaxEnt
1.5815
1.5567
1.5526
1.5541
1.5557
For purposes of comparing our results, the root mean squared error (RMSE) is often useful. The RMSE values
provided by ARMA(1,1) and MinMaxEnt Models are calculated as 0.356044 and 0.231898, respectively. It is
seen that the RMSE of MinMaxEnt model is lower than the RMSE of ARMA(1,1) model. Therefore, for these
models, one would prefer the MinMaxEnt model over the ARMA(1,1) specification.
Furthermore, in order to evaluate the performance of MinMaxEnt Model, data sets consist of 100 time series
which have each sample size of 50 observations is simulated from AR(1) and AR(2) models by the use of
program written in MATLAB. For each time series, two forecasts are determined by employing the MinMaxEnt
model and the process is repeated 100 times. In the next step, forecast errors are calculated by the difference
between the actual value and the forecast value for the corresponding period and mean square errors (MSE)
given in Table 2 are obtained.
Table 2. Simulation results for forecasting via MinMaxEnt Model
Model
Coefficients
AR(1)
AR(1)
AR(2)
AR(2)
0.2
0.8
0.8 0.2
0.4 -0.7
0.04
0.04
0.5
0.5
MinMaxEnt
MSE
0.0429
0.0796
0.6229
0.7535
These simulation results confirmed that the MinMaxEnt Model is a useful tool for forecasting in stationary times
series. In the future study, simulation studies can be developed to the different autoregressive and moving
average models for the larger the sample size.
1312
Load Rveri.mat
N=size(x,1);
x=log(x)
m=2; %the number of autocovariances
l=N %the position of estimated value
a=vpa(zeros(1,m+1));
lamda=vpa(zeros(1,m+1));
r=vpa(zeros(1,N-1));
q=vpa(zeros(1,N));
c0x=0;
cx=vpa(zeros(1,m));
syms ort
syms t;%
top=(sum(x)-x(l));
ort=(top+t)/N;
syms y
for i=1:N
if i==l
y(l)=t;
else
y(i)=x(i);
end
end
for i=1:N
c0x=c0x+(y(i)-ort)^2;
end
for k=1:m
for i=1:N-k
cx(k)=cx(k)+((y(i)-ort)*(y(i+k)-ort));
end
end
r0=c0x/N;
r(1)=vpa(cx(1)/N);
r(2)=vpa(cx(2)/N);
R=vpa(zeros(m+1,m+1));
for i=1:m+1
for j=1:m+1
k=abs(j-i);
if j~=i
R(i,j)=r(k);
else
R(i,j)=r0;
end
end
end
D=inv(R);
gkare=D(1,1);
1313
a(1)=1;
a(2)=D(2,1)/D(1,1);
a(3)=D(3,1)/D(1,1);
for k=0:m
for j=0:m-k
lamda(k+1)=lamda(k+1)+a(j+1)*a(j+k+1);
end
lamda(k+1)=lamda(k+1)*gkare;
end
for j=0:N-1
for k=1:m
q(j+1)=q(j+1)+2*lamda(k+1)*cos((2*pi*k*j)/N);
end
q(j+1)=q(j+1)+lamda(1);
end
dett=prod(q);%entropy optimization function
%minimization of entropy optimization function
fid=fopen('mesaa.m', 'w');
fprintf(fid,'function fu=mesaa(y)\n');
fprintf(fid,'t=y;\n');
s = char(dett);
fprintf(fid,'fu=-(%s);\n', s);
fprintf(fid,'return;\n\n');
fclose(fid);
[tt,fval]=fminbnd(@mesaa, min(x), max(x));
clear mesaa.m
clear(s)
%Entropy of MinMaxEnt Distribuiton
Hminmax=(N/2)*log(2*pi*exp(1))+1/2*log(1/-fval);
tt % 0
exp(tt)%forecast
4 Conclusion
This paper explored how to apply MinMaxEnt Modeling to forecast time series data under Maximum Entropy
Distribution. MinMaxEnt distributions (models) are obtained on the basis of MaxEnt distributions dependent on
parameter and it is shown that the proposed method can be used to forecast in time series. In order to evaluate
the performance of this method is applied on real time series consisting of the unemployment rate of the United
States and it is required a computer program the written in MATLAB to do computations. In addition to this, via
simulation studies, it is shown that the forecast error which is the difference between the actual value and the
forecast value for the corresponding period is very low and it is seen that MinMaxEnt modeling provides reliable
and efficient forecasting outputs.
1314
References
1.
Kapur, J.N. and Kesavan, H.K. 1992. Entropy Optimization Principles with Applications. New York: Academic
Press.
2. Wei, W.S. 2006. Time Series Analysis, Univariate and Multivariate Methods. U.S.A: Pearson.
3. Pfaff, B. 2008. Analysis of Integrated and Cointegrated Time Series with R. U.S.A.: Springer.
4. Shamilov, A. (2006). A Development of Entropy Optimization Methods. WSEAS Transaction on Mathematics,
5(5): 568-575.
5. Shamilov, A. (2007). Generalized entropy optimization problems and the existence of their solutions. Physica A
382: 465-472.
6. Shamilov, A. (2010). Generalized entropy optimization problems with finite moment functions sets. Journal of
Statistics and Management Systems 3(13): 595-603.
7. Shamilov, A. and Giriftinoglu, C. (2010). Generalized Entropy Optimization Distributions Dependent on Parameter
in Time Series. WSEAS TRANSACTIONS on INFORMATION SCIENCE &APPLICATIONS 1(7): 102-111.
8. Shamilov, A. and Giriftinolu, . at all (2008). A New Concept of Relative Suitability of Moment Function Sets.
Applied Mathematics and Computation (SCI) 206: 521-529.
9. Shamilov, A. and Giriftinolu, . (2009). Some relationships between entropy values of entropy optimization
distributions for time series. The Scientific and Pedagogical News of Odlar Yurdu University 28: 10-17.
10. Shamilov, A. and Giriftinolu, . (2009). A new method for estimating the missing value of observed time series
and time series forecasting. The Scientific and Pedagogical News of Odlar Yurdu University 28: 17-23.
Biographies
Aladdin Shamilov Aladdin Shamilov was born on 10 March, 1943, in Azerbaijan. In 1963, He graduated from Gence
Pedagogical University in Physics and Mathematics. He received a PhD diploma from Baku State University in 1972. In
1984, He completed the second PhD from the Ukraine Academy of Science. He received the title of professor at Baku
State University in 1987. He has worked as a professor since 2001 in Anadolu University, at Department of Statistics,
Eskisehir, Turkey.
He has published over 100 international conference papers and journals in his research areas. He is interested in differential
equations, information theory, stochastic processes, time series analysis and generalized entropy optimization principles.
Cigdem Giriftinoglu Cigdem Giriftinoglu was born on 19 August, 1979, in Izmir, Turkey. She graduated from Department
of Statistics of Anadolu University in 2001. She finished her Master Degree at Statistics in 2005. She completed her PhD in
2009, currently works as a research assistant in Anadolu University Department of Statistics.
She has published over 15 international conference papers and journals in her research areas. She is interested in stochastic
processes, time series analysis and entropy optimization principles.
1315
Abstract.Analysis of stability of slopes in terms of factor of safety (Fs) is important to define the stability of a
slope. Most stability analysis is carried out under static loading. However, in a seismically active region,
earthquakes are one of the important forces that can cause the failures of slopes. In this study, attempts were
made to estimate of the (Fs) values of three slopes with different heights subjected to earthquake forces. To
achieve this, a computer program based on Simplified Bishop and Simplified Janbu Method was prepared in
the Matlab programming environment. With the help of this program, the (Fs) values of three slopes were
estimated for different earthquake parameters by using both methods and comparisons were made between
the Fs values obtained from the methods. It has been found that Fs values obtained from the simplified Bishop
method are less than those obtained from the simplified Janbu method.
Keywords:slope stability, simplified Bishop method, simplified Janbu method
1 Introduction
Analysis of stability of slopes in terms of factor of safety (Fs) is important to define the stability of a slope. Most
stability analysis is carried out under static loading. However, in a seismically active region, earthquakes are one
of the important forces that can cause the failures of slopes. A slope becomes unstable when the shear stresses on
a potential failure plane exceed the shear resistance of the soil. The additional stresses due to earthquake further
increases the stresses on these planes and decreases the Fs value further.
Limit equilibrium methods have been widely adopted for slope stability analysis [1]. A potential sliding surface
is assumed prior to the analysis and a limit equilibrium analysis is then performed with respect to the soil mass
above the potential slip surface. Many methods based on this approach are available. For example, Ordinary [2],
Simplified Bishop [3], Simplified Janbu [4], Morgenstern and Price [5], Spencer [6], and Sarma [7].
In this study, attempts were made to estimate of the Fs values of three slopes with different heights subjected to
earthquake forces. To achieve this, a computer program based on Simplified Bishop and Simplified Janbu
Method was prepared in the Matlab programming environment. [8] Then, the Fs values of three slopes were
estimated by using both methods and comparisons were made between the Fs values.
1316
2 Literature review
Earthquakes are a major trigger for instability of natural and man-made slopes. Often the instability of slopes
due to an earthquake causes more destruction and kills more people than the actual earthquake itself [9]. More
recently, the magnitude 7.5 Guatemala Earthquake of 1976 is reported to have generated at least 10.000
landslides and slope failures [10]. The slope instabilities and resulting disasters after the El Salvador Earthquakes
in 2001 hit the front pages of the papers worldwide [9]. Natural and artificial slopes may become equally
unstable during an earthquake. To calculate how a slope will behave during an earthquake or to design a slope
that is also stable under earthquake conditions is not that easy [9].
The factor of safety (Fs) based on an appropriate geotechnical model as an index of stability, is required in
order to evaluate slope stability. However, the stability analysis of slopes is difficult because of many
uncertainties. The problem is statistically indeterminate, and some simplifying assumptions are necessary in
order to obtain a critical factor of safety (Fs) [11]. For slope stability analysis, the limit equilibrium method is
widely used by engineers and researchers and this is a traditional and well established method. [1] Alkasawneh
et.al. has shown that limit equilibrium methods are reliable and can be used with confidence to investigate the
stability of slopes.[12] Due to the differences in the assumptions, several limit equilibrium methods such as
Ordinary [2], Simplified Bishop [3], Simplified Janbu [4], Morgenstern and Price [5], Spencer [6], and Sarma [7]
have been proposed.
2.2 Methods
In this study, while the properties of the soil involved in the slope given in Figure 1 were kept constant, the
effects of the slope height and the magnitude of the earthquake (M), the distance between the center of the
earthquake and the location of the slope (R) and the frequency of the cyclic loading (N) on slope stability were
investigated. The slope of the soil was taken constant as 2 (horizontal) / 1 (vertical) while the height of the slope
(h) was taken as 5, 10 and 15 m. The geometry of the slope and the properties of the soil involved in the slope
are given in Fig. 1. In order to examine the effect of ground water the geometry of the ground water layer was
taken as the mid-height of the slope as shown in Figure 1.
Fig. 1. The geometry and the properties of the soil involved in the slope.
To observe the effect of the earthquake forces to the slope, the parameters were chosen as follows: The M
value was selected as 6, 7 and 8; the R value was selected as 5, 10, 20, 50 and 100 km; and the N value was
selected as 5, 10, 30, and 60. The factor of safety, Fs, for each case was then calculated by using a computer
program developed in the Matlab programming environment [8] based on the Simplified Bishop Method (Eq. 1)
and the Simplified Janbu Method (Eq. 2)
W u b tan c b / m
W sin a W
n
Fs
'
i 1
i 1
'
i i
i i
(2)
i 1
1317
W u b tan c b / n
W tan a W
n
Fs f 0
'
i 1
i 1
'
i i
i i
(2)
i 1
where c is the cohesion, is the angle of internal friction, u is the pore water pressure at the base, f0 is the
'
correction factor for the Simplified Janbu Method, ap is the peak acceleration value of the earthquake and
and
ni
mi
(3)
ni cos mi
(4)
The pore pressure in the slope increases with the repeated shear stress during the earthquake and this increase
depends on the earthquake and soil properties. Seed and Idriss [14] gave the following equation for the
calculation of the additional increase of the pore water pressure,
u 0,915 LogN 0,108 0,65 a p 1 0,0075 H cot
(5)
where u is the pore water pressure at the base,and ap is the peak acceleration value of the earthquake and can be
expressed from the Equation (6) :
1, 09
(6)
It can be seen from the Equation (5), the effect of h on the increase of the pore water pressure is ignorable.
Therefore, the Equation (5) changes as follows:
u 0,6 a p cot log N
(7)
2.3 Results
In order to examine the influence of R in the Fs value, the Fs value calculated from both methods without
considering ground water was plotted against the R value for the chosen M values, as shown in Figures 2 to 4 for
the h value of 5, 10 and 15m, respectively. It can be seen from Figure 2, the Fs value was increased with an
increase in the R value, and decreased with an increase in the M value for each R value. The similar trend was
also observed for the other slope heights (Figures 3 and 4). Also, it can be noted that the Fs value decreased with
an increase in the h value. It can be also observed that the Fs value calculated from the simplified Bishop method
was less than those from the simplified Janbu method.
1318
Bishop M=6
Janbu M=6
Fs
Bishop M=7
Janbu M=7
Bishop M=8
Janbu M=8
R (km)
Fig. 2. The effect of R on the Fs value depending on M, Bishop versus Janbu of h=5 m
Bishop M=6
Janbu M=6
Fs
Bishop M=7
Janbu M=7
Bishop M=8
Janbu M=8
R (km)
Fig. 3. The effect of R on the Fs value depending on M, Bishop versus Janbu of h=10 m
Bishop M=6
Janbu M=6
Fs
Bishop M=7
Janbu M=7
Bishop M=8
R (km)
Janbu M=8
Fig. 4. The effect of R on the Fs value depending on M, Bishop versus Janbu of h=15 m
The influence of R and M in the Fs value from both methods with also considering ground water had similar
trend as described above for the h value of 5, 10 and 15m. In order to examine the influence of N in the Fs value,
the Fs value calculated from both methods with ground water was plotted against the N value for the chosen M
values, as shown in Figures 5 to 7 for the h value of 5, 10 and 15m, , respectively. In these figures, the R value
was chosen as 10 km. It can be seen from Figure 5, the Fs value was decreased with an increase in the N value
for each M value. The similar trend was also observed for the other slope heights (Figures 6 and 7). Also, it can
be noted that the Fs value decreased with an increase in the h value. It can be also observed that the Fs value
calculated from the simplified Bishop method was less than those from the simplified Janbu method.
1319
Bishop M=6
Janbu M=6
Fs
Bishop M=7
Janbu M=7
Bishop M=8
Janbu M=8
N
Fig. 5. The effect of N on the Fs value depending on M, Bishop versus Janbu of h=5 m
Bishop M=6
Janbu M=6
Fs
Bishop M=7
Janbu M=7
Bishop M=8
Janbu M=8
N
Fig. 6. The effect of N on the Fs value depending on M, Bishop versus Janbu of h=10 m
Bishop M=6
Janbu M=6
Fs
Bishop M=7
Janbu M=7
Bishop M=8
Janbu M=8
N
Fig. 7. The effect of N on the Fs value depending on M, Bishop versus Janbu of h=15 m
3 Conclusions
The following conclusions have been obtained from this study:
Low factor of safety (Fs) values were obtained in both methods for the earthquakes with high magnitudes.
1320
When considered the influence of the earthquake properties on the Fs value, R and M have more effects in
both methods on the slope stability safety factors as compared to N. Lower Fs values were obtained for the
higher values of the R and M parameters.
In both methods, in addition to the earthquake effects, the existence of ground water also decreased the Fs
value by considerable amount. That is why the ground water should be avoided in the artificial slopes.
Also in both methods, the increase in the slope height has negative effect on the the Fs value.
When compared the simplified Bishop Method with the Simplified Janbu Method, the simplified Bishop
method gives lower Fs values than the simplified Janbu method.
References
1.
2.
3.
4.
5.
6.
7.
8.
9.
10.
11.
12.
13.
14.
Cheng, Y.M., Lansivaara, T., Wei, W.B. (2007). Two-dimensional Slope Stability Analysis By Limit Equilibrium
And Strength Reduction Methods.
Fellenius, W. (1936). Calculation of the stability of earth dams, In Transactions of the 2nd International Congress
on Large Dams, Washington, USA, 445-462,
Bishop, A. W. (1955). The use of the slip circle in the stability analysis of slopes, Gotechnique5(1), 7-17.
Janbu, N. (1954). Application of composite slip surface for stability analysis, In Proceedings of European
Conference on Stability of Earth Slopes, Stockholm, Sweden, 43-49.
Morgenstern, N. R., and Price, V. E. (1965). The Analysis of the Stability of General Slip Surfaces. Geotechnique,
Vol 15, No. 1, pp 79-93.
Spencer, E. (1967). A method of analysis of the stability of embankments assuming parallel interslice forces.
Gotechnique, 17(1): 1126
Sarma, S. K. (1979). Stability Analysis of Embankments and Slopes, Journal of the Geotechnical Engineering
Division, American Society of Civil Engineers, ASCE, Vol. 105, No. GT 12, pp. 1511-1524.
Cetin, T. (2010). Developing a computer program for analysis of slope stability and comparing different analysis
methods, MSc. Thesis, Celal Bayar University Manisa, (In Turkish).
Hack, R., Alkema, D., Kruse, G.A.M., Lenders, N., Luzi, L. (2007.) Influence of earthquakes on the stability of
slopes.
Anon. (1997). Report on Early Warning Capabilities for Geological Hazards. Convener of International Working
Group, and first author: Dr. Robert Hamilton Chairman, International Decade for Natural Disaster Reduction
(IDNDR). Early Warning Programme, IDNDR Scientific and Technical Committee Washington, D.C.USA.
Document 12085, IDNDR Secretariat, Geneva. 35 pp.
Tan, . (2006). Investigation of soil parameters affecting the stability of homogeneous slopes using the Taguchi
methods, Eurasian Soil Science, 39, 1248-1254.
Alkasawneh, W., Malkawi, A.I.H., Nusairat, J.H., Albataineh, N. (2008) A Comparative Study of Various
Commercially Available Programs in Slope Stability Analysis.
Matasovic, N. (1991). Selection of Methods for Seismic Slope Stability Analysis.
Seed, H.B., Idriss, I.M. (1971). Simplified Procedure for Evaluating Soil Liquefaction Potential, Journal Soil
Mechanics and Foundations Division, ASCE, Volume 97(9), pp. 1249-1273.
Biographies
Tlin etin is a lecturer in the Vocational School of Turgutlu at CelalBayarUniversity, Manisa. She received her B.Sc.
degrees in computer engineering from EgeUniversity, Izmir, Turkey in 1992 and in civil engineering from
CelalBayarUniversity, Manisa, Turkey in 2004. She is now a Ph.D student in the Department of Civil Engineering at
CelalBayarUniversity, Manisa, Turkey in the hydraulic field (2010-), from where she received her M.Sc. degree in
geotechnical engineering field in 2010.
Her research interests include: computer programming, softwares in civil engineering, slope stability, hydraulic.
Yusuf Erzin is an assistant Professor in the Department of Civil Engineering at CelalBayarUniversity, Manisa, Turkey. He
received his B.Sc. degree from SelukUniversity, Konya, Turkey. Also, he received his M.Sc and Ph.D degrees from Middle
EastTechnicalUniversity, Ankara, Turkey.
His research interests include: artificial neural networks, swelling soils, slope stability and crushing strength of sands.
1321
Abstract. In the present study, MaxMaxEnt and MinMaxEnt distributions obtained on the basis of MaxEnt
distribution dependent on parameter for time series are comprised in the sense of entropy. It is proved that as
number of autocovariances k goes on increasing; the entropy of both MaxMaxEnt and MinMaxEnt
distributions goes on decreasing. MinMaxEnt distributions also possesses this property that entropy of
MinMaxEnt distributions form a decreasing sequence when the number of autocovariances increases.
Comparing MaxMaxEnt and MinMaxEnt distributions generated by different autocovariance vectors, it is
possible to obtain autocovariance vectors generating distributions neighboring in the sense of entropy. If
MaxMaxEnt and MinMaxEnt distributions generated by different autocovariance vectors are neighboring in
the sense of entropy, then both numbers , generating these distributions can be accepted as estimators
to missing or future values. It should be noted that MinMaxEnt distribution is more suitable for estimation to
missing value or future values in time series.
Keywords: MaxMaxEnt distribution, MinMaxEnt distribution, time series, entropy, information
1 Introduction
In [1] by considering entropy optimization measure on the given set of moment vector functions special
functional according to each measure is defined. Via the moment vector functions giving the least and the
greatest values to the mentioned functional, Generalized Entropy Optimization Distributions (GEOD) are
obtained. Each of this distributions is closest (or remotest) to the given a priori distribution in the corresponding
measure. The process of obtaining GEOD is formulated as a development of entropy optimization methods.
In [2], it is proposed sufficient conditions for the solutions of GEOD. In [3,4,5,6] and [4] are considered some
development and applications of GEOD. Furthermore, in [7] GEOD for stationary time series are defined on the
basis of MaxEnt distribution dependent on parameter.
It is important to establish relations between entropies of MaxMaxEnt and MinMaxEnt distributions
according to different autocovariance vectors of stationary time series. This problem has also a consequential
rule in the sense of information theory. For this reason in the present study MaxMaxEnt and MinMaxEnt
modeling according to different autocovariance vectors of time series are investigated in the sense of entropy and
compared.
1322
y0 ,
, ys 1 , , ys 1 ,
, rk of given
min y j , max y j and entropy value of this distribution be H Prk y . MaxMaxEnt and
js
js
Let k k 0,1,
i.e.
(3)
Let k k 0,1,
i.e.
(2)
respectively. Then, between entropy values of these distributions, the following inequalities are satisfied:
H Y0 H Y1 H Y2
H Y0 H Y1 H Y2
H Yk H Yk , k 0,1,
Proof:
The
fact
that
entropy
H Yk
H Ym .
(3)
H Ym
(4)
(5)
,m
of
model
Yk
generated
by
autocovariance
vector
rk k r0 k ,
rk k r0 k ,
Indeed according to (1) and (2) the inequality (5) follows. For any fixed k , the inequality
H Prk 1 y H Prk y
(6)
Therefore
(7)
1323
H Pr y H Pr y H Yk
k 1
k
k
k
(8)
H Yk 1 H Yk k 0,1,
,m
(4)
,m
(3)
H Yk 1 H Yk k 0,1,
These complete the proof of theorem.
H Y0 H Ym H Y1 H Ym1
H Yk H Ymk
H Ym H Y0 .
(9)
Remark 2. If m1 m2 then,
or
H Ym1 H Ym2 .
(40)
Therefore in order to use MaxMaxEnt distribution Ym1 as model for estimation missing or future values, it is
sufficient to obtain numbers m1 and m2 for which
H Y H Y
m1
(51)
m2
Ym2 neighbouring. The inequality (10) indicated that in solving estimation problems MinMaxEnt models are
more suitable than MaxMaxEnt models. Because, when m2 is sufficiently large, the MinMaxEnt model Ym2
gives good results in the sense of entropy or information. But according to (10), MaxMaxEnt model cannot
provide mentioned results.
1324
4 Conclusion
In this study, generalized entropy optimization distributions as MaxMaxEnt and MinMaxEnt distributions are
compared in the sense of entropy. It is shown that mentioned models for stationary time series constitute
decreasing sequences in the sense of entropy when the number of autocovariances generating these distributions
increases. Between entropy values of as MaxMaxEnt and MinMaxEnt distributions generated by the same
autocovariances, there are inequality that entropy value of MaxMaxEnt model greater than the corresponding of
MinMaxEnt model.
It is shown that in solving estimation problems in time series MinMaxEnt models are more suitable than
MaxMaxEnt models. Because, when the number of autocovariances is sufficiently large the MinMaxEnt model
gives good results in the sense of entropy or information, but MaxMaxEnt model cannot be provide mentioned
results.
References
1.
Biographies
Aladdin Shamilov Aladdin Shamilov was born on 10 March, 1943, in Azerbaijan. In 1963, He graduated from Gence
Pedagogical University in Physics and Mathematics. He received a PhD diploma from Baku State University in 1972. In
1984, He completed the second PhD from the Ukraine Academy of Science. He received the title of professor at Baku
State University in 1987. He has worked as a professor since 2001 in Anadolu University, at Department of Statistics,
Eskisehir, Turkey.
He has published over 100 international conference papers and journals in his research areas. He is interested in differential
equations, information theory, stochastic processes, time series analysis and entropy optimization principles.
Cigdem Giriftinoglu Cigdem Giriftinoglu was born on 19 August, 1979, in Izmir, Turkey. She graduated from Department
of Statistics of Anadolu University in 2001. She finished her Master Degree at Statistics in 2005. She completed her PhD in
2009, currently works as a research assistant in Anadolu University Department of Statistics.
She has published over 15 international conference papers and journals in her research areas. She is interested in stochastic
processes, time series analysis and entropy optimization principles.
1325
1 Introduction
Migraine headache is a disease known for century. The most important factor in the diagnosis of migraine,
the patients medical history, age at onset of pain, location of pain, characteristics, and the course of pain,
accompanying symptoms and examining of neurological disorders is a great importance in terms of diagnosis. In
parallel to developing technology, for interrogated of neurogical functions and put forward a clear diagnosis is
widely used neurology examination tools such as Magnetic Rezonans(MR), Brain Tomography and
Electroencephalographic(EEG) in clinics[4,2].
Potentials of the EEG measured head through the surface of the electrodes is the sum of the potentials from the
underlying celebral cortex and a wide area. There is a very complex structure of EEG signals and difficult to
interpret [3]. Potentials of the EEG measured head through the surface have peak amplitude range of 1-100 V
and frequency band is 0-100 Hz. Despite EEG signals have a wide frequency in 0-100 Hz, in the interpretation
and analysis of EEG signals is very important frequency range of 0-30 Hz. O-30 Hz frequency range includes
four basic EEG frequency range. This frequency band ranges, respectively, is 0,5-4 Hz (Delta), 4-8 Hz (Theta),
8-13 Hz (Alpha), 13-30 Hz (Beta) frequency ranges.
EEG signals are not periodic and frequency, phase, amplitude varies continuously. When EEG signals are
interpreted by experts because of both long-term records and time axis, visual analysis of this signs is deficiency.
In Parallel with todays technology, to overcome this deficiency and obtain meaningful result that EEG signals
developed different signal processing techniques and methods of statistical analysis. In the literature, there are no
finding migraine-specific of EEG. Migraine, migraine with aura and migraine without aura is basically classified
into two groups. Severe pain in attack migraine without aura lasting 4-72 hours. In migraine with aura, the pain
duration is shorter than 60 minutes also it is examined that what sort of changes occurs on migraine at pregnancy
term.
1326
2. Methodology
2.1. EEG Data Collection and Methodology
In this paper, EEG signals of people diagnosis with migraine, people in the during pregnancy and healty persons
examined with spectral analysis methods and whether diagnosis of migraine by EEG tools can be observed.
EEG data used in this study selected female patients ranging in ages 18 to 35 from Cumhuriyet University
Research and Training Hospital Department of Neurology Branch and study with three different experimental
groups (17 healthy people, 9 migraine patients and 5 migraine patients during pregnancy) were studied. EEG
data is used for migraine are those of persons diagnosis with migraine without aura in migraine attacks and in the
process to apply to the clinic with complaints of headache was obtained. According to international 10-20
electrode system for analysis, all channels shown that in the Figure 1.1 are selected as in monopoles and sampled
with 200 Hz sampling frequency.
(1)
Welchs method, signal is divide into intervals of overlapping and dividing that these ranges sing calculated by
windowing. Last in, calculated that this periodogram is calculated average[8]. As a result, Welchs method for
the estimation of power spectral density;
1327
Is given by the expression.Where L is length of time series. Short data records are consolidated and nonrectangular windows with the investigation reduces the predictive resolution. If Welch method is low signal to
ratio, it allows us to achieve a solution better than periyodograma [5-6].
2.3. Yule-Walker Method
The output of p series which have AR model, white noise input the sum of previous p linear combination of
outputs. Weight on the outputs of the previous p, autoregressive mean square estimation error is minimized[5-9].
If x[n] and y[n] output of existing value is zero mean of white noise input, AR(p) model would be;
, - ,
, -
(3)
( )
)|
(4)
Fig. 3.1.EEG signal in time axis(respectively healthy person, migraine patients and migraine patient during
pregnancy)
1328
In order to obtain a diagnosis EEG signals uses in wide range of signal frequency spectrum. EEG frequency
spectrum contains four basic frequency ranges. This frequency range is frequency ranges 0-4 Hz(Delta), 4-8
Hz(Theta), 8-13 Hz(Alpha) and 13-30 Hz(Beta). Matlab 7.0 is used for analysis results.
According to the standard international 10-20 electrode system was used for analysis of all channels. For
analysis non-parametric Welch method and parametric Burg, Yule-Walker, Covariance and Modified
Covariance methods for analysis are used. As a result of analyzing from the channels calculated the Power
Spectral Density (PSD).
With healthy persons, migraine patients, the migraine patients during pregnancy of the tree experimental groups
including power spectral density were obtained by calculating the average power spectral density for each
channel. Average power spectral density speared that meaningful to EEG frequency ranges of 0-4 Hz(Delta), 4-8
Hz(Theta), 8-13 Hz(Alpha) and 13-30 Hz(Beta) and the tree experimental groups compared with statistical t-test
and threshold was chosen of 0.05(p values given in tables). The fallowing statistical tables are obtained as a
result the comparison;
1329
the
Yule-Walker
method
compare
to
PSD
1330
Covariance
method
compare
to
PSD
1331
Figure.3.2. PSD calculated from channels Fp1 chart with Modified Covariance Method between 0-35
frequencies a) Delta b) Teta c) Alfa d) Beta (Blue: Healthy persons, Red: migraine patients, Green: migraine
patients during pregnancy)
1332
References
[1] eker, M., Tokmak, M., Asyal M.H., Semen, H., Examining EEG Signals with Parametric and Non-Parametric
Analyses Methods in Migraine Patients during Pregnancy, Biyomut 2010, Antalya, Turkey
[2] Akn, M., Kymk, M.K., Arserim, M.A., Trkolu, .,
Separation of Brain Signals Using FFT and Neural Networks,
Biyomut 2000, stanbul, Turkey
[3] Dahlf, C., Linde M., One-year prevelance of migraine in sweden: a population- based study in adult. Cephelalgia 2001;
664-671
[4] Lleci, A., Maltepe lesi Dourganlk andaki Kadnlarda Migren Prevalansnn aratrlmas,TC Salk Bakanl
Dr Ltfi Krdar Kartal Eitim Ve Aratrma Hastanesi Nroloji Klinii , Uzmanlk Tezi., stanbul, 2004
[5] Proakis, J.G., Manolakis, D.G., Digital Signal Processing Principles, Algorithms, and Applications , Prentice-Hall, New
Jersey, 1996.
[6] beyli, E. D., Gler, I., Selection of optimal AR spectral estimation method for internal carotid arterial Doppler signals
using Cramer-Rao bound. Comput.Electr. Eng, 30,491508, 2004
[7] beyli, E. D, Gler ., Atardamarlardaki Daralma ve Tkankln Maksimum Olabilirlik Kestiriminin AR metodu ile
ncelenmesi, Gazi niversitesi Fen Bilimleri Dergisi, 375-385, 2003
[8] Zhenmei Li, Jin Shen, Bouxe Tan, Liju Yin, Power System nterharmonic Monitoring Based on Wavelet Transform and
AR Model, IEEE Pasific- Asia Workshop on Computational Intelligence and Industrial Aplication, 2008
[9] Alkan, A., Suba, A., Kymk, M.K., Epilepsi Tansnda MUSIC ve AR Yntemlerinin Karlatrlmas, IEEE 13. Sinyal
leme Uygulamalar Kurultay(SU05) 2005, Kayseri,Trkiye
[10]Bronzino J. D. 1996., The Biomedical Engineering handbook, IEEE Pres, 3erd edition
Biographies
Mustafa EKER was born in stanbul, Turkey in 1981. He is continuing Ph.D education in electric and electronic
engineering from the nn University of science institute since 2010 and has been working Cumhuriyet University Divrii
Nuri Demira Vocationaly High School since 2006.
1333
Abstract.We study the overdamped motion of a particle in a double well potential under fractional Gaussian
noises with positive and negative power law correlations and with ordinary Gaussian noise. The equation of
motion is solved numerically by a fourth order Runge Kutta method. We notice that the recrossing times have
exponential distribution while they still keep long range correlations.
1 Introduction
The Brownian motion in a double well potential is applicable in a number of fields. Usually an additive Gaussian
white noise is assumed in these processes. The assumption of a white noise force, where the correlation function
of the noise is proportional to the delta function, is sometimes not fully valid. In this paper, we investigate the
double-well system with a power law correlated Gaussian noise force.
2 The Noise
We generate fractional Brownian motion (fBm) trajectories, ( ), by the midpoint replacement method (Ref. [1],
[2], [3]). The process takes
steps. To get the results close to the real fBm, must be large enough. Figure 1
shows the process for
. In step zero, all values are zero. In the first step, random values are given for
initial time
, middle time
and final time
(full circles) and the average of each couple of
values are written at their middle times,
and
(empty circles) as shown in the figure. In the
second step, random numbers are added to the all previous nonzero values and again averaged values are
written at the middle of each couple. This repeats at each step until the final step where, for
, random
numbers are added to all times
. The randomly selected numbers have Gaussian distribution
with zero average but the variance get smaller at each step. The variance for the step is
1334
where
(6)
()
(7)
The auto-correlation functions are power law functions for both of fGn as seen in Figure 2. They are plotted in
normal and log-log scales to see the values in short and long times. The relationship between the power index
( )>
and scaling is
for fGn, which can be obtained by simply taking the second derivative of
.The super-diffusive fGn has positive auto-correlation with
and sub-diffusive fGn has negative autocorrelation with
. The ordinary Gaussian noise is white noise correlated with delta function, which is not
shown.
The power spectrums for both cases are shown in Figure 3. The spectrum is proportional to
for
and it is proportional to for
as expected from the Fourier transform of power law function,
( ) .
1335
Figure 2 Auto-corelation functions of the fGn for positive (square) and negative (triangle) correlations and the
fitting formulas (dashed and solid lines). (The negative function is taken absolute value to show in the
logarithmic scale). The small graph is the linear plot.
Figure 3 The power spectrum of the fGn that have negative (down) and positive (up) auto-correlations.
3 Runge Kutta
We have a particle in the potential
( )
(8)
is the mass, V(x) is the double well potential and is the noise with intensity
(10)
1336
The noise is added only at the integer times. The equations of motion, is solved numerically without the noise
between each integer times, using a forth order Runge Kutta method with time difference
. For each
integer time , the values of (
) is obtained from ( ) by
(
where ( )
and
(11)
. Finally, noise
).
Figure 4 An example trajectory of x(t) using the noises that has negative (bottom), delta (middle) and positive
(top) poverlaw auto-correlations.
1337
Figure 5 Distribution of the recrossing times of the trajectory ( ) using the noises that have negative (triangle),
delta (circle) and positive (square) power law auto-correlation functions.
Figure 6 Auto-correlation of the recrossing times of the trajectory ( ) using the noises that have negative
(triangle), delta (circle) and positive (square) power law auto-correlation functions.
References
1.
2.
3.
4.
5.
6.
7.
Voss, R. F.; Earnshaw R. A. (Editor) (1985). Fundamental Algorithms for Computer Graphics, Random Fractal
Forgeries. Berlin: Springer-Verlag.
Peitgen H.; Jurgens H.; and Saupe D. (1992). Chaos and Fractals, New Frontiers of Science. New York. SpringerVerlag.
Fournier A.; Fussell D.; Carpenter L. (1982). Computer rendering of stochastic models, Communications of the
ACM 25(6): 371-384.
Mandelbrot B. B.; Van Ness J. (1968). The Fractional Brownian motions, fractional noises, and applications, SIAM
Rev. 10(4): 422-437.
Feder J.(1988). The Fractals. New York. Plenum Press.
Cakir R.; Grigolini P.;and Krokhin A. A. (2006). Dynamical origin of memory and renewal, Phys. Rev. E 74(2):
021108-1.
Romero A. H.; Sancho J. M.; Lindenberg K. (2002). First Passage Time Statistics For Systems Driven By Long
Rage Correlated Gaussian Noises,Fluctuation and Noise Letters 2(2): L79:L100.
1338
Biographies
Rait akr Dr. akr was born in stanbul, Turkey in 1975. He graduated from the Physics Department of the Middle
East Technical University, Ankara, Turkey with a B.Sc. in 1998, and M.Sc. in 2001. He got his Ph.D. from University of
North Texas, Denton, TX in 2007. His research includes stochastic process, complexity. Currently he is a faculty member in
the Department of Physics at Ar brahim een University.
Osman alar Akn Born in 1971 in Mersin Turkey, Osman alar Akn received his B.Sc. in Physics from Middle East
Technical University in 1994. He received his M.Sc. from METU Physics Department in 1997 with a thesis on the design
considerations of an IFEL accelerator. He received his Ph.D. degree from University of North Texas in 2008 with his work
on perturbation of renewal processes. Currently he is a faculty member in the Department of Physics at Fatih University.
1339
Literature Review
Crash rates have been found to be higher on low-volume roads than on other roads. In a study on a sample of
nearly 5,000 miles of paved two-lane rural roads in seven states, Zegeer et al. (1994) computed a crash rate of
3.5 per million vehicle miles (MVM) on low-volume roads (defined in the study as roads with 2,000 average
daily traffic [ADT]), versus a crash rate of 2.4 per MVM on all high-volume roads. Zegeer et al. determined that
fixed object crashes, rollover crashes, and other run-off-road crashes were more frequent on the lower-volume
roads.
Additionally, some studies have found that crash severity is relatively high on low-volume roads. Caldwell
and Wilson (1997) compared the injury crash rates on unpaved county road sections in Albany County,
Wyoming, to injury crash rates on all roads in the state. Injury risk on county roads was determined to be more
than five times higher than on all roads. However, this report was based on a very small sample of road sections
1340
and crashes. The present study has covered seven years of crash data from all of the ninety-nine counties in the
state of Iowa.
Methodology
Descriptive statistics for seven years of available GIS crash data (2001-2007, later updated to 2008), were
prepared for rural low-volume (400 Average Daily Traffic) secondary road crashes, as well as for all other 2lane rural road crashes for comparison purposes. The descriptive statistics included a broad range of both crashand driver/vehicle-level attributes. Iowa DOT Geographic Information Management Systems (GIMS) roadway
data were integrated with crash data for analysis. System-wide crash data were utilized to investigate trends in
frequency, rate, and severity with respect to the various crash characteristics and general road types. This also
facilitated a test of proportions analysis.
1341
Figures 2.a and 3.b: Comparison data for various road classes
A test of proportions was employed to determine various crash characteristics which were overrepresented
on different low-volume road categories. Ten different secondary road categories were compared to comparison
groups (e.g., paved and unpaved secondary roads in varying AADT ranges compared to all undivided two-lane
rural primary roads).
The proportion of various characteristics of crashes on roads of different jurisdictions, traffic volume ranges,
and surface types were computed. These proportions were then statistically tested to determine if differences
between pairs of proportion were statistically significant. Given adequate sample sizes and assumptions of
independence of the proportions, the z-statistic for a standard Normal random variable was utilized for the test.
A significance level of 0.05 was selected for the test. The results point to the types of crashes that are particularly
problematic for low-volume roads. These characteristics were also used to develop a crash-based statistical
model.
The test of proportions identified crash characteristics that were overrepresented on low-volume rural roads as
compared to the comparison group. One of the categories found to be overrepresented on low-volume rural roads
was crashes involving younger drivers (drivers 19 and younger). One mitigation strategy that is a prime
candidate for action is to provide young drivers with practical experience driving on unpaved roads and to focus
on the dangers of, and recovery from, excessive speed.
1342
CRASH SEVERITY
Fatal
Major Injury
Minor Injury
Possible/Unknown
Property Damage Only
Frequency
% Frequency
%
235 1.5%
563 1.7%
933 5.8%
1,410 4.1%
2,750 17.1%
3,433 10.1%
2,929 18.3%
5,142 15.1%
9,193 57.3%
23,492 69.0%
CRASH TYPE
MV (No Ani/Obj on Road)**
SV ROR (No Ani/Obj on Road)*
SV Other (No Ani/Obj on Road)*
SV ROR (Ani/Obj on Road)*
SV Other (Ani/Obj on Road)*
Hit Animal
Frequency
% Frequency
%
2,940 18.3%
9,539 28.0%
9,121 56.9%
7,366 21.6%
507 3.2%
1,460 4.3%
984 6.1%
491 1.4%
35 0.2%
109 0.3%
2,453 15.3%
15,075 44.3%
P-Value
Rank:
Ratio Ratio Diff.
Rank:
Diff.
0.000001
0.000001
0.000001
1.40
1.70
1.21
68
50
85
1.7%
7.1%
3.2%
44
20
31
0.000001
2.63
29 35.2%
0.000001
4.25
12
4.7%
24
MANNER OF CRASH/COLLISION
Frequency
% Frequency
%
MV Head-on**
378 2.4%
855 2.5%
MV Read-end**
524 3.3%
3,211 9.4%
MVFigure
Oncoming
left turn**
206 1.3%
692 unpaved
2.0%
4: Sample
table showing the test of proportions
results (0-100
secondary vs 2 lane primary)
MV Broadside**
1,007 6.3%
1,896 5.6% 0.001555
1.13
94 0.7%
68
MV Sideswipe-Same Dir.**
217 1.4%
1,377 4.0%
MV Sideswipe-Oppos. Dir.**
391 2.4%
1,035 3.0%
VARIOUS CHARACTERISTICS OF CRASHES ON UNDIVIDED RURAL ROADS BY ROAD SYSTEM, VOLUME RANGE, AND SURFACE TYPE
MV
Other**
217
1.4%
473 1.4%
Statistical Comparison of Proportions
SV Coll.
with(2001-2007)
Fixed Obj.*
4,050 25.2%
3,841 11.3% 0.000001
2.24
36 14.0%
10
State
of Iowa
Top
20 highlighted in each column
Comparison
SV Overturn/Rollover*
3,332 20.8%
2,279 6.7% 0.000001
3.10
19 14.1%
9
Group
SV
CollwFixObj+Over/Roll*
2,574
16.0%
1,519
4.5%
0.000001
3.60
13
11.6%
14
Road System:
Secondary Secondary Secondary Secondary Secondary Secondary Secondary Primary-ALL
SV Other*
691 4.3%
1,787
AADT
Range:
0-100
0-100
101-400
101-400 5.2%0-400
0-400
0-400
70-12200
Surface
Type:
Unpaved
Paved 15.3%
Unpaved 15,075
Paved 44.3%
Unpaved
Paved
All
All Paved
Hit Animal
2,453
Total Crashes
Total Daily VMT
SPEED/WEATHER-SURFACE
RELATION
Total
Centerline-Miles
Weighted
Speed RelAverage of ADT
Crash Rate (per 100M VMT)
Weather/Surface Rel
Crash Density (crash per mile)
Speed&Weather/Surface Rel
Other/Unknown
CRASH SEVERITY
Fatal
LIGHTInjury
CONDITIONS
Major
Minor
Day Injury
Possible/Unknown
Dark, dusk
or dawn
Property
Damage
Only
16,040
470
5,502
7,603
21,542
8,073
29,615
34,040
2,447,000
72,080
677,000
1,911,000
3,124,000
1,983,000
5,107,000 13,460,000
Frequency
% 4,345
Frequency7,421 % 71,245
66,900
1,063
8,484
79,729
5,963
37
44
230
6,894 6843.0% 160 4,991 260
14.7% 0.000001
2.93
23 64
28.3% 2,300
3
257
255
318
156
270
159
227
99
1,143 0 7.1%
3,053 9.0%
0
1
1
0
0.95
0
6
1,619 10.1%
68
50
85
Frequency
%
44
42
40
78
20
46
2148.8%
54
7,829
31
75
Frequency
59
56
34 %74
48
56
2012,891
61
20
51 0.000001
20
59
37.9%
26
82
29
6,559 40.9%
12,552 36.9% 0.000001
1,652 10.3%
8,597 25.3%
Other/Unknown
CRASH TYPE
MV (No Ani/Obj on Road)**
SV ROR (No Ani/Obj on Road)*
SV Other (No Ani/Obj on Road)*
SV ROR (Ani/Obj on Road)*
SV Other (Ani/Obj on Road)*
Hit Animal
Figure
MANNER OF CRASH/COLLISION
MV Head-on**
MV Read-end**
MV Oncoming left turn**
MV Broadside**
MV Sideswipe-Same Dir.**
MV Sideswipe-Oppos. Dir.**
MV Other**
SV Coll. with Fixed Obj.*
SV Overturn/Rollover*
SV CollwFixObj+Over/Roll*
SV Other*
3,739 11.0%
68
43
7
52 75 20
10.9%7
87
31
4
97
4.0%
29
30
26
25
28
25
29
12
24
10
23
13
28
10
46
27
84
11
24
10
26
10
26
7
1
98
83
34
20
14
8
10
13
7
7
7
3
68
86
64
98
67
36
19
14
9
10
12
1343
36
19
13
10
9
14
28
31
27
44
4
15
17
29
36
23
12
8
12
13
26
30
15
74
4
12
11
54
27
30
17
69
4
13
12
47
15
27
(Note: Ratio is the proportion of crashes in the low volume road category divided by the proportion of crashes
for rural primary 2-lane; these ratios are ranked and the top 20 for each LVR category are shaded. Similarly, the
differences in proportions are ranked (the second column in each LVR category).
Figure 6: A scatterplot matrix of correlations between various variables (Created using the JMP software)
1344
1345
Table 1: Explanation of variables used in the model and their signs in the output
X3
X9
X10
X11
X15
X16
X19
X31
X32
X33
X34
X35
X36
X37
X39
X40
X41
X42
X43
X47
X50
X55
Num. Obsrv.
Iterations
Num. Var.
L.L.F.
R.L.L.F.
Ro-sqrd
Corr. Ro-sqrd
Chi-sqrd
SRFCTYPE_
FLATTRRN
RLLGTRRN
HLLYTRRN
SUMMER
WEEKEND
NIGHT
COLFIXOB
HITANIML
OVERROLL
MVBRDSID
IMPAIRED
ENVCOCIR_
DARK
SPEEDREL
WESURREL
YOUNGINV
OLDERINV
TOTRUPO
VMTPRUPO
CONATINT
CRSHORDT
16942
34
22
-14828.75
-15802.50
0.0616200
0.0602278
1947.508
+
+
+
+
+
+
+
-
As seen in Table 1, the R-squared, corrected R-squared, and Chi-squared numbers were monitored as
goodness of fit measures.
The thresholds for outputs of crash severity levels:
y = 1 (fatal or serious injury crash)
Z 2.38039631
Z > 3.78842872
1346
Findings
The test of proportions compared more than 200 aspects of crashes. More than 60 of these factors were
statistically significantly more risky on low-volume roads. Since the test of proportions included various pairs
of comparison groups, it was hard to tell which of these 60+ factors were the most important. The crash-level
statistical model (the ordered probit model) has supplemented the test of proportions analysis, accounted for the
effect of multiple causal factors and explanatory variables, and helped determine the top risk factors of a more
limited number. For low-volume rural roads (400 vpd or less, excluding intersections with roads carrying any
higher traffic) the following factors were found to increase the severity of crashes:
1.
Paved surfaces
2.
Spring/Summer months (April thru September)
3.
Weekends
4.
Fixed objects struck
5.
Overturn/rollover crashes are more severe
6.
Multi-vehicle broadside collisions are more severe
7.
Impaired driving, including both alcohol and/or drug involvement
8.
Daytime
9.
Speeding
10.
Younger (19) and older (65) driver involvement
11.
Counties with less total rural population* and lower vehicle miles traveled (VMT) per rural population*
12.
Counties with positive traffic control at intersections* (information was available for 73 of Iowas 99
counties).
* Counterintuitive results that may warrant further study
The following factors were found NOT to increase the severity of crashes
1.
2.
Animal collisions
Weather, environment, and surface-related factors
Conclusions(Implementation Readiness)
The findings of this study should permit local agencies to prioritize mitigative action in the most cost-effective
manner. Results indicate that county engineers, law enforcement, and educators should especially consider safety
needs on unpaved local roads with traffic volumes of 101 to 400 AADT. By considering safety needs on these
relatively limited mileage roads, agencies can achieve cost-effective results when deploying low-cost safety
improvements such as signing upgrades, higher maintenance levels, and spot improvements. Especially effective
may be targeted signing programs for curves on gravel roads.
Law enforcement agencies will want to consider directing appropriate resources to local roads with high
incidence of crashes involving impaired, speed-related, and other driver behaviorrelated incidents. Driver
educators could present crash statistics for novice drivers and offer instruction for driving on unpaved surfaces,
in an effort to reduce speed and vehicle controlrelated crashes. Parents could also be encouraged to provide
supervised experience on these roads. Finally, the Iowa Department of Transportation and the Institute for
Transportations (InTrans) Iowa Traffic Safety Data Service (ITSDS) could step up outreach and crash data
sharing to local agencies, focusing especially on 101400 AADT unpaved local roads where cost-effectiveness is
expected to be high.
1347
Acknowledgement
Special thanks to Dr. Reginald Souleyrette, Dr. Nadia Gkritza, and Zachary Hans
References
1. Caldwell, R. C., Wilson, E. M., A Safety Improvement Program for Rural Unpaved Roads. MPC Report No. 97-70.
January 1997
2. FHWA and NHI, Guide to Safety Features for Local Roads and Streets. 1992.
3. Iowa Department of Transportation, Office of Traffic and Safety, Engineering Bureau, Highway Division, Historical
Summary of Travel, Crashes, Fatalities, and Rates (1970-2007) - State of Iowa. Updated on October 8, 2008.
4. Khattak, A. J., Pawlovich, M. D., Souleyrette, R. R., and Hallmark S. L., Factors Related to More Severe Older Driver
Traffic Crash Injuries. Journal of Transportation Engineering, May-June 2002.
5. Ksaibati, K. and Evans, B.,Wyoming Rural Roads Safety Program. TRB 88th Annual Meeting, Washington, D.C.,
November 2008.
6. Liu, L. and Dissanayake, S.,Speed Limit-Related Issues on Gravel Roads. Kansas State University, August 2007.
7. Liu, L. and Dissanayake, S.,Examination of Factors Affecting Crash Severity on Gravel Roads. Kansas State University,
November 17, 2008.
8. Madsen, M.,Farm Equipment Crashes on Upper Midwest Roads. Midwest Rural Agricultural Safety and Health Forum
(MRASH), 2008.
9. Manual Uniform Traffic Control Devices for Streets and Highways (MUTCD), Part 5: Traffic Control Devices for LowVolume Roads. FHWA, U.S. Department of Transportation, 2003 Edition.
10. Neenan, D.,Driver Education Survey Results on Rural Roadway Driving - The National Education Center for
Agricultural Safety. Midwest Rural Agricultural Safety and Health Forum (MRASH), 2008.
11. Roadway Safety Tools for Local Agencies -A Synthesis of Highway Practice, NCHRP Synthesis 321. 2003.
12. Russell, E. R., Smith, B. L., and Brondell, W. , Traffic Safety Assessment Guide. Kansas State University, Civil
Engineering Department, April 1996.
13. Souleyrette, R. R., Guidelines for Removal of Traffic Control Devices in Rural Areas. 2005.
14. Tate III, J., Wilson, E., Adapting Road Safety Audits to Local Rural Roads. MPC Report No. 98-96B. October 1998.
15. The American Traffic Safety Services Association, Low Cost Local Roads Safety Solutions, 2006.
16. US Department of Transportation Rural Transportation Initiative. May 23, 2006.
http://www.communityinvestmentnetwork.org/nc/single-news-item-states/article/us-department-of-transportation-ruraltransportation-initiative/terrain%2C%20faster%20speeds/?tx_ttnews[backPid]=782&cHash=e4906f03c4, Last reviewed:
June 27, 2009.
17. Zegeer, C. V., Stewart, R., Council, F., Neuman, T. R., Accident Relationship of Roadway Width on Low-Volume
Roads. TRR 1445, 1994.
18. Zegeer, C. V., Stewart, R., Council, F., Neuman, T.R., Roadway Widths for Low-Traffic-Volume Roads (NCHRP Report
362). 1994.
Note
The study has yielded to an official project report titled Safety Analysis of Low Volume Rural Roads in
Iowa which was presented to the Iowa Department of Transportation (the sponsor) by Institute for
Transportation at Iowa State University, Ames, Iowa, USA in December 2010. www.intrans.iastate.edu can be
visited for color PDF files of this and other research reports.
1348
Abstract. The electroencephalogram (EEG) is a simple and efficient method to present the electrical activity
of the brain as recorded by placing a number of electrodes on the scalp. The EEG is widely used for
diagnostic evaluation of various brain disorders such as determining the type and location of the activity
observed especially during an epileptic seizure or for studying sleep disorders.
Dichotic listening (DL) literally means presenting two auditory stimuli simultaneously, one in each ear, and
the standard experiment requires that the subject report which of the two stimuli was perceived best. During
this procedure, the EEG signals of the subjects were recorded by a 64 channel cap.
In this study ZM statistics is used to find the most similar responses. Again by using ZM statistics, magnitude
and behavior similarity of signals in different electrodes and stimuli are investigated by manipulating the
usage of ZM statistics.
Keywords: EEG, signal similarity, Zm statistics, dichotic listening.
1 Introduction
Brain is one of the most important elements of human beings. The limits of brain functions and capabilities are
still not completely known and provide a large area of research. There are several methods for following the
brain activity such as CAT (Computer Axial Tomography), PET (Positron Emission Tomography), MRI
(Magnetic Resonance Imaging), MEG (Magneto Encephalography) and EEG (ElectroEncephaloGraphy).
EEG is a noninvasive clinical tool in better understanding the human brain and for diagnosing various
functional brain disturbances. It is a simple and efficient way of measuring electrical activity of brain. By
conducting some electrodes on the scalp, the voltage values can be measured on different locations of the brain
[1].
Dichotic listening is one of the favorite studies on human brain monitoring. In dichotic listening, the subjects
are alerted by dichotic stimuli which are meaningless, consonant vowel syllables on both ears. The subjects then
present the syllable they heard through a 6 button keypad. Dichotic listening has been used in hundreds of
research and clinical reports related to language processing, emotional arousal, hypnosis and altered states of
consciousness, stroke patients, psychiatric disorders and child disorders, including dyslexia and congenital
hemiplegia [2].
Measures of similarity are required in a wide range of radar sonar, communications, remote sensing, artificial
intelligence and medical applications, where one signal or image is compared with another. In this study a new
measure of similarity (ZM), suitable for both signal processing and image processing applications is presented [3]
and applied to detect similarities in EEG signals.
1349
The application is based on EEG signals received during a dichotic listening test from 2 different subjects. The
signal similarities are investigated between electrodes, stimuli and ear advantages. ZM statistical test is applied to
detect similarities on amplitude. Because of the weakness of the test in detecting signal behavior similarity
detection, a manipulation is made. The results are compared.
1350
3 Dichotic Listening
One of the favorite studies on human brain monitoring is dichotic listening. In cognitive psychology, dichotic
listening is a procedure commonly used to investigate selective attention in the auditory system. In dichotic
listening, two different auditory stimuli (usually speech) are presented to the participant simultaneously, one to
each ear, normally using a set of headphones. Participants are asked to attend to one or (in a divided-attention
experiment) both of the messages. They may later be asked about the content of either message.
Dichotic listening literally means presenting two auditory stimuli simultaneously, one in each ear, and the
standard experiment requires that the subject report which of the two stimuli was perceived best.
Dichotic listening has been used in hundreds of research and clinical reports related to language processing,
emotional arousal, hypnosis and altered states of consciousness, stroke patients, psychiatric disorders and child
disorders, including dyslexia and congenital hemiplegia.
1351
direction of the source is set equal to the angular coordinate for which the null hypothesis is least likely, i.e. the
test statistic is maximized.
The delay-and-sum beam-former is applied as
(1)
where xm(n) is the nth sample output from the mth delay channel and y(n) is the beam-formed output. By
applying some statistical relation rules under certain assumptions, the following equation is derived:
(2)
Knowing that under the null hypothesis the ZM statistic is F distributed, allows a detection threshold to be
computed to give the desired probability of falsely rejecting the null hypothesis when it is indeed true (a false
alarm). If the computed ZM value exceeds the threshold then a localizable signal is instead assumed to be present
[3].
A disadvantage of the ZM measure of similarity is that it requires the signal power levels in each channel to be
approximately equal.
1352
(3)
where M=2 (for 2 electrodes compared), n=300 (300 msecs of the signal), xm(n) is the measured value of mth
electrode and y(n) is the sum of voltage values of electrode-1 and electrode-2 at time=n. The resulting tables are
given in Figure 2 and Figure 3.
Fig. 2.ZM statistics of Data Set 1 (Left Ear Advantaged). Higher ZM means higher similarity.
Fig. 3.ZM statistics of Data Set 2 (Right Ear Advantaged). Higher ZM means higher similarity.
Fig. 4.ZM statistics of REA, LEA and HOM signals in two data sets.
EEG signal averages of subject with Left Ear Advantage has the greatest similarity in CZ and C4 electrodes on
right ear responses and CZ and C3 on left ear responses. Other subject, which has a Right Ear Advantage, has the
greatest similarity values in CZ and C4 electrodes on the right ear responses (same as the other subject) but CZ
and C4 in left ear responses, different from the other subject.
Figure 4 shows the signal similarities in REA, LEA and HOM signal averages on both subjects. Left Ear
Advantaged has the greatest similarity in REA and HOM signals, but Right Ear Advantaged subject has the
greatest in LEA and HOM signals.
1353
Fig. 5.x(n) and x(n)+k signals. Different signals in amplitude but same in shape.
In this example ZM statistics of two signals are calculated as 0.426 which means these signals are not similar,
they are just noises.
In order to catch the behavior of the signals similarity, a new vector of z is obtained by taking the differences
between t and t + 1 and ZM statistics is calculated on these difference vectors. (z(t) = x(t + 1) x(t)). Analyzing
the ZM statistics on difference vectors, it was observed that CZ-F4 similarity on REA of Data Set 1, and CZ-F3
and CZ-F4 on REA and CZ-F3, CZ-F4 and C4-F3 on LEA of Data Set 2 were more similar in behavior than in
amplitude. On the other side, while ZM value of C4-T8 on LEA of Data Set 2 was 14.28 (which means high
similarity in amplitude), ZM_Differencevalue of the same electrode pairs was calculated ad 0.99 (which means there
is no significance similarity). Also C3-T7 and T7-T8 electrode pairs have the same situation of similarity in
amplitude but no similarity in signal shape.
REA-LEA (ZM_Difference)
w=50 msec Data Set 1 Data Set 2
CZ
150-200
150-200
C3
200-250
0-50
C4
250-300
250-300
F3
200-250
350-400
F4
100-150
150-200
T7
100-150
100-150
T8
250-300
300-350
1354
Table 2. Most similar time intervals with 100 msec. time interval
REA-LEA (ZM)
w=100 msec Data Set 1 Data Set 2
CZ
400-500
200-300
C3
100-200
200-300
C4
400-500
200-300
F3
100-200
300-400
F4
100-200
100-200
T7
100-200
400-500
T8
100-200
200-300
REA-LEA (ZM_Difference)
w=100 msec Data Set 1 Data Set 2
CZ
100-200
100-200
C3
0-100
100-200
C4
200-300
300-400
F3
0-100
100-200
F4
100-200
200-300
T7
100-200
400-500
T8
200-300
400-500
As seen in the table values, data sets show a particular pattern in ZM statistics but most similar time intervals
vary in ZM_Difference values.
6 Conclusion
EEG signals recorded during dichotic listening are in concern of this study. Responses given with right and
left ear are averaged and compared to detect the signal similarity. ZM is used as the signal similarity measure and
signals on different electrodes are examined. It was observed that the similarities extremely differ in left ear
responses in two data sets.
The weakness of ZM statistics is tried to be avoided by manipulating on the data. A new series is constructed
by taking the differences between contiguous seconds. By this technique, signal similarity in shape is detected
efficiently. Doing so, some other differences and similarities are observed differing from ZM results.
Another study was on detecting most similar time intervals of the related signals. Again ZM and ZM_Difference
statistics are used and the results are displayed.
References
1.
2.
3.
4.
5.
6.
7.
8.
Srnmo, L., Laguna, P. (2005), Bioelectrical Signal Processing in Cardiac and Neurological Applications
Biomedical Engineering, London: Elsevier Academic Press
Hugdahl, K. (2005), Symmetry and asymmetry in the human brain. Academia Europaea, European Review, Vol.
13, Supp. No. 2, p: 119133
Hugh L. Kennedy (2007), A New Statistical Measure of Signal Similarity. Information, Decision and
Control,Digital Object Identifier: 10.1109/ IDC.2007.374535
Vahaplar, A., elikolu, C. C., zgren, M. (2011), Entropy in Dichotic Listening EEG Recordings, Mathematical
and Computational Applications, Vol. 16, No. 1, pp.43-52, 2011
Bayazt, O., niz, A., Gntrkn, O., Hahn, C., and zgren, M. (2008), Dichotic listening revisited: Trial-by-trial
ERP analyses reveal intra- and interhemispheric differences. Neuropsychologia, Article in Press.
Agilent Technologies, (2000), The Fundamentals of Signal Analysis, Document No: 6/00 5952-8898E
Rangaraj M. Rangayyan, (2002), Biomedical Signal Analysis, Calgary: Wiley and Sons Inc.,
Akay, M. (2000), Nonlinear Biomedical Signal Processing Volume I, NY: IEEE Press
1355
Abstract.In this study, incident database of Ministry of National Education was investigated to provide
valuable structured data for improving service quality. A data mining process, based on IBM SPSS Modeler
Professional Clementine, a professional data mining tool, was applied to data in order to discover the insights
in data, classifying the information to interpret data and make decision on how to improve service quality. In
this paper, incident database of Ministry of National Education and Business Process are introduced and the
valuable knowledge was produced and reported to institution to be used for setting next year objectives on
service level agreement with stakeholders in auctions by using. The study results indicate that general service
level for every type incident category especially Furniture and Infrastructure and some locations is low.
Keywords: Data mining, classification, decision tree, incident database, service desk; service level
agreement
1 Introduction
In recent years, especially in Turkey, government agencies and big business organizations have built their own
unified contact center and service desk for giving good and effective service to their employees and customers.
The service desk can be mentioned with some other titles such as help desk, contact center and call center.
However, all missions and goals of the organizations are nearly similar, their strategies, operations, organizations
and technologies may differ in their contact center and service desk architecture. Despite their missions and
goals are quite similar, firms use different strategies, operations, organizations and technologies in their contact
center and service desk architecture. Because, in this architecture, organizations define their own communication
channels such as 7/24 call center operator service, e-mail, web, mobile, service and service levels, work
processes, workflow models and roles in organization for employees professions and experiences.
In this study, incident database of The Ministry of National Education was investigated to provide valuable
structured data for improving service quality. A data mining process, based on IBM SPSS Modeler Professional
Clementine, a professional data mining tool, was applied to data in order to discover the insights in data,
classifying the information to interpret data and make decision on how to improve the service quality.
Data mining is an extension of statistical analysis to identify valuable hidden patterns and subtle relationships
by extracting information from a collection of data through various techniques such as decision trees, classifying,
clustering analysis, rule induction, and statistical approaches. Data mining uses a combination of machine
learning, statistical analysis, modeling techniques and database technology.
1356
2 Literature Review
In study of Anand and others, a subset of incidents from facilities in Harris County, TX, extracted from the
National Response Center Database was investigated. By classifying the information into groups and using data
mining (DM) techniques, interesting patterns of incidents according to the characteristics such as type of
equipment involved, type of chemical released and causes involved were revealed and further these were used to
modify the annual failure probabilities of equipment [1]. The results of data mining indicate customer
relationship, customer segmentation, customer valuation, such as credit rating and marker base analysis [2-5].
The other study conducted by S. C. Hui and G. Jha, investigates how to apply data mining techniques to
extract knowledge from the database to support two kinds of customer service activities: decision support and
machine fault diagnosis. A data mining process, based on the data mining tool DBMiner, was investigated to
provide structured management data for decision support. In addition, a data mining technique that integrates
neural network, case-based reasoning, and rule-based reasoning have been proposed; it would search the
unstructured customer service records for machine fault diagnosis. The proposed technique has been
implemented to support intelligent fault diagnosis over the World Wide Web [6].
The use of computer technology in decision support is now widespread and pervasive across a wide range of
business and industry. This has been resulted in the capture and availability of data in immense volume and
proportion [7]. There are many examples that can be cited. Mainly, decision support system based on KDD
(Knowledge discovery in databases). KKD can be defined as non-trivial process of identifying valid, novel,
potentially useful, and ultimately understandable patterns in data [8]. In this case, data mining is a phase also
central step in the KDD process concerned with applying computational techniques (DM algorithms
implemented as computer programs) to actually find patterns in the data [9]. Thanks to data mining software,
data from many different dimensions is analyzed to categorize and summarize the relationships identified. Also,
it allows to find correlations and patterns among dozens of fields in large relational databases [10].
1357
Fig. 1. Tree diagram of incident solution time, cities and error categories
The analyzes of the effect of location and error categories on solution time is evaluated. The results of the
analysis indicate that data mining function results at maximum three levels, tree depth in that location and the
error categories were taken as independent for dependent solution time. The nodes 8 and 13s solution time is
highly more than other nodes, followed them in groups of nodes 19, 14, 9 and 11. It is obvious that there are
service level problems at every incident error category in locations of node 8, these are Hakkari and Sirnak.
So, that supports not to take on interest of incidents coming from this location. The Nodes 9, 11, 13, 17 and 19
describe that Furniture and Infrastructure type incidents solution time is more than Computer and
Peripheral type incidents. Also, incidents in every category coming from node 3s locations have nearly the
highest solution time. This study illustrates the Ministry of National Education for reevaluation by operation
1358
experts of this project and managers to help improving incident management process and setting next years
objectives.
References
ANAND, S., KEREN, N., TRETTER, M. J., WANG, Y., OCONNOR, T. M., and MANNAN, S. M. (2006),
Harnessing Data Mining to Explore Incident Databases, Journal of Hazardous Materials, 130, 3341
2. RUD O.,Data Mining Cookbook: Modelling Data for Marketing, Risk and Customer
Relationship
Management, John Wiley & Sons, Canada, U.S.A, 2001.
3. BERRY M, Linoff G., Data Mining Techniques For Marketing, Sales And Customer Support, John Wiley &
Sons , New Jersey, U.S.A, 1997.
4. RYGIELSKI C., WANG J., YEN D., Data Mining Techniques for Customer Relationship
Management,Technology in Society, 24:483-502, 2002.
5. HWANG H., JUNG T., SUH E., An LTV Model And Customer Segmentation Based On Customer Value: A
Case Study On The Wireless Telecommunication ndustry, Expert System with Application, 26: 181-188, 2004.
6. HUI, S.C., and JHA, G. (2000), Data Mining for Customer Service Support, Information & Management, 38, 113.
7. APTE, C., and WEISS, S. (1997), Data Mining with Decision Trees and Decision Rules, Future Generation
Computer Systems, 13, 197-210.
8. FAYYAD, U., PIATETSKY-SHAPIRO, G., and SMYTH, P., (1996), From Data
Mining
to
Knowledge
Discovery: an Overview, Advances in Knowledge Discovery and Data Mining, pp. 134. Cambridge, MA: MIT
Press.
9. DZEROSKI, S. (2008), Introduction Data Mining Tasks Patterns Data Mining Algorithms, Elsevier B.V.
10. LUO, Q. (2008), Advancing Knowledge Discovery and Data Mining, Workshop on Knowledge Discovery and
Data Mining.
1.
Biographies
Filiz ERSZBS. in Statistics, AnadoluUniversity, Eskiehir, Turkey, 1989. MS. in Biostatistics, Ankara University, 1992
and Ph.D. in Biostatistics, Ankara University, Ankara, Turkey, 1998. Ersoz had worked as a statistician in the Turkish
Statistical Institute for 12 year and Operational Research Expert in the Headquarters of Land Forces Command for 5 years.
She is presently working in Defense Sciences Institute of Turkish Military Academy for 5 years. Ersz have experience in
applied statistics (Multivariate Statistics Analysis, Parametrics and Non-Parametrics Methods, Categorical Data Analysis,
Survey Methodology). Dr.Filiz ERSZs memberships in professional societies; The Turkish Statistics Association,
International Statistics Institute (ISI).
Ahmet HAYRAN BS. in Computer Education and Instructional Technology, Middle EastTechnologyUniversity, Ankara,
Turkey, 2007. MS in Information Systems, Gazi University, 2010, Ankara, Turkey. Ahmet have experience in software
engineering and BT Service Management. Ahmet had worked as a software developer in some informatics company for 2
years. He is presently working as BT Service Management Department Manager in Signum TTE since 2 years.
Cevriye GENCER She received the B.Eng. degree in Industrial Engineering from the University of Gazi, Ankara, TR, in
1987, and the M.S., Ph.D. degree from the same university, in 1993. She is a Professor in Industrial Engineering at
GaziUniversity. Prof.Dr.Cevriye GENCERs research interests include location problem, decision support systems, integer
programming, project management, data mining, knowledge-based systems, computational intelligence, and rule-based
classification.
1359
1 Introduction
Accident prediction models are important tools for estimating road safety with regards to roadway, weather and
accidents conditions. There are different empirical equations developed for accident prediction models.
However, new regression techniques have recently found application opportunities in this area. It is obvious that
the model development and subsequently the model results are strongly affected by the choice of the regression
technique used. This study evaluates the influence of roadway, weather and accidents conditions, and type of
traffic control on accident severity (number of person killed) using regression models. Information on accident
severity and roadway and weather conditions was obtained from the Michigan Department of Transportation
Accident Database. Negative Binomial (NB) and Poisson regression models were deployed to measure the
association between accident severity and roadway, weather and accidents conditions.
1360
sitesallowing for additional dispersion (variance) of the crash data. Although the Poisson and NB regression
models possess desirable distributional properties to describe motor vehicle accidents, these models are not
without limitations. One problem that often arises with crash data is the problem of excess zeroes, which often
leads to dispersion above that described by even the negative binomial model. Excess does not mean too
many in the absolute sense, it is a relative comparison that merely suggests that the Poisson and/or negative
binomial distributions predict fewer zeroes than present in the data. As discussed in Lord et al. (2004), the
observance of a preponderance of zero crashes results from low exposure (i.e. train frequency and/or traffic
volumes), high heterogeneity in crashes, observation periods that are relatively small, and/or under-reporting of
crashes, and not necessarily a dual state process which underlies the zero-inflated model. Thus, the
motivation to fit zero-inflated probability models accounting for excess zeroes often arises from the need to find
better fitting models which from a statistical standpoint is justified; unfortunately, however, the zero-inflated
model comes also with excess theoretical baggage that lacks theoretical appeal (see Lord et al., 2004). Another
problem not often observed with crash data is underdispersionwhere the variance of the data is less than the
expected variance under an assumed probability model (e.g. the Poisson). One manifestation might be too few
zeroes, but this is not a formal description. Underdispersion is a phenomenon which has been less convenient to
model directly than over-dispersion mainly because it is less common observed. Winkelman's gamma probability
count model offers an approach for modeling underdispersed (or overdispersed) count data (Winkelmann and
Zimmermann, 1995), and therefore may offer an alternative to the zero-inflated family of models for modeling
overdispersed data as well as provide a tool for modeling underdispersion.
1361
Frequency
15066
26370
50442
46370
37888
56740
142108
374984
Valid Percent
4.0
7.0
13.5
12.4
10.1
15.1
37.9
100.0
2000
Frequency
18169
30937
55396
51068
40927
59905
169576
425978
Valid Percent
4.3
7.3
13.0
12.0
9.6
14.1
39.8
100.0
Frequency
39427
175398
158737
1422
Valid Percent
10.5
46.8
42.3
0.4
374984
100.0
2000
Frequency
35529
212698
176706
1909
426842
Valid Percent
8.3
49.8
41.4
0.4
100.0
Table 3 present the accident occurrences by more detailed area type. the most interesting number is related to
straight road sections. These sections constitute over 40% of all crashes. It is also interesting to note that
railroad crossing crashes are only 0.2 or 0.3% of all.
1362
2004
Frequency
Valid Percent
12255
3,3
34795
9,5
56553
15,4
19184
5,2
2000
Frequency
Valid Percent
13221
3,1
35960
8,6
73341
17,5
24551
5,8
37623
10,3
42968
10,2
161754
12909
14311
8016
952
7606
44,2
3,5
3,9
2,2
,3
2,1
175574
14269
17982
10589
697
10575
41,8
3,4
4,3
2,5
,2
2,5
160
366118
8866
374984
,0
100,0
317
420044
6798
426842
,1
100,0
Table 4 present the relationship of accident occurrences to road sections. It is very normal that 85% of all crashes
occur on the road. However, considerable number of crashes also occur outside of the should/curb (about 8%).
These crashes must be occurred in urban as well as rural areas at nighttime by intoxicated drivers.
Table 4. Relationship of accident occurrences to road sections
On the Road
In the Median
On the Shoulder
Outside of
Shoulder / Curb
In the Gore
Unknown
Total
Missing
Total
2004
Frequency
305722
2924
15786
30507
800
6022
361761
13223
374984
Valid Percent
84.5
0.8
4.4
8.4
0.2
1.7
100.0
2000
Frequency
364102
3212
17503
32096
884
7310
425107
1735
426842
Valid Percent
85.6
0.8
4.1
7.6
0.2
1.7
100.0
Table 5 present the relationship of accident occurrences to road types. The majority of crashes occurred on
county roads and city streets (about 60%). M-routes also take a considerable amount of crashes (19%). Interstate
and US routes have relatively lower crashes by 9 and 7%.
1363
Route Class
Not Located
Interstate Route
US Route
M Route
I-State Business
Loop or Spur
US Business Route
2004
Frequency
Valid Percent
4777
1,3
34820
9,3
27737
7,4
72152
19,2
5731
1,5
M Business Route
Connector
County Road or City
Street
Total
2000
Frequency
Valid Percent
39402
9,2
31638
7,4
27725
6,5
78107
18,3
6842
1,6
4098
1,1
4746
1,1
105
,0
120
,0
721
224843
,2
60,0
742
237520
,2
55,6
374984
100,0
426842
100,0
Table 6 shows the relationship of accident occurrences to speed limits. It should be noted that this table does not
present the risk of speed limits since the road with speed limits higher than 55 mph show lower accident
occurrences. City streets with speed limits equal or less than 25 mph take about 20% of all accidents. M, US and
Interstate routes with speed limits higher than 50 mph have the highest percent (over 30%).
Table 6. Relationship of accident occurrences to speed limits
Speed Categories
Equal or less than 25
mph
26-30 mph
31-35 mph
36-40 mph
41-45- mph
46-50 mph
51-55 mph
56-60 mph
61-65 mph
66-70 mph
Greater than 70 mph
Missing
Total
2004
Frequency
Valid Percent
74688
19,9
19608
47028
26004
53995
12909
107214
971
6807
25612
148
0
374984
5,2
12,5
6,9
14,4
3,4
28,6
0,3
1,8
6,8
0,0
100,0
2000
Frequency
Valid Percent
79867
20,7
25070
58910
30115
58121
13006
113883
306
7421
17
--40126
426842
6,5
15,2
7,8
15,0
3,4
29,4
0,1
1,9
0,0
--100,0
1364
properties of the negative binomial are derived and GLM algorithms are developed for both the canonical and
log form. The log forms of both may be effectively used to model types of Poisson-overdispersed count data
(Hilbe, 1993). It is not recommended that negative binomial models be applied to small samples. What
constitutes a small sample does not seem to be clearly defined in the literature though (UCLA, 2011). Poisson
regression, also a member of the class of models known as generalized linear models (GLM), is the standard
method used to analyze count data. However, many real data situations violate the assumptions upon which the
Poisson model is based. For instance, the Poisson model assumes that the mean and variance of the response are
identical. This means that events occur within a period of observation at a constant rate; an event is equally
likely at any point within the period. When there is heterogeneity in the data, it is likely that the Poisson model is
overdispersed. Such overdispersion is indicated if the variance of the response is greater than its mean. One may
also check for model overdispersion by submitting the data to a Poisson model and observing the Chi2-based or
Deviance-based dispersion statistic. The model is Poisson-overdispersed if the dispersion value is greater than
unity. Log negative binomial regression can rather effectively be used to model count data in which the response
variance is greater than that of the mean (Hilbe, 1993).
Model Results. In order to apply Poisson regression model, we first need to check is there is heterogeneity in
the data. Table 7 presents the descriptive statistics and it seen that the variance of the response variable (number
of persons killed in crash) is little bit higher than the mean. Thus, the results of the Poisson regression must be
approached cautiously.
Table 7. Descriptive statistics of the response variable
Response variable:
Number of Persons
Killed in Crash
2004
2000
Minimu
m
Maxi
mum
Mean
Std.
Deviation
Variance
0.003
0.062
0.004
0.003
0.064
0.004
37498
4
42684
2
Table 8 presents the case processing summary. As seen about 90% of the cases included in regression
modeling. Regarding the measure of goodness of fit, the value of deviance/df is 0.012 and lower than 1. This
means that the model is not Poisson-overdispersed.
Table 8. Case processing summary
models
2004
Cases
for regression
2000
Included
n
349268
Percent
93.1%
N
382367
Percent
89.6%
Excluded
25716
6.9%
44475
10.4%
374984
100.0%
426842
100.0%
Total
Table 9 presents the parameter estimates using NB regression model. Significance of the parameters for 2004
and 2000 are similar except for the variables of day and surface. The variable of day is significant at 0.05 level
with year 2000 data and the variable of surface is significant at 0.10 level with year 2004 data.
1365
Parameter
(Intercept)
Month
Day
Hour
Weekday
NumVeh
CrshType
NumInj
NumUnInj
NumOcc
Weather
Light
Surface
TControl
Std.
B
Error
-4.518 .2127
.000 .0098
.003 .0038
-.008 .0051
-.011 .0169
.818 .0362
-.112 .0065
-3.687 .0591
-4.312 .0494
3.673 .0412
-.092 .0273
.041 .0197
.057 .0310
-.018 .0317
2000
Hypothesis Test
Sig.@
Wald
0.10
Chi2
Sig. level?
451.365
.000
yes
.001
.976
no
.460
.498
no
2.614
.106
no
.426
.514
no
510.613
.000
yes
294.156
.000
yes
3890.27
.000
yes
7626.77
.000
yes
7935.97
.000
yes
11.320
.001
yes
4.347
.037
yes
3.359
.067
yes
.334
.563
no
Std.
B
Error
-5.552 .2317
.000 .0100
.012 .0040
.004 .0051
.003 .0173
.734 .0417
-.079 .0063
-3.385 .0579
-4.304 .0531
3.425 .0412
-.067 .0313
.066 .0204
.022 .0341
.054 .0341
Hypothesis Test
Sig.@
Wald
0.10
Chi2
Sig. level?
574.247
.000
yes
.000
.996
no
8.848
.003
yes
.765
.382
no
.025
.875
no
309.797
.000
yes
154.388
.000
yes
3422.32
.000
yes
6565.80
.000
yes
6912.91
.000
yes
4.574
.032
yes
10.458
.001
yes
.433
.510
no
2.558
.110
no
Table 10 presents the parameter estimates using Poisson regression model. Significance of the parameters for
2004 and 2000 are similar except for the variables of month, day, weather and surface. These variables are
significant at 0.05 level with year 2004 data.
Table 10. Parameter estimates using Poisson regression
2004
2000
Hypothesis Test
Parameter
(Intercept)
Month
Day
Hour
Weekday
NumVeh
CrshType
NumInj
NumUnInj
NumOcc
Weather
Light
Surface
TControl
B
-4.410
-.044
.025
-.034
.005
.867
-.102
-2.731
-3.585
2.898
-.083
.036
.074
-.045
Std.
Error Wald Chi2
.1979
496.542
.0093
22.089
.0036
50.400
.0050
46.191
.0157
.099
.0284
934.627
.0059
301.378
.0388 4949.289
.0400 8038.972
.0280 10728.93
.0259
10.140
.0188
3.633
.0294
6.244
.0300
2.255
Sig.
.000
.000
.000
.000
.753
.000
.000
.000
.000
.000
.001
.057
.012
.133
Hypothesis Test
Sig.@
0.010
level
yes
yes
yes
yes
no
yes
yes
yes
yes
yes
yes
yes
yes
no
B
-4.691
.014
-.001
-.019
-.016
.841
-.079
-2.483
-3.465
2.513
-.037
.098
-.058
-.022
Std.
Error Wald Chi2
.2231
442.253
.0098
2.064
.0038
.030
.0048
15.263
.0168
.933
.0310
737.370
.0059
178.777
.0482 2652.439
.0456 5782.059
.0243 10733.83
.0303
1.464
.0203
23.039
.0361
2.556
.0303
.515
1366
Sig.
.000
.151
.863
.000
.334
.000
.000
.000
.000
.000
.226
.000
.110
.473
Sig.@
0.010
level
yes
no
no
yes
no
yes
yes
yes
yes
yes
no
yes
no
no
Acknowledgement
The author acknowledges Dr. Dale Lighthizer of the Michigan Department of Transportation and Prof. Dr.
Richard Lyles of Michigan State University for providing the crash data for the sole purpose of doing academic
research and writing papers.
References
1.
El-Basyouny, K. and Sayed T. (2006). Comparison of two Negative Binomial Regression Techniques in
Developing Accident Prediction Models. Transportation Research Record: Journal of the Transportation Research
Board, volume 1950, pp 9-16.
2. Joshua, S.C. and Garber, N.J. (1990). Estimating truck accident rate and involvements using linear and Poisson
regression models. Transport. Planning Technol. 15 (1990) (1), pp. 4158.
3. Hilbe, J.M. (1993). Log-Negative Binomial Regression as a Generalized Linear Model. Technical Report COS
93/94-5-26. Department of Sociology and Graduate College Committee on Statistics.
4. Lord, D. (2000). The prediction of accidents on digital networks: characteristics and issues related to the
application of accident prediction models. Ph.D. Dissertation, Department of Civil Engineering, University of
Toronto, Toronto.
5. Lord, D.; Washington, S.; and Ivan, J. (2004). Poisson, Poisson-gamma, and zero-inflated regression models of
motor vehicle crashes: balancing statistical fit and theory. Accident Analysis and Prevention, Pergamon
Press/Elsevier Science, 2004.
6. Lord, D. (2006). Modeling motor vehicle crashes using Poisson-gamma models: examining the effects of low
sample mean values and small sample size on the estimation of the fixed dispersion parameter. Accident Analysis
and Prevention, 38 (2006) (4), pp. 751766.
7. Miaou, S.P. and Lum, H. (1993). Modeling vehicle accidents and highway geometric design relationships. Accident
Analysis and Prevention, 25 (1993), pp. 689709.
8. Oh, J.; Washington, S.P.; and Nam, D (2006). Accident prediction model for railway-highway interfaces. Accident
Analysis and Prevention, Volume 38, Issue 2, pp. 346-356.
9. UCLA: Academic Technology Services, Statistical Consulting Group. SPSS Data Analysis Examples: Negative
Binomial Regression. http://www.ats.ucla.edu/stat/Spss/dae/spss_neg_binom_DAE.htm (accessed in April 20011)
10. Washington, S.; Karlaftis, M.; and Mannering, F (2003). Statistical and Econometric Methods for Transportation
Data Analysis, Chapman Hall/CRC, Boca Raton, FL.
11. Winkelmann, R. and Zimmermann, K. (1995). Recent developments in count data modeling: theory and
applications. J. Econ. Surveys 9, pp. 124.
1367
Biography
Dr. Darn Akn Born in Germany in 1966. He received his BS degree in civil engineering from Dokuz Eylul University,
Izmir, Turkey in 1987 as valedictorian. Respectively, he received his MS and Ph.D degrees in transportation engineering
from Dokuz Eylul and Michigan State Universities (E. Lansing, MI, USA) in 1992 and 2000.
During his Ph.D. study, he was involved in traffic engineering research projects including work zone speed study, study of
pedestrian behavior at signalized and unsignalized crossings, and simulation of campus network. He was also worked as parttime at TriCounty Regional Planning Commission and Michigan Department of Transportation (MDOT), where he worked
on bicycle network planning and statewide long-range plan. Lastly, before he graduated, he worked as a full-time transport
planner at MDOT. He maintained highway networks for citys long range plans. He completed his Ph.D. in May 2000 and
returned home to accept a faculty position at Gebze Institute of Technology (GIT) in the department of City and Regional
Planning. At GIT he has been teaching urban transport policies, urban transport systems, urban planning and travel
modeling. Outside of GIT, he taught at several well-known institutes including Bogazici and Istanbul Technical Universities
at graduate as well as undergraduate programs. Dr. Akn received his tenure and the title of associate professor in planning in
Feb 2010. Dr. Akn has many papers and publications at several national and international conferences and symposiums. He
has been a reviewer at several periodicals and international conferences including TRB of USA.
Dr. Akin is the member of Turkish Road Association since 2005, Chambers of Civil Engineers since 1987, Institute of
Transportation Engineers (ITE) of USA between 1996 and 2000, American Society of Civil Engineers (ASCE) between 1996
and 2000, World Conference on Transport Research Society (WCTRS) from 2004-6 to 2010-13, and Member of the Board of
Trustee of EMIT Research Platform since 2011.
1368
Chapter 9
CITY PLANNING
AND
CIVIL ENGINEERING
1369
muharrema@sakarya.edu.tr , 2elifd@sakarya.edu.tr
1,2
Sakarya University/Civil Engineering
Abstract. Investigation of the behavior of reinforced beams, both experimental studies are conducted and
confirmation of the solutions for the nonlinear behaviors in conjunction with the experimental results is
attempted by designing appropriate computer models. However, nonlinear behavior regarding the composite
beams with concrete slabs is not convenient for computer modeling. However, the same behavior could be
simulated by computer modeling via the replacement of the concrete slab on the composite beam by a steel
plate to act as the pressure cap. Thus, the composite beam with a concrete slab could be simulated by a steel
plated beam. In this study, the behavior of IPE160 steel profile was investigated by applying concrete slab
and steel plate on it.
1 Introduction
In this day and age, the need for strengthening steel beams may arise in order to obviate cross-section losses
caused by corrosion, to remedy cross-section deficiencies due to traffic loads or to cope with design errors.
During the process of strengthening and repairing, the primary concern should be safety and economical
objectives should be secondary. Among the strengthening techniques, replacement of the cross-section with a
new one, welding or bolting a new component to the existing structure or the implementation of materials such
as FRP could be counted. Comparing these strengthening techniques, the use of FRP, which possesses various
strength levels and is easier to implement, could be recommended instead of both welding and bolting, in which
the existing structure is damaged by heat treatment and the opening of bolt holes, respectively (1). Welding may
also lead to unidentified collapses as it erodes the metal structures of the main components (2). Hence, FRP
strengthening received widespread usage in the past few years, especially in highway bridges,
telecommunications industry and even marine vessels for repairing and strengthening purposes (3)(6). The
lightness of this material is also preferred as it introduces a negligible additional load on the structure. In this
context, computer models are widely used in determining the load capacities of bridge beams and addressing
their strengthening needs.
Simulation of the concrete parts of composite beams with steel plates is highly critical with respect to time.
Such a simulation can be conducted by steel plates at various sizes, which may then be utilized as testing beams
for future experiments (4). This criterion is especially taken into consideration in the choice of material for large
scale experiments (5).
In practice, composite beams are generally formed by the collective utilization of concrete paving slabs or the
steel beams in the bridge deck slabs. The compression stresses induced by the bending moment in the composite
1370
beam cross-sections are born by the concrete slab, while the tensile stresses are born by the steel beams tension
chord. Consequently, the lower flange of cross-section is influential in the conveyance of tensile stresses. The
usage of composite materials for bridge beams and concrete for upper flanges of the cross-sections largely
affects the designed moment carrying capacity of the cross-section. Moment carrying capacities are critical with
respect to determination of the load bearing limits. Simulation of the concrete slabs by steel plates is feasible in
case they exhibit the same moment carrying capacity.
2 Material Properties
The composite beam is 3000 mm tall and its concrete section is 50 mm high and 4500 mm wide. The
composite beam is formed by cross installment of the square reinforcing mesh at a size of 150x150 mm and the
installment of 12x30 mm key coupling element such that it would lie 50 mm in horizontal direction and 100 mm
in vertical direction. In order to facilitate the adsorption of the concrete section to the steel, the mould was
prepared from steel sheet.
During the concrete casting, 3 specimens at sizes of 15x15 cm were taken. Completion of curing was achieved
as the beams were immediately wrapped in clothing to ensure that hydration of the concrete fully takes its
course. Additionally, the 3 specimens were placed in a curing pool after 24 hours. At the end of 28 days, these
specimens were subjected to compression tests by the 200 ton press machine located in the Civil Engineering
Construction Materials Laboratory at Sakarya University Fig(1).
a) Sample 1
b) Sample 2
c) Sample 3
3 Experimental Setup
Two experiments at actual sizes were conducted (Fig 2). Four point tests were conducted on the composite beam
with the concrete upper flange and the I beams with steel upper flanges by the frame with a capacity of 25-30
tons (Fig 3 and Fig 4). All the tests were conducted in the Civil Engineering Construction Materials Laboratory
at Sakarya University. The frame in which the specimens were placed is HI-TECH brand and has a capacity of
1371
15 tons, which was increased to 25-30 tons by the implementation of vertical and cross reinforcements. In order
to perform data readings during the tests, a 4-channnel Di4b and an 8-channel Ai8b data collection system were
used. The dislocations were measured by an OpkonLinfar Resistive Position Transducer brand LPS 100-B-5K
potentiometric ruler. The horizontal dislocations were measured by a Mitutoyo Absolute Digimatic Indicator
brand ID-C 1050B device, and a Kyowa brand KFG-10-120-C1-11 model tension shim was used to measure
tension forces. Finally, an Enerpac piston was used as the load generator. During the tests, all changes occurring
in the specimens were recorded by affixing the above mentioned devices onto the specially produced
apparatuses.
4 Methodology
4.1 Flexural experiment of composite steel beams
1372
4.2. Creating steel beams by replacing concrete and conducting a flexural test;
1373
Fig. 6. Experimental results of the composite section beam and the steel plate beam from the full-scale test.
5. Conclusions
A comparison of the test results for the I beam reinforced with steel plate and the I beam reinforced with
concrete slab against the FE model are given in Fig (6).
References
1.
2.
3.
4.
5.
6.
D.Schnerch, K.Stanford, E.Summer, S.Rizkalla (2005). Bond behavior of CFRP strengthened steel bridges and
structure. Proceedings of International Symposium on Bond Behavior of FRP in Structures.
Pric, A. And R.J. MouldsKauvar(1991).Repair and strengthening of structures using plate bonding. Construction
and Building Materials(U.K),v.5, n. 4,December 1991,pp 189-192
D Schnerch, S.Rizkalla(2005). Strengthening of scaled steel-concrete composite girders and steel monopole towers
with
CFRP.
FRP
Composites
in
Civil
Engineering-CICE
2004-SEracino(ed)
2005
Taylor,FrancisGroup,London,ISBN 90 5809 638
Sami Rizkalla; Mina Dawood(2006). High modulus carbon fiber materials for retrofit of steel structures and
bridges. Developments in composites; Advanced infrastructural, natural and Nano composites. UNSW, Sydney,
Australia.11-14-July 2006
Sami Rizkalla, Mina Dawood, David SChnerch(2008). Development of a carbon fiber reinforced polymer system
for strengthening steel. Structures Composites: Part A 39(2008)388-397
Glen Smith, Tarek Hassan And Sami Rizkalla. Bond Characteristics and Quantifications of adhesive form marine
Applications and steel pipe repair. North Carolina State University, USA
1374
Abstract. City planning is experiencing fundamental changes on the use of computer -based models in practice and
education. However, with this increased use, Geographic Information Systems (GIS) alone cannot serve all of the
needs of city planning. By utilizing integrated modeling approaches, spatial analysis and computational techniques
could be used to deal better with the tasks/problems in city planning process. In this paper, Case Based Reasoning
(CBR) an artificial intelligence (AI) technique- and GIS geographic analysis, data management and visualization
technique- are used to build a Case Based Model (CBM) for knowledge representation on an operational study.
This study tried to find out whether an integrated model which employs CBR & GIS could support hu man decision
making in a city planning task. While the CBM met many predefined goals of the study, both advantages and
limitations have been realized from findings when applied to the complex domain such as city planning.
1.
Introduction
City planning process is experiencing fundamental changes by emerging impacts of using computer-based
models in planning practice and education. However, besides this increased use, it has been realized that GIS
alone cannot serve all of the needs of planning. Planners have to strengthen existing GIS tools to meet their
demands.
In this paper, CBR technique is used to build a Case Based Model (CBM: CBR & GIS) approach. By the
development of a system that integrates CBR as an AI reasoner and GIS as a spatial analyst, could be very
helpful to planners for reaching a knowledge acquisition on a special purpose in planning process. In such an
integrated system, CBR will provide a retrieval method by using previous experiences in proposing a solution to
a new problem or providing relevant experiences to the planners.
The goals of this study are (1) to demonstrate the feasibility and usefulness of CBR technique, (2) to
combine knowledge inference capability of CBR with analytical, management and visualization capability of
GIS within an operational pilot study, (3) to examine whether these techniques could benefit the city planning
process.
1375
CBR
Planning
Support
&
GIS
Background
1376
Similarity Assessment
CBR uses matching and ranking to derive similarity. Matching is achieved through index and weights, while
ranking is the total of the match score. CBR also searches and matches the entire database not just by comparing
two values. Most CBR systems and also this study use the nearest neighbor 1 matching technique for retrieval.
Nearest neighbor algorithm (k-NN) is represented as follows:
(1)
where:
S: Similarity Score
Wi: Feature Weights
n: Feature Number
finput: feature value of input case
fstored : feature value of stored case
Similarity Function for 3 features:
(
(2)
(3)
or
An important issue in similarity assessment is how to determine the right features to compare. Similarity
function determines the measure of similarity between the new problem and each stored case. Most similarity
measures use a numeric value to indicate the level of similarity. This numeric value is the result of matching and
ranking techniques to provide a similarity/match score.
1The k-nearest neighbor algorithm (k-NN) is a method for classifying objects based on closest training examples in the
feature space. An object is classified by a majority vote of its neighbors, with the object being assigned to the class most
common amongst its k-nearest neighbors.
1377
3.
Methodology
Previous city planning experiences could be used to make new plan decisions/recommendations.
Consequently, a digital archive and a working reasoning model are needed to verify this hypothesis on a real
case.
In this paper, it is tried to provide a spatial decision support model. When developing minor/incremental
planning implementations (assumed as a new case), we can find most similar recent cases and use their
knowledge in decision making. There are no universal CBR methods suitable for every domain of application. In
this respect, the following subsections provide model implementation process in eme Peninsula.
3.1. Selection of the Study Area
eme is located on the western coasts of Turkey within the borders of the Aegean Region, surrounded by
the Aegean Sea and affiliated to Izmir province.
eme Peninsula is a special place for both having natural-cultural heritage and urban development
potentials. eme is under pressure of continuous tourism investment and urban development demands. There
are variety of modifications on archeological and natural sites and variety of plan revisions since 1980s in the
planning history of the region.
There has been a critical balance between conservation decisions and planning applications about the
eme region which cause uncertainty in preservation versus development aspects. Due to the existing
administrative, structural and planning factors, it is decided that the geographic/spatial extent of the project area
should cover eme Municipality boundary.
3.2. Building the CBM in eme
The major aim of developing this pilot project is to estimate and interpret the possible incremental/minor
applications of eme Peninsula by utilizing former city planning experiences. In this respect, an experimental
and partly hypothetical planning process is simulated. Firstly cases are generated semi-automatically within a
GIS environment by cellular representation. The proposed model retrieves the cases through tabular data and
link the inference to spatial and exterior data. For instance, after the similar case(s) found, we can reach their
GIS maps, reports, photos, etc. Hence, after selecting the similar cases from the case library, to enable an
extended connection to further data (about the similar cases) is considered to provide knowledge acquisition.
1378
ArcGIS software is used for preparing and handling geospatial data and jCOLIBRI2 software is used to
utilize the GIS database as a case library, indexing of the cases and retrieval process. Firstly, spatial data and
non-spatial attributes are generated by using ArcGIS as a tabular input for the CBR software jCOLIBRI.
Fig. 4. Prototype of the CBM architecture(Adapted from Lanter, 1992; Shi and Yeh, 1999)
3.2.1. Creating and Using Geo-Data
The most common approach for structuring the geography of the real world in the GIS environment is to use
a layered structure. Therefore, to build up a spatial representation of a study area, multiple layers are used.
Data Layers: Three data frames are used to build main case library: land use and roads, geographic analyses
and conservation zones. These analyses were generated during the preparation period of the structure plan of
eme by Izmir Institute of Technology (IYTE), Department of City and Regional Planning in 2002.
Classifications of these layers are as follows:
Coordinate System: Projected Coordinate System-metric
Projection: Universal Transverse Mercator (WGS1984, UTM Zone 35N)
LAND USE: Existing land use pattern and major roads are used.
GEOGRAPHIC ANALYSES: Ownership, Geological Formation, Slope and Seashore Buffer Zone layers are
used.
CONSERVATION ZONES: The latest (2006) archeological and natural site zones and classifications are used.
Abbreviations are used to symbolize conservation types in legend illustrations and database:
A: Archeological Conservation
D: Natural Conservation
K: Urban Conservation
Numbers are used for hierarchical order. For instance D1 is superior to D2 about conservation criteria.
3.2.2.
Although the domain of city planning contains many cases, it is difficult to acquire complete (due to the lack
of analyses, incomplete data, out of date information, etc.), digital and well documented real cases. So it is
decided that an experimental case generation process could be organized to overcome this limitation and semiautomatic case generation process is devoted to develop a unique case and their unity. To represent the spatial
features by cellular representation is not a new idea in urban studies. The way may be conceptual and has
operational difficulties, but much of the labor can be obviated by computerized methods. This method includes
the input and output of material on a spatial basis via a matrix of a map based coordinate grid.
2jCOLIBRI: Java Cases and Ontology Libraries Integration for Building Reasoning Infrastructures is an object-oriented
framework which has been developed in Java programming language.
1379
600 m
GRID Layer: It is assumed that, 800m X 600m (48 hectares each) dimensioned rectangular cell which was
derived from 1/1000 scale map units (Turkish planning practice) could build the spatial extent of the unique
case in the CBM.
800 m
There are 284 cells (rectangular polygons) joined together to generate the grid layer that overlays the eme
Municipality boundary. Grid layer is created by the vector grid function of an ArcGIS extension; ET
Geowizards LT (ET Spatial Techniques, 2008). In the next step, only required cells are selected and extracted by
select by location (which completely within the eme municipality boundary) function of ArcGIS.
1380
Fig.7. Intersection of Grid + Land-Use (deployed areas within each case unit)
After the intersection process completed, non-spatial attribute database is produced by joining the necessary
columns of all intersected layers. Another simple geoprocessing model (to join data fields) is utilized to perform
this joining operation to create final database.
The descriptive features (columns) are critical parts of the case and they represent the key feature of a case,
such as Land_Use, Road_Type, Conservation Type, Ownership, Geological Formation, Slope
Analysis and Seashore Buffer. The CaseID is a number that uniquely identifies the case. Feature values are
stored in a table file cell.dbf (dbase data file format) in ArcGIS software but this table file type is independent
from both ArcGIS and jCOLIBRI.
3.2.3. Developing the Customized Graphic User Interface
Developments in software engineering such as object oriented programming languages have enabled the
increased reuse and improved modularity of code written for specific purposes. Customizations and
modifications may include the development of GUI and specialized tools relating to a particular application (for
instance GIS).
jCOLIBRI 2 platform offer a framework for implementing customizations through java programming
language. Thus an effective customized GUI which is based on jCOLIBRI 2 core architecture is developed in
java development environment Eclipse 3. This GUI is also developed for seamless integration and user
interaction.
GIS
(Input)
DBF
Query
(Output1)
Retrieved
Cases
CBR
(Output2)
External
Hyperlink
(jCOLIBRI2 Core)
Similarity
Graph
1381
One of the distinguishing capabilities of the modified GUI is about the dbase file connector. This GUI
utilizes the dbf file (which is generated by GIS software) as a case library. At the beginning, this tool obtains the
user preferences using a form (uses form-filling) in query (input), and then it computes nearest neighbor scoring
(whose similarities are above a predetermined threshold or elicitation of desired number of top similarity scored
ones) for retrieval in CBR core and presents most similar cases on the first template (output1) and presents extra
media about these cases on another template (output2). Additionally, a similarity evaluation graphic is presented
after retrieval.
If the customized GUI is used, user can easily access to the extra content and no manual effort or links are
needed. Consequently, this is the main advantage of the customized GUI when compared to default execution
and result (simplified textual output) of the jCOLIBRI 1 software. Briefly customized GUI differentiates from
default jCOLIBRI platforms by using dbase file format, domain specific similarity matrices and extended library
connection.
4. Applying the Model
4.1. Retrieval Process
When testing the CBM, assume that we have a semi-structured scenario and we need to develop a new case
-an incremental implementation- in one of the cell within or adjacency of the study area eme Peninsula and we
demand to find recent similar applications to use their knowledge in decision support.
1382
Another noticeable feature of the customized GUI is extended case library connection. User can reach to
extra information about the retrieved cases (user should select one case from the list) by triggering Extra
Content button on the first template. An extended part of the case library which contains project files of cellular
cases is shown in another template.
User can reach to reports, aerial/satellite images, GIS analysis, maps, plans by simply clicking the extra
content buttons. The directory which contains the files of selected case (e.g. case133) is accessed by hitting the
Open Folder button.
1383
1384
can be used in order to transfer expertise and render advice or recommendations, much like a human expert.
Urban planning often requires the experiences and expertise of planners and assistance of planning models and
analytical methods in its complicated decision-making process. By integrating CBR technique with powerful
capabilities of GIS can facilitate this process.
References
1.
2.
3.
4.
5.
6.
7.
8.
9.
10.
11.
12.
13.
14.
15.
16.
17.
18.
19.
20.
21.
22.
23.
24.
25.
26.
Aamodt, A., & Plaza, E. (1994). Case-Based Reasoning: Foundational Issues, Methodological Variations, and System
Approaches. AI Communications. IOS Press, 7 (1), 39-59.
Abidi,&Manickam. (2002). Leveraging XML-based electronic medical records to extract experiential clinical
knowledge: An automated approach to generate cases for medical CBR systems. International Journal of Medical
Informatics, 68 (1-3), 187-203.
Batty, M. (2003). Planning Support Systems: Technologies that are Driving Planning. In Geertman, S., & Stillwell, J. C.
H. (Eds.), Planning support systems in practice (pp. v-viii). Berlin ; New York: Springer.
Benedikt, &Reinberg, &Reidl. (2002). A GIS Application to Enhance Cell Based Information Modeling. Information
Sciences, 142 (1-4), 151-160.
Benwell, G.L., & Holt, A. (1999). Applying case-based reasoning techniques in GIS. International Journal of
Geographical Information Science, 13 (1), 9-25.
Bhogaraju, P. (1996). A Case-Based Reasoner for Evaluating Crop Rotations in Whole-Farm Planning. MSci Thesis,
Virginia Polytechnic Institute and State University, USA.
Brail, R. K., &Klosterman, R. E. (2001). Planning support systems: integrating geographic information systems,
models, and visualization tools. Redlands, Calif.: ESRI Press.
Case Based Reasoning Homepage: AI-CBR. (2007). Retrieved June 2009,
from http://www.ai-cbr.org/
Chadwick, G. F. (1971). A systems view of planning; towards a theory of the urban and regional planning process (1st
ed.). Oxford, New York,:Pergamon Press.
Chan, & Chen, &Geng. (2000). Knowledge Engineering for an Intelligent Case-based System for Help Desk Operations.
Expert Systems with Applications, 18 (2), 125-132.
Chang, & Cheng, & Su. (2004). Using Case-Based Reasoning to Establish a Continuing Care Information System of
Discharge Planning. Expert Systems with Applications, 26 (4), 601-613.
Cognitive Computing Lab. (2009). Retrieved June 2009, from Georgia Tech http://home.cc.gatech.edu/ccl
Dikmen, I., &Birgnl M.T., &Gr A.K. (2007). A Case-Based Decision Support Tool for Bid Mark-Up Estimation of
International Construction Projects. Automation in Construction, 17 (1), 30-44.
Domeshek, E., &Kolodner, J. (1993). Using the Points of Large Cases. AI EDAM, 7 (2), 87-96.
EgePlanlama. (2006). 1/100000 lekliManisa-Ktahya-zmir PlanlamaBlgesievreDzeniPlan. Ankara.
ET SpatialTechniques - ArcGIS and ArcView extensions. (2008). Retrieved June 2009, from http://www.ian-ko.com
Geertman, S., & Stillwell, J. (2004). PSS: An Inventory of Current Practice. Computers, Environment and Urban
Systems, 28 (4), 291-310.
Hammond, K. (1988). Case-based Planning. Proceedings of Workshop on Case-Based Reasoning, 17-20.
Heylighen, A., &Neuckermans, H. (2001). A Case Base of Case Based Design Tools for Architecture. Computer-Aided
Design, 33 (14), 1111-1122.
Holt, A., &Benwell, G. (1996). Case Based Reasoning and Spatial Analysis. Urisa Journal, 8 (1-2), 27-36.
IYTE, Department of City and Regional Planning. (2002). 1/25000 lekliemelesievreDzeniPlanRaporu. Izmir:
IYTE.
jCOLIBRI. (2008). Retrieved June 2009 from GAIA - Group for Artificial Intelligence Applications,
http://gaia.fdi.ucm.es/grupo/projects/jcolibri/
Kaster, D.S., & Medeiros, C.B., & Rocha, H.V. (2005). Supporting modeling and problem solving from precedent
experiences: the role of workflows and case-based reasoning. Environmental Modeling & Software, 20 (6), 689-704.
Kolodner, J. L. (1993). Case-based reasoning. San Mateo, CA: Morgan Kaufmann Publishers.
Lanter, D. (1992). Intelligent Assistants for Filling Critical Gaps in GIS. Research Report. National Center for
Geographic Information and Analysis, Department of Geography, University of California at Santa Barbara.
Leake, D. B. (1996). Case-based reasoning: experiences, lessons & future directions. Menlo Park, Calif. Cambridge,
Mass, AAAI Press.
1385
1386
Chapter 10
INVITED
SPEAKERS
1387
1. INTRODUCTION
Images and image sequences are among the most important information carriers in todays society, and have
applications in a wide variety of fields such as industry, commerce, entertainment and medicine. The power of
images is that they can provide a lot of information in the blink of an eye. Due to bad acquisition, transmission
or recording, the images are however often corrupted by noise. A preprocessing module to denoise the images
then becomes necessary. For example, satellite images have to be denoised before ground structures can be
detected, and surveillance images have to be denoised before face recognition algorithms can be applied. Hence,
noise reduction is a very active research area.
The detection of noise is not a crisp problem, but a challenge that has to deal with uncertainties and
imprecision. It is not always straightforward to detect whether a pixel is contaminated with noise or not, and
if it is contaminated then one has to detect to what degree in order to apply a good correction. Since fuzzy
set theory and fuzzy logic are mathematical tools to deal with uncertainty and imprecision, they can also have
potential in the field of image processing. Our Laboratory works on the topic of image and video noise filtering
for nearly a decade, and it is our experience that so-called fuzzy filters outperform their classical counterparts,
both in terms of numerical and visual evaluation. This is mainly due to the fact that they allow the processing
of linguistic information (together with numerical information) and are capable of approximate reasoning (by
using fuzzy rules, see Section 2). We refer to Fig. 1 for a graphical representation.
In this contribution we discuss different aspects of two specific fuzzy filters. First we consider the earlier
mentioned FIDRM filter. We will not go into technical detail, but instead we will discuss the results of a large
comparative study that was undertaken to confirm the good results of this filter. In general, such comparative
studies are very important to position a newly developed filter in the variety of existing filters, but quite often
one notices that comparative studies are rather limited. Next, we consider the FRINV-G filter, for which we
will briefly discuss the principles of its construction and show some visual results. The presence of a temporal
component makes the noise removal process even more challenging, which explains why a multi-step approach is
used here. Both applications should provide the reader with some insight in the usefullness of fuzzy set theory
and fuzzy logic in image processing.
We will first start with a short note on the basics of fuzzy set theory and fuzzy logic.
1388
Figure 1. Fuzzy filters not only use numerical information to filter out the noise in images, but can also work with
linguistic information. Furthermore, fuzzy logic allows us to reason with this linguistic information and enables us to
better approximate human reasoning.
small
Figure 2. A possible membership function of the fuzzy set small (the parameter K can be chosen by the user, depending
on the application). The closer an element is to 0, the higher its membership value is.
The extension from crisp to fuzzy sets comes along with an extension of the underlying binary logical framework to fuzzy logic. In fuzzy logic, expressions can be true or false to a certain degree, and consequently we
should be able to connect such expressions (with the logical NOT, AND, OR, . . .) using fuzzy logical operators
that extend their binary counterparts. This can be achieved by using fuzzy logical operators, such as negators
(NOT), conjunctors (AND) and disjunctors (OR).
Formally,2 a negator N is a decreasing [0, 1] [0, 1] mapping that satisfies N (0) = 1 and N (1) = 0, a
conjunctor C is an increasing [0, 1] [0, 1] [0, 1] mapping that satisfies C(0, 0) = C(1, 0) = C(0, 1) = 0 and
C(1, 1) = 1, and a disjunctor D is an increasing [0, 1] [0, 1] [0, 1] mapping that satisfies D(1, 1) = D(1, 0) =
D(0, 1) = 1 and D(0, 0) = 0. The boundary conditions ensure that these fuzzy operators are real extensions of
the binary NOT, AND and OR. Popular examples are Ns (a) = 1 a, CM (a, b) = min(a, b) and CP (a, b) = a b,
DM (a, b) = max(a, b) and DP (a, b) = a + b a b, respectively, with a, b [0, 1].
Having fuzzy sets to model linguistic values and fuzzy logic to reason with them, fuzzy rules can be used to
model human reasoning and to derive new (imprecise) knowledge from given (imprecise) knowledge. An example
of a fuzzy rule is an expression of the form
IF ((p is P AND q is Q) OR (r is NOT R)), THEN (s is S),
with P, Q, R, S fuzzy sets (modeling linguistic values) and p, q, r, s elements from the corresponding universes.
The degree S(s) to which s is S (e.g., to which a pixel is considered noisy) is given by the degree to which the
1389
antecedent of the rule (i.e., the IF-part) is true. This degree is given by
S(s) = D(C(P (p), Q(q)), N (R(r))),
using a disjunctor D, a conjunctor C and a negator N .
With the above tools we are able to create a mathematical model for human reasoning with imprecise
knowledge.
1390
Figure 3. Noise removal from the Mandrill image: top = part of the original image, middle = image contaminated with
50% impulse noise (salt & pepper noise), bottom = denoised result with the FIDRM filter.
1391
Noisy
MF
WF
AWF
WIENER
GAUS
EMF
FMF
TMED
ATMED
GMED
FIDRM
WFM
FWM
AWFM
AWFM2
CK
FDDF
TMAV
ATMAV
DWMAV
GMAV
MPASS
FMMF
FIRE
DSFIRE
PWLFIRE1
PWLFIRE2
IFCF
MIFCF
EIFCF
SFCF
SSFCF
GOA
HAF
FSB1
FSB2
FSB1R
FSB2R
10%
1731.4
80.93
305.2
1734.06
1857.91
1714.73
13.24
40.01
83.74
95.95
81.73
12.1
130.03
550.88
121.96
69.32
1489.27
843.04
87.49
133.19
305.2
3441.11
123.37
996.78
88.7
269.61
112.32
49.76
92.97
95.62
93.2
111.68
104.52
265.46
89.4
87.64
88.52
107.2
130.63
20%
3309.4
127.17
540.48
3287.44
1819.8
3276.44
44.48
102.31
134.85
116.29
128.03
30.57
136.42
907.78
130.7
74.96
2954.2
2146.31
144.76
144.57
540.48
5669.26
211.66
1981.24
258.42
296.72
385.29
263.57
137.41
167.65
141.43
214.79
199.16
411.51
94.76
137.34
136.2
134.53
154.86
30%
4719.34
232.86
758.05
4679.23
2002.45
4670.75
118.23
212.53
248.44
145.58
348.88
51.1
146.13
1292.56
139.64
84.46
4313.4
3510.19
267.64
145.82
758.05
508.46
376.51
3033.74
560.76
344.26
798.03
603.99
217.4
292.64
224.32
419.16
401.28
536.37
101.72
253.14
258.41
177.57
193.94
Table 1. MSE values for 10%, 20% and 30% impulse noise. The denoised image is compared to the original noisefree Lena
image.
1392
In this expression, In (x, y, t) and Io (x, y, t) stand for the grey value of the pixel with spatial coordinates (x, y) in
the t-th frame of the noisy and original sequence and pr [0, 1] denotes the probabilty that a pixel is corrupted
and replaced by a random noise value (x, y, t) coming from a uniform distribution.
It is quite clear that video makes things much more complicated: the introduction of a temporal component
implies that good filtering can only be achieved if the temporal aspect is taken well into account, i.e., if motion
information is incorporated in the filter design. Filters for color video are even more complex, because the
temporal information has to be combined with the color information.
As most video filters in literature are designed for sequences corrupted by gaussian noise, much less video
filters exist for the removal of impulse noise. Existing 2D filters for impulse noise can be extended to video by
applying the 2D filter to each frame of the video sequence, but temporal inconsistencies will arise when motion
is neglected in the filtering process. Existing 3D filters on the other hand often suffer from detail loss because
too many pixels are filtered or too many pixels are left unfiltered in order to preserve the image details.
Our FRINV-G filter, which is specifically designed to remove random impulse noise from grayscale image
sequences, tries to overcome these difficulties by using a multi-step approach. Instead of removing all the supposed
noise at once, the FRINV-G filter consists of three different successive filtering steps and a fourth refinement
step. In each filtering step, only the pixels that are detected as being noisy (a detection that is achieved by fuzzy
rules) are filtered and to exploit the temporal information detected pixels are filtered in a motion compensated
way. This multi-step approach might take more time than a single-step approach, but it allows to detect and
remove the noise in a more structured way.
To show the general principles behind each of these filtering steps we limit ourselves to the first step, as the
other steps are more or less similar. In this first step, both a degree to which a pixel is considered noisefree
and noisy is calculated. If the noisy degree is larger than the noisefree degree, the pixel is filtered in a motion
compensated way.
To obtain the noisefree degree of a pixel we adopt the following reasoning. A pixel can be considered noisefree
if it is similar to the pixel at the same spatial location in the previous and preprevious frame and has also some
similar neighbours. If the noisefree situation of the pixel is not confirmed in time, possibly due to motion, we
will require more spatial confirmation (similar neighbours). We can translate this into the following fuzzy rule
to obtain the noisefree degree.
Fuzzy Rule 1.
IF( the absolute differences |In (x, y, t)If2 (x, y, t1)| and |In (x, y, t)If (x, y, t2)| are NOT LARGE POSITIVE AND
there are two neighbours (x + k, y + l, t) (2 k, l 2 and (k, l) 6= (0, 0)) for which |In (x, y, t) In (x + k, y + l, t)|
is NOT LARGE POSITIVE)
OR there are four neighbours (x+k, y+l, t) (2 k, l 2 and (k, l) 6= (0, 0)) for which |In (x, y, t)In (x+k, y+l, t)|
is NOT LARGE POSITIVE
THEN the pixel at position (x, y, t) is considered NOISEFREE.
In a similar way, we adopt the following reasoning to detect noisy pixels. A pixel can be considered noisy
if the difference in grey value compared to the pixel at the same spatial location in the filtered previous frame
is large positive or large negative and if this does not hold for its neighbours (then we can assume that the
difference is not caused by motion) and if this is confirmed spatially by the fact that there is a direction in
which the differences in grey level between the considered pixel and the two respective neighbours are both large
positive or large negative and if the difference in grey value of those two neighbours is not large positive (then
we can assume that the pixel is an impulse between two pixels that are expected to belong to the same object).
We can translate this into the following fuzzy rule to obtain the noise degree.
1393
Figure 4. 20-th frame of the Deadline video sequence: top = original frame, middle = noisy frame (25% random impulse
noise), bottom = result of our FRINV-G filter.
Fuzzy Rule 2.
IF((the difference In (x, y, t) If2 (x, y, t 1) is LARGE
POSITIVE AND NOT for five neighbours (x + k, y + l, t) (2 k, l 2 and (k, l) =
6 (0, 0)) the difference In (x +
k, y + l, t) If2 (x + k, y + l, t 1) is LARGE POSITIVE)
OR (the difference In (x, y, t) If2 (x, y, t 1) is LARGE
NEGATIVE AND NOT for five neighbours (x + k, y + l, t) (2 k, l 2 and (k, l) =
6 (0, 0)) the difference In (x +
k, y + l, t) If2 (x + k, y + l, t 1) is LARGE NEGATIVE))
AND (( in one of the four directions the differences In (x, y, t) In (x + k, y + l, t) and In (x, y, t) In (x k, y l, t)
((k, l) {(1, 1), (1, 0), (1, 1), (0, 1)}) are both LARGE
POSITIVE OR both LARGE NEGATIVE) AND the absolute difference |In (x+k, y+l, t)In (xk, yl, t)| is NOT LARGE)
THEN the pixel at position (x, y, t) is considered noisy.
1394
To apply these rules one only needs a membership function to represent the linguistic value large positive
(which is in this case achieved by a piecewise linear function); its reflection is the membership function of the
linguistic value large negative. The degree to which a pixel belongs to the fuzzy set noisefree corresponds to the
degree to which the antecedent in fuzzy rule 1 is true. The degree to which a pixel belongs to the fuzzy set noisy
corresponds to the degree to which the antecedent in fuzzy rule 2 is true.
Finally, all pixels for which the noisy degree is higher than the noisefree degree are filtered, the other pixels
remain unchanged. The filtering is performed in a motion compensated way (block matching).
Experimental results clearly demonstrate that the FRINV-G filter outperforms other state-of-the-art filters.
The performance of the proposed method was compared to several 3D state-of-the-art filters, such as the 3D
rational filter,24 the adaptive 3D median filter (A3DM) and the weighted 3D median filter (W3DM),25 the
adaptive 3D LUM smoother (LUM)26 and the peak-and-valley filter (PAV).27 An example of the powerful visual
result of the FRINV-G filter is shown in Figure 4.
5. CONCLUDING REMARKS
Fuzzy set theory and fuzzy logic are tools to model human (approximate) reasoning, using linguistic variables in
the reasoning process. This ability is a major advantage in applications where uncertainty and imprecision are
naturally present. Noise removal in images and video is such a application, as it is usually uncertain whether a
pixel is contaminated with noise, and if so, to what degree. In this contribution we have just briefly touched this
subject by discussing two fuzzy filters. The goal here was not to provide all the technical details (we refer to the
bibliography below), but to give a quick look on some construction and performance aspects, and to show that
fuzzy set theory and fuzzy logic can be very usefull tools in image processing applications.
REFERENCES
[1] L.A. Zadeh, Fuzzy Sets, in: Information and Control, Vol. 8, 1965, pp. 338-353.
[2] X. Wang, D. Ruan, E.E. Kerre, Mathematics of Fuzziness - Basic Issues, in: Studies in Fuzz. and Soft
Comp., Vol. 245, Springer, 2009.
[3] S. Schulte, M. Nachtegael, V. De Witte, D. Van der Weken, E.E. Kerre, A Fuzzy Impulse Noise Detection
and Reduction Method, in: IEEE Trans. on Image Proc., Vol. 15(5), 2006, pp. 1153-1162.
[4] K. Arakawa, Median filter based on fuzzy rules and its application to image restoration, in: Fuzzy Sets and
Systems, Vol. 77, 1996, pp. 3-13.
[5] K. Arakawa, Fuzzy rule-based image processing with optimization, in: Fuzzy Techniques in Image Processing
(Kerre E.E. & Nachtegael M., editors), Springer-Verlag, 2000, pp. 222-247.
[6] H.K. Kwan, Fuzzy filters for noise reduction in images, in: Fuzzy Filters for Image Processing (Nachtegael
M., Van der Weken D., Van De Ville D. & Etienne E.E., editors), Springer-Verlag, 2002, pp. 25-53.
[7] C.S. Lee, Y.H. Kuo, P.T. Yu, Weighted fuzzy mean filters for image processing, in: Fuzzy Sets and Systems,
Vol. 89, 1997, pp. 157-180.
[8] Lee C.S. & Kuo Y.H., Adaptive fuzzy filter and its application to image enhancement, in: Fuzzy Techniques
in Image Processing (Kerre E.E. & Nachtegael M., editors), Springer-Verlag, 2000, pp. 172-193.
[9] Y. Choi, R. Krishnapuram, Image enhancement based on fuzzy logic, in: IEEE Proceedings, 1995, pp.167171.
[10] M. Mancuso, R. De Luca, R. Poluzzi, G.G. Rizzotto, A fuzzy decision directed filter for impulsive noise
reduction, in: Fuzzy Sets and Systems, Vol. 77, 1996, pp. 111-116.
[11] M.G. Forero-Vargas, L.J. Delgado-Rangel, Fuzzy filters for noise reduction, in: Fuzzy Filters for Image
Processing (Nachtegael M., Van der Weken D., Van De Ville D. & Etienne E.E., editors), Springer-Verlag,
2002, pp. 3-24.
[12] F. Russo, G. Ramponi, A noise smoother using cascade FIRE filters, in: Proceedings of the 4th FUZZ-IEEE
Conference, 1995, pp. 351-358.
[13] J.Y. Jiu, Multilevel median filter based on fuzzy decision, DSP IC Design Lab E.E. NTU., 1996.
[14] F. Russo, G. Ramponi, A fuzzy filter for images corrupted by impulse noise, in: IEEE Sign. Proc. Letters,
Vol.3, No. 6, 1996, pp. 168-170.
1395
[15] F. Russo, G. Ramponi, Removal of impulse noise using a FIRE filter, in: IEEE Proceedings, 1996, pp.
975-978.
[16] F. Russo, FIRE operators for image processing, in: Fuzzy Sets and Systems, Vol. 103, 1999, pp. 265-275.
[17] F. Farbiz, M.B. Menhaj, A fuzzy logic control based approach for image filtering, in: Fuzzy Techniques in
Image Processing (Kerre E.E. & Nachtegael M., editors), Springer-Verlag, 2000, pp. 194-221.
[18] F. Farbiz, M.B. Menhaj, S.A. Motamedi, Edge preserving image filtering based on fuzzy logic, in: Proceedings of the Sixth EUFIT conference (Aken, Duitsland), 1998, pp. 1417-1421.
[19] D. Van De Ville, M. Nachtegael, D. Van der Weken, E.E. Kerre, W. Philips, I. Lemahieu, Noise Reduction
by Fuzzy Image Filtering, in: IEEE Trans. on Fuzzy Systems, Vol. 11(4), 2003 pp. 429-436.
[20] J.H. Wang, H.C. Chiu, HAF: An adaptive fuzzy filter for restoring highly corrupted images by histogram
estimation, in: Proc. Natl. Sci. Counc. ROC(A), Vol. 23, No. 5, 1999. pp. 630-643.
[21] G. Tolt, I. Kalaykov, Fuzzy-similarity-based noise cancellation for real-time image processing, in: Proceedings of the 10th FUZZ-IEEE Conference, 2001, pp. 15-18.
[22] G. Tolt, I. Kalaykov, Real-time image noise cancellation based on fuzzy similarity, in: Fuzzy Filters for Image
Processing (Nachtegael M., Van der Weken D., Van De Ville D. & Etienne E.E., editors), Springer-Verlag,
2002, pp. 54-71.
[23] T. Melange, M. Nachtegael, E.E. Kerre, A Fuzzy Filter for Random Impulse Noise Removal From Video,
In: Proc. of NAFIPS 2010, 2010, pp. 204-209.
[24] F. Cocchia, S. Carrato, G. Ramponi, Design and real-time implementation of a 3-D rational filter for edge
preserving smoothing, in: IEEE Trans. on Consumer Electronics, Vol. 43(4), 1997, pp. 1291-1300.
[25] J.-S. Kim, H.W. Park, Adaptive 3-D median filtering for restoration of an image sequence corrupted by
impulse noise, in: Signal Proc.: Image Communication, Vol. 16, 2001, pp. 657-668.
[26] R. Lukac, S. Marchevsky, LUM smoother with smooth control for noisy image sequences, in: EURASIP J.
on Applied Signal Proc. 2001(2), pp. 110 120.
[27] P.S. Windyga, Fast impulsive noise removal, in: IEEE Trans. Image Proc., Vol. 10(1), 2001, pp. 173-179.
1396