Escolar Documentos
Profissional Documentos
Cultura Documentos
30
International Conference on Advances in Computer Applications (ICACA) 2012
Proceedings published by International Journal of Computer Applications® (IJCA)
Biological systems are one of the complex system examples. 3.6 PSEUDORANDOM NUMBER
Neural network [42][43][44] is a part of biological system GENERATORS
[45][46] [47] domain. Human brain has a complicated
architecture. It is built of large number of neurons that interact Cellular automata is also used to generate Pseudorandom
with each other mutually. Brain engineering [48][49]is the Number [46][2]. We use controllable CA for this method. We
latest attempt to understand the brain-architecture. The base of can obtain the real random numbers by some physical
Brain engineering system is cellular automata, that manage phenomenon. Rule 30 [60] of cellular automata is use to
the growth and communication between neurons. Bayesian generate the random numbers [60]. The sequence of numbers
theory [27][50] is one of the approach that we follow in in PNRG does not always random, it can be determined by
biological systems. It is based on statistical classification. We initial values of small cells. CCA (controllable CA) PNRG
can combine the output of multiple networks for decision- and PCA (programmable CA) PNRG are the two approaches
making in higher-level systems for generating random numbers. PCA is non-uniform,
different rules are allowed at different time stamps. Transition
role of cell play a very major role in this whether it is uniform
or non-uniform CA.
31
International Conference on Advances in Computer Applications (ICACA) 2012
Proceedings published by International Journal of Computer Applications® (IJCA)
5. DISCUSSIONS All the above classifiers are implemented in Weka. IBK (Lazy
Classifier) is a Cellular Automata based classifier. I studied
We can create an effective classifier with cellular automata different classifiers output and try to analyze their
that use a pure local decision. If we use simple cells in a performance. Table 1 shows the result of different classifiers.
collective behavior then it achieves the classification Their execution time and accuracy. On the basis of above
performance. Each cell has its own local decision and work table , we can conclude that the most accurate classifiers are
with the neighbor’s information. We put the cells in the grid J48 and SMO because they correctly classified 96% instances.
space in such a way that each attribute of the classified dataset But SMO and J48 takes 0.19 seconds and 0.02 seconds to
will be in its own dimension. But we have to consider the grid execute. So, on the basis of execution time J48 is better than
size because it will increase with number of attributes of the SMO. On the other hand, IBK takes minimum time to execute
dataset. So the k- nearest neighborhood method is one of the that is 0 second and it is 95.3333% accurate. So, on the basis
best approach for this type of classification. of execution time IBK(Lazy Classifier) shows comparatively
better result and that is Cellular Autamata based Classifier.
32
International Conference on Advances in Computer Applications (ICACA) 2012
Proceedings published by International Journal of Computer Applications® (IJCA)
B. Comparison and Analysis of Cellular Automata It contains the cellular automata behavior with the instances
(CA) Based Classifiers in Data mining (Iris Dataset) of the class, how they behave, what are the characteristics of
object and their respective neighbors. It also describes the
uses of cellular automata in other areas such in biological,
CA Executi Correc Incorre TP Rate FP Rate ecological and in parallel computing. We can say that with the
based on tly ctly (Weigh (Weigh pattern recognition approach and following the k-nearest
Classifi time Classifi Classifi ted ted neighborhood algorithm it is possible to achieve the
ers ed ed Avg.) Avg.) classification through cellular automata that define a fusion of
(Lazy Instan Instanc data mining with cellular automata. I think this paper will
Classifi ces es really help to cellular automata researcher in coming future.
er)
8. REFERENCES
IB1 0 95.333 4.6667 0.953 0.023
second 3% % [1] Tom Fawcett (Center for the Study ), “Data mining with
s cellular automata,” pp. 1-10, 2007.
[2] N. Ganguly, B. K. Sikdar, A. Deutsch, G. Canright, and
IBK 0 95.333 4.6667 0.953 0.023
P. P. Chaudhuri, “A Survey on Cellular Automata∗,”
second 3% %
Engineering, pp. 1-30.
s
[3] P. Gu and Y. Zheng, “A New Data Mining Model for
KStar 0 94.666 5.3333 0.947 0.027 Forest-Fire Cellular Automata,” 2012 Fifth International
second 7% % Conference on Intelligent Computation Technology and
s Automation, pp. 37-40, Jan. 2012.
LWL 0 93.333 6.6667 0.933 0.033 [4] H. Nishio, “Changing the Neighborhood of Cellular
second 3% % Automata,” no. September, pp. 10-13, 2007.
s
[5] S. Physics, R. October, K. E. Y. Words, and A. Study,
Table 3 “Two-Dimensional Cellular Automata,” vol. 38, 1985.
[6] C. D. Brummitt, H. Delventhal, and M. Retzlaff,
“Packard Snowflakes on the von Neumann
Lazy Classifier is a CA based classifier. IB1, IBK, KStar and Neighborhood,” Science, vol. 3, pp. 57-79, 2008.
LWL are different types of Lazy Classifiers. I performed the
comparative analysis work on different CA based classifiers [7] P. M. Wocjan and C. Science, “Hamiltonian Cellular
in Weka. On the basis of Table 2 we conclude that all lazy Automata in Joint work with Daniel Nagaj ( MIT ),”
classifiers execution time is almost same i.e zero seconds. Electrical Engineering.
Their accuracy is also almost same. But the least accurate one
[8] A. Seth, S. Bandyopadhyay, and U. Maulik,
is the LWL which correctly classified only 93.3333%
“Probabilistic Analysis of Cellular Automata Rules and
instances.
its Application in Pseudo Random Pattern Generation,”
C. Comparison and analysis of CA based Classifier on International Journal, no. November, 2008.
different datasets
[9] B. Meshkboo and M. Kangavari, “Video Data Mining
with Learning Cellular Automata,” pp. 1-9.
Dataset IBK (Lazy Classifier) [10] A. Sleit, A. Dalhoum, I. Al-dhamari, and A. Awwad,
Execution Time “Efficient Enhancement on Cellular Automata for Data
Mining,” Technology, pp. 616-620.
Abalone 0 seconds
[11] L. Zhou and M. Yang, “A Classifier Build Around
Adult 0.01 seconds Cellular Automata for Distributed Data Mining,” 2008
International Conference on Computer Science and
Anneal 0 seconds Software Engineering, pp. 312-315, 2008.
Iris 0 seconds [12] T. N. Phyu, “Survey of Classification Techniques in Data
Mining,” Computer, vol. I, 2009.
Table 4 Table 3
[13] R. M. Hamou, A. Lehireche, A. C. Lokbani, and M.
Table 3 shows the performance of CA based classifier i.e
Rahmani, “Text Clustering by 2D Cellular Automata
Lazy classifier (IBK) on different datasets. On the basis of
Based on the N-Grams,” 2010 First ACIS International
Table 3 we can conclude that IBK takes almost equal time (0
Symposium on Cryptography, and Network Security,
seconds) on every dataset.
Data Mining and Knowledge Discovery, E-Commerce
7. CONCLUSIONS and Its Applications, and Embedded Systems, pp. 271-
277, Oct. 2010.
This paper elaborates the fusion of data mining with cellular
automata. It tells the brief history of cellular automata and [14] M. P. and P. M. M. S. of G. U. of N. N. U. K. N. 2RD,
data mining systems. It also includes the different type of “DECISION TREE BASED CLASSIFICATION OF
classification algorithm of data mining and their brief REMOTELY SENSED DATA,” no. November, pp. 5-8,
associative approaches. 2001.
33
International Conference on Advances in Computer Applications (ICACA) 2012
Proceedings published by International Journal of Computer Applications® (IJCA)
[15] P. Yin, A. Criminisi, J. Winn, and I. Essa, “Tree-based [32] T. Weise, “Global Optimization Algorithms – Theory
Classifiers for Bilayer Video Segmentation,” 2007 IEEE and Application –,” 2009.
Conference on Computer Vision and Pattern
Recognition, pp. 1-8, Jun. 2007. [33] M. V. Butz and M. Pelikan, “Studying XCS / BOA
Learning in Boolean Functions : Structure Encoding and
[16] J. N. Al-karaki and A. E. Kamal, “Routing Techniques in Random Boolean Functions Categories and Subject
Wireless Sensor Networks : A,” Computer Engineering, Descriptors,” Cognitive Psychology.
pp. 1-37.
[34] S. W. Wilson, “State of XCS Classifier System
[17] M. Modeling, “Different Routing Techniques in Research,” Environment, no. 99, pp. 1-21, 1999.
VANET.”
[35] P. I. Armstrong, S. X. Day, J. P. McVay, and J. Rounds,
[18] U. S. Office and P. Management, “THE CLASSIFIER ’ “Holland’s RIASEC model as an integrative framework
S HANDBOOK Table of Contents ( Also See The for individual differences.,” Journal of Counseling
Introduction to the Position Classification Standards .) Psychology, vol. 55, no. 1, pp. 1-18, 2008.
THE CLASSIFIER ' S HANDBOOK Table of Contents (
Continued ),” Management, vol. 41, no. August, pp. 1- [36] M. Abedini and M. Kirley, “Guided Rule Discovery in
45, 1991. XCS for High-dimensional Classification Problems,”
Engineering.
[19] K. P. Murphy, “Naive Bayes classifiers Generative
classifiers,” Bernoulli, pp. 1-8, 2006. [37] A. Orriols-puig, D. E. Goldberg, E. Bernad, and K.
Sastry, “Modeling XCS in Class Imbalances : Population
[20] J. V. Marcos, R. Hornero, D. Alvarez, F. del Campo, M. Size and Parameter Settings,” no. January, 2007.
López, and C. Zamarrón, “Radial basis function
classifiers to help in the diagnosis of the obstructive sleep [38] A. Gandhe, S.-hsin Yu, and R. E. Smith, “XCS for
apnoea syndrome from nocturnal oximetry.,” Medical & Fusing Multi-Spectral Data in Automatic Target
biological engineering & computing, vol. 46, no. 4, pp. Recognition.”
323-32, May 2008. [39] Q. Jackson and D. Landgrebe, “No Title,” Office, pp. 1-
[21] Y.-J. Oyang, S.-C. Hwang, Y.-Y. Ou, C.-Y. Chen, and 35, 2001.
Z.-W. Chen, “Data classification with radial basis [40] P. Sarkar, “A brief history of cellular automata,” ACM
function networks based on a novel kernel density Computing Surveys, vol. 32, no. 1, pp. 80-107, Mar.
estimation algorithm.,” IEEE transactions on neural 2000.
networks / a publication of the IEEE Neural Networks
Council, vol. 16, no. 1, pp. 225-36, Jan. 2005. [41] “Game of Life,” Physics, 1970.
[22] A. G. Bors, “Introduction of the Radial Basis Function ( [42] G. P. Zhang, “Neural Networks for Classification : A
RBF ) Networks,” Science, pp. 1-7. Survey,” vol. 30, no. 4, pp. 451-462, 2000.
[23] C. J. C. Burges, “A Tutorial on Support Vector Machines [43] A. S. Introduction, “Neural Networks,” Neural Networks,
for Pattern Recognition,” Data Mining and Knowledge 1996.
Discovery, vol. 43, pp. 1-43, 1997. [44] C. Gershenson, “Artificial Neural Networks for
[24] S. Tong and D. Koller, “with Applications to Text Beginners,” Networks, pp. 1-8.
Classification,” Journal of Machine Learning Research, [45] M. Daven, A. Kheyfits, G. Zeleke, MSam, and Elkamu,
pp. 45-66, 2001. “An introduction to Cellular Automata and their
[25] W. S. Noble and P. Street, “What is a support vector applications Pascal ’ s Triangle,” Group, pp. 1-23, 2001.
machine ?,” Nature Biotechnology, vol. 24, no. 12, pp. [46] N. Yu, M. Li, and X. Ruan, “Applications of cellular
1565-1567, 2006. automata in complex system study,” Education, vol. 1,
[26] “A LAZY LEARNING CONTROL METHOD USING no. 3, pp. 302-310, 2005.
SUPPORT VECTOR REGRESSION Masayuki [47] A. P. Carlos, “Quantum Cellular Automata : Theory and
Kobayashi, Yasuo Konishi and Hiroyuki Ishigaki,” Applications by,” 2007.
Information and Control, vol. 3, no. 6, p. 2007, 2007.
[48] J. R. Wolpaw, D. J. McFarland, and T. M. Vaughan,
[27] Z. Zheng and G. I. Webb, “Lazy Learning of Bayesian “Brain-computer interface research at the Wadsworth
Rules,” vol. 87, pp. 53-87, 2000. Center.,” IEEE transactions on rehabilitation
[28] M.-ling Zhang and Z.-hua Zhou, “Ml-knn : A Lazy engineering : a publication of the IEEE Engineering in
Learning Approach to Multi-Label Learning,” Medicine and Biology Society, vol. 8, no. 2, pp. 222-6,
Technology. Jun. 2000.
[29] G.-xun Yuan, C.-hua Ho, and C.-jen Lin, “Recent [49] B. R. Friedman, “No Title,” Simulation, 2009.
Advances of Large-scale Linear Classification,” [50] K. D. Anthony, “Introduction to Causal Modeling ,
Computer, pp. 1-25. Bayesian Theory and Major Bayesian Modeling Tools
[30] R.-en Fan, X.-rui Wang, and C.-jen Lin, “LIBLINEAR : for the Intelligence Analyst,” European Journal Of
A Library for Large Linear Classification,” Corpus, vol. Operational Research, pp. 1-31.
9, pp. 1871-1874, 2008. [51] P. S. Albin, “Rationality :”
[31] M. Dredze, “Confidence-Weighted Linear [52] “Theory and Applications of Nonlinear Cellular
Classification,” Update, 2008. Automata In VLSI Design Theory and Applications of
34
International Conference on Advances in Computer Applications (ICACA) 2012
Proceedings published by International Journal of Computer Applications® (IJCA)
Nonlinear Cellular Automata In VLSI Design,” [63] A. Ault, X. Zhong, and E. J. Coyle, “K-Nearest-Neighbor
Engineering and Science. Analysis of Received Signal Strength Distance
Estimation Across Environments,” Environment.
[53] I. Processing, “Image Processing,” Image Processing, pp.
404-465. [64] A. D. Lattner, “Instance-Based Learning and Information
Extraction for the Generation of Metadata,”
[54] “What is Pattern Recognition ?,” vol. 25, no. 1, 2003. Management, pp. 472-479, 2003.
[55] D. Talia, “Cellular Automata + Parallel Computing = [65] D. W. Aha, D. Kibler, and M. K. Albert, “Instance-based
Computational Simulation,” Computer. learning algorithms,” Machine Learning, vol. 6, no. 1,
[56] J. Suthakorn and G. S. Chirikjian, “A Semi-Autonomous pp. 37-66, Jan. 1991.
Replicating Robotic System,” Mechanical Engineering, [66] A. Y. Al-Omary and M. S. Jamil, “A new approach of
pp. 776-781, 2003. clustering based machine-learning algorithm,”
[57] A. Smith, P. Turney, and R. Ewaschuk, “Self-replicating Knowledge-Based Systems, vol. 19, no. 4, pp. 248-258,
machines in continuous space with virtual physics.,” Aug. 2006.
Artificial life, vol. 9, no. 1, pp. 21-40, Jan. 2003. [67] I. H. Witten and E. Frank, “Machine Learning
[58] I. Electricista, M. Automática, and U. Valle, Algorithms in Java Nuts and bolts : Machine,” Machine
“Implementation of a Self-Replicating Héctor Fabio Learning, 2000.
Restrepo-García,” vol. 2457, 2001. [68] G. Schmidberger, “Practical Data Mining Tutorial 2 :
[59] “NP-complete problems,” Search. Nearest Neighbor Learning and Decision Trees,”
Learning, 2008.
[60] D. Gage, E. Laub, and B. Mcgarry, “Cellular automata: is
rule 30 random?,” pp. 1-10. [69] T. V. Inglesby and D. a Henderson, “Introduction.,”
Biosecurity and bioterrorism : biodefense strategy,
[61] G. Guo, H. Wang, D. Bell, Y. Bi, and K. Greer, “KNN practice, and science, vol. 10, no. 1, p. 5, Mar. 2012.
Model-Based Approach in Classification,” Idea.
[62] J. A. Gamez and P. Puerta, Jose MBermejo, “Improving
KNN-based e-mail classification into folders generating
class-balanced datasets .,” pp. 529-536, 2008.
35