Escolar Documentos
Profissional Documentos
Cultura Documentos
www.elsevier.com/locate/dsw
Abstract
This paper demonstrates the ¯exibility of neural networks for modeling and solving diverse mathematical problems.
Advantages of using neural networks to solve problems include clear visualization, powerful computation and easy to
be made into hardware. In this paper, the well-known exclusive OR (XOR) problem is ®rst introduced. Then, two
examples are discussed in order to show how to use neural networks to represent dierent problems. One problem is
Taylor series expansion and the other is Weierstrass's ®rst approximation theorem. The neural representation of linear
programming and the neural representation of fuzzy linear programming are also discussed. Ó 2000 Elsevier Science
B.V. All rights reserved.
Keywords: Linear programming; Neural networks; Mathematical neural networks; Functional-link networks; Learning
algorithms; Fuzzy linear programming
0377-2217/00/$ - see front matter Ó 2000 Elsevier Science B.V. All rights reserved.
PII: S 0 3 7 7 - 2 2 1 7 ( 9 9 ) 0 0 3 7 6 - 8
H.-X. Li, X.L. Da / European Journal of Operational Research 124 (2000) 224±234 225
make a network more complicated so that the which can simulate XOR function. In fact, taking
learning speed of a network moves strikingly. w1 w2 1; w3 2, and h 1, Eq. (1) becomes
However, Pao (1989) prefers idea 2. It is interest- the following:
ing to learn that both ideas represent the same
concept. y u
x1 x2 2x1 x2 ÿ 1:
2
Example 1. So-called `functional-link' means that If we do 45° rotation of axes and the new coor-
signal transmission between neurons is designed as dinates are denoted by x01 and x02 , then the curve
function transformation in a neural network so
x1 x2 2x1 x2 1 can be expressed as follows:
that the grade of non-linearity is heightened. We
know that the network shown in Fig. 1 can not ÿ 0 p 2 ÿ 0 2
solve XOR problem. But if a higher order term x1 ÿ
2=2 x
ÿ ÿp 2 ÿp2 2 1;
3
x1 x2 is added in the network, then network 2=2 2=2
shown in Fig. 2 can deal with XOR problem.
which is a hyperbola. It can classify the two pat-
Pao (1989) suggests, by using functional-link, terns (as inputs) into two classes, i.e., 0- and 1-class
many problems can be solved by one-layer net- (as outputs) (see Fig. 3).
work. For instance, in Fig. 2, to add the higher
order term x1 x2 means to add a `new' input. But
the input is actually not new. It is only one forming Example 2. Above example solves XOR problem
by x1 and x2 , which can be regarded as interaction by just using idea 2. Now we consider the idea 1 to
of x1 and x2 . Then the output of the network is as solve XOR problem.
follows:
We design a 2-layer perceptron shown in Fig. 4,
y u
x1 w1 x2 w2 x1 x2 w3 ÿ h;
1 where the activation function of neuron fi is
ui
i 1; 2; 3, and the activation function of
neuron h is u. We de®ne u1 and u2 to be identity
functions (i.e., u1
u u, u2
u u, u3 to be a
hyperbolic function (i.e., u3
u; v uv), and u to
be step-up function (i.e., u
u 1, u P 0;
u
u 0, u < 0).
This comes back to Eq. (1). In other words, the
network shown in Fig. 2 is equivalent to the one
shown in Fig. 4.
Fig. 2. Add a higher order term x1 x2 . Fig. 3. The hyperbola for solving XOR problem.
226 H.-X. Li, X.L. Da / European Journal of Operational Research 124 (2000) 224±234
Remark 1. Clearly the neurons f1 and f2 in Ex- Fig. 7. A network for accomplishing the network Fig. 6.
ample 2 only play the role of `passageways', since
u1 and u2 are identity functions. So we can remove
the two `stations' f1 and f2 (see Fig. 5). y u
w1 u1
x1 w2 u2
x1 ; x2 w3 u3
x2
w4 x1 w5 x2 ÿ h
The activation functions of the neurons f and h u
w1 x21 w2 x1 x2 w3 x22 w4 x1 w5 x2 ÿ h;
are denoted by w and u, and taken for hyperbolic
function and step-up function. So we also have
which is the same as Eq. (4). This means that a
one-layer network with higher order terms can be
y u
w1 x1 w2 x2 w3 x1 x2 ÿ h: always accomplished by a multi-layer network,
vice versa.
We would like to say that the idea of a one-
Remark 2. In Example 1, we can add some higher layer network problem is easier than a multi-layer
order terms (see Fig. 6). Then the output of net- network problem seems not to be true. On the
work is the following: contrary, multi-layer networks can resolve the
complexity of one-layer networks so that they are
y u
w1 x1 w2 x2 w3 x1 x2 w4 x21 w5 x22 ÿ h: clearer and can be easily treated. As for the
learning algorithms of multi-layer networks, they
4
may not be more complicated than the ones of
single layer. Generally speaking, layer number of a
Chosen adequately w1 ; . . . ; w5 and h, it can also network and complexity of the network are not of
simulate XOR function. Of course, the one-layer direct ratio.
network in Fig. 6 can be accomplished by a 2-layer
network shown in Fig. 7, where the activation
function of the neurons f1 , f2 , f3 and h is respec-
tively u1 ; u2 ; u3 and u; u2 is taken for a hyperbolic 2. Mathematical essence of functional-link neural
function, u1 and u3 are all taken for quadratic networks
function (i.e., u1
u u2 ; u2
u u2 , and u as a
step-up function. Then the output of the network We begin with some examples. Pao (1989) gave
is as follows: an interesting example of a functional-link net-
H.-X. Li, X.L. Da / European Journal of Operational Research 124 (2000) 224±234 227
y w00 w01 sin px w02 cos px w03 sin px Example 4. We consider a general form of func-
tional-link networks (see Fig. 10, an example of
w04 cos 2px w05 sin 4px:
6
one-output is given; as for multi-output, the pre-
sentation is similar). The output of the network is
Clearly, it just is a triangular interpolation func-
the following:
tion.
y u
w1 x1 wn xn wn1 g1
x1 ; . . . ; xn
Example 3. Fig. 9 gives a functional-link (also
function expansion) network and its output is as wnm gm
x1 ; . . . ; xn ÿ h:
8
follows:
Clearly, the functional-link networks discussed
y u
w1 x w2 x2 w3 x3 wn xn ÿ h:
7 above are all special cases of this network.
Fig. 9. Another function expansion network. y u
w1 z1 w2 z2 wnm znm ÿ h:
9
228 H.-X. Li, X.L. Da / European Journal of Operational Research 124 (2000) 224±234
Then the output of the network is as follows Then the output of the network is the following:
!
Xn X
m
!
yu wi xi wnj gj
x1 ; . . . ; xn ÿ h ; X
nm
i1 j1 yu wj uj
x1 ; . . . ; xn ÿ h
i1
which is Eq. (8). This is to say the function that !
X
n X
nm
forms in a functional-link network can be perfectly u wi xi wj uj
x1 ; . . . ; xn ÿ h ;
i1 jn1
and their threshold values are b0 ; b1 ; . . . ; bm ; the If the set of feasible solutions
Pn is not empty, we
weight values in the network are cj and aij should take b0 > max j1 cj xj that is just the
i 1; 2; . . . ; m; j 1; 2; . . . ; n. For given input threshold value that we need.
x1 ; x2 ; . . . ; xn , the components of the output Problem 2: Under the same condition as in
y0 ; y1 ; . . . ; ym are as follows: Problem 1, we want to ®nd the threshold value b0
H.-X. Li, X.L. Da / European Journal of Operational Research 124 (2000) 224±234 231
Li, H.X., 1990. Multifactorial functions in fuzzy sets theory. Minsky, M., Papert, S., 1990. In: Naturally Intelligent Systems.
Fuzzy Sets and Systems 35, 69±84. Caudill, M., Butler, C. (Eds.), The MIT Press, Cambridge,
Li, H.X., 1995a. To see the success of fuzzy logic from MA.
mathematical essence of fuzzy control: On The paradoxical Minsky, M., Papert, S., 1991. In: Introduction to the Theory of
success of fuzzy logic. Fuzzy Systems and Mathematics 9 Neural Computation. Hertz, J., Krogh, A., Palmer, R.G.
(4), 1±14. (Eds.), Addison-Wesley, New York.
Li, H.X., 1997. Fuzzy Information Processing and Fuzzy Minsky, M., Papert, S., 1992. In: Neural Networks and Fuzzy
Computers. Science Press, New York, Beijing. Systems. Kosko, B. (Ed.), Prentice-Hall Inc., Englewood
Li, H.X., Li, L., 1999. Representing diverse mathematical Clis, NJ.
problems using neural networks in hybrid intelligent Pao, Y.H., 1989. Adaptive Pattern Recognition and Neural
systems. Expert Systems 16, 262±272. Networks. Addison-Wesley, New York.
Li, H.X., Yen, V.C., 1995b. Fuzzy Sets and Fuzzy Decision- Rosenblatt, 1990. In: Naturally Intelligent Systems. Caudill,
Making. CRC Press, Boca Raton, FL. M., Butler, C. (Eds.), The MIT Press, Cambridge, MA.
Li, H.X., Chen, C.L.P., Huang, H.P., 1999. Fuzzy Neural Rosenblatt, 1991. In: Introduction to the Theory of Neural
Intelligent Systems: Mathematical Foundation and the Computation. Hertz, J., Krogh, A., Palmer, R.G. (Eds.),
Applications in Engineering. CRC press, Boca Raton, Addison-Wesley, New York.
FL. Rosenblatt, 1992. In: Neural Networks and Fuzzy Systems.
Malik, S.C., 1984. Mathematical Analysis. Wiley, New York. Kosko, B. (Ed.), Prentice-Hall Inc., Englewood Clis, NJ.