Escolar Documentos
Profissional Documentos
Cultura Documentos
Neurons
used:
Hopfield
networks
The above is just a general picture of what neural
networks can do in real life. There are many
creative uses of neural networks that arises from
these general applications. One example is image
compression using association networks; another is
solving the Travelling Salesman's Problem using
clustering networks.
The perceptron
The perceptron is a mathematical model of a
biological neuron. While in actual neurons the
dendrite receives electrical signals from the axons
of other neurons, in the perceptron these electrical
signals are represented as numerical values. At the
synapses between the dendrite and axons,
electrical signals are modulated in various
amounts. This is also modeled in the perceptron by
multiplying each input value by a value called the
weight. An actual neuron fires an output signal only
when the total strength of the input signals exceed
a certain threshold. We model this phenomenon in
a perceptron by calculating the weighted sum of
the inputs to represent the total strength of the
input signals, and applying a step function on the
weight values of
each perceptron
weigh
are collectively
t
called the weight
vector
vector of that
perceptron.
What can a perceptron do?
As mentioned above, a perceptron calculates the
weighted sum of the input values. For simplicity, let
us assume that there are two input values, x and y
for a certain perceptron P. Let the weights for x and
y be A and B for respectively, the weighted sum
could be represented as: A x + B y.
Since the perceptron outputs an non-zero value
only when the weighted sum exceeds a certain
threshold C, one can write down the output of this
perceptron as follows:
Outpu
t of P {1 if A x + B y > C
=
{0 if A x + B y < =
C
Recall that A x + B y > C and A x + B y < C are the
two regions on the xy plane separated by the line A
x + B y + C = 0. If we consider the input (x, y) as a
point on a plane, then the perceptron actually tells
us which region on the plane to which this point
(Fig. 3) Graphs
showing linearly
separable logic
functions
In the above graphs, the two axes are the inputs
which can take the value of either 0 or 1, and the
numbers on the graph are the expected output for
a particular input. Using an appropriate weight
vector for each case, a single perceptron can
perform all of these functions.
However, not all logic operators are linearly
separable. For instance, the XOR operator is not
linearly separable and cannot be achieved by a
single perceptron. Yet this problem could be
overcome by using more than one perceptron
arranged in feed-forward networks.
(Fig.
4)
Since
it is
impos
sible
to
draw
a line
to
divide
the
region
s
contai
ning
either
1 or 0,
the
XOR
functi
on is
not
linearl
y
separ
able.