Escolar Documentos
Profissional Documentos
Cultura Documentos
Terminal Branches
Dendrites
of Axon
x1
w1
x2
w2
x3 w3
Axon
wn
xn
Learning in biological systems involves adjustments to the synaptic connections that exist
between the neurons. This is true of ANNs as well i.e learning occurs via changes in value of the
connection weights. ANNs can process information at a great speed owing to their highly
massive parallelism.
The building blocks of artificial neural networks are artificial neurons. Artificial neurons have
the same basic components as biological neurons. Three basic elements of the neural model are
A. A set of synapses, or connecting links, each of which is characterized by a weight or
strength of its own.
B. An adder for summing the input signals,weighted by the respective synaptic strengths of
the neuron; the operations described here constitute a linear combiner.
C. An activation function for limiting the amplitude of the output of a neuron.
The simplest ANNs consist of a set of McCulloch-Pitts neuron was originally proposed by
McCulloch and Pitts (McCulloch & Pitts 1943) which is the attempt to model the computing
process of the biological neuron. The McCulloch-Pitts model is composed of one neuron only,
limited computing capability and no learning capability.
Equation-1
A negative value for a weight indicates an inhibitory connection while a positive value
indicates an excitatory one. Although in biological neurons, θ has a negative value, it
may be assigned a positive value in artificial neuron models. If θ is positive, it is usually
Artificial neural networks are among the most powerful learning models. They have the
versatility to approximate a wide range of complex functions representing multidimensional
input-output maps. Neural networks also have inherent adaptability, and can perform robustly
even in noisy environments.
Artificial neural network is a computational model inspired by the structure, processing method
and learning ability of a biological brain. It is composed of a number of interconnected simple
processing elements called neurons or nodes. The neurons interconnected to form a group and
process information using a connectionist.
Neural networks, with their remarkable ability to derive meaning from complicated or imprecise
data, can be used to extract patterns and detect trends that are too complex to be noticed by either
humans or other computer techniques. A trained neural network can be thought of as an "expert"
in the category of information it has been given to analyse. This expert can then be used to
provide projections given new situations of interest and answer "what if" questions.
Other advantages include:
A. Adaptive learning: An ability to learn how to do tasks based on the data given for training
or initial experience.
B. Self-Organisation: An ANN can create its own organization or representation of the
information it receives during learning time.
C. Real Time Operation: ANN computations may be carried out in parallel, and special
hardware devices are being designed and manufactured which take advantage of this
capability.
D. Fault Tolerance via Redundant Information Coding: Partial destruction of a network
leads to the corresponding degradation of performance. However, some network
capabilities may be retained even with major network damage.
There are a wide variety of networks depending on the nature of information processing carried
out at individual nodes, the topology of the links, and the algorithm for adaptation of link
weights. In other words, a specific neural network model is determined by its topology, learning
paradigm and learning algorithm.
The manner in which the neurons of a neural network are structured is intimately linked with the
learning algorithm used to train the network. We may therefore speak of learning algorithms
(rules) used in the design of neural networks as being structured.
In general, according to the structure/topology of the connections, fundamentally ANN can be
categorized in to three different classes of network architectures:
A. Feed forward Neural networks
B. Recurrent Neural Networks
1. Feed forward Neural networks:
In feed forward neural networks, the neurons are organized in the form of layers. The neurons in
a layer get input from the previous layer and feed their output to the next layer. In this kind of
networks connections to the neurons in the same or previous layers are not permitted.
The last layer of neurons is called the output layerand the layers between the input and output
layers are called the hidden layers. The input layer is made up of special input neurons,
transmitting only the applied external input to their outputs.
Links in feed forward can only go in one direction and each unit is linked only in the unit
in next layer. i.e no units are linked between the same layer, back to the previous layer or
skipping a layer.
Computations can proceed uniformly from input to output units.
No feed-back connections
Figure 2.4: Single layer network Figure 2.5: Multilayer neural network
Learning/ Training in ANN is a process of teaching the neural net to learn hidden patterns from
training dataset. During the training process, NN parameters are adjusted in order to make
the actual outputs (predicated) close to the target (measured) outputs of the network.
There are three learning approaches in learning of ANN. These are:
Each activation function is characterized by its shape, output range, and derivative function. In
general, activation functions can be categorized as binary or continuous (differentiable)
activation functions. There are different types of activation functions, the two types are threshold
function and sigmoid function. Other commonly used activation function are: Tan-Sigmoid, Log-
Sigmoid and linear (purelin) activation function, ramp and Gaussian function. See below figure.
Perceptron consists of a single neuron with multiple inputs and a single output. It has restricted
information processing capability. The information processing is done through a transfer
function which is either linear or non-linear.
Figure 2.10: hyperplane (in this example, a straight line) as decision boundary for a two-
dimensional, two-class pattern-classification problem.
A perceptron can learn only examples that are called “linearly separable”. These are examples
that can be perfectly separated by a hyperplane. Some sets of positive and negative examples
cannot be separated by any hyperplane and those examples are called non linearly separable
examples.
0 1 -1 -0.3
1 0 -1 -0.3
1 1 1 0.2
x2
- +
- -
x1
-0.8 + 0.5 x1 + 0.5 x2 = 0
Figure 2.12: Two-Dimensional decision boundary for AND function
Even we can prove using the graph or decision boundary on the above figure, the examples can
be perfectly separated by a hyper plane (the positive one is on one side and negative ones is on
Test result x2
X1 X2 Output(desired) ∑WiXi
0 0 -1 -0.3
+ +
0 1 1 0.2
1 0 1 0.2
1 1 1 0.7 - +
x1
-0.3 + 0.5 x1 + 0.5 x2 = 0
Therefore, OR function also can be represented by perceptron using the graph and equation.
<Training examples> x2
x1 x2 output
0 0 -1 + -
0 1 1
- +
1 0 1 x1
1 1 -1
2.5.1 Perceptron Learning
Learning a perceptron means finding the right values for W (weight). The hypothesis space of a
perceptron is the space of all weight vectors. The perceptron learning algorithm can be stated as
below.
1. Assign random values to the weight vector
2. Apply the weight update rule to every training example
3. Are all training examples correctly classified?
A. Yes. Quit
B. No. Go back to Step 2.
Learning is to determine a weight vector that causes the perceptron to produce the correct +1 or –
1 for each of the given training examples. There are two popular weight update rules(algorithms)
A. The perceptron rule
Equ-1
E . Equ-2.
Notice E is itself a vector, whose components are the partial derivatives of E with respect to
each of the wi. When interpreted as a vector in weight space, the gradient specifies the direction
that produces the steepest increase in E. The negative of this vector therefore gives the direction
of steepest decrease.
The delta rule is as follows:
For a new training example X = (x1, x2, …, xn), update each weight according to this rule:
wi = wi + Δwi Equ-3
Multi-layered Perceptron (MLP): It has a layered architecture consisting of input, hidden and
output layers. Each layer consists of a number of perceptrons(neurons). A neuron in any layer of
the network is connected to all the neurons (nodes) in the previous layer. Signal flow through the
network progresses in a forward direction, from left to right and on a layer-by-layer basis. The
output of each layer is transmitted to the input of nodes in other layers through weighted links.
Usually, this transmission is done only to nodes of the next layer, leading to what are known as
feed forward networks.
Three points highlight the basic features of multilayer perceptrons are:
A. The model of each neuron in the network includes a nonlinear activation function that is
differentiable.
B. The network contains one or more layers that are hidden from both the input and output
nodes.
C. The network exhibits a high degree of connectivity, the extent of which is determined by
synaptic weights of the network.
These same characteristics, however, are also responsible for the deficiencies in our knowledge
on the behavior of the network. First, the presence of a distributed form of nonlinearity and the
high connectivity of the network make the theoretical analysis of a multilayer perceptron
difficult to undertake. Second, the use of hidden neurons makes the learning process harder to
visualize. In an implicit sense, the learning process must decide which features of the input
pattern should be represented by the hidden neurons. The learning process is therefore made
more difficult because the search has to be conducted in a much larger space of possible
functions, and a choice has to be made between alternative representations of the input pattern.
The typical architecture of a multi-layer perceptron (MLP) is shown below.
Figure 2.14: Architectural graph of a multilayer perceptron with two hidden layers.
Two kinds of signals are identified in this network:
1. Function Signals. A function signal is an input signal (stimulus) that comes in at the
input end of the network, propagates forward (neuron by neuron) through the network,