Thanks to visit codestin.com
Credit goes to www.scribd.com

0% found this document useful (0 votes)
4 views37 pages

Supervised Learning of Neural Networks

Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PPTX, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
4 views37 pages

Supervised Learning of Neural Networks

Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PPTX, PDF, TXT or read online on Scribd
You are on page 1/ 37

Supervised Learning

of Neural Networks
P R E P A R E D B Y
M R S . D . T E J A S W I , I T
A N I T S S
Contents

• Introduction
• Perceptrons
• Adaline Back Propagation Multilayer Perceptrons
• Back Propagation Learning Rules
• Methods of Speeding
• Radial Basis Function Networks
• Functional Expansion Networks
Introduction

• As the name suggests, supervised learning takes place under the


supervision of a teacher.
• This learning process is dependent.
• During the training of ANN under supervised learning, the input
vector is presented to the network, which will produce an output
vector.
• This output vector is compared with the desired/target output vector.
• An error signal is generated if there is a diff erence between the
actual output and the desired/target output vector.
• On the basis of this error signal, the weights would be adjusted until
the actual output is matched with the desired output.
Can a single neuron learn a task?

• In 1958, Frank Rosenblatt introduced a training


algorithm that provided the first procedure for
training a simple ANN : a Perceptron.
• The perceptron is the simplest form of a neural
n/w.
• It consists of a single neuron with adjustable
synaptic weights and a hard limiter
The Main aim of perceptron is classify inputs, x1,x2,x3……….xn, into
one of two classes , say A1 and A2
Components:
1. Input Layer
2. Weights
3.Bias
4. Activation Function
5. Output
6. Training Algorithm
Types of Perceptrons

1. Single Layer Perceptron model: The main


objective of the single-layer perceptron model
is to analyze the linearly separable objects with
binary outcomes. A Single-layer perceptron can
learn only linearly separable patterns.
2. Multi-Layered Perceptron model: It is mainly
similar to a single-layer perceptron model but
has more hidden layers.
Single Layer Perceptron
model
Single Layer Perceptron model
Algorithm(For each Epoch)
Step 1: The perceptron model begins with
multiplying all input values and their
weights, then adds these values to create
the weighted sum
Step 2: Further, this weighted sum is applied
∑wi*xi = x1*w1 + x2*w2 + x3*w3+……..x4*w4
to the activation function ‘f’ to obtain the
desired output.
Y=f(∑wi*xi + b)

Step -3
Step 4: Update the weights of the Perceptrons

Step 5 :Alter the weights and bias using the Perceptron learning run
the show.
Example:
Implement a single-layer perceptron to learn the AND logical
operation. You are given two binary inputs, X1​and X2​, and the
initial weights and bias are set to zero (w1​=0, w2​=0, b=0) and
learning rate α=1
X1 X2 Target
(t)
1 1 1
1 -1 -1
-1 1 -1
-1 -1 -1
Solution
• Epotch-1

Inputs Target Y(in) Expect ∆w1 ∆w ∆b1 Updated


Output ed O/P 2 Weights
(W1=0,
W2=0,b
=0)
X1 X2 t Y WW b
12
1 1 1 0 0 1 1 1 11 1
1 -1 -1 1 1 -1 1 -1 02 0
-1 1 -1 2 1 1 -1 -1 11 -1
-1 -1 -1 -3 -1 0 0 0 11 -1
Continue
• Epotch-2

Inputs Target Yn) Expecte Updated


Output d O/P Weights
(W1=1,W2=1,b
=-1)
X1 X2 t Y W1 W2 b
1 1 1
1 -1 -1
-1 1 -1
-1 -1 -1
ADVANTAGES DRAWBACKS

• Simplicity and • Linear


Ease of Separability
Understanding
Ex: XOR Problem
• Low
• Limited to Binary
Computational
Classifi cation
Cost
• Lack of Hidden
• Fast Training
Layers
• Binary
• Over fi tting
Classifi cation
• Generalization
ADALINE Multi perceptron
• ADALINE is an acronym for ADAptive LINear Element (or
ADAptive LInear NEuron). It was developed by Bernard
Widrow and Marcian Hoff (1960).
• The Adaline learning rule (also known as the least-
mean-squares rule, the delta rule, and the Widrow-Hoff
rule) is a training rule that minimizes the output error
using (approximate) gradient descent.
• After each training pattern I p is presented, the
correction to apply to the weights is proportional to the
error.
• The correction is calculated before the thresholding
step, using err ij (p)=T p -W ij I p :
ADALINE Multi perceptron
ADALINE Multi perceptron
PEDAGOGY
Example- Design a OR Gate using
Adaline

Assume W1,W2,b=0.1
Leaning rate α=0.1
Error Rate Er=2
Solution
• Epotch-1

Inputs Target Y(in)= Expecte Updated Error


Output b+∑xi d O/P Weights
wi (W1=0.1,W2=0.
1, b=0.1)
X1 X2 t (t-yin) W1 W2 b (t-
yin)2
1 1 1
1 -1 1
-1 1 1
-1 -1 -1
Solution
• Epotch-2

Inputs Target Y(in)= Expecte Updated Error


Output b+∑xi d O/P Weights
wi (W1=0.26,W2=
0.27, b=0.27)
X1 X2 t (t-yin) W1 W2 b (t-
yin)2
1 1 1
1 -1 1
-1 1 1
-1 -1 -1
EXAMPLE -2
Continue the same process

Continue the same


process untill you satify
with error rate
Radial Bias Function in Neural network
RBFNN-Introduction
• Data can be either linearly/ non-linearly
separable.
• Single layer perceptron can be used for
classifying linearly separable data.
• Multi layer perceptron can be used for
classifying non-linearly separable data.
RBFNN-Introduction
• Radian Bias Function is a type of multi layer
perceptron which has one input layer, one
output layer and strictly one hidden layer.
•The hidden layer uses a non-
linear radial bias function as
the activation function, Which
converts the i/p parameter into
high dimension space which is
then fed into the n/w to
linearly separate the problem.
• Typical RBF’s are
1. Gaussian RBF

2. Multiquadaric RBF
Example: Design XOR Function Gaussian RBF
Thank you

You might also like