DHSCH 6
DHSCH 6
DHSCH 6
Classification
• Introduction
• Feedforward Operation and Classification
• Backpropagation Algorithm
3
Introduction
• Goal: Classify objects by learning nonlinearity
• There are many problems for which linear
discriminants are insufficient for minimum error
0 y1 = +1
x1 + x2 + 0.5 = 0
< 0 y1 = -1
0 y2 = +1
x1 + x2 -1.5 = 0
< 0 y2 = -1
nH d
gk ( x ) z k f w kj f w ji x i w j 0 w k 0 (1)
j 1 i 1
(k 1,...,c)
• Hidden units enable us to express more complicated nonlinear functions
and thus extend the classification
• We can allow the activation in the output layer to be different from the
activation function in the hidden layer or have different activation for each
individual unit
2 n 1
g( x ) j ij ( xi )
j 1
x I n ( I [ 0 ,1 ]; n 2 )
• Feedforward
The feedforward operations consists of presenting a
pattern to the input units and passing (or feeding) the
signals through the network in order to get outputs
units (no cycles!)
• Learning
The supervised learning consists of presenting an input
pattern and modifying the network parameters
(weights) to reduce distances between the computed
output and the desired output
Pattern Classification, Chapter 6
2
0
2
tz
• The total training error is the sum over the errors of n individual
patterns
n
J Jp (1)
p1
• Before training starts, the error on the training set is high; through
the learning process, the error becomes smaller
• The error per pattern depends on the amount of training data and
the expressive power (such as the number of weights) in the
network
EXERCISES
• Exercise #1.
Explain why a MLP (multilayer perceptron) does not
learn if the initial weights and biases are all zeros