SlideShare a Scribd company logo
1 of 31
Artificial Neural Networks
- Introduction Peter Andras
peter.andras@ncl.ac.uk
Overview
1. Biological inspiration
2. Artificial neurons and neural networks
3. Learning processes
4. Learning with artificial neural networks
Biological inspiration
Animals are able to react adaptively to changes in their
external and internal environment, and they use their nervous
system to perform these behaviours.

An appropriate model/simulation of the nervous system
should be able to produce similar responses and behaviours in
artificial systems.

The nervous system is build by relatively simple units, the
neurons, so copying their behavior and functionality should be
the solution.
Biological inspiration
Dendrites

Soma (cell body)
Axon
Biological inspiration
axon

dendrites

synapses

The information transmission happens at the synapses.
Biological inspiration
The spikes travelling along the axon of the pre-synaptic
neuron trigger the release of neurotransmitter substances
at the synapse.
The neurotransmitters cause excitation or inhibition in the
dendrite of the post-synaptic neuron.
The integration of the excitatory and inhibitory signals
may produce spikes in the post-synaptic neuron.
The contribution of the signals depends on the strength of
the synaptic connection.
Artificial neurons
Neurons work by processing information. They receive and
provide information in form of spikes.
x1
w1

x2
Inputs

w2

x3
…
xn-1
xn

n

z = ∑ wi xi ; y = H ( z )
i =1

..

w3
. wn-1
wn

The McCullogh-Pitts model

Output
y
Artificial neurons
The McCullogh-Pitts model:
• spikes are interpreted as spike rates;
• synaptic strength are translated as synaptic weights;
• excitation means positive product between the
incoming spike rate and the corresponding synaptic
weight;
• inhibition means negative product between the
incoming spike rate and the corresponding synaptic
weight;
Artificial neurons
Nonlinear generalization of the McCullogh-Pitts
neuron:

y = f ( x, w)
y is the neuron’s output, x is the vector of inputs, and w
is the vector of synaptic weights.
Examples:

y=

1
1+ e

y=e

T

−w x−a

|| x − w|| 2
−
2a 2

sigmoidal neuron

Gaussian neuron
Artificial neural networks

Inputs

Output

An artificial neural network is composed of many artificial
neurons that are linked together according to a specific
network architecture. The objective of the neural network
is to transform the inputs into meaningful outputs.
Artificial neural networks
Tasks to be solved by artificial neural networks:
• controlling the movements of a robot based on selfperception and other information (e.g., visual
information);
• deciding the category of potential food items (e.g.,
edible or non-edible) in an artificial world;
• recognizing a visual object (e.g., a familiar face);
• predicting where a moving object goes, when a robot
wants to catch it.
Learning in biological systems
Learning = learning by adaptation

The young animal learns that the green fruits are sour,
while the yellowish/reddish ones are sweet. The learning
happens by adapting the fruit picking behavior.

At the neural level the learning happens by changing of the
synaptic strengths, eliminating some synapses, and
building new ones.
Learning as optimisation
The objective of adapting the responses on the basis of the
information received from the environment is to achieve a
better state. E.g., the animal likes to eat many energy rich,
juicy fruits that make its stomach full, and makes it feel
happy.

In other words, the objective of learning in biological
organisms is to optimise the amount of available resources,
happiness, or in general to achieve a closer to optimal state.
Learning in biological neural
networks
The learning rules of Hebb:
• synchronous activation increases the synaptic strength;
• asynchronous activation decreases the synaptic
strength.
These rules fit with energy minimization principles.
Maintaining synaptic strength needs energy, it should be
maintained at those places where it is needed, and it
shouldn’t be maintained at places where it’s not needed.
Learning principle for
artificial neural networks
ENERGY MINIMIZATION

We need an appropriate definition of energy for artificial
neural networks, and having that we can use
mathematical optimisation techniques to find how to
change the weights of the synaptic connections between
neurons.

ENERGY = measure of task performance error
Neural network mathematics
Inputs

Output

1
 y1  2
2
 1  y1 = f ( y 1 , w12 )
 y3 
 2
2
3
y 1 = f ( x2 , w1 ) y 1 =  y 2  2
2
1
2
2
2
y =  y3  yOut = f ( y , w1 )
 1  y 2 = f ( y , w2 )
1
1
 2
y3 = f ( x3 , w3 )
 y3  2
 y3 
1
2
1  y 3 = f ( y , w3 )
 
y
1
1
 4
y 4 = f ( x4 , w4 )

1
1
y1 = f ( x1 , w1 )
Neural network mathematics
Neural network: input / output transformation

y out = F ( x, W )
W is the matrix of all weight vectors.
MLP neural networks
MLP = multi-layer perceptron
Perceptron:

yout = wT x

x

yout

MLP neural network:
1

y1 =
k

− w1 kT x − a1
k

1+ e
1
1
y 1 = ( y1 , y 1 , y3 ) T
2

, k = 1,2,3

1

2
yk =

2
− w 2 kT y 1 − a k

1+ e
2
y 2 = ( y12 , y 2 ) T
2

, k = 1,2

3 2
y out = ∑ wk y k = w3T y 2
k =1

x

yout
RBF neural networks
RBF = radial basis function

r ( x) = r (|| x − c ||)
Example:

4

f ( x) = e

2
y out = ∑ wk ⋅ e
k =1

|| x − w1,k || 2
−
2( ak ) 2

|| x − w|| 2
−
2a 2

Gaussian RBF

x

yout
Neural network tasks
• control
• classification
• prediction
• approximation

These can be reformulated
in general as
FUNCTION
APPROXIMATION
tasks.

Approximation: given a set of values of a function g(x)
build a neural network that approximates the g(x) values
for any input x.
Neural network approximation
Task specification:

Data: set of value pairs: (xt, yt), yt=g(xt) + zt; zt is random
measurement noise.

Objective: find a neural network that represents the input /
output transformation (a function) F(x,W) such that
F(x,W) approximates g(x) for every x
Learning to approximate
Error measure:
1
E=
N

N

∑ ( F ( xt ; W ) − y t ) 2
t =1

Rule for changing the synaptic weights:

∂E
∆wi = −c ⋅
(W )
j
∂wi
j

wi

j , new

= wi + ∆wi
j

j

c is the learning parameter (usually a constant)
Learning with a perceptron
Perceptron:

yout = wT x

1
2
N
Data: ( x , y1 ), ( x , y 2 ),..., ( x , y N )

E (t ) = ( y (t ) out − yt ) 2 = ( w(t ) T x t − yt ) 2
Error:
Learning:

∂ ( w(t )T x t − yt ) 2
∂E (t )
wi (t + 1) = wi (t ) − c ⋅
= wi (t ) − c ⋅
∂wi
∂wi
wi (t + 1) = wi (t ) − c ⋅ ( w(t )T x t − yt ) ⋅ xit
m

w(t ) x = ∑ w j (t ) ⋅ x tj
T

j =1

A perceptron is able to learn a linear function.
Learning with RBF neural
networks
M

2
RBF neural network: y out = F ( x, W ) = ∑ wk ⋅ e

|| x − w1,k || 2
−
2( ak ) 2

k =1

( x1 , y1 ), ( x 2 , y 2 ),..., ( x N , y N )
Data:
M

2
Error: E (t ) = ( y (t ) out − yt ) = (∑ wk (t ) ⋅ e
2

|| x t − w1,k || 2
−
2 ( ak ) 2

k =1

Learning:

− yt ) 2

∂E (t )
w (t + 1) = w (t ) − c ⋅
∂wi2
2
i

2
i

∂E (t )
= 2 ⋅ ( F ( x t , W (t )) − yt ) ⋅ e
∂wi2

−

|| x t − w1,i || 2
2 ( ai ) 2

Only the synaptic weights of the output neuron are modified.
An RBF neural network learns a nonlinear function.
Learning with MLP neural
networks
y1 =
k

− w1 kT x − a1
k

1+ e
1
y 1 = ( y1 ,..., y 1 1 ) T
M

MLP neural network:
with p layers
x

1

yout

2
yk =

, k = 1,..., M 1

1
2
− w 2 kT y 1 − a k

1+ e
2
y 2 = ( y12 ,..., y M 2 ) T

, k = 1,..., M 2

...
y out = F ( x;W ) = w pT y p −1

1 2 … p-1 p
( x1 , y1 ), ( x 2 , y 2 ),..., ( x N , y N )
Data:

Error: E (t ) = ( y (t ) out − yt ) 2 = ( F ( x t ;W ) − yt ) 2
It is very complicated to calculate the weight changes.
Learning with backpropagation
Solution of the complicated learning:
• calculate first the changes for the synaptic weights
of the output neuron;
• calculate the changes backward starting from layer
p-1, and propagate backward the local error terms.
The method is still relatively complicated but it
is much simpler than the original optimisation
problem.
Learning with general optimisation
In general it is enough to have a single layer of nonlinear
neurons in a neural network in order to learn to
approximate a nonlinear function.
In such case general optimisation may be applied without
too much difficulty.
Example: an MLP neural network with a single hidden layer:
M

y out = F ( x;W ) = ∑ w ⋅
k =1

2
k

1
1+ e

− w1,kT x − a k
Learning with general optimisation
Synaptic weight change rules for the output neuron:
∂E (t )
w (t + 1) = w (t ) − c ⋅
∂wi2
2
i

2
i

∂E (t )
1
= 2 ⋅ ( F ( x t , W (t )) − yt ) ⋅
1,iT t
∂wi2
1 + e − w x − ai

Synaptic weight change rules for the neurons of the
hidden layer: w (t + 1) = w (t ) − c ⋅ ∂E (t )
∂w
1, i
j

1, i
j

1, i
j

∂E (t )
∂
= 2 ⋅ ( F ( x t , W (t )) − yt ) ⋅ 1,i
∂w1j,i
∂w j
∂
∂w1j,i

1,iT

t

e − w x − ai

=

1,iT t
 1 + e − w x − ai

1


1,iT t

 1 + e − w x − ai

(

1


1,iT t

 1 + e − w x − ai

)

(

∂
− w1,iT x t − ai = − x tj
1, i
∂w j

)

2

⋅






(

∂
− w1,iT x t − ai
1, i
∂w j

1,iT

w (t + 1) = w (t ) − c ⋅ 2 ⋅ ( F ( x , W (t )) − yt ) ⋅
1, i
j

1, i
j

t

e−w

(1 + e

x t − ai

1,iT

−w

t

x − ai

)

2

)

⋅ (− x tj )
New methods for learning with
neural networks
Bayesian learning:
the distribution of the neural network
parameters is learnt

Support vector learning:
the minimal representative subset of the
available data is used to calculate the synaptic
weights of the neurons
Summary
• Artificial neural networks are inspired by the learning
processes that take place in biological systems.
• Artificial neurons and neural networks try to imitate the
working mechanisms of their biological counterparts.
• Learning can be perceived as an optimisation process.
• Biological neural learning happens by the modification
of the synaptic strength. Artificial neural networks learn
in the same way.
• The synapse strength modification rules for artificial
neural networks can be derived by applying
mathematical optimisation methods.
Summary
• Learning tasks of artificial neural networks can be
reformulated as function approximation tasks.
• Neural networks can be considered as nonlinear function
approximating tools (i.e., linear combinations of nonlinear
basis functions), where the parameters of the networks
should be found by applying optimisation methods.
• The optimisation is done with respect to the approximation
error measure.
• In general it is enough to have a single hidden layer neural
network (MLP, RBF or other) to learn the approximation of
a nonlinear function. In such cases general optimisation can
be applied to find the change rules for the synaptic weights.

More Related Content

What's hot

Neural network
Neural networkNeural network
Neural networkSilicon
 
Neural networks
Neural networksNeural networks
Neural networksSlideshare
 
Neural network
Neural networkNeural network
Neural networkmarada0033
 
Artificial Neural Network (draft)
Artificial Neural Network (draft)Artificial Neural Network (draft)
Artificial Neural Network (draft)James Boulie
 
Artificial neural networks
Artificial neural networksArtificial neural networks
Artificial neural networksmadhu sudhakar
 
Artificial Neural Networks Lect2: Neurobiology & Architectures of ANNS
Artificial Neural Networks Lect2: Neurobiology & Architectures of ANNSArtificial Neural Networks Lect2: Neurobiology & Architectures of ANNS
Artificial Neural Networks Lect2: Neurobiology & Architectures of ANNSMohammed Bennamoun
 
Artificial Neural Network
Artificial Neural NetworkArtificial Neural Network
Artificial Neural NetworkKnoldus Inc.
 
Neural network & its applications
Neural network & its applications Neural network & its applications
Neural network & its applications Ahmed_hashmi
 
Application of artificial_neural_network
Application of artificial_neural_networkApplication of artificial_neural_network
Application of artificial_neural_networkgabo GAG
 
(Artificial) Neural Network
(Artificial) Neural Network(Artificial) Neural Network
(Artificial) Neural NetworkPutri Wikie
 
Secondary structure prediction
Secondary structure predictionSecondary structure prediction
Secondary structure predictionsamantlalit
 
Artificial neural network for concrete mix design
Artificial neural network for concrete mix designArtificial neural network for concrete mix design
Artificial neural network for concrete mix designMonjurul Shuvo
 
Artificial neural networks (2)
Artificial neural networks (2)Artificial neural networks (2)
Artificial neural networks (2)sai anjaneya
 
Artificial Neural Networks Lect1: Introduction & neural computation
Artificial Neural Networks Lect1: Introduction & neural computationArtificial Neural Networks Lect1: Introduction & neural computation
Artificial Neural Networks Lect1: Introduction & neural computationMohammed Bennamoun
 
Introduction to Neural networks (under graduate course) Lecture 3 of 9
Introduction to Neural networks (under graduate course) Lecture 3 of 9Introduction to Neural networks (under graduate course) Lecture 3 of 9
Introduction to Neural networks (under graduate course) Lecture 3 of 9Randa Elanwar
 
Artificial neural networks
Artificial neural networksArtificial neural networks
Artificial neural networksernj
 

What's hot (20)

Neural network
Neural networkNeural network
Neural network
 
Neural networks
Neural networksNeural networks
Neural networks
 
Neural network
Neural networkNeural network
Neural network
 
Artifical Neural Network
Artifical Neural NetworkArtifical Neural Network
Artifical Neural Network
 
Artificial Neural Network (draft)
Artificial Neural Network (draft)Artificial Neural Network (draft)
Artificial Neural Network (draft)
 
Artificial neural networks
Artificial neural networksArtificial neural networks
Artificial neural networks
 
Artificial Neuron network
Artificial Neuron network Artificial Neuron network
Artificial Neuron network
 
Neural network and mlp
Neural network and mlpNeural network and mlp
Neural network and mlp
 
Artificial Neural Networks Lect2: Neurobiology & Architectures of ANNS
Artificial Neural Networks Lect2: Neurobiology & Architectures of ANNSArtificial Neural Networks Lect2: Neurobiology & Architectures of ANNS
Artificial Neural Networks Lect2: Neurobiology & Architectures of ANNS
 
Artificial Neural Network
Artificial Neural NetworkArtificial Neural Network
Artificial Neural Network
 
Neural network & its applications
Neural network & its applications Neural network & its applications
Neural network & its applications
 
Application of artificial_neural_network
Application of artificial_neural_networkApplication of artificial_neural_network
Application of artificial_neural_network
 
(Artificial) Neural Network
(Artificial) Neural Network(Artificial) Neural Network
(Artificial) Neural Network
 
Secondary structure prediction
Secondary structure predictionSecondary structure prediction
Secondary structure prediction
 
Artificial neural network for concrete mix design
Artificial neural network for concrete mix designArtificial neural network for concrete mix design
Artificial neural network for concrete mix design
 
Artificial neural networks (2)
Artificial neural networks (2)Artificial neural networks (2)
Artificial neural networks (2)
 
Neural Networks
Neural NetworksNeural Networks
Neural Networks
 
Artificial Neural Networks Lect1: Introduction & neural computation
Artificial Neural Networks Lect1: Introduction & neural computationArtificial Neural Networks Lect1: Introduction & neural computation
Artificial Neural Networks Lect1: Introduction & neural computation
 
Introduction to Neural networks (under graduate course) Lecture 3 of 9
Introduction to Neural networks (under graduate course) Lecture 3 of 9Introduction to Neural networks (under graduate course) Lecture 3 of 9
Introduction to Neural networks (under graduate course) Lecture 3 of 9
 
Artificial neural networks
Artificial neural networksArtificial neural networks
Artificial neural networks
 

Viewers also liked

Artificial Neural Network Implementation on FPGA – a Modular Approach
Artificial Neural Network Implementation on FPGA – a Modular ApproachArtificial Neural Network Implementation on FPGA – a Modular Approach
Artificial Neural Network Implementation on FPGA – a Modular ApproachRoee Levy
 
Artificial Neural Network Abstract
Artificial Neural Network AbstractArtificial Neural Network Abstract
Artificial Neural Network AbstractAnjali Agrawal
 
what is neural network....???
what is neural network....???what is neural network....???
what is neural network....???Adii Shah
 
Use of artificial neural network in pattern recognition
Use of artificial neural network in pattern recognitionUse of artificial neural network in pattern recognition
Use of artificial neural network in pattern recognitionkamalsrit
 
Artificial Neural Network
Artificial Neural NetworkArtificial Neural Network
Artificial Neural NetworkDessy Amirudin
 
Artificial neural network for misuse detection
Artificial neural network for misuse detectionArtificial neural network for misuse detection
Artificial neural network for misuse detectionLikan Patra
 
lecture07.ppt
lecture07.pptlecture07.ppt
lecture07.pptbutest
 
Back propagation
Back propagationBack propagation
Back propagationNagarajan
 
neural network
neural networkneural network
neural networkSTUDENT
 
Artificial neural networks
Artificial neural networksArtificial neural networks
Artificial neural networksstellajoseph
 
Artificial Intelligence
Artificial IntelligenceArtificial Intelligence
Artificial Intelligenceu053675
 
artificial intelligence
artificial intelligenceartificial intelligence
artificial intelligencevallibhargavi
 
Data Science - Part VIII - Artifical Neural Network
Data Science - Part VIII -  Artifical Neural NetworkData Science - Part VIII -  Artifical Neural Network
Data Science - Part VIII - Artifical Neural NetworkDerek Kane
 

Viewers also liked (14)

Artificial Neural Network Implementation on FPGA – a Modular Approach
Artificial Neural Network Implementation on FPGA – a Modular ApproachArtificial Neural Network Implementation on FPGA – a Modular Approach
Artificial Neural Network Implementation on FPGA – a Modular Approach
 
Abstract
AbstractAbstract
Abstract
 
Artificial Neural Network Abstract
Artificial Neural Network AbstractArtificial Neural Network Abstract
Artificial Neural Network Abstract
 
what is neural network....???
what is neural network....???what is neural network....???
what is neural network....???
 
Use of artificial neural network in pattern recognition
Use of artificial neural network in pattern recognitionUse of artificial neural network in pattern recognition
Use of artificial neural network in pattern recognition
 
Artificial Neural Network
Artificial Neural NetworkArtificial Neural Network
Artificial Neural Network
 
Artificial neural network for misuse detection
Artificial neural network for misuse detectionArtificial neural network for misuse detection
Artificial neural network for misuse detection
 
lecture07.ppt
lecture07.pptlecture07.ppt
lecture07.ppt
 
Back propagation
Back propagationBack propagation
Back propagation
 
neural network
neural networkneural network
neural network
 
Artificial neural networks
Artificial neural networksArtificial neural networks
Artificial neural networks
 
Artificial Intelligence
Artificial IntelligenceArtificial Intelligence
Artificial Intelligence
 
artificial intelligence
artificial intelligenceartificial intelligence
artificial intelligence
 
Data Science - Part VIII - Artifical Neural Network
Data Science - Part VIII -  Artifical Neural NetworkData Science - Part VIII -  Artifical Neural Network
Data Science - Part VIII - Artifical Neural Network
 

Similar to Artificial neural networks

JAISTサマースクール2016「脳を知るための理論」講義04 Neural Networks and Neuroscience
JAISTサマースクール2016「脳を知るための理論」講義04 Neural Networks and Neuroscience JAISTサマースクール2016「脳を知るための理論」講義04 Neural Networks and Neuroscience
JAISTサマースクール2016「脳を知るための理論」講義04 Neural Networks and Neuroscience hirokazutanaka
 
Artificial Neural Network
Artificial Neural Network Artificial Neural Network
Artificial Neural Network Iman Ardekani
 
Artificial Neural Network seminar presentation using ppt.
Artificial Neural Network seminar presentation using ppt.Artificial Neural Network seminar presentation using ppt.
Artificial Neural Network seminar presentation using ppt.Mohd Faiz
 
Annintro
AnnintroAnnintro
Annintroairsrch
 
SOFT COMPUTERING TECHNICS -Unit 1
SOFT COMPUTERING TECHNICS -Unit 1SOFT COMPUTERING TECHNICS -Unit 1
SOFT COMPUTERING TECHNICS -Unit 1sravanthi computers
 
Artificial neural network
Artificial neural networkArtificial neural network
Artificial neural networkIldar Nurgaliev
 
Backpropagation in Convolutional Neural Network
Backpropagation in Convolutional Neural NetworkBackpropagation in Convolutional Neural Network
Backpropagation in Convolutional Neural NetworkHiroshi Kuwajima
 
latest TYPES OF NEURAL NETWORKS (2).pptx
latest TYPES OF NEURAL NETWORKS (2).pptxlatest TYPES OF NEURAL NETWORKS (2).pptx
latest TYPES OF NEURAL NETWORKS (2).pptxMdMahfoozAlam5
 
ACUMENS ON NEURAL NET AKG 20 7 23.pptx
ACUMENS ON NEURAL NET AKG 20 7 23.pptxACUMENS ON NEURAL NET AKG 20 7 23.pptx
ACUMENS ON NEURAL NET AKG 20 7 23.pptxgnans Kgnanshek
 
Classification using perceptron.pptx
Classification using perceptron.pptxClassification using perceptron.pptx
Classification using perceptron.pptxsomeyamohsen3
 
Artificial Neural Network
Artificial Neural NetworkArtificial Neural Network
Artificial Neural Networkssuserab4f3e
 

Similar to Artificial neural networks (20)

Annintro
AnnintroAnnintro
Annintro
 
JAISTサマースクール2016「脳を知るための理論」講義04 Neural Networks and Neuroscience
JAISTサマースクール2016「脳を知るための理論」講義04 Neural Networks and Neuroscience JAISTサマースクール2016「脳を知るための理論」講義04 Neural Networks and Neuroscience
JAISTサマースクール2016「脳を知るための理論」講義04 Neural Networks and Neuroscience
 
Artificial Neural Network
Artificial Neural Network Artificial Neural Network
Artificial Neural Network
 
5.n nmodels i
5.n nmodels i5.n nmodels i
5.n nmodels i
 
Artificial Neural Network seminar presentation using ppt.
Artificial Neural Network seminar presentation using ppt.Artificial Neural Network seminar presentation using ppt.
Artificial Neural Network seminar presentation using ppt.
 
Annintro
AnnintroAnnintro
Annintro
 
SOFT COMPUTERING TECHNICS -Unit 1
SOFT COMPUTERING TECHNICS -Unit 1SOFT COMPUTERING TECHNICS -Unit 1
SOFT COMPUTERING TECHNICS -Unit 1
 
annintro.ppt
annintro.pptannintro.ppt
annintro.ppt
 
Artificial neural network
Artificial neural networkArtificial neural network
Artificial neural network
 
Backpropagation in Convolutional Neural Network
Backpropagation in Convolutional Neural NetworkBackpropagation in Convolutional Neural Network
Backpropagation in Convolutional Neural Network
 
10-Perceptron.pdf
10-Perceptron.pdf10-Perceptron.pdf
10-Perceptron.pdf
 
latest TYPES OF NEURAL NETWORKS (2).pptx
latest TYPES OF NEURAL NETWORKS (2).pptxlatest TYPES OF NEURAL NETWORKS (2).pptx
latest TYPES OF NEURAL NETWORKS (2).pptx
 
Max net
Max netMax net
Max net
 
tutorial.ppt
tutorial.ppttutorial.ppt
tutorial.ppt
 
6
66
6
 
ACUMENS ON NEURAL NET AKG 20 7 23.pptx
ACUMENS ON NEURAL NET AKG 20 7 23.pptxACUMENS ON NEURAL NET AKG 20 7 23.pptx
ACUMENS ON NEURAL NET AKG 20 7 23.pptx
 
Classification using perceptron.pptx
Classification using perceptron.pptxClassification using perceptron.pptx
Classification using perceptron.pptx
 
071bct537 lab4
071bct537 lab4071bct537 lab4
071bct537 lab4
 
19_Learning.ppt
19_Learning.ppt19_Learning.ppt
19_Learning.ppt
 
Artificial Neural Network
Artificial Neural NetworkArtificial Neural Network
Artificial Neural Network
 

More from Institute of Technology Telkom

More from Institute of Technology Telkom (20)

Econopysics
EconopysicsEconopysics
Econopysics
 
Science and religion 100622120615-phpapp01
Science and religion 100622120615-phpapp01Science and religion 100622120615-phpapp01
Science and religion 100622120615-phpapp01
 
Konvergensi sains dan_spiritualitas
Konvergensi sains dan_spiritualitasKonvergensi sains dan_spiritualitas
Konvergensi sains dan_spiritualitas
 
Matematika arah kiblat mikrajuddin abdullah 2017
Matematika arah kiblat   mikrajuddin abdullah 2017Matematika arah kiblat   mikrajuddin abdullah 2017
Matematika arah kiblat mikrajuddin abdullah 2017
 
Iau solar effects 2005
Iau solar effects 2005Iau solar effects 2005
Iau solar effects 2005
 
Hfmsilri2jun14
Hfmsilri2jun14Hfmsilri2jun14
Hfmsilri2jun14
 
Fisika komputasi
Fisika komputasiFisika komputasi
Fisika komputasi
 
Computer Aided Process Planning
Computer Aided Process PlanningComputer Aided Process Planning
Computer Aided Process Planning
 
Archimedes
ArchimedesArchimedes
Archimedes
 
Web and text
Web and textWeb and text
Web and text
 
Web data mining
Web data miningWeb data mining
Web data mining
 
Time series Forecasting using svm
Time series Forecasting using  svmTime series Forecasting using  svm
Time series Forecasting using svm
 
Timeseries forecasting
Timeseries forecastingTimeseries forecasting
Timeseries forecasting
 
Fuzzy logic
Fuzzy logicFuzzy logic
Fuzzy logic
 
World population 1950--2050
World population 1950--2050World population 1950--2050
World population 1950--2050
 
neural networks
 neural networks neural networks
neural networks
 
002 ray modeling dynamic systems
002 ray modeling dynamic systems002 ray modeling dynamic systems
002 ray modeling dynamic systems
 
002 ray modeling dynamic systems
002 ray modeling dynamic systems002 ray modeling dynamic systems
002 ray modeling dynamic systems
 
System dynamics majors fair
System dynamics majors fairSystem dynamics majors fair
System dynamics majors fair
 
System dynamics math representation
System dynamics math representationSystem dynamics math representation
System dynamics math representation
 

Recently uploaded

Micro-Scholarship, What it is, How can it help me.pdf
Micro-Scholarship, What it is, How can it help me.pdfMicro-Scholarship, What it is, How can it help me.pdf
Micro-Scholarship, What it is, How can it help me.pdfPoh-Sun Goh
 
SOC 101 Demonstration of Learning Presentation
SOC 101 Demonstration of Learning PresentationSOC 101 Demonstration of Learning Presentation
SOC 101 Demonstration of Learning Presentationcamerronhm
 
80 ĐỀ THI THỬ TUYỂN SINH TIẾNG ANH VÀO 10 SỞ GD – ĐT THÀNH PHỐ HỒ CHÍ MINH NĂ...
80 ĐỀ THI THỬ TUYỂN SINH TIẾNG ANH VÀO 10 SỞ GD – ĐT THÀNH PHỐ HỒ CHÍ MINH NĂ...80 ĐỀ THI THỬ TUYỂN SINH TIẾNG ANH VÀO 10 SỞ GD – ĐT THÀNH PHỐ HỒ CHÍ MINH NĂ...
80 ĐỀ THI THỬ TUYỂN SINH TIẾNG ANH VÀO 10 SỞ GD – ĐT THÀNH PHỐ HỒ CHÍ MINH NĂ...Nguyen Thanh Tu Collection
 
Interdisciplinary_Insights_Data_Collection_Methods.pptx
Interdisciplinary_Insights_Data_Collection_Methods.pptxInterdisciplinary_Insights_Data_Collection_Methods.pptx
Interdisciplinary_Insights_Data_Collection_Methods.pptxPooja Bhuva
 
Unit 3 Emotional Intelligence and Spiritual Intelligence.pdf
Unit 3 Emotional Intelligence and Spiritual Intelligence.pdfUnit 3 Emotional Intelligence and Spiritual Intelligence.pdf
Unit 3 Emotional Intelligence and Spiritual Intelligence.pdfDr Vijay Vishwakarma
 
ICT Role in 21st Century Education & its Challenges.pptx
ICT Role in 21st Century Education & its Challenges.pptxICT Role in 21st Century Education & its Challenges.pptx
ICT Role in 21st Century Education & its Challenges.pptxAreebaZafar22
 
Fostering Friendships - Enhancing Social Bonds in the Classroom
Fostering Friendships - Enhancing Social Bonds  in the ClassroomFostering Friendships - Enhancing Social Bonds  in the Classroom
Fostering Friendships - Enhancing Social Bonds in the ClassroomPooky Knightsmith
 
On_Translating_a_Tamil_Poem_by_A_K_Ramanujan.pptx
On_Translating_a_Tamil_Poem_by_A_K_Ramanujan.pptxOn_Translating_a_Tamil_Poem_by_A_K_Ramanujan.pptx
On_Translating_a_Tamil_Poem_by_A_K_Ramanujan.pptxPooja Bhuva
 
2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptx
2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptx2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptx
2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptxMaritesTamaniVerdade
 
Key note speaker Neum_Admir Softic_ENG.pdf
Key note speaker Neum_Admir Softic_ENG.pdfKey note speaker Neum_Admir Softic_ENG.pdf
Key note speaker Neum_Admir Softic_ENG.pdfAdmir Softic
 
This PowerPoint helps students to consider the concept of infinity.
This PowerPoint helps students to consider the concept of infinity.This PowerPoint helps students to consider the concept of infinity.
This PowerPoint helps students to consider the concept of infinity.christianmathematics
 
How to Add New Custom Addons Path in Odoo 17
How to Add New Custom Addons Path in Odoo 17How to Add New Custom Addons Path in Odoo 17
How to Add New Custom Addons Path in Odoo 17Celine George
 
Plant propagation: Sexual and Asexual propapagation.pptx
Plant propagation: Sexual and Asexual propapagation.pptxPlant propagation: Sexual and Asexual propapagation.pptx
Plant propagation: Sexual and Asexual propapagation.pptxUmeshTimilsina1
 
Kodo Millet PPT made by Ghanshyam bairwa college of Agriculture kumher bhara...
Kodo Millet  PPT made by Ghanshyam bairwa college of Agriculture kumher bhara...Kodo Millet  PPT made by Ghanshyam bairwa college of Agriculture kumher bhara...
Kodo Millet PPT made by Ghanshyam bairwa college of Agriculture kumher bhara...pradhanghanshyam7136
 
HMCS Vancouver Pre-Deployment Brief - May 2024 (Web Version).pptx
HMCS Vancouver Pre-Deployment Brief - May 2024 (Web Version).pptxHMCS Vancouver Pre-Deployment Brief - May 2024 (Web Version).pptx
HMCS Vancouver Pre-Deployment Brief - May 2024 (Web Version).pptxmarlenawright1
 
Holdier Curriculum Vitae (April 2024).pdf
Holdier Curriculum Vitae (April 2024).pdfHoldier Curriculum Vitae (April 2024).pdf
Holdier Curriculum Vitae (April 2024).pdfagholdier
 
Basic Civil Engineering first year Notes- Chapter 4 Building.pptx
Basic Civil Engineering first year Notes- Chapter 4 Building.pptxBasic Civil Engineering first year Notes- Chapter 4 Building.pptx
Basic Civil Engineering first year Notes- Chapter 4 Building.pptxDenish Jangid
 
How to setup Pycharm environment for Odoo 17.pptx
How to setup Pycharm environment for Odoo 17.pptxHow to setup Pycharm environment for Odoo 17.pptx
How to setup Pycharm environment for Odoo 17.pptxCeline George
 
ICT role in 21st century education and it's challenges.
ICT role in 21st century education and it's challenges.ICT role in 21st century education and it's challenges.
ICT role in 21st century education and it's challenges.MaryamAhmad92
 

Recently uploaded (20)

Micro-Scholarship, What it is, How can it help me.pdf
Micro-Scholarship, What it is, How can it help me.pdfMicro-Scholarship, What it is, How can it help me.pdf
Micro-Scholarship, What it is, How can it help me.pdf
 
SOC 101 Demonstration of Learning Presentation
SOC 101 Demonstration of Learning PresentationSOC 101 Demonstration of Learning Presentation
SOC 101 Demonstration of Learning Presentation
 
80 ĐỀ THI THỬ TUYỂN SINH TIẾNG ANH VÀO 10 SỞ GD – ĐT THÀNH PHỐ HỒ CHÍ MINH NĂ...
80 ĐỀ THI THỬ TUYỂN SINH TIẾNG ANH VÀO 10 SỞ GD – ĐT THÀNH PHỐ HỒ CHÍ MINH NĂ...80 ĐỀ THI THỬ TUYỂN SINH TIẾNG ANH VÀO 10 SỞ GD – ĐT THÀNH PHỐ HỒ CHÍ MINH NĂ...
80 ĐỀ THI THỬ TUYỂN SINH TIẾNG ANH VÀO 10 SỞ GD – ĐT THÀNH PHỐ HỒ CHÍ MINH NĂ...
 
Interdisciplinary_Insights_Data_Collection_Methods.pptx
Interdisciplinary_Insights_Data_Collection_Methods.pptxInterdisciplinary_Insights_Data_Collection_Methods.pptx
Interdisciplinary_Insights_Data_Collection_Methods.pptx
 
Unit 3 Emotional Intelligence and Spiritual Intelligence.pdf
Unit 3 Emotional Intelligence and Spiritual Intelligence.pdfUnit 3 Emotional Intelligence and Spiritual Intelligence.pdf
Unit 3 Emotional Intelligence and Spiritual Intelligence.pdf
 
ICT Role in 21st Century Education & its Challenges.pptx
ICT Role in 21st Century Education & its Challenges.pptxICT Role in 21st Century Education & its Challenges.pptx
ICT Role in 21st Century Education & its Challenges.pptx
 
Fostering Friendships - Enhancing Social Bonds in the Classroom
Fostering Friendships - Enhancing Social Bonds  in the ClassroomFostering Friendships - Enhancing Social Bonds  in the Classroom
Fostering Friendships - Enhancing Social Bonds in the Classroom
 
On_Translating_a_Tamil_Poem_by_A_K_Ramanujan.pptx
On_Translating_a_Tamil_Poem_by_A_K_Ramanujan.pptxOn_Translating_a_Tamil_Poem_by_A_K_Ramanujan.pptx
On_Translating_a_Tamil_Poem_by_A_K_Ramanujan.pptx
 
2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptx
2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptx2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptx
2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptx
 
Mehran University Newsletter Vol-X, Issue-I, 2024
Mehran University Newsletter Vol-X, Issue-I, 2024Mehran University Newsletter Vol-X, Issue-I, 2024
Mehran University Newsletter Vol-X, Issue-I, 2024
 
Key note speaker Neum_Admir Softic_ENG.pdf
Key note speaker Neum_Admir Softic_ENG.pdfKey note speaker Neum_Admir Softic_ENG.pdf
Key note speaker Neum_Admir Softic_ENG.pdf
 
This PowerPoint helps students to consider the concept of infinity.
This PowerPoint helps students to consider the concept of infinity.This PowerPoint helps students to consider the concept of infinity.
This PowerPoint helps students to consider the concept of infinity.
 
How to Add New Custom Addons Path in Odoo 17
How to Add New Custom Addons Path in Odoo 17How to Add New Custom Addons Path in Odoo 17
How to Add New Custom Addons Path in Odoo 17
 
Plant propagation: Sexual and Asexual propapagation.pptx
Plant propagation: Sexual and Asexual propapagation.pptxPlant propagation: Sexual and Asexual propapagation.pptx
Plant propagation: Sexual and Asexual propapagation.pptx
 
Kodo Millet PPT made by Ghanshyam bairwa college of Agriculture kumher bhara...
Kodo Millet  PPT made by Ghanshyam bairwa college of Agriculture kumher bhara...Kodo Millet  PPT made by Ghanshyam bairwa college of Agriculture kumher bhara...
Kodo Millet PPT made by Ghanshyam bairwa college of Agriculture kumher bhara...
 
HMCS Vancouver Pre-Deployment Brief - May 2024 (Web Version).pptx
HMCS Vancouver Pre-Deployment Brief - May 2024 (Web Version).pptxHMCS Vancouver Pre-Deployment Brief - May 2024 (Web Version).pptx
HMCS Vancouver Pre-Deployment Brief - May 2024 (Web Version).pptx
 
Holdier Curriculum Vitae (April 2024).pdf
Holdier Curriculum Vitae (April 2024).pdfHoldier Curriculum Vitae (April 2024).pdf
Holdier Curriculum Vitae (April 2024).pdf
 
Basic Civil Engineering first year Notes- Chapter 4 Building.pptx
Basic Civil Engineering first year Notes- Chapter 4 Building.pptxBasic Civil Engineering first year Notes- Chapter 4 Building.pptx
Basic Civil Engineering first year Notes- Chapter 4 Building.pptx
 
How to setup Pycharm environment for Odoo 17.pptx
How to setup Pycharm environment for Odoo 17.pptxHow to setup Pycharm environment for Odoo 17.pptx
How to setup Pycharm environment for Odoo 17.pptx
 
ICT role in 21st century education and it's challenges.
ICT role in 21st century education and it's challenges.ICT role in 21st century education and it's challenges.
ICT role in 21st century education and it's challenges.
 

Artificial neural networks

  • 1. Artificial Neural Networks - Introduction Peter Andras peter.andras@ncl.ac.uk
  • 2. Overview 1. Biological inspiration 2. Artificial neurons and neural networks 3. Learning processes 4. Learning with artificial neural networks
  • 3. Biological inspiration Animals are able to react adaptively to changes in their external and internal environment, and they use their nervous system to perform these behaviours. An appropriate model/simulation of the nervous system should be able to produce similar responses and behaviours in artificial systems. The nervous system is build by relatively simple units, the neurons, so copying their behavior and functionality should be the solution.
  • 6. Biological inspiration The spikes travelling along the axon of the pre-synaptic neuron trigger the release of neurotransmitter substances at the synapse. The neurotransmitters cause excitation or inhibition in the dendrite of the post-synaptic neuron. The integration of the excitatory and inhibitory signals may produce spikes in the post-synaptic neuron. The contribution of the signals depends on the strength of the synaptic connection.
  • 7. Artificial neurons Neurons work by processing information. They receive and provide information in form of spikes. x1 w1 x2 Inputs w2 x3 … xn-1 xn n z = ∑ wi xi ; y = H ( z ) i =1 .. w3 . wn-1 wn The McCullogh-Pitts model Output y
  • 8. Artificial neurons The McCullogh-Pitts model: • spikes are interpreted as spike rates; • synaptic strength are translated as synaptic weights; • excitation means positive product between the incoming spike rate and the corresponding synaptic weight; • inhibition means negative product between the incoming spike rate and the corresponding synaptic weight;
  • 9. Artificial neurons Nonlinear generalization of the McCullogh-Pitts neuron: y = f ( x, w) y is the neuron’s output, x is the vector of inputs, and w is the vector of synaptic weights. Examples: y= 1 1+ e y=e T −w x−a || x − w|| 2 − 2a 2 sigmoidal neuron Gaussian neuron
  • 10. Artificial neural networks Inputs Output An artificial neural network is composed of many artificial neurons that are linked together according to a specific network architecture. The objective of the neural network is to transform the inputs into meaningful outputs.
  • 11. Artificial neural networks Tasks to be solved by artificial neural networks: • controlling the movements of a robot based on selfperception and other information (e.g., visual information); • deciding the category of potential food items (e.g., edible or non-edible) in an artificial world; • recognizing a visual object (e.g., a familiar face); • predicting where a moving object goes, when a robot wants to catch it.
  • 12. Learning in biological systems Learning = learning by adaptation The young animal learns that the green fruits are sour, while the yellowish/reddish ones are sweet. The learning happens by adapting the fruit picking behavior. At the neural level the learning happens by changing of the synaptic strengths, eliminating some synapses, and building new ones.
  • 13. Learning as optimisation The objective of adapting the responses on the basis of the information received from the environment is to achieve a better state. E.g., the animal likes to eat many energy rich, juicy fruits that make its stomach full, and makes it feel happy. In other words, the objective of learning in biological organisms is to optimise the amount of available resources, happiness, or in general to achieve a closer to optimal state.
  • 14. Learning in biological neural networks The learning rules of Hebb: • synchronous activation increases the synaptic strength; • asynchronous activation decreases the synaptic strength. These rules fit with energy minimization principles. Maintaining synaptic strength needs energy, it should be maintained at those places where it is needed, and it shouldn’t be maintained at places where it’s not needed.
  • 15. Learning principle for artificial neural networks ENERGY MINIMIZATION We need an appropriate definition of energy for artificial neural networks, and having that we can use mathematical optimisation techniques to find how to change the weights of the synaptic connections between neurons. ENERGY = measure of task performance error
  • 16. Neural network mathematics Inputs Output 1  y1  2 2  1  y1 = f ( y 1 , w12 )  y3   2 2 3 y 1 = f ( x2 , w1 ) y 1 =  y 2  2 2 1 2 2 2 y =  y3  yOut = f ( y , w1 )  1  y 2 = f ( y , w2 ) 1 1  2 y3 = f ( x3 , w3 )  y3  2  y3  1 2 1  y 3 = f ( y , w3 )   y 1 1  4 y 4 = f ( x4 , w4 ) 1 1 y1 = f ( x1 , w1 )
  • 17. Neural network mathematics Neural network: input / output transformation y out = F ( x, W ) W is the matrix of all weight vectors.
  • 18. MLP neural networks MLP = multi-layer perceptron Perceptron: yout = wT x x yout MLP neural network: 1 y1 = k − w1 kT x − a1 k 1+ e 1 1 y 1 = ( y1 , y 1 , y3 ) T 2 , k = 1,2,3 1 2 yk = 2 − w 2 kT y 1 − a k 1+ e 2 y 2 = ( y12 , y 2 ) T 2 , k = 1,2 3 2 y out = ∑ wk y k = w3T y 2 k =1 x yout
  • 19. RBF neural networks RBF = radial basis function r ( x) = r (|| x − c ||) Example: 4 f ( x) = e 2 y out = ∑ wk ⋅ e k =1 || x − w1,k || 2 − 2( ak ) 2 || x − w|| 2 − 2a 2 Gaussian RBF x yout
  • 20. Neural network tasks • control • classification • prediction • approximation These can be reformulated in general as FUNCTION APPROXIMATION tasks. Approximation: given a set of values of a function g(x) build a neural network that approximates the g(x) values for any input x.
  • 21. Neural network approximation Task specification: Data: set of value pairs: (xt, yt), yt=g(xt) + zt; zt is random measurement noise. Objective: find a neural network that represents the input / output transformation (a function) F(x,W) such that F(x,W) approximates g(x) for every x
  • 22. Learning to approximate Error measure: 1 E= N N ∑ ( F ( xt ; W ) − y t ) 2 t =1 Rule for changing the synaptic weights: ∂E ∆wi = −c ⋅ (W ) j ∂wi j wi j , new = wi + ∆wi j j c is the learning parameter (usually a constant)
  • 23. Learning with a perceptron Perceptron: yout = wT x 1 2 N Data: ( x , y1 ), ( x , y 2 ),..., ( x , y N ) E (t ) = ( y (t ) out − yt ) 2 = ( w(t ) T x t − yt ) 2 Error: Learning: ∂ ( w(t )T x t − yt ) 2 ∂E (t ) wi (t + 1) = wi (t ) − c ⋅ = wi (t ) − c ⋅ ∂wi ∂wi wi (t + 1) = wi (t ) − c ⋅ ( w(t )T x t − yt ) ⋅ xit m w(t ) x = ∑ w j (t ) ⋅ x tj T j =1 A perceptron is able to learn a linear function.
  • 24. Learning with RBF neural networks M 2 RBF neural network: y out = F ( x, W ) = ∑ wk ⋅ e || x − w1,k || 2 − 2( ak ) 2 k =1 ( x1 , y1 ), ( x 2 , y 2 ),..., ( x N , y N ) Data: M 2 Error: E (t ) = ( y (t ) out − yt ) = (∑ wk (t ) ⋅ e 2 || x t − w1,k || 2 − 2 ( ak ) 2 k =1 Learning: − yt ) 2 ∂E (t ) w (t + 1) = w (t ) − c ⋅ ∂wi2 2 i 2 i ∂E (t ) = 2 ⋅ ( F ( x t , W (t )) − yt ) ⋅ e ∂wi2 − || x t − w1,i || 2 2 ( ai ) 2 Only the synaptic weights of the output neuron are modified. An RBF neural network learns a nonlinear function.
  • 25. Learning with MLP neural networks y1 = k − w1 kT x − a1 k 1+ e 1 y 1 = ( y1 ,..., y 1 1 ) T M MLP neural network: with p layers x 1 yout 2 yk = , k = 1,..., M 1 1 2 − w 2 kT y 1 − a k 1+ e 2 y 2 = ( y12 ,..., y M 2 ) T , k = 1,..., M 2 ... y out = F ( x;W ) = w pT y p −1 1 2 … p-1 p ( x1 , y1 ), ( x 2 , y 2 ),..., ( x N , y N ) Data: Error: E (t ) = ( y (t ) out − yt ) 2 = ( F ( x t ;W ) − yt ) 2 It is very complicated to calculate the weight changes.
  • 26. Learning with backpropagation Solution of the complicated learning: • calculate first the changes for the synaptic weights of the output neuron; • calculate the changes backward starting from layer p-1, and propagate backward the local error terms. The method is still relatively complicated but it is much simpler than the original optimisation problem.
  • 27. Learning with general optimisation In general it is enough to have a single layer of nonlinear neurons in a neural network in order to learn to approximate a nonlinear function. In such case general optimisation may be applied without too much difficulty. Example: an MLP neural network with a single hidden layer: M y out = F ( x;W ) = ∑ w ⋅ k =1 2 k 1 1+ e − w1,kT x − a k
  • 28. Learning with general optimisation Synaptic weight change rules for the output neuron: ∂E (t ) w (t + 1) = w (t ) − c ⋅ ∂wi2 2 i 2 i ∂E (t ) 1 = 2 ⋅ ( F ( x t , W (t )) − yt ) ⋅ 1,iT t ∂wi2 1 + e − w x − ai Synaptic weight change rules for the neurons of the hidden layer: w (t + 1) = w (t ) − c ⋅ ∂E (t ) ∂w 1, i j 1, i j 1, i j ∂E (t ) ∂ = 2 ⋅ ( F ( x t , W (t )) − yt ) ⋅ 1,i ∂w1j,i ∂w j ∂ ∂w1j,i 1,iT t e − w x − ai  =  1,iT t  1 + e − w x − ai 1   1,iT t   1 + e − w x − ai ( 1   1,iT t   1 + e − w x − ai ) ( ∂ − w1,iT x t − ai = − x tj 1, i ∂w j ) 2 ⋅     ( ∂ − w1,iT x t − ai 1, i ∂w j 1,iT w (t + 1) = w (t ) − c ⋅ 2 ⋅ ( F ( x , W (t )) − yt ) ⋅ 1, i j 1, i j t e−w (1 + e x t − ai 1,iT −w t x − ai ) 2 ) ⋅ (− x tj )
  • 29. New methods for learning with neural networks Bayesian learning: the distribution of the neural network parameters is learnt Support vector learning: the minimal representative subset of the available data is used to calculate the synaptic weights of the neurons
  • 30. Summary • Artificial neural networks are inspired by the learning processes that take place in biological systems. • Artificial neurons and neural networks try to imitate the working mechanisms of their biological counterparts. • Learning can be perceived as an optimisation process. • Biological neural learning happens by the modification of the synaptic strength. Artificial neural networks learn in the same way. • The synapse strength modification rules for artificial neural networks can be derived by applying mathematical optimisation methods.
  • 31. Summary • Learning tasks of artificial neural networks can be reformulated as function approximation tasks. • Neural networks can be considered as nonlinear function approximating tools (i.e., linear combinations of nonlinear basis functions), where the parameters of the networks should be found by applying optimisation methods. • The optimisation is done with respect to the approximation error measure. • In general it is enough to have a single hidden layer neural network (MLP, RBF or other) to learn the approximation of a nonlinear function. In such cases general optimisation can be applied to find the change rules for the synaptic weights.