SlideShare ist ein Scribd-Unternehmen logo
1 von 73
Lecture 7  Artificial neural networks: Supervised learning ,[object Object],[object Object],[object Object],[object Object],[object Object],[object Object],[object Object],[object Object]
Introduction, or how the brain works Machine learning involves adaptive mechanisms that enable computers to learn from experience, learn by example and learn by analogy.  Learning capabilities can improve the performance of an intelligent system over time. The most popular approaches to machine learning are  artificial neural networks  and  genetic algorithms .  This lecture is dedicated to neural networks.
[object Object],[object Object]
[object Object],[object Object]
Biological neural network
[object Object],[object Object],[object Object]
[object Object],[object Object],[object Object]
Architecture of a typical artificial neural network
Analogy between biological and  artificial neural networks
The neuron as a simple computing element Diagram of a neuron
[object Object],[object Object],[object Object]
Activation functions of a neuron
Can a single neuron learn a task? ,[object Object],[object Object]
Single-layer two-input perceptron
The Perceptron ,[object Object],[object Object]
[object Object],[object Object],[object Object],[object Object]
Linear separability in the perceptrons
This is done by making small adjustments in the weights to reduce the difference between the actual and desired outputs of the perceptron.  The initial weights are randomly assigned, usually in the range [  0.5, 0.5], and then updated to obtain the output consistent with the training examples. How does the perceptron learn its classification tasks?
[object Object],[object Object],[object Object],[object Object]
The perceptron learning rule where  p  = 1, 2, 3, . . .    is the  learning rate , a positive constant less than unity. The perceptron learning rule was first proposed by Rosenblatt  in 1960.  Using this rule we can derive  the perceptron training algorithm for classification  tasks.
Step 1 : Initialisation Set initial weights  w 1 ,  w 2 ,…,  w n  and threshold    to random numbers in the range [  0.5, 0.5].  If the error,  e ( p ), is positive, we need to increase perceptron output  Y ( p ), but if it is negative, we need to decrease  Y ( p ). Perceptron’s training algorithm
Step 2 : Activation Activate the perceptron by applying inputs  x 1 ( p ),  x 2 ( p ),…,  x n ( p ) and desired output  Y d  ( p ).  Calculate the actual output at iteration  p  = 1 where  n  is the number of the perceptron inputs, and  step  is a step activation function. Perceptron’s training algorithm (continued)
Step 3 : Weight training Update the weights of the perceptron where   w i ( p ) is the weight correction at iteration  p . The weight correction is computed by the  delta rule : Step 4 : Iteration Increase iteration  p  by one, go back to  Step 2  and repeat the process until convergence. Perceptron’s training algorithm (continued)
Example of perceptron learning: the logical operation  AND
Two-dimensional plots of basic logical operations A perceptron can learn the operations  AND  and  OR , but not  Exclusive-OR .
Multilayer neural networks ,[object Object],[object Object],[object Object]
Multilayer perceptron with two hidden layers
What does the middle layer hide? ,[object Object],[object Object]
Back-propagation neural network ,[object Object],[object Object],[object Object]
[object Object],[object Object],[object Object]
Three-layer back-propagation neural network
Step 1 : Initialisation Set all the weights and threshold levels of the network to random numbers uniformly distributed inside a small range: where  F i  is the total number of inputs of neuron  i  in the network.  The weight initialisation is done on a neuron-by-neuron basis. The back-propagation training algorithm
Step 2 : Activation Activate the back-propagation neural network by applying inputs  x 1 ( p ),  x 2 ( p ),…,  x n ( p ) and desired outputs  y d ,1 ( p ),  y d ,2 ( p ),…,  y d , n ( p ). ( a )  Calculate the actual outputs of the neurons in the hidden layer: where  n  is the number of inputs of neuron  j  in the hidden layer, and  sigmoid  is the  sigmoid  activation function.
( b )  Calculate the actual outputs of the neurons in the output layer: where  m  is the number of inputs of neuron  k  in the output layer. Step 2 : Activation (continued)
Step 3 : Weight training Update the weights in the back-propagation network propagating backward the errors associated with output neurons. ( a ) Calculate the error gradient for the neurons in the output layer: where Calculate the weight corrections: Update the weights at the output neurons:
( b )  Calculate the error gradient for the neurons in the hidden layer: Calculate the weight corrections: Update the weights at the hidden neurons: Step 3 : Weight training (continued)
Step 4 : Iteration Increase iteration  p  by one, go back to  Step 2  and repeat the process until the selected error criterion is satisfied. As an example, we may consider the three-layer back-propagation network.  Suppose that the network is required to perform logical operation  Exclusive-OR .  Recall that a single-layer perceptron could not do this operation.  Now we will apply the three-layer net.
Three-layer network for solving the Exclusive-OR operation
[object Object],[object Object],[object Object]
[object Object],[object Object],[object Object]
[object Object],[object Object],[object Object]
[object Object],[object Object]
[object Object],[object Object],5038 . 0 0038 . 0 5 . 0 13 13 13       w w w 8985 . 0 0015 . 0 9 . 0 14 14 14       w w w 4038 . 0 0038 . 0 4 . 0 23 23 23       w w w 9985 . 0 0015 . 0 0 . 1 24 24 24       w w w 2067 . 1 0067 . 0 2 . 1 35 35 35         w w w 0888 . 1 0112 . 0 1 . 1 45 45 45       w w w 7962 . 0 0038 . 0 8 . 0 3 3 3          0985 . 0 0015 . 0 1 . 0 4 4 4            3127 . 0 0127 . 0 3 . 0 5 5 5         
Learning curve for operation  Exclusive-OR
Final results of three-layer network learning
Network represented by McCulloch-Pitts model for solving the  Exclusive-OR  operation
( a ) Decision boundary constructed by hidden neuron 3; ( b ) Decision boundary constructed by hidden neuron 4;  ( c ) Decision boundaries constructed by the complete three-layer network Decision boundaries
Accelerated learning in multilayer neural networks ,[object Object],[object Object],[object Object],[object Object]
[object Object],[object Object],[object Object]
Learning with momentum for operation  Exclusive-OR
Learning with adaptive learning rate To accelerate the convergence and yet avoid the  danger of instability, we can apply two heuristics: Heuristic 1 If the change of the sum of squared errors has the same algebraic sign for several consequent epochs, then the learning rate parameter,   , should be increased. Heuristic 2 If the algebraic sign of the change of the sum of squared errors alternates for several consequent epochs, then the learning rate parameter,   , should be decreased.
[object Object],[object Object],[object Object]
Learning with adaptive learning rate
Learning with momentum and adaptive learning rate
[object Object],The Hopfield Network
[object Object],[object Object]
[object Object]
Single-layer  n -neuron Hopfield network
[object Object]
[object Object],[object Object],[object Object]
[object Object],[object Object]
Possible states for the three-neuron Hopfield network
[object Object],[object Object],[object Object],[object Object]
[object Object],[object Object],[object Object]
[object Object]
[object Object],[object Object],[object Object],[object Object]
[object Object],[object Object],Storage capacity of the Hopfield network
[object Object],[object Object],Bidirectional associative memory (BAM)
[object Object],[object Object]
BAM operation
The basic idea behind the BAM is to store pattern pairs so that when  n -dimensional vector  X  from set  A  is presented as input, the BAM recalls  m -dimensional vector  Y  from set  B , but when  Y  is presented as input, the BAM recalls  X .
[object Object],[object Object]
[object Object],[object Object],[object Object],Stability and storage capacity of the BAM

Weitere ähnliche Inhalte

Was ist angesagt?

Hebbian Learning
Hebbian LearningHebbian Learning
Hebbian LearningESCOM
 
Feedforward neural network
Feedforward neural networkFeedforward neural network
Feedforward neural networkSopheaktra YONG
 
Brief Introduction to Boltzmann Machine
Brief Introduction to Boltzmann MachineBrief Introduction to Boltzmann Machine
Brief Introduction to Boltzmann MachineArunabha Saha
 
Adaptive Resonance Theory
Adaptive Resonance TheoryAdaptive Resonance Theory
Adaptive Resonance TheoryNaveen Kumar
 
Backpropagation And Gradient Descent In Neural Networks | Neural Network Tuto...
Backpropagation And Gradient Descent In Neural Networks | Neural Network Tuto...Backpropagation And Gradient Descent In Neural Networks | Neural Network Tuto...
Backpropagation And Gradient Descent In Neural Networks | Neural Network Tuto...Simplilearn
 
Feed Forward Neural Network.pptx
Feed Forward Neural Network.pptxFeed Forward Neural Network.pptx
Feed Forward Neural Network.pptxNoorFathima60
 
Autoencoder
AutoencoderAutoencoder
AutoencoderHARISH R
 
Knowledge representation In Artificial Intelligence
Knowledge representation In Artificial IntelligenceKnowledge representation In Artificial Intelligence
Knowledge representation In Artificial IntelligenceRamla Sheikh
 
Fuzzy logic and application in AI
Fuzzy logic and application in AIFuzzy logic and application in AI
Fuzzy logic and application in AIIldar Nurgaliev
 
Introduction to Recurrent Neural Network
Introduction to Recurrent Neural NetworkIntroduction to Recurrent Neural Network
Introduction to Recurrent Neural NetworkKnoldus Inc.
 
Counter propagation Network
Counter propagation NetworkCounter propagation Network
Counter propagation NetworkAkshay Dhole
 
Artificial Neural Network Lect4 : Single Layer Perceptron Classifiers
Artificial Neural Network Lect4 : Single Layer Perceptron ClassifiersArtificial Neural Network Lect4 : Single Layer Perceptron Classifiers
Artificial Neural Network Lect4 : Single Layer Perceptron ClassifiersMohammed Bennamoun
 

Was ist angesagt? (20)

Hebbian Learning
Hebbian LearningHebbian Learning
Hebbian Learning
 
Feedforward neural network
Feedforward neural networkFeedforward neural network
Feedforward neural network
 
Neural Networks
Neural NetworksNeural Networks
Neural Networks
 
Brief Introduction to Boltzmann Machine
Brief Introduction to Boltzmann MachineBrief Introduction to Boltzmann Machine
Brief Introduction to Boltzmann Machine
 
Adaptive Resonance Theory
Adaptive Resonance TheoryAdaptive Resonance Theory
Adaptive Resonance Theory
 
Mc culloch pitts neuron
Mc culloch pitts neuronMc culloch pitts neuron
Mc culloch pitts neuron
 
Activation function
Activation functionActivation function
Activation function
 
Tsp is NP-Complete
Tsp is NP-CompleteTsp is NP-Complete
Tsp is NP-Complete
 
Backpropagation And Gradient Descent In Neural Networks | Neural Network Tuto...
Backpropagation And Gradient Descent In Neural Networks | Neural Network Tuto...Backpropagation And Gradient Descent In Neural Networks | Neural Network Tuto...
Backpropagation And Gradient Descent In Neural Networks | Neural Network Tuto...
 
Feed Forward Neural Network.pptx
Feed Forward Neural Network.pptxFeed Forward Neural Network.pptx
Feed Forward Neural Network.pptx
 
Neural networks introduction
Neural networks introductionNeural networks introduction
Neural networks introduction
 
Autoencoder
AutoencoderAutoencoder
Autoencoder
 
Knowledge representation In Artificial Intelligence
Knowledge representation In Artificial IntelligenceKnowledge representation In Artificial Intelligence
Knowledge representation In Artificial Intelligence
 
Max net
Max netMax net
Max net
 
Fuzzy logic and application in AI
Fuzzy logic and application in AIFuzzy logic and application in AI
Fuzzy logic and application in AI
 
Introduction to Recurrent Neural Network
Introduction to Recurrent Neural NetworkIntroduction to Recurrent Neural Network
Introduction to Recurrent Neural Network
 
Counter propagation Network
Counter propagation NetworkCounter propagation Network
Counter propagation Network
 
Artificial Neural Network Lect4 : Single Layer Perceptron Classifiers
Artificial Neural Network Lect4 : Single Layer Perceptron ClassifiersArtificial Neural Network Lect4 : Single Layer Perceptron Classifiers
Artificial Neural Network Lect4 : Single Layer Perceptron Classifiers
 
04 Multi-layer Feedforward Networks
04 Multi-layer Feedforward Networks04 Multi-layer Feedforward Networks
04 Multi-layer Feedforward Networks
 
Hopfield Networks
Hopfield NetworksHopfield Networks
Hopfield Networks
 

Ähnlich wie lecture07.ppt

SOFT COMPUTERING TECHNICS -Unit 1
SOFT COMPUTERING TECHNICS -Unit 1SOFT COMPUTERING TECHNICS -Unit 1
SOFT COMPUTERING TECHNICS -Unit 1sravanthi computers
 
Supervised Learning
Supervised LearningSupervised Learning
Supervised Learningbutest
 
Artificial neural networks (2)
Artificial neural networks (2)Artificial neural networks (2)
Artificial neural networks (2)sai anjaneya
 
Classification by back propagation, multi layered feed forward neural network...
Classification by back propagation, multi layered feed forward neural network...Classification by back propagation, multi layered feed forward neural network...
Classification by back propagation, multi layered feed forward neural network...bihira aggrey
 
Neural Networks Ver1
Neural  Networks  Ver1Neural  Networks  Ver1
Neural Networks Ver1ncct
 
Soft Computing-173101
Soft Computing-173101Soft Computing-173101
Soft Computing-173101AMIT KUMAR
 
Artificial Neural Networks ppt.pptx for final sem cse
Artificial Neural Networks  ppt.pptx for final sem cseArtificial Neural Networks  ppt.pptx for final sem cse
Artificial Neural Networks ppt.pptx for final sem cseNaveenBhajantri1
 
Artificial neural networks seminar presentation using MSWord.
Artificial neural networks seminar presentation using MSWord.Artificial neural networks seminar presentation using MSWord.
Artificial neural networks seminar presentation using MSWord.Mohd Faiz
 
ACUMENS ON NEURAL NET AKG 20 7 23.pptx
ACUMENS ON NEURAL NET AKG 20 7 23.pptxACUMENS ON NEURAL NET AKG 20 7 23.pptx
ACUMENS ON NEURAL NET AKG 20 7 23.pptxgnans Kgnanshek
 
Artificial neural network
Artificial neural networkArtificial neural network
Artificial neural networkmustafa aadel
 
Artificial Neural Network in Medical Diagnosis
Artificial Neural Network in Medical DiagnosisArtificial Neural Network in Medical Diagnosis
Artificial Neural Network in Medical DiagnosisAdityendra Kumar Singh
 
Artificial neural networks
Artificial neural networks Artificial neural networks
Artificial neural networks ShwethaShreeS
 
Perceptron Study Material with XOR example
Perceptron Study Material with XOR examplePerceptron Study Material with XOR example
Perceptron Study Material with XOR exampleGSURESHKUMAR11
 
Artificial Neural Network seminar presentation using ppt.
Artificial Neural Network seminar presentation using ppt.Artificial Neural Network seminar presentation using ppt.
Artificial Neural Network seminar presentation using ppt.Mohd Faiz
 

Ähnlich wie lecture07.ppt (20)

SOFT COMPUTERING TECHNICS -Unit 1
SOFT COMPUTERING TECHNICS -Unit 1SOFT COMPUTERING TECHNICS -Unit 1
SOFT COMPUTERING TECHNICS -Unit 1
 
Supervised Learning
Supervised LearningSupervised Learning
Supervised Learning
 
19_Learning.ppt
19_Learning.ppt19_Learning.ppt
19_Learning.ppt
 
Artificial neural networks (2)
Artificial neural networks (2)Artificial neural networks (2)
Artificial neural networks (2)
 
10-Perceptron.pdf
10-Perceptron.pdf10-Perceptron.pdf
10-Perceptron.pdf
 
Classification by back propagation, multi layered feed forward neural network...
Classification by back propagation, multi layered feed forward neural network...Classification by back propagation, multi layered feed forward neural network...
Classification by back propagation, multi layered feed forward neural network...
 
Neural Networks Ver1
Neural  Networks  Ver1Neural  Networks  Ver1
Neural Networks Ver1
 
20120140503023
2012014050302320120140503023
20120140503023
 
Soft Computing-173101
Soft Computing-173101Soft Computing-173101
Soft Computing-173101
 
ANN.pptx
ANN.pptxANN.pptx
ANN.pptx
 
Unit 2
Unit 2Unit 2
Unit 2
 
Artificial Neural Networks ppt.pptx for final sem cse
Artificial Neural Networks  ppt.pptx for final sem cseArtificial Neural Networks  ppt.pptx for final sem cse
Artificial Neural Networks ppt.pptx for final sem cse
 
Artificial neural networks seminar presentation using MSWord.
Artificial neural networks seminar presentation using MSWord.Artificial neural networks seminar presentation using MSWord.
Artificial neural networks seminar presentation using MSWord.
 
ACUMENS ON NEURAL NET AKG 20 7 23.pptx
ACUMENS ON NEURAL NET AKG 20 7 23.pptxACUMENS ON NEURAL NET AKG 20 7 23.pptx
ACUMENS ON NEURAL NET AKG 20 7 23.pptx
 
Artificial neural network
Artificial neural networkArtificial neural network
Artificial neural network
 
6
66
6
 
Artificial Neural Network in Medical Diagnosis
Artificial Neural Network in Medical DiagnosisArtificial Neural Network in Medical Diagnosis
Artificial Neural Network in Medical Diagnosis
 
Artificial neural networks
Artificial neural networks Artificial neural networks
Artificial neural networks
 
Perceptron Study Material with XOR example
Perceptron Study Material with XOR examplePerceptron Study Material with XOR example
Perceptron Study Material with XOR example
 
Artificial Neural Network seminar presentation using ppt.
Artificial Neural Network seminar presentation using ppt.Artificial Neural Network seminar presentation using ppt.
Artificial Neural Network seminar presentation using ppt.
 

Mehr von butest

EL MODELO DE NEGOCIO DE YOUTUBE
EL MODELO DE NEGOCIO DE YOUTUBEEL MODELO DE NEGOCIO DE YOUTUBE
EL MODELO DE NEGOCIO DE YOUTUBEbutest
 
1. MPEG I.B.P frame之不同
1. MPEG I.B.P frame之不同1. MPEG I.B.P frame之不同
1. MPEG I.B.P frame之不同butest
 
LESSONS FROM THE MICHAEL JACKSON TRIAL
LESSONS FROM THE MICHAEL JACKSON TRIALLESSONS FROM THE MICHAEL JACKSON TRIAL
LESSONS FROM THE MICHAEL JACKSON TRIALbutest
 
Timeline: The Life of Michael Jackson
Timeline: The Life of Michael JacksonTimeline: The Life of Michael Jackson
Timeline: The Life of Michael Jacksonbutest
 
Popular Reading Last Updated April 1, 2010 Adams, Lorraine The ...
Popular Reading Last Updated April 1, 2010 Adams, Lorraine The ...Popular Reading Last Updated April 1, 2010 Adams, Lorraine The ...
Popular Reading Last Updated April 1, 2010 Adams, Lorraine The ...butest
 
LESSONS FROM THE MICHAEL JACKSON TRIAL
LESSONS FROM THE MICHAEL JACKSON TRIALLESSONS FROM THE MICHAEL JACKSON TRIAL
LESSONS FROM THE MICHAEL JACKSON TRIALbutest
 
Com 380, Summer II
Com 380, Summer IICom 380, Summer II
Com 380, Summer IIbutest
 
The MYnstrel Free Press Volume 2: Economic Struggles, Meet Jazz
The MYnstrel Free Press Volume 2: Economic Struggles, Meet JazzThe MYnstrel Free Press Volume 2: Economic Struggles, Meet Jazz
The MYnstrel Free Press Volume 2: Economic Struggles, Meet Jazzbutest
 
MICHAEL JACKSON.doc
MICHAEL JACKSON.docMICHAEL JACKSON.doc
MICHAEL JACKSON.docbutest
 
Social Networks: Twitter Facebook SL - Slide 1
Social Networks: Twitter Facebook SL - Slide 1Social Networks: Twitter Facebook SL - Slide 1
Social Networks: Twitter Facebook SL - Slide 1butest
 
Facebook
Facebook Facebook
Facebook butest
 
Executive Summary Hare Chevrolet is a General Motors dealership ...
Executive Summary Hare Chevrolet is a General Motors dealership ...Executive Summary Hare Chevrolet is a General Motors dealership ...
Executive Summary Hare Chevrolet is a General Motors dealership ...butest
 
Welcome to the Dougherty County Public Library's Facebook and ...
Welcome to the Dougherty County Public Library's Facebook and ...Welcome to the Dougherty County Public Library's Facebook and ...
Welcome to the Dougherty County Public Library's Facebook and ...butest
 
NEWS ANNOUNCEMENT
NEWS ANNOUNCEMENTNEWS ANNOUNCEMENT
NEWS ANNOUNCEMENTbutest
 
C-2100 Ultra Zoom.doc
C-2100 Ultra Zoom.docC-2100 Ultra Zoom.doc
C-2100 Ultra Zoom.docbutest
 
MAC Printing on ITS Printers.doc.doc
MAC Printing on ITS Printers.doc.docMAC Printing on ITS Printers.doc.doc
MAC Printing on ITS Printers.doc.docbutest
 
Mac OS X Guide.doc
Mac OS X Guide.docMac OS X Guide.doc
Mac OS X Guide.docbutest
 
WEB DESIGN!
WEB DESIGN!WEB DESIGN!
WEB DESIGN!butest
 

Mehr von butest (20)

EL MODELO DE NEGOCIO DE YOUTUBE
EL MODELO DE NEGOCIO DE YOUTUBEEL MODELO DE NEGOCIO DE YOUTUBE
EL MODELO DE NEGOCIO DE YOUTUBE
 
1. MPEG I.B.P frame之不同
1. MPEG I.B.P frame之不同1. MPEG I.B.P frame之不同
1. MPEG I.B.P frame之不同
 
LESSONS FROM THE MICHAEL JACKSON TRIAL
LESSONS FROM THE MICHAEL JACKSON TRIALLESSONS FROM THE MICHAEL JACKSON TRIAL
LESSONS FROM THE MICHAEL JACKSON TRIAL
 
Timeline: The Life of Michael Jackson
Timeline: The Life of Michael JacksonTimeline: The Life of Michael Jackson
Timeline: The Life of Michael Jackson
 
Popular Reading Last Updated April 1, 2010 Adams, Lorraine The ...
Popular Reading Last Updated April 1, 2010 Adams, Lorraine The ...Popular Reading Last Updated April 1, 2010 Adams, Lorraine The ...
Popular Reading Last Updated April 1, 2010 Adams, Lorraine The ...
 
LESSONS FROM THE MICHAEL JACKSON TRIAL
LESSONS FROM THE MICHAEL JACKSON TRIALLESSONS FROM THE MICHAEL JACKSON TRIAL
LESSONS FROM THE MICHAEL JACKSON TRIAL
 
Com 380, Summer II
Com 380, Summer IICom 380, Summer II
Com 380, Summer II
 
PPT
PPTPPT
PPT
 
The MYnstrel Free Press Volume 2: Economic Struggles, Meet Jazz
The MYnstrel Free Press Volume 2: Economic Struggles, Meet JazzThe MYnstrel Free Press Volume 2: Economic Struggles, Meet Jazz
The MYnstrel Free Press Volume 2: Economic Struggles, Meet Jazz
 
MICHAEL JACKSON.doc
MICHAEL JACKSON.docMICHAEL JACKSON.doc
MICHAEL JACKSON.doc
 
Social Networks: Twitter Facebook SL - Slide 1
Social Networks: Twitter Facebook SL - Slide 1Social Networks: Twitter Facebook SL - Slide 1
Social Networks: Twitter Facebook SL - Slide 1
 
Facebook
Facebook Facebook
Facebook
 
Executive Summary Hare Chevrolet is a General Motors dealership ...
Executive Summary Hare Chevrolet is a General Motors dealership ...Executive Summary Hare Chevrolet is a General Motors dealership ...
Executive Summary Hare Chevrolet is a General Motors dealership ...
 
Welcome to the Dougherty County Public Library's Facebook and ...
Welcome to the Dougherty County Public Library's Facebook and ...Welcome to the Dougherty County Public Library's Facebook and ...
Welcome to the Dougherty County Public Library's Facebook and ...
 
NEWS ANNOUNCEMENT
NEWS ANNOUNCEMENTNEWS ANNOUNCEMENT
NEWS ANNOUNCEMENT
 
C-2100 Ultra Zoom.doc
C-2100 Ultra Zoom.docC-2100 Ultra Zoom.doc
C-2100 Ultra Zoom.doc
 
MAC Printing on ITS Printers.doc.doc
MAC Printing on ITS Printers.doc.docMAC Printing on ITS Printers.doc.doc
MAC Printing on ITS Printers.doc.doc
 
Mac OS X Guide.doc
Mac OS X Guide.docMac OS X Guide.doc
Mac OS X Guide.doc
 
hier
hierhier
hier
 
WEB DESIGN!
WEB DESIGN!WEB DESIGN!
WEB DESIGN!
 

lecture07.ppt

  • 1.
  • 2. Introduction, or how the brain works Machine learning involves adaptive mechanisms that enable computers to learn from experience, learn by example and learn by analogy. Learning capabilities can improve the performance of an intelligent system over time. The most popular approaches to machine learning are artificial neural networks and genetic algorithms . This lecture is dedicated to neural networks.
  • 3.
  • 4.
  • 6.
  • 7.
  • 8. Architecture of a typical artificial neural network
  • 9. Analogy between biological and artificial neural networks
  • 10. The neuron as a simple computing element Diagram of a neuron
  • 11.
  • 13.
  • 15.
  • 16.
  • 17. Linear separability in the perceptrons
  • 18. This is done by making small adjustments in the weights to reduce the difference between the actual and desired outputs of the perceptron. The initial weights are randomly assigned, usually in the range [  0.5, 0.5], and then updated to obtain the output consistent with the training examples. How does the perceptron learn its classification tasks?
  • 19.
  • 20. The perceptron learning rule where p = 1, 2, 3, . . .  is the learning rate , a positive constant less than unity. The perceptron learning rule was first proposed by Rosenblatt in 1960. Using this rule we can derive the perceptron training algorithm for classification tasks.
  • 21. Step 1 : Initialisation Set initial weights w 1 , w 2 ,…, w n and threshold  to random numbers in the range [  0.5, 0.5]. If the error, e ( p ), is positive, we need to increase perceptron output Y ( p ), but if it is negative, we need to decrease Y ( p ). Perceptron’s training algorithm
  • 22. Step 2 : Activation Activate the perceptron by applying inputs x 1 ( p ), x 2 ( p ),…, x n ( p ) and desired output Y d ( p ). Calculate the actual output at iteration p = 1 where n is the number of the perceptron inputs, and step is a step activation function. Perceptron’s training algorithm (continued)
  • 23. Step 3 : Weight training Update the weights of the perceptron where  w i ( p ) is the weight correction at iteration p . The weight correction is computed by the delta rule : Step 4 : Iteration Increase iteration p by one, go back to Step 2 and repeat the process until convergence. Perceptron’s training algorithm (continued)
  • 24. Example of perceptron learning: the logical operation AND
  • 25. Two-dimensional plots of basic logical operations A perceptron can learn the operations AND and OR , but not Exclusive-OR .
  • 26.
  • 27. Multilayer perceptron with two hidden layers
  • 28.
  • 29.
  • 30.
  • 32. Step 1 : Initialisation Set all the weights and threshold levels of the network to random numbers uniformly distributed inside a small range: where F i is the total number of inputs of neuron i in the network. The weight initialisation is done on a neuron-by-neuron basis. The back-propagation training algorithm
  • 33. Step 2 : Activation Activate the back-propagation neural network by applying inputs x 1 ( p ), x 2 ( p ),…, x n ( p ) and desired outputs y d ,1 ( p ), y d ,2 ( p ),…, y d , n ( p ). ( a ) Calculate the actual outputs of the neurons in the hidden layer: where n is the number of inputs of neuron j in the hidden layer, and sigmoid is the sigmoid activation function.
  • 34. ( b ) Calculate the actual outputs of the neurons in the output layer: where m is the number of inputs of neuron k in the output layer. Step 2 : Activation (continued)
  • 35. Step 3 : Weight training Update the weights in the back-propagation network propagating backward the errors associated with output neurons. ( a ) Calculate the error gradient for the neurons in the output layer: where Calculate the weight corrections: Update the weights at the output neurons:
  • 36. ( b ) Calculate the error gradient for the neurons in the hidden layer: Calculate the weight corrections: Update the weights at the hidden neurons: Step 3 : Weight training (continued)
  • 37. Step 4 : Iteration Increase iteration p by one, go back to Step 2 and repeat the process until the selected error criterion is satisfied. As an example, we may consider the three-layer back-propagation network. Suppose that the network is required to perform logical operation Exclusive-OR . Recall that a single-layer perceptron could not do this operation. Now we will apply the three-layer net.
  • 38. Three-layer network for solving the Exclusive-OR operation
  • 39.
  • 40.
  • 41.
  • 42.
  • 43.
  • 44. Learning curve for operation Exclusive-OR
  • 45. Final results of three-layer network learning
  • 46. Network represented by McCulloch-Pitts model for solving the Exclusive-OR operation
  • 47. ( a ) Decision boundary constructed by hidden neuron 3; ( b ) Decision boundary constructed by hidden neuron 4; ( c ) Decision boundaries constructed by the complete three-layer network Decision boundaries
  • 48.
  • 49.
  • 50. Learning with momentum for operation Exclusive-OR
  • 51. Learning with adaptive learning rate To accelerate the convergence and yet avoid the danger of instability, we can apply two heuristics: Heuristic 1 If the change of the sum of squared errors has the same algebraic sign for several consequent epochs, then the learning rate parameter,  , should be increased. Heuristic 2 If the algebraic sign of the change of the sum of squared errors alternates for several consequent epochs, then the learning rate parameter,  , should be decreased.
  • 52.
  • 53. Learning with adaptive learning rate
  • 54. Learning with momentum and adaptive learning rate
  • 55.
  • 56.
  • 57.
  • 58. Single-layer n -neuron Hopfield network
  • 59.
  • 60.
  • 61.
  • 62. Possible states for the three-neuron Hopfield network
  • 63.
  • 64.
  • 65.
  • 66.
  • 67.
  • 68.
  • 69.
  • 71. The basic idea behind the BAM is to store pattern pairs so that when n -dimensional vector X from set A is presented as input, the BAM recalls m -dimensional vector Y from set B , but when Y is presented as input, the BAM recalls X .
  • 72.
  • 73.