SlideShare ist ein Scribd-Unternehmen logo
1 von 47
Downloaden Sie, um offline zu lesen
Artificial Neural Networks
Torsten Reil
torsten.reil@zoo.ox.ac.uk
Outline
• What are Neural Networks?
• Biological Neural Networks
• ANN – The basics
• Feed forward net
• Training
• Example – Voice recognition
• Applications – Feed forward nets
• Recurrency
• Elman nets
• Hopfield nets
• Central Pattern Generators
• Conclusion
What are Neural Networks?
• Models of the brain and nervous system
• Highly parallel
– Process information much more like the brain than a serial
computer
• Learning
• Very simple principles
• Very complex behaviours
• Applications
– As powerful problem solvers
– As biological models
Biological Neural Nets
• Pigeons as art experts (Watanabe et al. 1995)
– Experiment:
• Pigeon in Skinner box
• Present paintings of two different artists (e.g. Chagall / Van
Gogh)
• Reward for pecking when presented a particular artist (e.g. Van
Gogh)
• Pigeons were able to discriminate between Van
Gogh and Chagall with 95% accuracy (when
presented with pictures they had been trained on)
• Discrimination still 85% successful for previously
unseen paintings of the artists
• Pigeons do not simply memorise the pictures
• They can extract and recognise patterns (the ‘style’)
• They generalise from the already seen to make
predictions
• This is what neural networks (biological and artificial)
are good at (unlike conventional computer)
ANNs – The basics
• ANNs incorporate the two fundamental components
of biological neural nets:
1. Neurones (nodes)
2. Synapses (weights)
• Neurone vs. Node
• Structure of a node:
• Squashing function limits node output:
• Synapse vs. weight
Feed-forward nets
• Information flow is unidirectional
• Data is presented to Input layer
• Passed on to Hidden Layer
• Passed on to Output layer
• Information is distributed
• Information processing is parallel
Internal representation (interpretation) of data
• Feeding data through the net:
(1  0.25) + (0.5  (-1.5)) = 0.25 + (-0.75) = - 0.5
0.3775
1
1
5.0

 e
Squashing:
• Data is presented to the network in the form of
activations in the input layer
• Examples
– Pixel intensity (for pictures)
– Molecule concentrations (for artificial nose)
– Share prices (for stock market prediction)
• Data usually requires preprocessing
– Analogous to senses in biology
• How to represent more abstract data, e.g. a name?
– Choose a pattern, e.g.
• 0-0-1 for “Chris”
• 0-1-0 for “Becky”
• Weight settings determine the behaviour of a network
 How can we find the right weights?
Training the Network - Learning
• Backpropagation
– Requires training set (input / output pairs)
– Starts with small random weights
– Error is used to adjust weights (supervised learning)
 Gradient descent on error landscape
• Advantages
– It works!
– Relatively fast
• Downsides
– Requires a training set
– Can be slow
– Probably not biologically realistic
• Alternatives to Backpropagation
– Hebbian learning
• Not successful in feed-forward nets
– Reinforcement learning
• Only limited success
– Artificial evolution
• More general, but can be even slower than backprop
Example: Voice Recognition
• Task: Learn to discriminate between two different
voices saying “Hello”
• Data
– Sources
• Steve Simpson
• David Raubenheimer
– Format
• Frequency distribution (60 bins)
• Analogy: cochlea
• Network architecture
– Feed forward network
• 60 input (one for each frequency bin)
• 6 hidden
• 2 output (0-1 for “Steve”, 1-0 for “David”)
• Presenting the data
Steve
David
• Presenting the data (untrained network)
Steve
David
0.43
0.26
0.73
0.55
• Calculate error
Steve
David
0.43 – 0 = 0.43
0.26 –1 = 0.74
0.73 – 1 = 0.27
0.55 – 0 = 0.55
• Backprop error and adjust weights
Steve
David
0.43 – 0 = 0.43
0.26 – 1 = 0.74
0.73 – 1 = 0.27
0.55 – 0 = 0.55
1.17
0.82
• Repeat process (sweep) for all training pairs
– Present data
– Calculate error
– Backpropagate error
– Adjust weights
• Repeat process multiple times
• Presenting the data (trained network)
Steve
David
0.01
0.99
0.99
0.01
• Results – Voice Recognition
– Performance of trained network
• Discrimination accuracy between known “Hello”s
– 100%
• Discrimination accuracy between new “Hello”’s
– 100%
• Demo
• Results – Voice Recognition (ctnd.)
– Network has learnt to generalise from original data
– Networks with different weight settings can have same
functionality
– Trained networks ‘concentrate’ on lower frequencies
– Network is robust against non-functioning nodes
Applications of Feed-forward nets
– Pattern recognition
• Character recognition
• Face Recognition
– Sonar mine/rock recognition (Gorman & Sejnowksi, 1988)
– Navigation of a car (Pomerleau, 1989)
– Stock-market prediction
– Pronunciation (NETtalk)
(Sejnowksi & Rosenberg, 1987)
Cluster analysis of hidden layer
FFNs as Biological Modelling Tools
• Signalling / Sexual Selection
– Enquist & Arak (1994)
• Preference for symmetry not selection for ‘good genes’, but
instead arises through the need to recognise objects
irrespective of their orientation
– Johnstone (1994)
• Exaggerated, symmetric ornaments facilitate mate recognition
(but see Dawkins & Guilford, 1995)
Recurrent Networks
• Feed forward networks:
– Information only flows one way
– One input pattern produces one output
– No sense of time (or memory of previous state)
• Recurrency
– Nodes connect back to other nodes or themselves
– Information flow is multidirectional
– Sense of time and memory of previous state(s)
• Biological nervous systems show high levels of
recurrency (but feed-forward structures exists too)
Elman Nets
• Elman nets are feed forward networks with partial
recurrency
• Unlike feed forward nets, Elman nets have a memory
or sense of time
Classic experiment on language acquisition and
processing (Elman, 1990)
• Task
– Elman net to predict successive words in sentences.
• Data
– Suite of sentences, e.g.
• “The boy catches the ball.”
• “The girl eats an apple.”
– Words are input one at a time
• Representation
– Binary representation for each word, e.g.
• 0-1-0-0-0 for “girl”
• Training method
– Backpropagation
• Internal representation of words
Hopfield Networks
• Sub-type of recurrent neural nets
– Fully recurrent
– Weights are symmetric
– Nodes can only be on or off
– Random updating
• Learning: Hebb rule (cells that fire together wire
together)
– Biological equivalent to LTP and LTD
• Can recall a memory, if presented with a
corrupt or incomplete version
 auto-associative or
content-addressable memory
Task: store images with resolution of 20x20 pixels
 Hopfield net with 400 nodes
Memorise:
1. Present image
2. Apply Hebb rule (cells that fire together, wire together)
• Increase weight between two nodes if both have same activity, otherwise decrease
3. Go to 1
Recall:
1. Present incomplete pattern
2. Pick random node, update
3. Go to 2 until settled
DEMO
• Memories are attractors in state space
Catastrophic forgetting
• Problem: memorising new patterns corrupts the memory of older
ones
 Old memories cannot be recalled, or spurious memories arise
• Solution: allow Hopfield net to sleep
• Two approaches (both using randomness):
– Unlearning (Hopfield, 1986)
• Recall old memories by random stimulation, but use an inverse
Hebb rule
‘Makes room’ for new memories (basins of attraction shrink)
– Pseudorehearsal (Robins, 1995)
• While learning new memories, recall old memories by random
stimulation
• Use standard Hebb rule on new and old memories
 Restructure memory
• Needs short-term + long term memory
• Mammals: hippocampus plays back new memories to neo-
cortex, which is randomly stimulated at the same time
RNNs as Central Pattern Generators
• CPGs: group of neurones creating rhythmic muscle activity for
locomotion, heart-beat etc.
• Identified in several invertebrates and vertebrates
• Hard to study
•  Computer modelling
– E.g. lamprey swimming (Ijspeert et al., 1998)
• Evolution of Bipedal Walking (Reil & Husbands, 2001)
0
0.1
0.2
0.3
0.4
0.5
0.6
0.7
0.8
0.9
1
1 3 5 7 9 11 13 15 17 19 21 23 25 27 29 31 33 35 37 39 41 43 45 47 49 51 53 55 57 59 61 63 65 67 69 71 73 75 77
time
activation
left hip lateral
left hip a/p
right hip lateral
right hip a/p
left knee
right knee
• CPG cycles are cyclic attractors in state space
Recap – Neural Networks
• Components – biological plausibility
– Neurone / node
– Synapse / weight
• Feed forward networks
– Unidirectional flow of information
– Good at extracting patterns, generalisation and
prediction
– Distributed representation of data
– Parallel processing of data
– Training: Backpropagation
– Not exact models, but good at demonstrating
principles
• Recurrent networks
– Multidirectional flow of information
– Memory / sense of time
– Complex temporal dynamics (e.g. CPGs)
– Various training methods (Hebbian, evolution)
– Often better biological models than FFNs
Online material:
http://users.ox.ac.uk/~quee0818

Weitere ähnliche Inhalte

Andere mochten auch

戴旭 --- 一位解放軍空軍上校沉 思錄
戴旭 --- 一位解放軍空軍上校沉 思錄戴旭 --- 一位解放軍空軍上校沉 思錄
戴旭 --- 一位解放軍空軍上校沉 思錄psjlew
 
中文萬歲
中文萬歲中文萬歲
中文萬歲psjlew
 
臺灣封面女郎合集
臺灣封面女郎合集臺灣封面女郎合集
臺灣封面女郎合集psjlew
 
一群自稱是中國人的日耳曼土著
一群自稱是中國人的日耳曼土著一群自稱是中國人的日耳曼土著
一群自稱是中國人的日耳曼土著psjlew
 
懷念臺灣 光陰隧道L呷飯
懷念臺灣 光陰隧道L呷飯懷念臺灣 光陰隧道L呷飯
懷念臺灣 光陰隧道L呷飯psjlew
 
120年前的中國 (很珍貴)
120年前的中國 (很珍貴)120年前的中國 (很珍貴)
120年前的中國 (很珍貴)psjlew
 
現存中國最早的照片
現存中國最早的照片現存中國最早的照片
現存中國最早的照片psjlew
 
Monitoring with Icinga @ SF Bay Area LSPE meetup
Monitoring with Icinga @ SF Bay Area LSPE meetupMonitoring with Icinga @ SF Bay Area LSPE meetup
Monitoring with Icinga @ SF Bay Area LSPE meetupmdxp
 
Macao world heritage
Macao world heritageMacao world heritage
Macao world heritagepsjlew
 
September 2010 September 15
September 2010 September 15September 2010 September 15
September 2010 September 15Ryan Livergood
 
Rubyのオブジェクト図をもう一度
Rubyのオブジェクト図をもう一度Rubyのオブジェクト図をもう一度
Rubyのオブジェクト図をもう一度ionis111
 
Camouflage
CamouflageCamouflage
Camouflagepsjlew
 
未來的生物戰爭
未來的生物戰爭未來的生物戰爭
未來的生物戰爭psjlew
 
So so-lovely
So so-lovelySo so-lovely
So so-lovelypsjlew
 
2011.07 陳文茜減肥祕笈
2011.07 陳文茜減肥祕笈2011.07 陳文茜減肥祕笈
2011.07 陳文茜減肥祕笈psjlew
 
Fun joke 解鬱集_一年份的笑話_52
Fun joke 解鬱集_一年份的笑話_52Fun joke 解鬱集_一年份的笑話_52
Fun joke 解鬱集_一年份的笑話_52psjlew
 
世界七大自然奇觀之一:維多利亞大瀑布
世界七大自然奇觀之一:維多利亞大瀑布世界七大自然奇觀之一:維多利亞大瀑布
世界七大自然奇觀之一:維多利亞大瀑布psjlew
 
印象.西湖.雨
印象.西湖.雨印象.西湖.雨
印象.西湖.雨psjlew
 

Andere mochten auch (20)

戴旭 --- 一位解放軍空軍上校沉 思錄
戴旭 --- 一位解放軍空軍上校沉 思錄戴旭 --- 一位解放軍空軍上校沉 思錄
戴旭 --- 一位解放軍空軍上校沉 思錄
 
中文萬歲
中文萬歲中文萬歲
中文萬歲
 
臺灣封面女郎合集
臺灣封面女郎合集臺灣封面女郎合集
臺灣封面女郎合集
 
一群自稱是中國人的日耳曼土著
一群自稱是中國人的日耳曼土著一群自稱是中國人的日耳曼土著
一群自稱是中國人的日耳曼土著
 
懷念臺灣 光陰隧道L呷飯
懷念臺灣 光陰隧道L呷飯懷念臺灣 光陰隧道L呷飯
懷念臺灣 光陰隧道L呷飯
 
120年前的中國 (很珍貴)
120年前的中國 (很珍貴)120年前的中國 (很珍貴)
120年前的中國 (很珍貴)
 
現存中國最早的照片
現存中國最早的照片現存中國最早的照片
現存中國最早的照片
 
Monitoring with Icinga @ SF Bay Area LSPE meetup
Monitoring with Icinga @ SF Bay Area LSPE meetupMonitoring with Icinga @ SF Bay Area LSPE meetup
Monitoring with Icinga @ SF Bay Area LSPE meetup
 
Macao world heritage
Macao world heritageMacao world heritage
Macao world heritage
 
September 2010 September 15
September 2010 September 15September 2010 September 15
September 2010 September 15
 
Rubyのオブジェクト図をもう一度
Rubyのオブジェクト図をもう一度Rubyのオブジェクト図をもう一度
Rubyのオブジェクト図をもう一度
 
Camouflage
CamouflageCamouflage
Camouflage
 
未來的生物戰爭
未來的生物戰爭未來的生物戰爭
未來的生物戰爭
 
So so-lovely
So so-lovelySo so-lovely
So so-lovely
 
2011.07 陳文茜減肥祕笈
2011.07 陳文茜減肥祕笈2011.07 陳文茜減肥祕笈
2011.07 陳文茜減肥祕笈
 
Fun joke 解鬱集_一年份的笑話_52
Fun joke 解鬱集_一年份的笑話_52Fun joke 解鬱集_一年份的笑話_52
Fun joke 解鬱集_一年份的笑話_52
 
世界七大自然奇觀之一:維多利亞大瀑布
世界七大自然奇觀之一:維多利亞大瀑布世界七大自然奇觀之一:維多利亞大瀑布
世界七大自然奇觀之一:維多利亞大瀑布
 
Production Shots
Production ShotsProduction Shots
Production Shots
 
印象.西湖.雨
印象.西湖.雨印象.西湖.雨
印象.西湖.雨
 
Storyboard
Storyboard Storyboard
Storyboard
 

Ähnlich wie 2013-1 Machine Learning Lecture 04 - Torsten Reil - Artificial Neural Netw…

Introduction to Neural Network
Introduction to Neural NetworkIntroduction to Neural Network
Introduction to Neural NetworkYan Xu
 
Artificial Neural Network Learning Algorithm.ppt
Artificial Neural Network Learning Algorithm.pptArtificial Neural Network Learning Algorithm.ppt
Artificial Neural Network Learning Algorithm.pptNJUSTAiMo
 
Neural net and back propagation
Neural net and back propagationNeural net and back propagation
Neural net and back propagationMohit Shrivastava
 
Deep Learning Sample Class (Jon Lederman)
Deep Learning Sample Class (Jon Lederman)Deep Learning Sample Class (Jon Lederman)
Deep Learning Sample Class (Jon Lederman)Jon Lederman
 
Artificial Neural Network (draft)
Artificial Neural Network (draft)Artificial Neural Network (draft)
Artificial Neural Network (draft)James Boulie
 
Short presentation about my thesis
Short presentation about my thesis Short presentation about my thesis
Short presentation about my thesis fourthplanet
 
Artificial Neural Network
Artificial Neural NetworkArtificial Neural Network
Artificial Neural Networkssuserab4f3e
 
Artificial Neural Network (ANN
Artificial Neural Network (ANNArtificial Neural Network (ANN
Artificial Neural Network (ANNAndrew Molina
 
How neural networks learn ii
How neural networks learn iiHow neural networks learn ii
How neural networks learn iiaiclub_slides
 
CS532L4_Backpropagation.pptx
CS532L4_Backpropagation.pptxCS532L4_Backpropagation.pptx
CS532L4_Backpropagation.pptxMFaisalRiaz5
 
Lecture artificial neural networks and pattern recognition
Lecture   artificial neural networks and pattern recognitionLecture   artificial neural networks and pattern recognition
Lecture artificial neural networks and pattern recognitionHưng Đặng
 
Lecture artificial neural networks and pattern recognition
Lecture   artificial neural networks and pattern recognitionLecture   artificial neural networks and pattern recognition
Lecture artificial neural networks and pattern recognitionHưng Đặng
 
The Art Of Backpropagation
The Art Of BackpropagationThe Art Of Backpropagation
The Art Of BackpropagationJennifer Prendki
 
33.-Multi-Layer-Perceptron.pdf
33.-Multi-Layer-Perceptron.pdf33.-Multi-Layer-Perceptron.pdf
33.-Multi-Layer-Perceptron.pdfgnans Kgnanshek
 

Ähnlich wie 2013-1 Machine Learning Lecture 04 - Torsten Reil - Artificial Neural Netw… (20)

Neural Computing
Neural ComputingNeural Computing
Neural Computing
 
L005.neural networks
L005.neural networksL005.neural networks
L005.neural networks
 
neuros
neurosneuros
neuros
 
Neutral Network
Neutral NetworkNeutral Network
Neutral Network
 
Introduction to Neural Network
Introduction to Neural NetworkIntroduction to Neural Network
Introduction to Neural Network
 
Artificial Neural Network Learning Algorithm.ppt
Artificial Neural Network Learning Algorithm.pptArtificial Neural Network Learning Algorithm.ppt
Artificial Neural Network Learning Algorithm.ppt
 
Neural net and back propagation
Neural net and back propagationNeural net and back propagation
Neural net and back propagation
 
Deep Learning Sample Class (Jon Lederman)
Deep Learning Sample Class (Jon Lederman)Deep Learning Sample Class (Jon Lederman)
Deep Learning Sample Class (Jon Lederman)
 
Artificial Neural Network (draft)
Artificial Neural Network (draft)Artificial Neural Network (draft)
Artificial Neural Network (draft)
 
Short presentation about my thesis
Short presentation about my thesis Short presentation about my thesis
Short presentation about my thesis
 
Artificial Neural Network
Artificial Neural NetworkArtificial Neural Network
Artificial Neural Network
 
Artificial Neural Network (ANN
Artificial Neural Network (ANNArtificial Neural Network (ANN
Artificial Neural Network (ANN
 
How neural networks learn ii
How neural networks learn iiHow neural networks learn ii
How neural networks learn ii
 
Artificial neural network
Artificial neural networkArtificial neural network
Artificial neural network
 
CS532L4_Backpropagation.pptx
CS532L4_Backpropagation.pptxCS532L4_Backpropagation.pptx
CS532L4_Backpropagation.pptx
 
Lecture artificial neural networks and pattern recognition
Lecture   artificial neural networks and pattern recognitionLecture   artificial neural networks and pattern recognition
Lecture artificial neural networks and pattern recognition
 
Lecture artificial neural networks and pattern recognition
Lecture   artificial neural networks and pattern recognitionLecture   artificial neural networks and pattern recognition
Lecture artificial neural networks and pattern recognition
 
7 nn1-intro.ppt
7 nn1-intro.ppt7 nn1-intro.ppt
7 nn1-intro.ppt
 
The Art Of Backpropagation
The Art Of BackpropagationThe Art Of Backpropagation
The Art Of Backpropagation
 
33.-Multi-Layer-Perceptron.pdf
33.-Multi-Layer-Perceptron.pdf33.-Multi-Layer-Perceptron.pdf
33.-Multi-Layer-Perceptron.pdf
 

Mehr von Dongseo University

Lecture_NaturalPolicyGradientsTRPOPPO.pdf
Lecture_NaturalPolicyGradientsTRPOPPO.pdfLecture_NaturalPolicyGradientsTRPOPPO.pdf
Lecture_NaturalPolicyGradientsTRPOPPO.pdfDongseo University
 
Evolutionary Computation Lecture notes03
Evolutionary Computation Lecture notes03Evolutionary Computation Lecture notes03
Evolutionary Computation Lecture notes03Dongseo University
 
Evolutionary Computation Lecture notes02
Evolutionary Computation Lecture notes02Evolutionary Computation Lecture notes02
Evolutionary Computation Lecture notes02Dongseo University
 
Evolutionary Computation Lecture notes01
Evolutionary Computation Lecture notes01Evolutionary Computation Lecture notes01
Evolutionary Computation Lecture notes01Dongseo University
 
Average Linear Selection Algorithm
Average Linear Selection AlgorithmAverage Linear Selection Algorithm
Average Linear Selection AlgorithmDongseo University
 
Lower Bound of Comparison Sort
Lower Bound of Comparison SortLower Bound of Comparison Sort
Lower Bound of Comparison SortDongseo University
 
Running Time of Building Binary Heap using Array
Running Time of Building Binary Heap using ArrayRunning Time of Building Binary Heap using Array
Running Time of Building Binary Heap using ArrayDongseo University
 
Proof By Math Induction Example
Proof By Math Induction ExampleProof By Math Induction Example
Proof By Math Induction ExampleDongseo University
 
Estimating probability distributions
Estimating probability distributionsEstimating probability distributions
Estimating probability distributionsDongseo University
 
2018-2 Machine Learning (Wasserstein GAN and BEGAN)
2018-2 Machine Learning (Wasserstein GAN and BEGAN)2018-2 Machine Learning (Wasserstein GAN and BEGAN)
2018-2 Machine Learning (Wasserstein GAN and BEGAN)Dongseo University
 
2018-2 Machine Learning (Linear regression, Logistic regression)
2018-2 Machine Learning (Linear regression, Logistic regression)2018-2 Machine Learning (Linear regression, Logistic regression)
2018-2 Machine Learning (Linear regression, Logistic regression)Dongseo University
 
2017-2 ML W9 Reinforcement Learning #5
2017-2 ML W9 Reinforcement Learning #52017-2 ML W9 Reinforcement Learning #5
2017-2 ML W9 Reinforcement Learning #5Dongseo University
 

Mehr von Dongseo University (20)

Lecture_NaturalPolicyGradientsTRPOPPO.pdf
Lecture_NaturalPolicyGradientsTRPOPPO.pdfLecture_NaturalPolicyGradientsTRPOPPO.pdf
Lecture_NaturalPolicyGradientsTRPOPPO.pdf
 
Evolutionary Computation Lecture notes03
Evolutionary Computation Lecture notes03Evolutionary Computation Lecture notes03
Evolutionary Computation Lecture notes03
 
Evolutionary Computation Lecture notes02
Evolutionary Computation Lecture notes02Evolutionary Computation Lecture notes02
Evolutionary Computation Lecture notes02
 
Evolutionary Computation Lecture notes01
Evolutionary Computation Lecture notes01Evolutionary Computation Lecture notes01
Evolutionary Computation Lecture notes01
 
Markov Chain Monte Carlo
Markov Chain Monte CarloMarkov Chain Monte Carlo
Markov Chain Monte Carlo
 
Simplex Lecture Notes
Simplex Lecture NotesSimplex Lecture Notes
Simplex Lecture Notes
 
Reinforcement Learning
Reinforcement LearningReinforcement Learning
Reinforcement Learning
 
Median of Medians
Median of MediansMedian of Medians
Median of Medians
 
Average Linear Selection Algorithm
Average Linear Selection AlgorithmAverage Linear Selection Algorithm
Average Linear Selection Algorithm
 
Lower Bound of Comparison Sort
Lower Bound of Comparison SortLower Bound of Comparison Sort
Lower Bound of Comparison Sort
 
Running Time of Building Binary Heap using Array
Running Time of Building Binary Heap using ArrayRunning Time of Building Binary Heap using Array
Running Time of Building Binary Heap using Array
 
Running Time of MergeSort
Running Time of MergeSortRunning Time of MergeSort
Running Time of MergeSort
 
Binary Trees
Binary TreesBinary Trees
Binary Trees
 
Proof By Math Induction Example
Proof By Math Induction ExampleProof By Math Induction Example
Proof By Math Induction Example
 
TRPO and PPO notes
TRPO and PPO notesTRPO and PPO notes
TRPO and PPO notes
 
Estimating probability distributions
Estimating probability distributionsEstimating probability distributions
Estimating probability distributions
 
2018-2 Machine Learning (Wasserstein GAN and BEGAN)
2018-2 Machine Learning (Wasserstein GAN and BEGAN)2018-2 Machine Learning (Wasserstein GAN and BEGAN)
2018-2 Machine Learning (Wasserstein GAN and BEGAN)
 
2018-2 Machine Learning (Linear regression, Logistic regression)
2018-2 Machine Learning (Linear regression, Logistic regression)2018-2 Machine Learning (Linear regression, Logistic regression)
2018-2 Machine Learning (Linear regression, Logistic regression)
 
2017-2 ML W11 GAN #1
2017-2 ML W11 GAN #12017-2 ML W11 GAN #1
2017-2 ML W11 GAN #1
 
2017-2 ML W9 Reinforcement Learning #5
2017-2 ML W9 Reinforcement Learning #52017-2 ML W9 Reinforcement Learning #5
2017-2 ML W9 Reinforcement Learning #5
 

Kürzlich hochgeladen

Introduction to Nonprofit Accounting: The Basics
Introduction to Nonprofit Accounting: The BasicsIntroduction to Nonprofit Accounting: The Basics
Introduction to Nonprofit Accounting: The BasicsTechSoup
 
Russian Escort Service in Delhi 11k Hotel Foreigner Russian Call Girls in Delhi
Russian Escort Service in Delhi 11k Hotel Foreigner Russian Call Girls in DelhiRussian Escort Service in Delhi 11k Hotel Foreigner Russian Call Girls in Delhi
Russian Escort Service in Delhi 11k Hotel Foreigner Russian Call Girls in Delhikauryashika82
 
Kisan Call Centre - To harness potential of ICT in Agriculture by answer farm...
Kisan Call Centre - To harness potential of ICT in Agriculture by answer farm...Kisan Call Centre - To harness potential of ICT in Agriculture by answer farm...
Kisan Call Centre - To harness potential of ICT in Agriculture by answer farm...Krashi Coaching
 
Key note speaker Neum_Admir Softic_ENG.pdf
Key note speaker Neum_Admir Softic_ENG.pdfKey note speaker Neum_Admir Softic_ENG.pdf
Key note speaker Neum_Admir Softic_ENG.pdfAdmir Softic
 
A Critique of the Proposed National Education Policy Reform
A Critique of the Proposed National Education Policy ReformA Critique of the Proposed National Education Policy Reform
A Critique of the Proposed National Education Policy ReformChameera Dedduwage
 
microwave assisted reaction. General introduction
microwave assisted reaction. General introductionmicrowave assisted reaction. General introduction
microwave assisted reaction. General introductionMaksud Ahmed
 
General AI for Medical Educators April 2024
General AI for Medical Educators April 2024General AI for Medical Educators April 2024
General AI for Medical Educators April 2024Janet Corral
 
Z Score,T Score, Percential Rank and Box Plot Graph
Z Score,T Score, Percential Rank and Box Plot GraphZ Score,T Score, Percential Rank and Box Plot Graph
Z Score,T Score, Percential Rank and Box Plot GraphThiyagu K
 
Measures of Dispersion and Variability: Range, QD, AD and SD
Measures of Dispersion and Variability: Range, QD, AD and SDMeasures of Dispersion and Variability: Range, QD, AD and SD
Measures of Dispersion and Variability: Range, QD, AD and SDThiyagu K
 
Activity 01 - Artificial Culture (1).pdf
Activity 01 - Artificial Culture (1).pdfActivity 01 - Artificial Culture (1).pdf
Activity 01 - Artificial Culture (1).pdfciinovamais
 
BAG TECHNIQUE Bag technique-a tool making use of public health bag through wh...
BAG TECHNIQUE Bag technique-a tool making use of public health bag through wh...BAG TECHNIQUE Bag technique-a tool making use of public health bag through wh...
BAG TECHNIQUE Bag technique-a tool making use of public health bag through wh...Sapna Thakur
 
Call Girls in Dwarka Mor Delhi Contact Us 9654467111
Call Girls in Dwarka Mor Delhi Contact Us 9654467111Call Girls in Dwarka Mor Delhi Contact Us 9654467111
Call Girls in Dwarka Mor Delhi Contact Us 9654467111Sapana Sha
 
Software Engineering Methodologies (overview)
Software Engineering Methodologies (overview)Software Engineering Methodologies (overview)
Software Engineering Methodologies (overview)eniolaolutunde
 
Ecosystem Interactions Class Discussion Presentation in Blue Green Lined Styl...
Ecosystem Interactions Class Discussion Presentation in Blue Green Lined Styl...Ecosystem Interactions Class Discussion Presentation in Blue Green Lined Styl...
Ecosystem Interactions Class Discussion Presentation in Blue Green Lined Styl...fonyou31
 
9548086042 for call girls in Indira Nagar with room service
9548086042  for call girls in Indira Nagar  with room service9548086042  for call girls in Indira Nagar  with room service
9548086042 for call girls in Indira Nagar with room servicediscovermytutordmt
 
BASLIQ CURRENT LOOKBOOK LOOKBOOK(1) (1).pdf
BASLIQ CURRENT LOOKBOOK  LOOKBOOK(1) (1).pdfBASLIQ CURRENT LOOKBOOK  LOOKBOOK(1) (1).pdf
BASLIQ CURRENT LOOKBOOK LOOKBOOK(1) (1).pdfSoniaTolstoy
 

Kürzlich hochgeladen (20)

Introduction to Nonprofit Accounting: The Basics
Introduction to Nonprofit Accounting: The BasicsIntroduction to Nonprofit Accounting: The Basics
Introduction to Nonprofit Accounting: The Basics
 
Advance Mobile Application Development class 07
Advance Mobile Application Development class 07Advance Mobile Application Development class 07
Advance Mobile Application Development class 07
 
Russian Escort Service in Delhi 11k Hotel Foreigner Russian Call Girls in Delhi
Russian Escort Service in Delhi 11k Hotel Foreigner Russian Call Girls in DelhiRussian Escort Service in Delhi 11k Hotel Foreigner Russian Call Girls in Delhi
Russian Escort Service in Delhi 11k Hotel Foreigner Russian Call Girls in Delhi
 
Kisan Call Centre - To harness potential of ICT in Agriculture by answer farm...
Kisan Call Centre - To harness potential of ICT in Agriculture by answer farm...Kisan Call Centre - To harness potential of ICT in Agriculture by answer farm...
Kisan Call Centre - To harness potential of ICT in Agriculture by answer farm...
 
Mattingly "AI & Prompt Design: Structured Data, Assistants, & RAG"
Mattingly "AI & Prompt Design: Structured Data, Assistants, & RAG"Mattingly "AI & Prompt Design: Structured Data, Assistants, & RAG"
Mattingly "AI & Prompt Design: Structured Data, Assistants, & RAG"
 
Key note speaker Neum_Admir Softic_ENG.pdf
Key note speaker Neum_Admir Softic_ENG.pdfKey note speaker Neum_Admir Softic_ENG.pdf
Key note speaker Neum_Admir Softic_ENG.pdf
 
A Critique of the Proposed National Education Policy Reform
A Critique of the Proposed National Education Policy ReformA Critique of the Proposed National Education Policy Reform
A Critique of the Proposed National Education Policy Reform
 
microwave assisted reaction. General introduction
microwave assisted reaction. General introductionmicrowave assisted reaction. General introduction
microwave assisted reaction. General introduction
 
General AI for Medical Educators April 2024
General AI for Medical Educators April 2024General AI for Medical Educators April 2024
General AI for Medical Educators April 2024
 
Z Score,T Score, Percential Rank and Box Plot Graph
Z Score,T Score, Percential Rank and Box Plot GraphZ Score,T Score, Percential Rank and Box Plot Graph
Z Score,T Score, Percential Rank and Box Plot Graph
 
Measures of Dispersion and Variability: Range, QD, AD and SD
Measures of Dispersion and Variability: Range, QD, AD and SDMeasures of Dispersion and Variability: Range, QD, AD and SD
Measures of Dispersion and Variability: Range, QD, AD and SD
 
Activity 01 - Artificial Culture (1).pdf
Activity 01 - Artificial Culture (1).pdfActivity 01 - Artificial Culture (1).pdf
Activity 01 - Artificial Culture (1).pdf
 
BAG TECHNIQUE Bag technique-a tool making use of public health bag through wh...
BAG TECHNIQUE Bag technique-a tool making use of public health bag through wh...BAG TECHNIQUE Bag technique-a tool making use of public health bag through wh...
BAG TECHNIQUE Bag technique-a tool making use of public health bag through wh...
 
Call Girls in Dwarka Mor Delhi Contact Us 9654467111
Call Girls in Dwarka Mor Delhi Contact Us 9654467111Call Girls in Dwarka Mor Delhi Contact Us 9654467111
Call Girls in Dwarka Mor Delhi Contact Us 9654467111
 
Software Engineering Methodologies (overview)
Software Engineering Methodologies (overview)Software Engineering Methodologies (overview)
Software Engineering Methodologies (overview)
 
INDIA QUIZ 2024 RLAC DELHI UNIVERSITY.pptx
INDIA QUIZ 2024 RLAC DELHI UNIVERSITY.pptxINDIA QUIZ 2024 RLAC DELHI UNIVERSITY.pptx
INDIA QUIZ 2024 RLAC DELHI UNIVERSITY.pptx
 
Ecosystem Interactions Class Discussion Presentation in Blue Green Lined Styl...
Ecosystem Interactions Class Discussion Presentation in Blue Green Lined Styl...Ecosystem Interactions Class Discussion Presentation in Blue Green Lined Styl...
Ecosystem Interactions Class Discussion Presentation in Blue Green Lined Styl...
 
9548086042 for call girls in Indira Nagar with room service
9548086042  for call girls in Indira Nagar  with room service9548086042  for call girls in Indira Nagar  with room service
9548086042 for call girls in Indira Nagar with room service
 
Mattingly "AI & Prompt Design: The Basics of Prompt Design"
Mattingly "AI & Prompt Design: The Basics of Prompt Design"Mattingly "AI & Prompt Design: The Basics of Prompt Design"
Mattingly "AI & Prompt Design: The Basics of Prompt Design"
 
BASLIQ CURRENT LOOKBOOK LOOKBOOK(1) (1).pdf
BASLIQ CURRENT LOOKBOOK  LOOKBOOK(1) (1).pdfBASLIQ CURRENT LOOKBOOK  LOOKBOOK(1) (1).pdf
BASLIQ CURRENT LOOKBOOK LOOKBOOK(1) (1).pdf
 

2013-1 Machine Learning Lecture 04 - Torsten Reil - Artificial Neural Netw…

  • 1. Artificial Neural Networks Torsten Reil torsten.reil@zoo.ox.ac.uk
  • 2. Outline • What are Neural Networks? • Biological Neural Networks • ANN – The basics • Feed forward net • Training • Example – Voice recognition • Applications – Feed forward nets • Recurrency • Elman nets • Hopfield nets • Central Pattern Generators • Conclusion
  • 3. What are Neural Networks? • Models of the brain and nervous system • Highly parallel – Process information much more like the brain than a serial computer • Learning • Very simple principles • Very complex behaviours • Applications – As powerful problem solvers – As biological models
  • 4. Biological Neural Nets • Pigeons as art experts (Watanabe et al. 1995) – Experiment: • Pigeon in Skinner box • Present paintings of two different artists (e.g. Chagall / Van Gogh) • Reward for pecking when presented a particular artist (e.g. Van Gogh)
  • 5.
  • 6.
  • 7.
  • 8. • Pigeons were able to discriminate between Van Gogh and Chagall with 95% accuracy (when presented with pictures they had been trained on) • Discrimination still 85% successful for previously unseen paintings of the artists • Pigeons do not simply memorise the pictures • They can extract and recognise patterns (the ‘style’) • They generalise from the already seen to make predictions • This is what neural networks (biological and artificial) are good at (unlike conventional computer)
  • 9. ANNs – The basics • ANNs incorporate the two fundamental components of biological neural nets: 1. Neurones (nodes) 2. Synapses (weights)
  • 11. • Structure of a node: • Squashing function limits node output:
  • 12. • Synapse vs. weight
  • 13. Feed-forward nets • Information flow is unidirectional • Data is presented to Input layer • Passed on to Hidden Layer • Passed on to Output layer • Information is distributed • Information processing is parallel Internal representation (interpretation) of data
  • 14. • Feeding data through the net: (1  0.25) + (0.5  (-1.5)) = 0.25 + (-0.75) = - 0.5 0.3775 1 1 5.0   e Squashing:
  • 15. • Data is presented to the network in the form of activations in the input layer • Examples – Pixel intensity (for pictures) – Molecule concentrations (for artificial nose) – Share prices (for stock market prediction) • Data usually requires preprocessing – Analogous to senses in biology • How to represent more abstract data, e.g. a name? – Choose a pattern, e.g. • 0-0-1 for “Chris” • 0-1-0 for “Becky”
  • 16. • Weight settings determine the behaviour of a network  How can we find the right weights?
  • 17. Training the Network - Learning • Backpropagation – Requires training set (input / output pairs) – Starts with small random weights – Error is used to adjust weights (supervised learning)  Gradient descent on error landscape
  • 18.
  • 19. • Advantages – It works! – Relatively fast • Downsides – Requires a training set – Can be slow – Probably not biologically realistic • Alternatives to Backpropagation – Hebbian learning • Not successful in feed-forward nets – Reinforcement learning • Only limited success – Artificial evolution • More general, but can be even slower than backprop
  • 20. Example: Voice Recognition • Task: Learn to discriminate between two different voices saying “Hello” • Data – Sources • Steve Simpson • David Raubenheimer – Format • Frequency distribution (60 bins) • Analogy: cochlea
  • 21. • Network architecture – Feed forward network • 60 input (one for each frequency bin) • 6 hidden • 2 output (0-1 for “Steve”, 1-0 for “David”)
  • 22. • Presenting the data Steve David
  • 23. • Presenting the data (untrained network) Steve David 0.43 0.26 0.73 0.55
  • 24. • Calculate error Steve David 0.43 – 0 = 0.43 0.26 –1 = 0.74 0.73 – 1 = 0.27 0.55 – 0 = 0.55
  • 25. • Backprop error and adjust weights Steve David 0.43 – 0 = 0.43 0.26 – 1 = 0.74 0.73 – 1 = 0.27 0.55 – 0 = 0.55 1.17 0.82
  • 26. • Repeat process (sweep) for all training pairs – Present data – Calculate error – Backpropagate error – Adjust weights • Repeat process multiple times
  • 27. • Presenting the data (trained network) Steve David 0.01 0.99 0.99 0.01
  • 28. • Results – Voice Recognition – Performance of trained network • Discrimination accuracy between known “Hello”s – 100% • Discrimination accuracy between new “Hello”’s – 100% • Demo
  • 29. • Results – Voice Recognition (ctnd.) – Network has learnt to generalise from original data – Networks with different weight settings can have same functionality – Trained networks ‘concentrate’ on lower frequencies – Network is robust against non-functioning nodes
  • 30. Applications of Feed-forward nets – Pattern recognition • Character recognition • Face Recognition – Sonar mine/rock recognition (Gorman & Sejnowksi, 1988) – Navigation of a car (Pomerleau, 1989) – Stock-market prediction – Pronunciation (NETtalk) (Sejnowksi & Rosenberg, 1987)
  • 31. Cluster analysis of hidden layer
  • 32. FFNs as Biological Modelling Tools • Signalling / Sexual Selection – Enquist & Arak (1994) • Preference for symmetry not selection for ‘good genes’, but instead arises through the need to recognise objects irrespective of their orientation – Johnstone (1994) • Exaggerated, symmetric ornaments facilitate mate recognition (but see Dawkins & Guilford, 1995)
  • 33. Recurrent Networks • Feed forward networks: – Information only flows one way – One input pattern produces one output – No sense of time (or memory of previous state) • Recurrency – Nodes connect back to other nodes or themselves – Information flow is multidirectional – Sense of time and memory of previous state(s) • Biological nervous systems show high levels of recurrency (but feed-forward structures exists too)
  • 34. Elman Nets • Elman nets are feed forward networks with partial recurrency • Unlike feed forward nets, Elman nets have a memory or sense of time
  • 35. Classic experiment on language acquisition and processing (Elman, 1990) • Task – Elman net to predict successive words in sentences. • Data – Suite of sentences, e.g. • “The boy catches the ball.” • “The girl eats an apple.” – Words are input one at a time • Representation – Binary representation for each word, e.g. • 0-1-0-0-0 for “girl” • Training method – Backpropagation
  • 37. Hopfield Networks • Sub-type of recurrent neural nets – Fully recurrent – Weights are symmetric – Nodes can only be on or off – Random updating • Learning: Hebb rule (cells that fire together wire together) – Biological equivalent to LTP and LTD • Can recall a memory, if presented with a corrupt or incomplete version  auto-associative or content-addressable memory
  • 38. Task: store images with resolution of 20x20 pixels  Hopfield net with 400 nodes Memorise: 1. Present image 2. Apply Hebb rule (cells that fire together, wire together) • Increase weight between two nodes if both have same activity, otherwise decrease 3. Go to 1 Recall: 1. Present incomplete pattern 2. Pick random node, update 3. Go to 2 until settled DEMO
  • 39. • Memories are attractors in state space
  • 40. Catastrophic forgetting • Problem: memorising new patterns corrupts the memory of older ones  Old memories cannot be recalled, or spurious memories arise • Solution: allow Hopfield net to sleep
  • 41. • Two approaches (both using randomness): – Unlearning (Hopfield, 1986) • Recall old memories by random stimulation, but use an inverse Hebb rule ‘Makes room’ for new memories (basins of attraction shrink) – Pseudorehearsal (Robins, 1995) • While learning new memories, recall old memories by random stimulation • Use standard Hebb rule on new and old memories  Restructure memory • Needs short-term + long term memory • Mammals: hippocampus plays back new memories to neo- cortex, which is randomly stimulated at the same time
  • 42. RNNs as Central Pattern Generators • CPGs: group of neurones creating rhythmic muscle activity for locomotion, heart-beat etc. • Identified in several invertebrates and vertebrates • Hard to study •  Computer modelling – E.g. lamprey swimming (Ijspeert et al., 1998)
  • 43. • Evolution of Bipedal Walking (Reil & Husbands, 2001) 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 1 3 5 7 9 11 13 15 17 19 21 23 25 27 29 31 33 35 37 39 41 43 45 47 49 51 53 55 57 59 61 63 65 67 69 71 73 75 77 time activation left hip lateral left hip a/p right hip lateral right hip a/p left knee right knee
  • 44.
  • 45. • CPG cycles are cyclic attractors in state space
  • 46. Recap – Neural Networks • Components – biological plausibility – Neurone / node – Synapse / weight • Feed forward networks – Unidirectional flow of information – Good at extracting patterns, generalisation and prediction – Distributed representation of data – Parallel processing of data – Training: Backpropagation – Not exact models, but good at demonstrating principles • Recurrent networks – Multidirectional flow of information – Memory / sense of time – Complex temporal dynamics (e.g. CPGs) – Various training methods (Hebbian, evolution) – Often better biological models than FFNs