SlideShare ist ein Scribd-Unternehmen logo
1 von 213
Downloaden Sie, um offline zu lesen
Dmytro Fishman (dmytrofishman@gmail.com)
Intro to Deep Learning: 100%
Tartu: 100%
BioInformatics and
Information Technology
research group
https://biit.cs.ut.ee/
Scientific
Software
Statistical Analysis
Biological
Imaging
OMICS
Teaching &
Training
Personalised
Medicine
Data
management
Machine
Learning
Biological
Imaging
Biological
Imaging
1
2
3
4 5
6
7
Microscopy imaging
Biological
Imaging
1
2
3
4 5
6
7
Microscopy imaging
Retinopathy
detection
Classification of
skin cancer
Slides are @ https://bit.ly/2LgrKRd
Machine
Learning
Machine
Learning
Supervised
Learning
Machine
Learning
Machine
Learning
Supervised
Learning
Unsupervised
Learning
Machine
Learning
Reinforcement
Learning
Supervised
Learning
Unsupervised
Learning
Machine
Learning
Supervised
Learning
Unsupervised
Learning
Reinforcement
Learning
Supervised
Learning
Machine
Learning
Unsupervised
Learning
Reinforcement
Learning
Machine
Learning
Supervised
Learning
Unsupervised
Learning
Reinforcement
Learning
Machine
Learning
Deep
Learning
Supervised
Learning
Unsupervised
Learning
Reinforcement
Learning
What is deep learning?
What is deep learning?
An adaptive non-linear mapping from one space
to another
Space 1 Space 2
What is deep learning?
Space 1 Space 2
3
An adaptive non-linear mapping from one space
to another
What is deep learning?
Space 1 Space 2
SETOSA
An adaptive non-linear mapping from one space
to another
What is deep learning?
Space 1 Space 2
“Hello world!”
An adaptive non-linear mapping from one space
to another
In practice
DL = Artificial Neural Networks with many layers
4 pixel
image
4 pixel image
4 pixel
image
Categorise images
4 pixel
image
Categorise images Solid
4 pixel
image
Categorise images Solid
Vertical
4 pixel
image
Categorise images Solid
Vertical
Diagonal
4 pixel
image
Categorise images Solid
Vertical
Diagonal
Horizontal
4 pixel
image
Simple rules won’t do the trick Solid
Vertical
Diagonal
Horizontal
4 pixel
image
4 pixel
image
4 pixel
image
-0.5-1.0 0 1.00.5
Scale of values
4 pixel
image
-0.5
0
1.0
-1.0
4 pixel
image
-0.5
0
1.0
-1.0
Receptive field of neuron
4 pixel
image
-0.5
0
1.0
-1.0
Input layer
4 pixel
image
-0.5
0
1.0
-1.0
Input layer
4 pixel
image
-0.5
0
1.0
-1.0
+
Input layer
Sum of inputs
4 pixel
image
-0.5
0
1.0
-1.0
+
Input layer
1.0
1.0
1.0
1.0
Weights
4 pixel
image
-0.5
0
1.0
-1.0
+
Input layer
1.0
1.0
1.0
1.0
= -0.5
-0.5 x 1.0
0 x 1.0
1.0 x 1.0
-1.0 x 1.0
Weights
4 pixel
image
-0.5
0
1.0
-1.0
+
Input layer
0.2
-0.3
0.0
-0.8
= 0.7
-0.5 x 0.2
0 x -0.3
1.0 x 0.0
-1.0 x -0.8
Weights
4 pixel
image
-0.5
0
1.0
-1.0
+
Input layer
0.2
-0.3
0.0
-0.8
= 0.7
-0.5 x 0.2
0 x -0.3
1.0 x 0.0
-1.0 x -0.8
Weights
4 pixel
image
-0.5
0
1.0
-1.0
+
Input layer
0.2
-0.3
0.0
-0.8
0.7
Activation function
-8 -6 -4 -2 2 4 6 8
0.2
0.4
0.6
0.8
1.0
x
sig(x)
Sigmoid activation function
-8 -6 -4 -2 2 4 6 8
0.2
0.4
0.6
0.8
1.0
x
sig(x)
sig(x) =
1
1 + e−x
Sigmoid activation function
-8 -6 -4 -2 2 4 6 8
0.2
0.4
0.6
0.8
1.0
x
sig(x)
sig(x) =
1
1 + e−x
Sigmoid activation function
0.7
-8 -6 -4 -2 2 4 6 8
0.2
0.4
0.6
0.8
1.0
x
sig(x)
sig(x) =
1
1 + e−x
Sigmoid activation function
0.7
0.66
-8 -6 -4 -2 2 4 6 8
0.2
0.4
0.6
0.8
1.0
x
sig(x)
sig(x) =
1
1 + e−x
Sigmoid activation function
-8 -6 -4 -2 2 4 6 8
0.2
0.4
0.6
0.8
1.0
x
sig(x)
sig(x) =
1
1 + e−x
Sigmoid activation function
0.88
-8 -6 -4 -2 4 6 82
0.2
0.4
0.6
0.8
1.0
x
sig(x)
sig(x) =
1
1 + e−x
Sigmoid activation function
0.99
-8 -6 -4 -2 4 6 82
0.2
0.4
0.6
0.8
1.0
sig(x) =
1
1 + e−x
x
sig(x)
Sigmoid activation function
0.99
No matter how big your x can grow, your sig(x)
will always be between 0 and 1
4 pixel
image
-0.5
0
1.0
-1.0
+
Input layer
0.2
-0.3
0.0
-0.8
0.7
Activation function
0.66
4 pixel
image
-0.5
0
1.0
-1.0
+
Input layer
0.2
-0.3
0.0
-0.8
Weighted sum + activation function
= neuron
0.66
Activation
function0.7
4 pixel
image
-0.5
0
1.0
-1.0
+
Input layer
0.2
-0.3
0.0
-0.8
Weighted sum + activation function
= neuron
input of the neuron
0.66
Activation
function0.7
4 pixel
image
-0.5
0
1.0
-1.0
+
Input layer
0.2
-0.3
0.0
-0.8
Weighted sum + activation function
= neuron
input of the neuron
output of
the neuron
0.66
Activation
function0.7
4 pixel
image
-0.5
0
1.0
-1.0
+
Input layer
0.2
-0.3
0.0
-0.8
Weighted sum + activation function
= neuron
input of the neuron
output of
the neuron
0.66
Activation
function0.7
4 pixel
image
-0.5
0
1.0
-1.0
+
Input layer
0.2
-0.3
0.0
-0.8
Weighted sum + activation function
= neuron
input of the neuron
output of
the neuron
0.66
Activation
function0.7
4 pixel
image
-0.5
0
1.0
-1.0
Input layer
0.2
-0.3
0.0
-0.8
Weighted sum + activation function
= neuron
0.66
Weighted sum usually is
not visualised explicitly
4 pixel
image
-0.5
0
1.0
-1.0
Input layer
1.0
-1.0
Let’s make more neurons,
connected by different weights
0.0(missing)
4 pixel
image
-0.5
0
1.0
-1.0
Input layer
4 pixel
image
-0.5
0
1.0
-1.0
Input layer More complex receptive fields
4 pixel
image
-0.5
0
1.0
-1.0
Input layer More complex receptive fields
4 pixel
image
-0.5
0
1.0
-1.0
Input layer More complex receptive fields
4 pixel
image
-0.5
0
1.0
-1.0
Input layer More complex receptive fields
4 pixel
image
-0.5
0
1.0
-1.0
Input layer
4 pixel
image
-0.5
0
1.0
-1.0
Input layer Receptive fields get even more complex
4 pixel
image
-0.5
0
1.0
-1.0
Input layer Adding one more layer
-8 -6 -4 -2 2 4 6 8
2
4
6
8
10
x
max(0, x)
Rectified Linear Units (ReLu)
-8 -6 -4 -2 2 4 6 8
2
4
6
8
10
R(x) = max(0, x)
x
max(0, x)
Rectified Linear Units (ReLu)
-8 -6 -4 -2 2 4 6 8
2
4
6
8
10
R(x) = max(0, x)
x
max(0, x)
Rectified Linear Units (ReLu)
If number is positive, keep it.
Otherwise it is zero.
4 pixel
image
-0.5
0
1.0
-1.0
Input layer Adding one more layer
4 pixel
image
-0.5
0
1.0
-1.0
Input layer Adding one more layer
4 pixel
image
-0.5
0
1.0
-1.0
Input layer Add output layer
Solid
Vertical
Diagonal
Horizontal
4 pixel
image
-1.0
-1.0
1.0
1.0
Input layer Example
Solid
Vertical
Diagonal
Horizontal
4 pixel
image
-1.0
-1.0
1.0
1.0
Input layer Example
Solid
Vertical
Diagonal
Horizontal
0.0
0.0
2.0
2.0
-8 -6 -4 -2 2 4 6 8
0.2
0.4
0.6
0.8
1.0
x
sig(x)
sig(x) =
1
1 + e−x
Sigmoid activation function
4 pixel
image
-1.0
-1.0
1.0
1.0
Input layer Example
Solid
Vertical
Diagonal
Horizontal
0.0
0.0
2.0
2.0
4 pixel
image
-1.0
-1.0
1.0
1.0
Input layer Example
Solid
Vertical
Diagonal
Horizontal
0.0
0.0
2.0
2.0
0.5
0.5
0.88
0.88
4 pixel
image
-1.0
-1.0
1.0
1.0
Input layer Example
Solid
Vertical
Diagonal
Horizontal
0.0
0.0
2.0
2.0
0.0
1.0
0.5
0.5
0.88
0.88
1.76
0.0
4 pixel
image
-1.0
-1.0
1.0
1.0
Input layer Example
Solid
Vertical
Diagonal
Horizontal
0.0
0.0
2.0
2.0
0.0
1.0
0.5
0.5 0.5
0.73
0.88
0.88
1.76
0.0
0.5
0.85
4 pixel
image
-1.0
-1.0
1.0
1.0
Input layer Example
Solid
Vertical
Diagonal
Horizontal
0.0
0.0
2.0
2.0
0.0
1.0
0.73
-0.73
0.5
0.5 0.5
0.5
-0.5
0.73
0.88
0.88
1.76
0.0
0.5
0.5
-0.5
0.85
0.85
-0.85
4 pixel
image
-1.0
-1.0
1.0
1.0
Input layer Example
Solid
Vertical
Diagonal
Horizontal
0.0
0.0
2.0
2.0
0.0
1.0
0.73
-0.73
0.5
0.5 0.5
0.5
-0.5
0.73
0.88
0.88
1.76
0.0
0.5
0.5
-0.5
0.85
0.85
-0.85
0.73
0.5
0.5
0.85
4 pixel
image
-1.0
-1.0
1.0
1.0
Input layer Example
Solid
Vertical
Diagonal
Horizontal
0.0
0.0
2.0
2.0
0.0
1.0
0.73
-0.73
0.5
0.5 0.5
0.5
-0.5
0.73
0.88
0.88
1.76
0.0
0.5
0.5
-0.5
0.85
0.85
-0.85
0.73
0.5
0.5
0.85
x
y
-0.8
0.2
-0.6
-0.4
-0.2
0.0
0.4
0.6
-0.75 -0.50 -0.25 0.00 0.25 0.50 0.75 1.00
Consider this 2D example
(0.1,0.05)
x
y
-0.8
0.2
-0.6
-0.4
-0.2
0.0
0.4
0.6
-0.75 -0.50 -0.25 0.00 0.25 0.50 0.75 1.00
Consider this 2D example
x
y
-0.8
0.2
-0.6
-0.4
-0.2
0.0
0.4
0.6
-0.75 -0.50 -0.25 0.00 0.25 0.50 0.75 1.00
(0.1,0.05)
Consider this 2D example
x
y
-0.8
0.2
-0.6
-0.4
-0.2
0.0
0.4
0.6
-0.75 -0.50 -0.25 0.00 0.25 0.50 0.75 1.00
(0.1,0.05)
Consider this 2D example
x
y
0.0
0.00
(0.1,0.05)
Consider this 2D example
.05
0.1
Point
coordinates
as input
x
y
0
0.0
(0.1,0.05)
y
x
.05
0.1
Point
coordinates
as input
Hidden
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
x
y
0
0.0
(0.1,0.05)
y
x 0.25(w3)
.05
0.1
Point
coordinates
as input
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.40 (w5)
0.55 (w8)
0.45(w6)
x
y
0
0.0
(0.1,0.05)
y
x 0.25(w3)
0.50(w7)
.05
0.1
Point
coordinates
as input
TruthAnswer Error
?
?
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.40 (w5)
0.55 (w8)
0.45(w6)
?
?x
y
0
0.0
(0.1,0.05)
y
x
1.0
0.00.0
0.25(w3)
0.50(w7)
.05
0.1
Adding one more type of neurons
TruthAnswer Error
?
?
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.40 (w5)
0.55 (w8)
0.45(w6)
?
?
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
0.00.0
Point
coordinates
as input
0.25(w3)
0.50(w7)
.05
0.1
TruthAnswer Error
?
?
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.40 (w5)
0.55 (w8)
0.45(w6)
?
?
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
0.00.0
Point
coordinates
as input
What is the role of bias in Neural Networks?
0.25(w3)
0.50(w7)
-8 -6 -4 -2 2 4 6 8
0.2
0.4
0.6
0.8
1.0
x
sig(x)
OutputInput
x
w1
sig(w1 * x)
What is the role of bias in Neural Networks?
Source: http://www.uta.fi/sis/tie/neuro/index/Neurocomputing2.pdf
-8 -6 -4 -2 2 4 6 8
0.2
0.4
0.6
0.8
1.0
x
sig(w1 * x)
OutputInput
x
w1
sig(w1 * x)
What is the role of bias in Neural Networks?
w1 =[0.5, 1.0, 2.0]
Source: http://www.uta.fi/sis/tie/neuro/index/Neurocomputing2.pdf
-8 -6 -4 -2 2 4 6 8
0.2
0.4
0.6
0.8
1.0
x
sig(w1 * x)
-8 -6 -4 -2 2 4 6 8
0.2
0.4
0.6
0.8
1.0
x
sig(x)
OutputInput
x
w1
sig(w1 * x)
What is the role of bias in Neural Networks?
w1 =[0.5, 1.0, 2.0]
OutputInput
x
w1
sig(w1 * x + b1)
b b1
Source: http://www.uta.fi/sis/tie/neuro/index/Neurocomputing2.pdf
OutputInput
x
w1
sig(w1 * x)
Bias helps to shift the resulting curve
w1 =[0.5, 1.0, 2.0]
OutputInput
x
w1
sig(w1 * x + b1)
b1
= [-4.0, 0.0, 4.0]
b b1
w1=[1.0]
-8 -6 -4 -2 2 4 6 8
0.2
0.4
0.6
0.8
1.0
x
sig(w1 * x + b1)
-8 -6 -4 -2 2 4 6 8
0.2
0.4
0.6
0.8
1.0
x
sig(w1 * x)
b1 b1
Source: http://www.uta.fi/sis/tie/neuro/index/Neurocomputing2.pdf
OutputInput
x
w1
sig(w1 * x)
Bias helps to shift the resulting curve
w1 =[0.5, 1.0, 2.0]
OutputInput
x
w1
sig(w1 * x + b1)
b1
= [-4.0, 0.0, 4.0]
b b1
w1=[1.0]
sig(x) =
1
1 + e−x
-8 -6 -4 -2 2 4 6 8
0.2
0.4
0.6
0.8
1.0
x
sig(w1 * x + b1)
-8 -6 -4 -2 2 4 6 8
0.2
0.4
0.6
0.8
1.0
x
sig(w1 * x)
b1 b1
Source: http://www.uta.fi/sis/tie/neuro/index/Neurocomputing2.pdf
OutputInput
x
w1
sig(w1 * x)
Bias helps to shift the resulting curve
w1 =[0.5, 1.0, 2.0]
OutputInput
x
w1
sig(w1 * x + b1)
b1
= [-4.0, 0.0, 4.0]
b b1
w1=[1.0]
sig(x) =
1
1 + e−(wx+b)
-8 -6 -4 -2 2 4 6 8
0.2
0.4
0.6
0.8
1.0
x
sig(w1 * x + b1)
-8 -6 -4 -2 2 4 6 8
0.2
0.4
0.6
0.8
1.0
x
sig(w1 * x)
b1 b1
Source: http://www.uta.fi/sis/tie/neuro/index/Neurocomputing2.pdf
.05
0.1
TruthAnswer Error
?
?
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.40 (w5)
0.55 (w8)
0.45(w6)
0.50(w7)
?
?
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
0.00.0
Point
coordinates
as input
Let’s calculate scores for each class
Feed-forward path
0.0.05
0.1
TruthAnswer Error
?
?
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.40 (w5)
0.55 (w8)
0.45(w6)
0.50(w7)
?
?
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
Point
coordinates
as input
input(h1) = ?
Input of the first neuron in the
hidden layer?
Feed-forward path
0.0.05
0.1
TruthAnswer Error
?
?
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.40 (w5)
0.55 (w8)
0.45(w6)
0.50(w7)
?
?
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
Point
coordinates
as input
Input of the first neuron in the
hidden layer?
input(h1) = x * w1 + y * w2 + b1
input(h1) = ?
Feed-forward path
0.0.05
0.1
TruthAnswer Error
?
?
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.40 (w5)
0.55 (w8)
0.45(w6)
0.50(w7)
?
?
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
Point
coordinates
as input
input(h1) = 0.3775
Input of the first neuron in the
hidden layer?
input(h1) = x * w1 + y * w2 + b1
input(h1) = 0.05 * 0.15 + 0.1 * 0.2 + 0.35 = 0.3775
Feed-forward path
0.0.05
0.1
TruthAnswer Error
?
?
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.40 (w5)
0.55 (w8)
0.45(w6)
0.50(w7)
?
?
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
Point
coordinates
as input
input(h1) = 0.3775
Input of the first neuron in the
hidden layer?
input(h1) = x * w1 + y * w2 + b1
input(h1) = 0.05 * 0.15 + 0.1 * 0.2 + 0.35 = 0.3775
What about output of this
neuron?
out(h1) = ?
out(h1) = sig(input(h1))
Feed-forward path
0.0.05
0.1
TruthAnswer Error
?
?
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.40 (w5)
0.55 (w8)
0.45(w6)
0.50(w7)
?
?
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
Point
coordinates
as input
input(h1) = 0.3775
Input of the first neuron in the
hidden layer?
input(h1) = x * w1 + y * w2 + b1
input(h1) = 0.05 * 0.15 + 0.1 * 0.2 + 0.35 = 0.3775
What about output of this
neuron?
out(h1) = ?
out(h1) = sig(0.3775)
Feed-forward path
0.0.05
0.1
TruthAnswer Error
?
?
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.40 (w5)
0.55 (w8)
0.45(w6)
0.50(w7)
?
?
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
Point
coordinates
as input
input(h1) = 0.3775
Input of the first neuron in the
hidden layer?
input(h1) = x * w1 + y * w2 + b1
input(h1) = 0.05 * 0.15 + 0.1 * 0.2 + 0.35 = 0.3775
What about output of this
neuron?
out(h1) = ?
-8 -6 -4 -2 2 4 6 8
0.2
0.4
0.6
0.8
1.0
x
sig(x)
0.3775
out(h1) = sig(0.3775)
Feed-forward path
0.0.05
0.1
TruthAnswer Error
?
?
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.40 (w5)
0.55 (w8)
0.45(w6)
0.50(w7)
?
?
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
Point
coordinates
as input
input(h1) = 0.3775
Input of the first neuron in the
hidden layer?
input(h1) = x * w1 + y * w2 + b1
input(h1) = 0.05 * 0.15 + 0.1 * 0.2 + 0.35 = 0.3775
What about output of this
neuron?
out(h1) = ?
-8 -6 -4 -2 2 4 6 8
0.2
0.4
0.6
0.8
1.0
x
sig(x)
0.3775
out(h1) = sig(0.3775) = 0.5933
Feed-forward path
.05
0.1
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.40 (w5)
0.55 (w8)
0.45(w6)
0.50(w7)
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
Point
coordinates
as input
out(h1) = 0.5933 0.0
TruthAnswer Error
?
?
?
?1.0
input(h1) = 0.3775
Feed-forward path
.05
0.1
TruthAnswer Error
?
?
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.40 (w5)
0.55 (w8)
0.45(w6)
0.50(w7)
?
?
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
0.00.0
Point
coordinates
as input
0.5933 out(h1)
Feed-forward path
0.0.05
0.1
TruthAnswer Error
?
?
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.40 (w5)
0.55 (w8)
0.45(w6)
0.50(w7)
?
?
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
Point
coordinates
as input
Input of the second neuron in
the hidden layer?
0.5933 out(h1)
input(h2) = ?
Feed-forward path
0.0.05
0.1
TruthAnswer Error
?
?
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.40 (w5)
0.55 (w8)
0.45(w6)
0.50(w7)
?
?
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
Point
coordinates
as input
Input of the second neuron in
the hidden layer?
0.5933 out(h1)
input(h2) = ?
input(h2) = ? * ? + ? * ? + ?
Feed-forward path
0.0.05
0.1
TruthAnswer Error
?
?
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.40 (w5)
0.55 (w8)
0.45(w6)
0.50(w7)
?
?
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
Point
coordinates
as input
Input of the second neuron in
the hidden layer?
0.5933 out(h1)
input(h2) = ?
input(h2) = x * w3 + y * w4 + b1
Feed-forward path
0.0.05
0.1
TruthAnswer Error
?
?
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.40 (w5)
0.55 (w8)
0.45(w6)
0.50(w7)
?
?
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
Point
coordinates
as input
Input of the second neuron in
the hidden layer?
0.5933 out(h1)
input(h2) = 0.3925
input(h2) = 0.05 * 0.25 + 0.1 * 0.3 + 0.35 = 0.3925
input(h2) = x * w3 + y * w4 + b1
Feed-forward path
0.0.05
0.1
TruthAnswer Error
?
?
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.40 (w5)
0.55 (w8)
0.45(w6)
0.50(w7)
?
?
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
Point
coordinates
as input
Input of the second neuron in
the hidden layer?
0.5933 out(h1)
input(h2) = 0.3925
input(h2) = 0.05 * 0.25 + 0.1 * 0.3 + 0.35 = 0.3925
input(h2) = x * w3 + y * w4 + b1
out(h2) = ?
out(h2) = sig(0.3925) = 0.5969
Output of the second neuron is
Feed-forward path
0.0.05
0.1
TruthAnswer Error
?
?
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.40 (w5)
0.55 (w8)
0.45(w6)
0.50(w7)
?
?
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
Point
coordinates
as input
Input of the second neuron in
the hidden layer?
0.5933 out(h1)
input(h2) = 0.3925
input(h2) = 0.05 * 0.25 + 0.1 * 0.3 + 0.35 = 0.3925
input(h2) = x * w3 + y * w4 + b1
out(h2) = sig(0.3925) = 0.5969
Output of the second neuron is
out(h2) = 0.5969
Feed-forward path
.05
0.1
TruthAnswer Error
?
?
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.40 (w5)
0.55 (w8)
0.45(w6)
0.50(w7)
?
?
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
0.00.0
Point
coordinates
as input
0.5933 out(h1)
0.5969 out(h2)
Feed-forward path
0.0.05
0.1
TruthAnswer Error
?
?
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.40 (w5)
0.55 (w8)
0.45(w6)
0.50(w7)
?
?
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
Point
coordinates
as input
0.5933 out(h1)
0.5969 out(h2)
input(o1) = ?
Feed-forward path
0.0.05
0.1
TruthAnswer Error
?
?
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.40 (w5)
0.55 (w8)
0.45(w6)
0.50(w7)
?
?
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
Point
coordinates
as input
0.5933 out(h1)
0.5969 out(h2)
input(o1) = ?
Input of the first neuron
in the output layer?
Feed-forward path
0.0.05
0.1
TruthAnswer Error
?
?
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.40 (w5)
0.55 (w8)
0.45(w6)
0.50(w7)
?
?
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
Point
coordinates
as input
0.5933 out(h1)
0.5969 out(h2)
input(o1) = ?
Input of the first neuron
in the output layer?
input(o1) = out(h1) * w5 + out(h2) * w6 + b2
Feed-forward path
0.0.05
0.1
TruthAnswer Error
?
?
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.40 (w5)
0.55 (w8)
0.45(w6)
0.50(w7)
?
?
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
Point
coordinates
as input
0.5933 out(h1)
0.5969 out(h2)
input(o1) = ?
Input of the first neuron
in the output layer?
input(o1) = out(h1) * w5 + out(h2) * w6 + b2
input(o1) = 0.5933 * 0.4 + 0.5969 * 0.45 + 0.6 = 1.105
Feed-forward path
0.0.05
0.1
TruthAnswer Error
?
?
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.40 (w5)
0.55 (w8)
0.45(w6)
0.50(w7)
?
?
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
Point
coordinates
as input
0.5933 out(h1)
0.5969 out(h2)
input(o1) = 1.105
Input of the first neuron
in the output layer?
input(o1) = out(h1) * w5 + out(h2) * w6 + b2
input(o1) = 0.5933 * 0.4 + 0.5969 * 0.45 + 0.6 = 1.105
Feed-forward path
0.0.05
0.1
TruthAnswer Error
?
?
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.40 (w5)
0.55 (w8)
0.45(w6)
0.50(w7)
?
?
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
Point
coordinates
as input
0.5933 out(h1)
0.5969 out(h2)
input(o1) = 1.105
Input of the first neuron
in the output layer?
input(o1) = out(h1) * w5 + out(h2) * w6 + b2
input(o1) = 0.5933 * 0.4 + 0.5969 * 0.45 + 0.6 = 1.105
out(o1) = sig(input(o1))
Output
out(o1) = ?
Feed-forward path
0.0.05
0.1
TruthAnswer Error
?
?
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.40 (w5)
0.55 (w8)
0.45(w6)
0.50(w7)
?
?
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
Point
coordinates
as input
0.5933 out(h1)
0.5969 out(h2)
input(o1) = 1.105
Input of the first neuron
in the output layer?
input(o1) = out(h1) * w5 + out(h2) * w6 + b2
input(o1) = 0.5933 * 0.4 + 0.5969 * 0.45 + 0.6 = 1.105
out(o1) = sig(1.105) = 0.7511
Output
out(o1) = 0.7511
Feed-forward path
.05
0.1
TruthAnswer Error
?
?
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.40 (w5)
0.55 (w8)
0.45(w6)
0.50(w7)
?
?
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
0.00.0
Point
coordinates
as input
0.5933 out(h1)
0.5969 out(h2)
0.7511 out(o1)
Feed-forward path
.05
0.1
TruthAnswer Error
?
?
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.40 (w5)
0.55 (w8)
0.45(w6)
0.50(w7)
?
?
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
0.00.0
Point
coordinates
as input
0.5933 out(h1)
0.5969 out(h2)
0.7511 out(o1)
0.7732 out(o2)
Feed-forward path
.05
0.1
TruthAnswer Error
0.751
0.773
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.40 (w5)
0.55 (w8)
0.45(w6)
0.50(w7)
?
?
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
0.00.0
Point
coordinates
as input
0.5933 out(h1)
0.5969 out(h2)
0.7511 out(o1)
0.7732 out(o2)
Feed-forward path
.05
0.1
TruthAnswer Error
0.751
0.773
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.40 (w5)
0.55 (w8)
0.45(w6)
0.50(w7)
?
?
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
0.00.0
Point
coordinates
as input
0.5933 out(h1)
0.5969 out(h2)
0.7511 out(o1)
0.7732 out(o2)
How good are these predictions?
Feed-forward path
.05
0.1
TruthAnswer Error
0.751
0.773
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.40 (w5)
0.55 (w8)
0.45(w6)
0.50(w7)
?
?
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
0.00.0
Point
coordinates
as input
0.5933 out(h1)
0.5969 out(h2)
0.7511 out(o1)
0.7732 out(o2)
In order to find out how good
are these predictions, we
should compare them to the
expected outcomes
.05
0.1
TruthAnswer Error
0.751
0.773
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.40 (w5)
0.55 (w8)
0.45(w6)
0.50(w7)
?
?
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
0.00.0
Point
coordinates
as input
0.5933 out(h1)
0.5969 out(h2)
0.7511 out(o1)
0.7732 out(o2)
In order to find out how good
are these predictions, we
should compare them to the
expected outcomes
Eo1
=
1
2
(truth − out(o1))2
.05
0.1
TruthAnswer Error
0.751
0.773
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.40 (w5)
0.55 (w8)
0.45(w6)
0.50(w7)
?
?
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
0.00.0
Point
coordinates
as input
0.5933 out(h1)
0.5969 out(h2)
0.7511 out(o1)
0.7732 out(o2)
In order to find out how good
are these predictions, we
should compare them to the
expected outcomes
Eo1
=
1
2
(truth − out(o1))2
=
1
2
(0.751 − 0)2
.05
0.1
TruthAnswer Error
0.751
0.773
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.40 (w5)
0.55 (w8)
0.45(w6)
0.50(w7)
?
?
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
0.00.0
Point
coordinates
as input
0.5933 out(h1)
0.5969 out(h2)
0.7511 out(o1)
0.7732 out(o2)
In order to find out how good
are these predictions, we
should compare them to the
expected outcomes
Eo1
=
1
2
(truth − out(o1))2
=
1
2
(0.751 − 0)2
=
1
2
(−0.751)2
.05
0.1
TruthAnswer Error
0.751
0.773
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.40 (w5)
0.55 (w8)
0.45(w6)
0.50(w7)
0.282
?
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
0.00.0
Point
coordinates
as input
0.5933 out(h1)
0.5969 out(h2)
0.7511 out(o1)
0.7732 out(o2)
In order to find out how good
are these predictions, we
should compare them to the
expected outcomes
Eo1
=
1
2
(truth − out(o1))2
=
1
2
(0.751 − 0)2
= 0.282
0.0.05
0.1
TruthAnswer Error
0.751
0.773
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.40 (w5)
0.55 (w8)
0.45(w6)
0.50(w7)
?
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
Point
coordinates
as input
0.5933 out(h1)
0.5969 out(h2)
0.7511 out(o1)
0.7732 out(o2)
In order to find out how good
are these predictions, we
should compare them to the
expected outcomes
0.282
Eo1
=
1
2
(truth − out(o1))2
=
1
2
(0.751 − 0)2
= 0.282
Eo2
=
1
2
(truth − out(o2))2
=
1
2
(0.773 − 1)2
= 0.025
0.0.05
0.1
TruthAnswer Error
0.751
0.773
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.40 (w5)
0.55 (w8)
0.45(w6)
0.50(w7)
0.025
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
Point
coordinates
as input
0.5933 out(h1)
0.5969 out(h2)
0.7511 out(o1)
0.7732 out(o2)
In order to find out how good
are these predictions, we
should compare them to the
expected outcomes
0.282
Eo1
=
1
2
(truth − out(o1))2
=
1
2
(0.751 − 0)2
= 0.282
Eo2
=
1
2
(truth − out(o2))2
=
1
2
(0.773 − 1)2
= 0.025
.05
0.1
TruthAnswer Error
0.751
0.773
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.40 (w5)
0.55 (w8)
0.45(w6)
0.50(w7)
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
0.00.0
Point
coordinates
as input
0.5933 out(h1)
0.5969 out(h2)
0.7511 out(o1)
0.7732 out(o2)
How good are these predictions?
0.025
0.282
.05
0.1
TruthAnswer Error
0.751
0.773
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.40 (w5)
0.55 (w8)
0.45(w6)
0.50(w7)
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
0.00.0
Point
coordinates
as input
0.5933 out(h1)
0.5969 out(h2)
0.7511 out(o1)
0.7732 out(o2)
How good are these predictions?
0.025
0.282
Well, they are this good.
.05
0.1
TruthAnswer Error
0.751
0.773
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.40 (w5)
0.55 (w8)
0.45(w6)
0.50(w7)
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
0.00.0
Point
coordinates
as input
0.5933 out(h1)
0.5969 out(h2)
0.7511 out(o1)
0.7732 out(o2)
0.025
0.282
Overall the error is: Etotal = Eo1
+ Eo2
.05
0.1
TruthAnswer Error
0.751
0.773
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.40 (w5)
0.55 (w8)
0.45(w6)
0.50(w7)
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
0.00.0
Point
coordinates
as input
0.5933 out(h1)
0.5969 out(h2)
0.7511 out(o1)
0.7732 out(o2)
0.025
0.282
Overall the error is: Etotal = Eo1
+ Eo2
= 0.282 + 0.025 = 0.307
Etotal = 0.307
.05
0.1
TruthAnswer Error
0.751
0.773
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.40 (w5)
0.55 (w8)
0.45(w6)
0.50(w7)
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
0.00.0
Point
coordinates
as input
0.5933 out(h1)
0.5969 out(h2)
0.7511 out(o1)
0.7732 out(o2)
0.025
0.282
Our goal is to reduce the total error ( )!Etotal
Etotal = 0.307
0.0.05
0.1
TruthAnswer Error
0.751
0.773
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.40 (w5)
0.55 (w8)
0.45(w6)
0.50(w7)
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
Point
coordinates
as input
0.5933 out(h1)
0.5969 out(h2)
0.7511 out(o1)
0.7732 out(o2)
0.025
0.282
Etotal = 0.307
Our goal is to reduce the total error ( )!Etotal
The only thing we can do is change weights
0.40 (w5)
0.0.05
0.1
TruthAnswer Error
0.751
0.773
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.55 (w8)
0.45(w6)
0.50(w7)
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
Point
coordinates
as input
0.5933 out(h1)
0.5969 out(h2)
0.7511 out(o1)
0.7732 out(o2)
0.025
0.282
Etotal = 0.307
How does w5 influence the total error?
0.0.05
0.1
TruthAnswer Error
0.751
0.773
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
(w5)
0.55 (w8)
0.45(w6)
0.50(w7)
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
Point
coordinates
as input
0.5933 out(h1)
0.5969 out(h2)
0.7511 out(o1)
0.7732 out(o2)
0.025
0.282
Etotal = 0.307
0.45
0.50
0.35
0.30
We can change it randomly
(increase or decrease) and
see what happens to
How does w5 influence the total error?
0.55
0.25
Etotal
0.40
0.0.05
0.1
TruthAnswer Error
0.757
0.773
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
(w5)
0.55 (w8)
0.45(w6)
0.50(w7)
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
Point
coordinates
as input
0.5933 out(h1)
0.5969 out(h2)
0.7511 out(o1)
0.7732 out(o2)
0.025
0.286
Etotal = 0.311
0.45
0.50
0.35
0.30
We can change it randomly
(increase or decrease) and
see what happens to
How does w5 influence the total error?
0.55
0.25
Etotal
0.40
0.0.05
0.1
TruthAnswer Error
0.751
0.773
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
(w5)
0.55 (w8)
0.45(w6)
0.50(w7)
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
Point
coordinates
as input
0.5933 out(h1)
0.5969 out(h2)
0.7511 out(o1)
0.7732 out(o2)
0.025
0.282
Etotal = 0.307
0.45
0.50
0.35
0.30
We can change it randomly
(increase or decrease) and
see what happens to
How does w5 influence the total error?
0.55
0.25
Etotal
0.40
0.0.05
0.1
TruthAnswer Error
0.745
0.773
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
(w5)
0.55 (w8)
0.45(w6)
0.50(w7)
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
Point
coordinates
as input
0.5933 out(h1)
0.5969 out(h2)
0.7511 out(o1)
0.7732 out(o2)
0.025
0.278
Etotal = 0.303
0.45
0.50
0.35
0.30
We can change it randomly
(increase or decrease) and
see what happens to
How does w5 influence the total error?
0.55
0.25
Etotal
0.40
0.0.05
0.1
TruthAnswer Error
0.740
0.773
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
(w5)
0.55 (w8)
0.45(w6)
0.50(w7)
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
Point
coordinates
as input
0.5933 out(h1)
0.5969 out(h2)
0.7511 out(o1)
0.7732 out(o2)
0.025
0.273
Etotal = 0.299
0.45
0.50
0.35
0.30
We can change it randomly
(increase or decrease) and
see what happens to
How does w5 influence the total error?
0.55
0.25
Etotal
0.40
0.0.05
0.1
TruthAnswer Error
0.773
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
(w5)
0.55 (w8)
0.45(w6)
0.50(w7)
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
Point
coordinates
as input
0.5933 out(h1)
0.5969 out(h2)
0.7511 out(o1)
0.7732 out(o2)
0.025
Etotal = 0.299
0.45
0.50
0.35
0.30
We can change it randomly
(increase or decrease) and
see what happens to
How does w5 influence the total error?
0.55
0.25
Etotal
0.40
w5
EtotalEtotal
Higher weightLower weight
0.740 0.273
0.0.05
0.1
TruthAnswer Error
0.773
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
(w5)
0.55 (w8)
0.45(w6)
0.50(w7)
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
Point
coordinates
as input
0.5933 out(h1)
0.5969 out(h2)
0.7511 out(o1)
0.7732 out(o2)
0.025
Etotal = 0.299
0.45
0.50
0.35
0.30
We can change it randomly
(increase or decrease) and
see what happens to
How does w5 influence the total error?
0.55
0.25
Etotal
0.40
w5
EtotalEtotal
Higher weightLower weight
The problem is that every time we
need to recalculate all the neuron
values again.
0.740 0.273
0.0.05
0.1
TruthAnswer Error
0.773
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
(w5)
0.55 (w8)
0.45(w6)
0.50(w7)
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
Point
coordinates
as input
0.5933 out(h1)
0.5969 out(h2)
0.7511 out(o1)
0.7732 out(o2)
0.025
Etotal = 0.299
0.45
0.50
0.35
0.30
0.55
0.25
0.40
0.740 0.273
0.0.05
0.1
TruthAnswer Error
0.773
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
(w5)
0.55 (w8)
0.45(w6)
0.50(w7)
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
Point
coordinates
as input
0.5933 out(h1)
0.5969 out(h2)
0.7511 out(o1)
0.7732 out(o2)
0.025
Etotal = 0.299
0.45
0.50
0.35
0.30
0.55
0.25
0.40
0.740 0.273
0.6 (b2)0.60
0.55
0.35 (b1)0.35
0.30
0.20(w2)
0.250.200.15
0.25(w3)
0.50(w7)
0.30
0.25
0.20
0.55
0.50
0.45
0.45(w6)
0.500.450.40
0.0.05
0.1
TruthAnswer Error
0.773
Hidden
layer
0.15 (w1)
0.30 (w4)
(w5)
0.55 (w8)
b b
x
y
0
0.0
(0.1,0.05)
1.0
Point
coordinates
as input
0.5933 out(h1)
0.5969 out(h2)
0.7511 out(o1)
0.7732 out(o2)
0.025
Etotal = 0.299
0.35
0.30
0.25
0.740 0.273
0.20
0.10
0.15
0.60
0.50
0.55
0.35
0.25
0.30
0.650.40
Also tweaking each weight separately
would be a nightmare
0.6 (b2)0.60
0.55
0.35 (b1)0.35
0.30
0.20(w2)
0.250.200.15
0.25(w3)
0.50(w7)
0.30
0.25
0.20
0.55
0.50
0.45
0.45(w6)
0.500.450.40
0.0.05
0.1
TruthAnswer Error
0.773
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
(w5)
0.55 (w8)
b b
x
y
0
0.0
(0.1,0.05)
1.0
Point
coordinates
as input
0.5933 out(h1)
0.5969 out(h2)
0.7511 out(o1)
0.7732 out(o2)
0.025
Etotal = 0.299
0.35
0.30
0.25
0.740 0.273
0.20
0.10
0.15
0.60
0.50
0.55
0.35
0.25
0.30
0.650.40
Also tweaking each weight separately
would be a nightmare
weight
Etotal
weight + 1
We want a way to efficiently update all our weights so
that total error decreases most substantially
weight - 1
weight
Etotal
weight + 1
We want a way to efficiently update all our weights so
that total error decreases most substantially
We want to compute the gradient
weight - 1
weight
Etotal
weight + 1
We want to compute the gradient
Gradient shows the direction of the greatest increase of the
function
weight - 1
weight
Etotal
weight + 1
We want to compute the gradient
∂E
∂w
Gradient shows the direction of the greatest increase of the
function
weight - 1
weight
Etotal
weight + 1
We want to compute the gradient
−
∂E
∂w
weight - 1
∂E
∂w
Gradient shows the direction of the greatest increase of the
function
w5
Etotal
weight + 1
It looks OK when there is just one weight to take care of
We want to compute the gradient
−
∂E
∂w
∂E
∂w
Etotal
weight + 1
∂E
∂w−
∂E
∂w
w5
It starts to look a lot scarier when we try to optimise all
weights
w4
w3
w2
w1
w6 w7
w8
Etotal
weight + 1
∂E
∂w
w5
Gradient still can be computed efficiently using
backpropagation algorithm
It starts to look a lot scarier when we try to optimise all
weights
w4
w3
w2
w1
w6 w7
w8
−
∂E
∂w
0.40 (w5)
0.0.05
0.1
TruthAnswer Error
0.751
0.773
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.55 (w8)
0.45(w6)
0.50(w7)
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
Point
coordinates
as input
0.5933 out(h1)
0.5969 out(h2)
0.7511 out(o1)
0.7732 out(o2)
0.025
0.282
Etotal = 0.307
How does w5 influence the total error?
Backpropagation
algorithm
0.40 (w5)
0.0.05
0.1
TruthAnswer Error
0.751
0.773
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.55 (w8)
0.45(w6)
0.50(w7)
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
Point
coordinates
as input
0.5933 out(h1)
0.5969 out(h2)
0.7511 out(o1)
0.7732 out(o2)
0.025
0.282
Etotal = 0.307
How does w5 influence the total error?
∂Etotal
∂w5
= . . .
Backpropagation
algorithm
∂Etotal
∂w5
=
∂Etotal
∂Eo1
* . . .
0.40 (w5)
0.0.05
0.1
TruthAnswer Error
0.751
0.773
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.55 (w8)
0.45(w6)
0.50(w7)
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
Point
coordinates
as input
0.5933 out(h1)
0.5969 out(h2)
0.7511 out(o1)
0.7732 out(o2)
0.025
0.282
Etotal = 0.307
How does w5 influence the total error?
How does error on out(o1) influence the total error?
Backpropagation
algorithm
∂Etotal
∂w5
=
∂Etotal
∂Eo1
*
∂Eo1
∂outo1
* . . .
0.40 (w5)
0.0.05
0.1
TruthAnswer Error
0.751
0.773
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.55 (w8)
0.45(w6)
0.50(w7)
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
Point
coordinates
as input
0.5933 out(h1)
0.5969 out(h2)
0.7511 out(o1)
0.7732 out(o2)
0.025
0.282
Etotal = 0.307
How does w5 influence the total error?
How does out(o1) influence the error on out(o1)?
Backpropagation
algorithm
∂Etotal
∂w5
=
∂Etotal
∂Eo1
*
∂Eo1
∂outo1
*
∂outo1
∂ino1
* . . .
0.40 (w5)
0.0.05
0.1
TruthAnswer Error
0.751
0.773
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.55 (w8)
0.45(w6)
0.50(w7)
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
Point
coordinates
as input
0.5933 out(h1)
0.5969 out(h2)
0.7511 out(o1)
0.7732 out(o2)
0.025
0.282
Etotal = 0.307
How does w5 influence the total error?
How does input(o1) influence the out(o1)?
Backpropagation
algorithm
∂Etotal
∂w5
=
∂Etotal
∂Eo1
*
∂Eo1
∂outo1
*
∂outo1
∂ino1
*
∂ino1
∂w5
0.40 (w5)
0.0.05
0.1
TruthAnswer Error
0.751
0.773
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.55 (w8)
0.45(w6)
0.50(w7)
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
Point
coordinates
as input
0.5933 out(h1)
0.5969 out(h2)
0.7511 out(o1)
0.7732 out(o2)
0.025
0.282
Etotal = 0.307
How does w5 influence the total error?
How does w5 influence the input(o1)?
Backpropagation
algorithm
∂Etotal
∂w5
=
∂Etotal
∂Eo1
*
∂Eo1
∂outo1
*
∂outo1
∂ino1
*
∂ino1
∂w5
∂Etotal
∂Eo1
Etotal = Eo1
+ Eo2
0.40 (w5)
0.0.05
0.1
TruthAnswer Error
0.751
0.773
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.55 (w8)
0.45(w6)
0.50(w7)
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
Point
coordinates
as input
0.5969 out(h2)
0.7511 out(o1)
0.7732 out(o2)
0.025
0.282
Etotal = 0.307
How does w5 influence the total error?
0.5933 out(h1)
Backpropagation
algorithm
∂Etotal
∂w5
=
∂Etotal
∂Eo1
*
∂Eo1
∂outo1
*
∂outo1
∂ino1
*
∂ino1
∂w5
∂Etotal
∂Eo1
∂Etotal
∂Eo1
= 1
Etotal = Eo1
+ Eo2
0.40 (w5)
0.0.05
0.1
TruthAnswer Error
0.751
0.773
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.55 (w8)
0.45(w6)
0.50(w7)
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
Point
coordinates
as input
0.5933 out(h1)
0.5969 out(h2)
0.7511 out(o1)
0.7732 out(o2)
0.025
0.282
Etotal = 0.307
How does w5 influence the total error?
0.5933 out(h1)
Backpropagation
algorithm
∂Etotal
∂w5
= 1 *
∂Eo1
∂outo1
*
∂outo1
∂ino1
*
∂ino1
∂w5
∂Etotal
∂Eo1
= 1
Etotal = Eo1
+ Eo2
0.40 (w5)
0.0.05
0.1
TruthAnswer Error
0.751
0.773
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.55 (w8)
0.45(w6)
0.50(w7)
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
Point
coordinates
as input
0.5933 out(h1)
0.5969 out(h2)
0.7511 out(o1)
0.7732 out(o2)
0.025
0.282
Etotal = 0.307
How does w5 influence the total error?
0.5933 out(h1)
Backpropagation
algorithm
∂Etotal
∂w5
= 1 *
∂Eo1
∂outo1
*
∂outo1
∂ino1
*
∂ino1
∂w5
∂Eo1
∂outo1
0.40 (w5)
0.0.05
0.1
TruthAnswer Error
0.751
0.773
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.55 (w8)
0.45(w6)
0.50(w7)
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
Point
coordinates
as input
0.5933 out(h1)
0.5969 out(h2)
0.7511 out(o1)
0.7732 out(o2)
0.025
0.282
Etotal = 0.307
How does w5 influence the total error?
0.5933 out(h1)
Backpropagation
algorithm
∂Etotal
∂w5
= 1 *
∂Eo1
∂outo1
*
∂outo1
∂ino1
*
∂ino1
∂w5
∂Eo1
∂outo1
Eo1
=
1
2
(truth − outo1
)2
0.40 (w5)
0.0.05
0.1
TruthAnswer Error
0.751
0.773
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.55 (w8)
0.45(w6)
0.50(w7)
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
Point
coordinates
as input
0.5933 out(h1)
0.5969 out(h2)
0.7511 out(o1)
0.7732 out(o2)
0.025
0.282
Etotal = 0.307
How does w5 influence the total error?
∂Eo1
∂outo1
= . . .
0.5933 out(h1)
Backpropagation
algorithm
∂Etotal
∂w5
= 1 *
∂Eo1
∂outo1
*
∂outo1
∂ino1
*
∂ino1
∂w5
∂Eo1
∂outo1
Eo1
=
1
2
(truth − outo1
)2
∂Eo1
∂outo1
= 2 *
1
2
(truth − outo1
) * (−1) = . . .
0.40 (w5)
0.0.05
0.1
TruthAnswer Error
0.751
0.773
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.55 (w8)
0.45(w6)
0.50(w7)
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
Point
coordinates
as input
0.5933 out(h1)
0.5969 out(h2)
0.7511 out(o1)
0.7732 out(o2)
0.025
0.282
Etotal = 0.307
How does w5 influence the total error?
0.5933 out(h1)
Backpropagation
algorithm
∂Etotal
∂w5
= 1 *
∂Eo1
∂outo1
*
∂outo1
∂ino1
*
∂ino1
∂w5
∂Eo1
∂outo1
Eo1
=
1
2
(truth − outo1
)2
∂Eo1
∂outo1
= 2 *
1
2
(truth − outo1
) * (−1) = 0.751 − 0 = 0.751
0.40 (w5)
0.0.05
0.1
TruthAnswer Error
0.751
0.773
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.55 (w8)
0.45(w6)
0.50(w7)
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
Point
coordinates
as input
0.5933 out(h1)
0.5969 out(h2)
0.7511 out(o1)
0.7732 out(o2)
0.025
0.282
Etotal = 0.307
How does w5 influence the total error?
0.5933 out(h1)
Backpropagation
algorithm
∂Etotal
∂w5
= 1 * 0.751 *
∂outo1
∂ino1
*
∂ino1
∂w5
∂outo1
∂ino1
∂outo1
∂ino1
=
sig(x) =
1
1 + e−x
0.40 (w5)
0.0.05
0.1
TruthAnswer Error
0.751
0.773
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.55 (w8)
0.45(w6)
0.50(w7)
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
Point
coordinates
as input
0.5933 out(h1)
0.5969 out(h2)
0.7511 out(o1)
0.7732 out(o2)
0.025
0.282
Etotal = 0.307
How does w5 influence the total error?
0.5933 out(h1)
Backpropagation
algorithm
∂outo1
∂ino1
= outo1
* (1 − outo1
) = 0.751 * (1 − 0.751) = 0.186
∂Etotal
∂w5
= 1 * 0.751 *
∂outo1
∂ino1
*
∂ino1
∂w5
∂outo1
∂ino1
sig(x) =
1
1 + e−x
0.40 (w5)
0.0.05
0.1
TruthAnswer Error
0.751
0.773
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.55 (w8)
0.45(w6)
0.50(w7)
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
Point
coordinates
as input
0.5933 out(h1)
0.5969 out(h2)
0.7511 out(o1)
0.7732 out(o2)
0.025
0.282
Etotal = 0.307
How does w5 influence the total error?
0.5933 out(h1)
Backpropagation
algorithm
∂Etotal
∂w5
= 1 * 0.751 * 0.186 *
∂ino1
∂w5
∂ino1
∂w5
∂ino1
∂w5
=
in01
= . . .
0.40 (w5)
0.0.05
0.1
TruthAnswer Error
0.751
0.773
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.55 (w8)
0.45(w6)
0.50(w7)
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
Point
coordinates
as input
0.5933 out(h1)
0.5969 out(h2)
0.7511 out(o1)
0.7732 out(o2)
0.025
0.282
Etotal = 0.307
How does w5 influence the total error?
0.5933 out(h1)
Backpropagation
algorithm
∂Etotal
∂w5
= 1 * 0.751 * 0.186 *
∂ino1
∂w5
∂ino1
∂w5
∂ino1
∂w5
=
in01
= outh1
* w5 + outh2
* w6 + b2
0.40 (w5)
0.0.05
0.1
TruthAnswer Error
0.751
0.773
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.55 (w8)
0.45(w6)
0.50(w7)
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
Point
coordinates
as input
0.5933 out(h1)
0.5969 out(h2)
0.7511 out(o1)
0.7732 out(o2)
0.025
0.282
Etotal = 0.307
How does w5 influence the total error?
0.5933 out(h1)
Backpropagation
algorithm
∂ino1
∂w5
= outh1
= . . .
∂Etotal
∂w5
= 1 * 0.751 * 0.186 *
∂ino1
∂w5
∂ino1
∂w5
in01
= outh1
* w5 + outh2
* w6 + b2
0.40 (w5)
0.0.05
0.1
TruthAnswer Error
0.751
0.773
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.55 (w8)
0.45(w6)
0.50(w7)
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
Point
coordinates
as input
0.5933 out(h1)
0.5969 out(h2)
0.7511 out(o1)
0.7732 out(o2)
0.025
0.282
Etotal = 0.307
How does w5 influence the total error?
0.5933 out(h1)
Backpropagation
algorithm
∂ino1
∂w5
= outh1
= 0.5933
∂Etotal
∂w5
= 1 * 0.751 * 0.186 *
∂ino1
∂w5
∂ino1
∂w5
in01
= outh1
* w5 + outh2
* w6 + b2
0.40 (w5)
0.0.05
0.1
TruthAnswer Error
0.751
0.773
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.55 (w8)
0.45(w6)
0.50(w7)
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
Point
coordinates
as input
0.5933 out(h1)
0.5969 out(h2)
0.7511 out(o1)
0.7732 out(o2)
0.025
0.282
Etotal = 0.307
How does w5 influence the total error?
0.5933 out(h1)
Backpropagation
algorithm
∂Etotal
∂w5
= 1 * 0.751 * 0.186 * 0.5933
0.40 (w5)
0.0.05
0.1
TruthAnswer Error
0.751
0.773
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.55 (w8)
0.45(w6)
0.50(w7)
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
Point
coordinates
as input
0.5933 out(h1)
0.5969 out(h2)
0.7511 out(o1)
0.7732 out(o2)
0.025
0.282
Etotal = 0.307
How does w5 influence the total error?
0.5933 out(h1)
Backpropagation
algorithm
∂Etotal
∂w5
= 1 * 0.751 * 0.186 * 0.5933 = 0.082
0.40 (w5)
0.0.05
0.1
TruthAnswer Error
0.751
0.773
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.55 (w8)
0.45(w6)
0.50(w7)
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
Point
coordinates
as input
0.5933 out(h1)
0.5969 out(h2)
0.7511 out(o1)
0.7732 out(o2)
0.025
0.282
Etotal = 0.307
How does w5 influence the total error?
Backpropagation
algorithm 0.5933 out(h1)
If value of the gradient is positive, then if we increase the
w5, the total error will …?
∂Etotal
∂w5
= 1 * 0.751 * 0.186 * 0.5933 = 0.082
0.40 (w5)
0.0.05
0.1
TruthAnswer Error
0.751
0.773
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.55 (w8)
0.45(w6)
0.50(w7)
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
Point
coordinates
as input
0.5933 out(h1)
0.5969 out(h2)
0.7511 out(o1)
0.7732 out(o2)
0.025
0.282
Etotal = 0.307
How does w5 influence the total error?
Backpropagation
algorithm 0.5933 out(h1)
If value of the gradient is positive, then if we increase the
w5, the total error will increase!
∂Etotal
∂w5
= 1 * 0.751 * 0.186 * 0.5933 = 0.082
0.40 (w5)
0.0.05
0.1
TruthAnswer Error
0.751
0.773
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.55 (w8)
0.45(w6)
0.50(w7)
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
Point
coordinates
as input
0.5933 out(h1)
0.5969 out(h2)
0.7511 out(o1)
0.7732 out(o2)
0.025
0.282
Etotal = 0.307
How does w5 influence the total error?
Backpropagation
algorithm 0.5933 out(h1)
w5new
= w5old
−
∂Etotal
∂w5
So we update w5:
If value of the gradient is positive, then if we increase the
w5, the total error will increase!
∂Etotal
∂w5
= 1 * 0.751 * 0.186 * 0.5933 = 0.082
0.40 (w5)
0.0.05
0.1
TruthAnswer Error
0.751
0.773
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.55 (w8)
0.45(w6)
0.50(w7)
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
Point
coordinates
as input
0.5933 out(h1)
0.5969 out(h2)
0.7511 out(o1)
0.7732 out(o2)
0.025
0.282
Etotal = 0.307
How does w5 influence the total error?
Backpropagation
algorithm 0.5933 out(h1)
w5new
= w5old
− 0.082
So we update w5:
If value of the gradient is positive, then if we increase the
w5, the total error will increase!
∂Etotal
∂w5
= 1 * 0.751 * 0.186 * 0.5933 = 0.082
0.40 (w5)
0.0.05
0.1
TruthAnswer Error
0.751
0.773
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.55 (w8)
0.45(w6)
0.50(w7)
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
Point
coordinates
as input
0.5933 out(h1)
0.5969 out(h2)
0.7511 out(o1)
0.7732 out(o2)
0.025
0.282
Etotal = 0.307
How does w5 influence the total error?
Backpropagation
algorithm 0.5933 out(h1)
So we update w5:
If update by a lot,
we can miss to optimum
w5
Etotal
If value of the gradient is positive, then if we increase the
w5, the total error will increase!
w5new
= w5old
− 0.082
∂Etotal
∂w5
= 1 * 0.751 * 0.186 * 0.5933 = 0.082
0.40 (w5)
0.0.05
0.1
TruthAnswer Error
0.751
0.773
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.55 (w8)
0.45(w6)
0.50(w7)
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
Point
coordinates
as input
0.5933 out(h1)
0.5969 out(h2)
0.7511 out(o1)
0.7732 out(o2)
0.025
0.282
Etotal = 0.307
How does w5 influence the total error?
Backpropagation
algorithm 0.5933 out(h1)
So we update w5:
Thus, we make a bit smaller step
w5
Etotal
step size
If value of the gradient is positive, then if we increase the
w5, the total error will increase!
w5new
= w5old
− η * 0.082
w5new
= 0.4 − 0.5 * 0.083
∂Etotal
∂w5
= 1 * 0.751 * 0.186 * 0.5933 = 0.082
0.40 (w5)
0.0.05
0.1
TruthAnswer Error
0.751
0.773
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.55 (w8)
0.45(w6)
0.50(w7)
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
Point
coordinates
as input
0.5933 out(h1)
0.5969 out(h2)
0.7511 out(o1)
0.7732 out(o2)
0.025
0.282
Etotal = 0.307
How does w5 influence the total error?
Backpropagation
algorithm 0.5933 out(h1)
So we update w5:
Thus, we make a bit smaller step
w5
Etotal
If value of the gradient is positive, then if we increase the
w5, the total error will increase!
∂Etotal
∂w5
= 1 * 0.751 * 0.186 * 0.5933 = 0.082
0.40 (w5)
0.0.05
0.1
TruthAnswer Error
0.751
0.773
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.55 (w8)
0.45(w6)
0.50(w7)
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
Point
coordinates
as input
0.5933 out(h1)
0.5969 out(h2)
0.7511 out(o1)
0.7732 out(o2)
0.025
0.282
Etotal = 0.307
How does w5 influence the total error?
Backpropagation
algorithm 0.5933 out(h1)
So we update w5:
Thus, we make a bit smaller step
w5
Etotal
If value of the gradient is positive, then if we increase the
w5, the total error will increase!
w5new
= 0.4 − 0.5 * 0.083 = 0.3585
0.359 (w5)
0.0.05
0.1
TruthAnswer Error
0.773
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.55 (w8)
0.45(w6)
0.50(w7)
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
Point
coordinates
as input
0.5933 out(h1)
0.5969 out(h2)
out(o1)
0.7732 out(o2)
0.025
Etotal = . . .
How does w5 influence the total error?
Backpropagation
algorithm 0.5933 out(h1)
So we update w5:
Thus, we make a bit smaller step
w5
Etotal
w5new
= 0.4 − 0.5 * 0.083 = 0.3585
0.0.05
0.1
TruthAnswer Error
0.611
0.773
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.55 (w8)
0.45(w6)
0.50(w7)
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
Point
coordinates
as input
0.5933 out(h1)
0.5969 out(h2)
0.6112 out(o1)
0.7732 out(o2)
0.025
How does w5 influence the total error?
Backpropagation
algorithm 0.5933 out(h1)
So we update w5:
Thus, we make a bit smaller step
w5
Etotal
Etotal = . . .
0.359 (w5)
w5new
= 0.4 − 0.5 * 0.083 = 0.3585
0.0.05
0.1
TruthAnswer Error
0.611
0.773
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.55 (w8)
0.45(w6)
0.50(w7)
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
Point
coordinates
as input
0.5933 out(h1)
0.5969 out(h2) 0.7732 out(o2)
0.025
0.187
How does w5 influence the total error?
Backpropagation
algorithm 0.5933 out(h1)
So we update w5:
Thus, we make a bit smaller step
w5
Etotal
Etotal = . . .
0.6112 out(o1)
0.359 (w5)
w5new
= 0.4 − 0.5 * 0.083 = 0.3585
Etotal = 0.303
0.2780.745 0.0.05
0.1
TruthAnswer Error
0.773
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.55 (w8)
0.45(w6)
0.50(w7)
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
Point
coordinates
as input
0.5933 out(h1)
0.5969 out(h2) 0.7732 out(o2)
0.025
How does w5 influence the total error?
Backpropagation
algorithm 0.5933 out(h1)
So we update w5:
Thus, we make a bit smaller step
w5
Etotal
0.6112 out(o1)
0.359 (w5)
w5new
= 0.4 − 0.5 * 0.083 = 0.3585
0.0.05
0.1
TruthAnswer Error
0.773
Hidden
layer
Output
layer
0.15 (w1)
0.30 (w4)
0.20(w2)
0.25(w3)
0.55 (w8)
0.45(w6)
0.50(w7)
b
0.35 (b1)
b
0.6 (b2)
x
y
0
0.0
(0.1,0.05)
1.0
Point
coordinates
as input
0.5933 out(h1)
0.5969 out(h2) 0.7732 out(o2)
0.025
How does w5 influence the total error?
Backpropagation
algorithm 0.5933 out(h1) 0.6112 out(o1)
After updating all remaining weights total error = 0.131
Repeating 1000 times decreases it to 0.001
Etotal = 0.303
0.2780.745
0.359 (w5)
Why we should understand backpropagation?
https://medium.com/@karpathy/yes-you-should-understand-backprop-e2f06eab496b
Why we should understand backpropagation?
-8 -6 -4 -2 2 4 6 8
0.2
0.4
0.6
0.8
1.0
x
sig(x)
Vanishing gradients on sigmoids
https://medium.com/@karpathy/yes-you-should-understand-backprop-e2f06eab496b
Why we should understand backpropagation?
-8 -6 -4 -2 2 4 6 8
0.2
0.4
0.6
0.8
1.0
x
sig(x)
-8 -6 -4 -2 2 4 6 8
0.2
0.4
0.6
0.8
1.0
x
sig′(x)
Vanishing gradients on sigmoids
https://medium.com/@karpathy/yes-you-should-understand-backprop-e2f06eab496b
Why we should understand backpropagation?
-8 -6 -4 -2 2 4 6 8
0.2
0.4
0.6
0.8
1.0
x
sig(x)
-8 -6 -4 -2 2 4 6 8
0.2
0.4
0.6
0.8
1.0
x
sig′(x)
Vanishing gradients on sigmoids
Derivative is zero at tails
https://medium.com/@karpathy/yes-you-should-understand-backprop-e2f06eab496b
Why we should understand backpropagation?
-8 -6 -4 -2 2 4 6 8
0.2
0.4
0.6
0.8
1.0
x
sig(x)
-8 -6 -4 -2 2 4 6 8
0.2
0.4
0.6
0.8
1.0
x
sig′(x)
Vanishing gradients on sigmoids
Derivative is zero at tails
https://medium.com/@karpathy/yes-you-should-understand-backprop-e2f06eab496b
Recall the chain rule:
∂Etotal
∂w5
=
∂Etotal
∂Eo1
*
∂Eo1
∂outo1
*
∂outo1
∂ino1
*
∂ino1
∂w5
Why we should understand backpropagation?
-8 -6 -4 -2 2 4 6 8
0.2
0.4
0.6
0.8
1.0
x
sig(x)
-8 -6 -4 -2 2 4 6 8
0.2
0.4
0.6
0.8
1.0
x
sig′(x)
Vanishing gradients on sigmoids
Derivative is zero at tails
https://medium.com/@karpathy/yes-you-should-understand-backprop-e2f06eab496b
Recall the chain rule:
∂Etotal
∂w5
=
∂Etotal
∂Eo1
*
∂Eo1
∂outo1
*
∂outo1
∂ino1
*
∂ino1
∂w5
So, if x is too high (e.g. >8) or too low (e.g. < -8) then
∂outo1
∂ino1
= 0 and thus
∂Etotal
∂w5
= 0
Why we should understand backpropagation?
-8 -6 -4 -2 2 4 6 8
0.2
0.4
0.6
0.8
1.0
x
sig(x)
-8 -6 -4 -2 2 4 6 8
0.2
0.4
0.6
0.8
1.0
x
sig′(x)
Vanishing gradients on sigmoids
2 4 6 8-8 -6 -4 -2
2
4
6
8
10
x
Dying ReLus
max(0, x)
Derivative is zero at tails
https://medium.com/@karpathy/yes-you-should-understand-backprop-e2f06eab496b
Why we should understand backpropagation?
-8 -6 -4 -2 2 4 6 8
0.2
0.4
0.6
0.8
1.0
x
sig(x)
-8 -6 -4 -2 2 4 6 8
0.2
0.4
0.6
0.8
1.0
x
sig′(x)
Vanishing gradients on sigmoids
2 4 6 8-8 -6 -4 -2
2
4
6
8
10
x
Dying ReLus
max(0, x)
2 4 6 8-8 -6 -4 -2
0.2
0.4
0.6
0.8
1.0
x
max′(0, x)
Derivative is zero at tails
https://medium.com/@karpathy/yes-you-should-understand-backprop-e2f06eab496b
Why we should understand backpropagation?
-8 -6 -4 -2 2 4 6 8
0.2
0.4
0.6
0.8
1.0
x
sig(x)
-8 -6 -4 -2 2 4 6 8
0.2
0.4
0.6
0.8
1.0
x
sig′(x)
Vanishing gradients on sigmoids
2 4 6 8-8 -6 -4 -2
2
4
6
8
10
x
Dying ReLus
max(0, x)
2 4 6 8-8 -6 -4 -2
0.2
0.4
0.6
0.8
1.0
x
max′(0, x)
Derivative is zero at tails
Derivative is zero
https://medium.com/@karpathy/yes-you-should-understand-backprop-e2f06eab496b
x
y
0.0
0.00
Now let’s practice
(0.1,0.05)
x
y
-0.8
0.2
-0.6
-0.4
-0.2
0.0
0.4
0.6
-0.75 -0.50 -0.25 0.00 0.25 0.50 0.75 1.00
Now let’s practice
x
y
-0.8
0.2
-0.6
-0.4
-0.2
0.0
0.4
0.6
-0.75 -0.50 -0.25 0.00 0.25 0.50 0.75 1.00
Now let’s practice
x
y
-0.8
0.2
-0.6
-0.4
-0.2
0.0
0.4
0.6
-0.75 -0.50 -0.25 0.00 0.25 0.50 0.75 1.00
Now let’s practice
https://goo.gl/cX2dGd
http://www.emergentmind.com/neural-network
Training Neural Networks
(part I)
http://playground.tensorflow.org/
Training Neural Networks
(part II)
Resources:
Brandon Rohrer’s youtube video: How Convolutional Neural Networks
work (https://youtu.be/FmpDIaiMIeA)
Brandon Rohrer’s youtube video: How Deep Neural Networks Work
(https://youtu.be/ILsA4nyG7I0)
Stanford University CS231n Convolutional Neural Networks for Visual
Recognition (github.io version): http://cs231n.github.io/
Matt Mazur’s: A Step by Step Backpropagation Example (https://
mattmazur.com/2015/03/17/a-step-by-step-backpropagation-example/)
Andrej Karphaty’s blog post: Yes you should understand backprop
(https://medium.com/@karpathy/yes-you-should-understand-backprop-
e2f06eab496b)
Raul Vicente’s lecture: From brain to Deep Learning and back (https://
www.uttv.ee/naita?id=23585)
Mayank Agarwal’s blog: Back Propagation in Convolutional Neural
Networks — Intuition and Code (https://becominghuman.ai/back-
propagation-in-convolutional-neural-networks-intuition-and-
code-714ef1c38199)
Mordor riddle
http://fouryears.eu/2009/11/21/machine-learning-in-mordor/
https://biit.cs.ut.ee/
Thank you!
Introduction to Deep Learning

Weitere ähnliche Inhalte

Ähnlich wie Introduction to Deep Learning

Deep Learning - Overview of my work II
Deep Learning - Overview of my work IIDeep Learning - Overview of my work II
Deep Learning - Overview of my work IIMohamed Loey
 
Deep learning simplified
Deep learning simplifiedDeep learning simplified
Deep learning simplifiedLovelyn Rose
 
A New Algorithm for Digital Colour Image Encryption and Decryption
A New Algorithm for Digital Colour Image Encryption and DecryptionA New Algorithm for Digital Colour Image Encryption and Decryption
A New Algorithm for Digital Colour Image Encryption and DecryptionIRJET Journal
 
An introduction to Deep Learning with Apache MXNet (November 2017)
An introduction to Deep Learning with Apache MXNet (November 2017)An introduction to Deep Learning with Apache MXNet (November 2017)
An introduction to Deep Learning with Apache MXNet (November 2017)Julien SIMON
 
Week2- Deep Learning Intuition.pptx
Week2- Deep Learning Intuition.pptxWeek2- Deep Learning Intuition.pptx
Week2- Deep Learning Intuition.pptxfahmi324663
 
Machine Learning Tutorial Part - 1 | Machine Learning Tutorial For Beginners ...
Machine Learning Tutorial Part - 1 | Machine Learning Tutorial For Beginners ...Machine Learning Tutorial Part - 1 | Machine Learning Tutorial For Beginners ...
Machine Learning Tutorial Part - 1 | Machine Learning Tutorial For Beginners ...Simplilearn
 
Visual Quality for both Images and Display of Systems by Visual Enhancement u...
Visual Quality for both Images and Display of Systems by Visual Enhancement u...Visual Quality for both Images and Display of Systems by Visual Enhancement u...
Visual Quality for both Images and Display of Systems by Visual Enhancement u...IJMER
 
Gesture Recognition Based Video Game Controller
Gesture Recognition Based Video Game ControllerGesture Recognition Based Video Game Controller
Gesture Recognition Based Video Game ControllerIRJET Journal
 
CE344L-200365-Lab5.pdf
CE344L-200365-Lab5.pdfCE344L-200365-Lab5.pdf
CE344L-200365-Lab5.pdfUmarMustafa13
 
Convolution Neural Network Lecture Slides
Convolution Neural Network Lecture SlidesConvolution Neural Network Lecture Slides
Convolution Neural Network Lecture SlidesAdnanHaider234505
 
Deep Neural Networks for Computer Vision
Deep Neural Networks for Computer VisionDeep Neural Networks for Computer Vision
Deep Neural Networks for Computer VisionAlex Conway
 
IRJET - Skin Disease Predictor using Deep Learning
IRJET - Skin Disease Predictor using Deep LearningIRJET - Skin Disease Predictor using Deep Learning
IRJET - Skin Disease Predictor using Deep LearningIRJET Journal
 
Deep Learning Tutorial | Deep Learning Tutorial For Beginners | What Is Deep ...
Deep Learning Tutorial | Deep Learning Tutorial For Beginners | What Is Deep ...Deep Learning Tutorial | Deep Learning Tutorial For Beginners | What Is Deep ...
Deep Learning Tutorial | Deep Learning Tutorial For Beginners | What Is Deep ...Simplilearn
 
Rock Paper Scissors with MLNET.pptx
Rock Paper Scissors with MLNET.pptxRock Paper Scissors with MLNET.pptx
Rock Paper Scissors with MLNET.pptxLuis Beltran
 
Chapter 1 introduction (Image Processing)
Chapter 1 introduction (Image Processing)Chapter 1 introduction (Image Processing)
Chapter 1 introduction (Image Processing)Varun Ojha
 
Development of wearable object detection system &amp; blind stick for visuall...
Development of wearable object detection system &amp; blind stick for visuall...Development of wearable object detection system &amp; blind stick for visuall...
Development of wearable object detection system &amp; blind stick for visuall...Arkadev Kundu
 
Visualizing the Evolution of Working Sets
Visualizing the Evolution of Working SetsVisualizing the Evolution of Working Sets
Visualizing the Evolution of Working SetsRoberto Minelli
 
Simple APIs and innovative documentation
Simple APIs and innovative documentationSimple APIs and innovative documentation
Simple APIs and innovative documentationPyDataParis
 

Ähnlich wie Introduction to Deep Learning (20)

Deep Learning - Overview of my work II
Deep Learning - Overview of my work IIDeep Learning - Overview of my work II
Deep Learning - Overview of my work II
 
Deep learning simplified
Deep learning simplifiedDeep learning simplified
Deep learning simplified
 
A New Algorithm for Digital Colour Image Encryption and Decryption
A New Algorithm for Digital Colour Image Encryption and DecryptionA New Algorithm for Digital Colour Image Encryption and Decryption
A New Algorithm for Digital Colour Image Encryption and Decryption
 
An introduction to Deep Learning with Apache MXNet (November 2017)
An introduction to Deep Learning with Apache MXNet (November 2017)An introduction to Deep Learning with Apache MXNet (November 2017)
An introduction to Deep Learning with Apache MXNet (November 2017)
 
Week2- Deep Learning Intuition.pptx
Week2- Deep Learning Intuition.pptxWeek2- Deep Learning Intuition.pptx
Week2- Deep Learning Intuition.pptx
 
Machine Learning Tutorial Part - 1 | Machine Learning Tutorial For Beginners ...
Machine Learning Tutorial Part - 1 | Machine Learning Tutorial For Beginners ...Machine Learning Tutorial Part - 1 | Machine Learning Tutorial For Beginners ...
Machine Learning Tutorial Part - 1 | Machine Learning Tutorial For Beginners ...
 
Visual Quality for both Images and Display of Systems by Visual Enhancement u...
Visual Quality for both Images and Display of Systems by Visual Enhancement u...Visual Quality for both Images and Display of Systems by Visual Enhancement u...
Visual Quality for both Images and Display of Systems by Visual Enhancement u...
 
Gesture Recognition Based Video Game Controller
Gesture Recognition Based Video Game ControllerGesture Recognition Based Video Game Controller
Gesture Recognition Based Video Game Controller
 
CE344L-200365-Lab5.pdf
CE344L-200365-Lab5.pdfCE344L-200365-Lab5.pdf
CE344L-200365-Lab5.pdf
 
Convolution Neural Network Lecture Slides
Convolution Neural Network Lecture SlidesConvolution Neural Network Lecture Slides
Convolution Neural Network Lecture Slides
 
Multimedia Steganography
Multimedia SteganographyMultimedia Steganography
Multimedia Steganography
 
Deep Neural Networks for Computer Vision
Deep Neural Networks for Computer VisionDeep Neural Networks for Computer Vision
Deep Neural Networks for Computer Vision
 
ML基本からResNetまで
ML基本からResNetまでML基本からResNetまで
ML基本からResNetまで
 
IRJET - Skin Disease Predictor using Deep Learning
IRJET - Skin Disease Predictor using Deep LearningIRJET - Skin Disease Predictor using Deep Learning
IRJET - Skin Disease Predictor using Deep Learning
 
Deep Learning Tutorial | Deep Learning Tutorial For Beginners | What Is Deep ...
Deep Learning Tutorial | Deep Learning Tutorial For Beginners | What Is Deep ...Deep Learning Tutorial | Deep Learning Tutorial For Beginners | What Is Deep ...
Deep Learning Tutorial | Deep Learning Tutorial For Beginners | What Is Deep ...
 
Rock Paper Scissors with MLNET.pptx
Rock Paper Scissors with MLNET.pptxRock Paper Scissors with MLNET.pptx
Rock Paper Scissors with MLNET.pptx
 
Chapter 1 introduction (Image Processing)
Chapter 1 introduction (Image Processing)Chapter 1 introduction (Image Processing)
Chapter 1 introduction (Image Processing)
 
Development of wearable object detection system &amp; blind stick for visuall...
Development of wearable object detection system &amp; blind stick for visuall...Development of wearable object detection system &amp; blind stick for visuall...
Development of wearable object detection system &amp; blind stick for visuall...
 
Visualizing the Evolution of Working Sets
Visualizing the Evolution of Working SetsVisualizing the Evolution of Working Sets
Visualizing the Evolution of Working Sets
 
Simple APIs and innovative documentation
Simple APIs and innovative documentationSimple APIs and innovative documentation
Simple APIs and innovative documentation
 

Mehr von Dmytro Fishman

DOME: Recommendations for supervised machine learning validation in biology
DOME: Recommendations for supervised machine learning validation in biologyDOME: Recommendations for supervised machine learning validation in biology
DOME: Recommendations for supervised machine learning validation in biologyDmytro Fishman
 
Tips for effective presentations
Tips for effective presentationsTips for effective presentations
Tips for effective presentationsDmytro Fishman
 
Autonomous Driving Lab - Simultaneous Localization and Mapping WP
Autonomous Driving Lab - Simultaneous Localization and Mapping WPAutonomous Driving Lab - Simultaneous Localization and Mapping WP
Autonomous Driving Lab - Simultaneous Localization and Mapping WPDmytro Fishman
 
Introduction to Machine Learning for Taxify/Bolt
Introduction to Machine Learning for Taxify/BoltIntroduction to Machine Learning for Taxify/Bolt
Introduction to Machine Learning for Taxify/BoltDmytro Fishman
 
Introduction to Gaussian Processes
Introduction to Gaussian ProcessesIntroduction to Gaussian Processes
Introduction to Gaussian ProcessesDmytro Fishman
 
Detecting Nuclei from Microscopy Images with Deep Learning
Detecting Nuclei from Microscopy Images with Deep LearningDetecting Nuclei from Microscopy Images with Deep Learning
Detecting Nuclei from Microscopy Images with Deep LearningDmytro Fishman
 
Deep Learning in Healthcare
Deep Learning in HealthcareDeep Learning in Healthcare
Deep Learning in HealthcareDmytro Fishman
 
5 Introduction to neural networks
5 Introduction to neural networks5 Introduction to neural networks
5 Introduction to neural networksDmytro Fishman
 
4 Dimensionality reduction (PCA & t-SNE)
4 Dimensionality reduction (PCA & t-SNE)4 Dimensionality reduction (PCA & t-SNE)
4 Dimensionality reduction (PCA & t-SNE)Dmytro Fishman
 
3 Unsupervised learning
3 Unsupervised learning3 Unsupervised learning
3 Unsupervised learningDmytro Fishman
 
What does it mean to be a bioinformatician?
What does it mean to be a bioinformatician?What does it mean to be a bioinformatician?
What does it mean to be a bioinformatician?Dmytro Fishman
 
Machine Learning in Bioinformatics
Machine Learning in BioinformaticsMachine Learning in Bioinformatics
Machine Learning in BioinformaticsDmytro Fishman
 

Mehr von Dmytro Fishman (14)

DOME: Recommendations for supervised machine learning validation in biology
DOME: Recommendations for supervised machine learning validation in biologyDOME: Recommendations for supervised machine learning validation in biology
DOME: Recommendations for supervised machine learning validation in biology
 
Tips for effective presentations
Tips for effective presentationsTips for effective presentations
Tips for effective presentations
 
Autonomous Driving Lab - Simultaneous Localization and Mapping WP
Autonomous Driving Lab - Simultaneous Localization and Mapping WPAutonomous Driving Lab - Simultaneous Localization and Mapping WP
Autonomous Driving Lab - Simultaneous Localization and Mapping WP
 
Introduction to Machine Learning for Taxify/Bolt
Introduction to Machine Learning for Taxify/BoltIntroduction to Machine Learning for Taxify/Bolt
Introduction to Machine Learning for Taxify/Bolt
 
Introduction to Gaussian Processes
Introduction to Gaussian ProcessesIntroduction to Gaussian Processes
Introduction to Gaussian Processes
 
Biit group 2018
Biit group 2018Biit group 2018
Biit group 2018
 
Detecting Nuclei from Microscopy Images with Deep Learning
Detecting Nuclei from Microscopy Images with Deep LearningDetecting Nuclei from Microscopy Images with Deep Learning
Detecting Nuclei from Microscopy Images with Deep Learning
 
Deep Learning in Healthcare
Deep Learning in HealthcareDeep Learning in Healthcare
Deep Learning in Healthcare
 
5 Introduction to neural networks
5 Introduction to neural networks5 Introduction to neural networks
5 Introduction to neural networks
 
4 Dimensionality reduction (PCA & t-SNE)
4 Dimensionality reduction (PCA & t-SNE)4 Dimensionality reduction (PCA & t-SNE)
4 Dimensionality reduction (PCA & t-SNE)
 
3 Unsupervised learning
3 Unsupervised learning3 Unsupervised learning
3 Unsupervised learning
 
1 Supervised learning
1 Supervised learning1 Supervised learning
1 Supervised learning
 
What does it mean to be a bioinformatician?
What does it mean to be a bioinformatician?What does it mean to be a bioinformatician?
What does it mean to be a bioinformatician?
 
Machine Learning in Bioinformatics
Machine Learning in BioinformaticsMachine Learning in Bioinformatics
Machine Learning in Bioinformatics
 

Kürzlich hochgeladen

Unveiling Insights: The Role of a Data Analyst
Unveiling Insights: The Role of a Data AnalystUnveiling Insights: The Role of a Data Analyst
Unveiling Insights: The Role of a Data AnalystSamantha Rae Coolbeth
 
Generative AI on Enterprise Cloud with NiFi and Milvus
Generative AI on Enterprise Cloud with NiFi and MilvusGenerative AI on Enterprise Cloud with NiFi and Milvus
Generative AI on Enterprise Cloud with NiFi and MilvusTimothy Spann
 
RA-11058_IRR-COMPRESS Do 198 series of 1998
RA-11058_IRR-COMPRESS Do 198 series of 1998RA-11058_IRR-COMPRESS Do 198 series of 1998
RA-11058_IRR-COMPRESS Do 198 series of 1998YohFuh
 
Low Rate Call Girls Bhilai Anika 8250192130 Independent Escort Service Bhilai
Low Rate Call Girls Bhilai Anika 8250192130 Independent Escort Service BhilaiLow Rate Call Girls Bhilai Anika 8250192130 Independent Escort Service Bhilai
Low Rate Call Girls Bhilai Anika 8250192130 Independent Escort Service BhilaiSuhani Kapoor
 
(PARI) Call Girls Wanowrie ( 7001035870 ) HI-Fi Pune Escorts Service
(PARI) Call Girls Wanowrie ( 7001035870 ) HI-Fi Pune Escorts Service(PARI) Call Girls Wanowrie ( 7001035870 ) HI-Fi Pune Escorts Service
(PARI) Call Girls Wanowrie ( 7001035870 ) HI-Fi Pune Escorts Serviceranjana rawat
 
Brighton SEO | April 2024 | Data Storytelling
Brighton SEO | April 2024 | Data StorytellingBrighton SEO | April 2024 | Data Storytelling
Brighton SEO | April 2024 | Data StorytellingNeil Barnes
 
Schema on read is obsolete. Welcome metaprogramming..pdf
Schema on read is obsolete. Welcome metaprogramming..pdfSchema on read is obsolete. Welcome metaprogramming..pdf
Schema on read is obsolete. Welcome metaprogramming..pdfLars Albertsson
 
100-Concepts-of-AI by Anupama Kate .pptx
100-Concepts-of-AI by Anupama Kate .pptx100-Concepts-of-AI by Anupama Kate .pptx
100-Concepts-of-AI by Anupama Kate .pptxAnupama Kate
 
BabyOno dropshipping via API with DroFx.pptx
BabyOno dropshipping via API with DroFx.pptxBabyOno dropshipping via API with DroFx.pptx
BabyOno dropshipping via API with DroFx.pptxolyaivanovalion
 
B2 Creative Industry Response Evaluation.docx
B2 Creative Industry Response Evaluation.docxB2 Creative Industry Response Evaluation.docx
B2 Creative Industry Response Evaluation.docxStephen266013
 
Carero dropshipping via API with DroFx.pptx
Carero dropshipping via API with DroFx.pptxCarero dropshipping via API with DroFx.pptx
Carero dropshipping via API with DroFx.pptxolyaivanovalion
 
VidaXL dropshipping via API with DroFx.pptx
VidaXL dropshipping via API with DroFx.pptxVidaXL dropshipping via API with DroFx.pptx
VidaXL dropshipping via API with DroFx.pptxolyaivanovalion
 
Data-Analysis for Chicago Crime Data 2023
Data-Analysis for Chicago Crime Data  2023Data-Analysis for Chicago Crime Data  2023
Data-Analysis for Chicago Crime Data 2023ymrp368
 
꧁❤ Greater Noida Call Girls Delhi ❤꧂ 9711199171 ☎️ Hard And Sexy Vip Call
꧁❤ Greater Noida Call Girls Delhi ❤꧂ 9711199171 ☎️ Hard And Sexy Vip Call꧁❤ Greater Noida Call Girls Delhi ❤꧂ 9711199171 ☎️ Hard And Sexy Vip Call
꧁❤ Greater Noida Call Girls Delhi ❤꧂ 9711199171 ☎️ Hard And Sexy Vip Callshivangimorya083
 
Call me @ 9892124323 Cheap Rate Call Girls in Vashi with Real Photo 100% Secure
Call me @ 9892124323  Cheap Rate Call Girls in Vashi with Real Photo 100% SecureCall me @ 9892124323  Cheap Rate Call Girls in Vashi with Real Photo 100% Secure
Call me @ 9892124323 Cheap Rate Call Girls in Vashi with Real Photo 100% SecurePooja Nehwal
 
代办国外大学文凭《原版美国UCLA文凭证书》加州大学洛杉矶分校毕业证制作成绩单修改
代办国外大学文凭《原版美国UCLA文凭证书》加州大学洛杉矶分校毕业证制作成绩单修改代办国外大学文凭《原版美国UCLA文凭证书》加州大学洛杉矶分校毕业证制作成绩单修改
代办国外大学文凭《原版美国UCLA文凭证书》加州大学洛杉矶分校毕业证制作成绩单修改atducpo
 
April 2024 - Crypto Market Report's Analysis
April 2024 - Crypto Market Report's AnalysisApril 2024 - Crypto Market Report's Analysis
April 2024 - Crypto Market Report's Analysismanisha194592
 

Kürzlich hochgeladen (20)

Unveiling Insights: The Role of a Data Analyst
Unveiling Insights: The Role of a Data AnalystUnveiling Insights: The Role of a Data Analyst
Unveiling Insights: The Role of a Data Analyst
 
Generative AI on Enterprise Cloud with NiFi and Milvus
Generative AI on Enterprise Cloud with NiFi and MilvusGenerative AI on Enterprise Cloud with NiFi and Milvus
Generative AI on Enterprise Cloud with NiFi and Milvus
 
RA-11058_IRR-COMPRESS Do 198 series of 1998
RA-11058_IRR-COMPRESS Do 198 series of 1998RA-11058_IRR-COMPRESS Do 198 series of 1998
RA-11058_IRR-COMPRESS Do 198 series of 1998
 
Low Rate Call Girls Bhilai Anika 8250192130 Independent Escort Service Bhilai
Low Rate Call Girls Bhilai Anika 8250192130 Independent Escort Service BhilaiLow Rate Call Girls Bhilai Anika 8250192130 Independent Escort Service Bhilai
Low Rate Call Girls Bhilai Anika 8250192130 Independent Escort Service Bhilai
 
(PARI) Call Girls Wanowrie ( 7001035870 ) HI-Fi Pune Escorts Service
(PARI) Call Girls Wanowrie ( 7001035870 ) HI-Fi Pune Escorts Service(PARI) Call Girls Wanowrie ( 7001035870 ) HI-Fi Pune Escorts Service
(PARI) Call Girls Wanowrie ( 7001035870 ) HI-Fi Pune Escorts Service
 
Brighton SEO | April 2024 | Data Storytelling
Brighton SEO | April 2024 | Data StorytellingBrighton SEO | April 2024 | Data Storytelling
Brighton SEO | April 2024 | Data Storytelling
 
Schema on read is obsolete. Welcome metaprogramming..pdf
Schema on read is obsolete. Welcome metaprogramming..pdfSchema on read is obsolete. Welcome metaprogramming..pdf
Schema on read is obsolete. Welcome metaprogramming..pdf
 
VIP Call Girls Service Charbagh { Lucknow Call Girls Service 9548273370 } Boo...
VIP Call Girls Service Charbagh { Lucknow Call Girls Service 9548273370 } Boo...VIP Call Girls Service Charbagh { Lucknow Call Girls Service 9548273370 } Boo...
VIP Call Girls Service Charbagh { Lucknow Call Girls Service 9548273370 } Boo...
 
100-Concepts-of-AI by Anupama Kate .pptx
100-Concepts-of-AI by Anupama Kate .pptx100-Concepts-of-AI by Anupama Kate .pptx
100-Concepts-of-AI by Anupama Kate .pptx
 
BabyOno dropshipping via API with DroFx.pptx
BabyOno dropshipping via API with DroFx.pptxBabyOno dropshipping via API with DroFx.pptx
BabyOno dropshipping via API with DroFx.pptx
 
B2 Creative Industry Response Evaluation.docx
B2 Creative Industry Response Evaluation.docxB2 Creative Industry Response Evaluation.docx
B2 Creative Industry Response Evaluation.docx
 
Carero dropshipping via API with DroFx.pptx
Carero dropshipping via API with DroFx.pptxCarero dropshipping via API with DroFx.pptx
Carero dropshipping via API with DroFx.pptx
 
VidaXL dropshipping via API with DroFx.pptx
VidaXL dropshipping via API with DroFx.pptxVidaXL dropshipping via API with DroFx.pptx
VidaXL dropshipping via API with DroFx.pptx
 
Data-Analysis for Chicago Crime Data 2023
Data-Analysis for Chicago Crime Data  2023Data-Analysis for Chicago Crime Data  2023
Data-Analysis for Chicago Crime Data 2023
 
꧁❤ Greater Noida Call Girls Delhi ❤꧂ 9711199171 ☎️ Hard And Sexy Vip Call
꧁❤ Greater Noida Call Girls Delhi ❤꧂ 9711199171 ☎️ Hard And Sexy Vip Call꧁❤ Greater Noida Call Girls Delhi ❤꧂ 9711199171 ☎️ Hard And Sexy Vip Call
꧁❤ Greater Noida Call Girls Delhi ❤꧂ 9711199171 ☎️ Hard And Sexy Vip Call
 
꧁❤ Aerocity Call Girls Service Aerocity Delhi ❤꧂ 9999965857 ☎️ Hard And Sexy ...
꧁❤ Aerocity Call Girls Service Aerocity Delhi ❤꧂ 9999965857 ☎️ Hard And Sexy ...꧁❤ Aerocity Call Girls Service Aerocity Delhi ❤꧂ 9999965857 ☎️ Hard And Sexy ...
꧁❤ Aerocity Call Girls Service Aerocity Delhi ❤꧂ 9999965857 ☎️ Hard And Sexy ...
 
Call me @ 9892124323 Cheap Rate Call Girls in Vashi with Real Photo 100% Secure
Call me @ 9892124323  Cheap Rate Call Girls in Vashi with Real Photo 100% SecureCall me @ 9892124323  Cheap Rate Call Girls in Vashi with Real Photo 100% Secure
Call me @ 9892124323 Cheap Rate Call Girls in Vashi with Real Photo 100% Secure
 
Delhi 99530 vip 56974 Genuine Escort Service Call Girls in Kishangarh
Delhi 99530 vip 56974 Genuine Escort Service Call Girls in  KishangarhDelhi 99530 vip 56974 Genuine Escort Service Call Girls in  Kishangarh
Delhi 99530 vip 56974 Genuine Escort Service Call Girls in Kishangarh
 
代办国外大学文凭《原版美国UCLA文凭证书》加州大学洛杉矶分校毕业证制作成绩单修改
代办国外大学文凭《原版美国UCLA文凭证书》加州大学洛杉矶分校毕业证制作成绩单修改代办国外大学文凭《原版美国UCLA文凭证书》加州大学洛杉矶分校毕业证制作成绩单修改
代办国外大学文凭《原版美国UCLA文凭证书》加州大学洛杉矶分校毕业证制作成绩单修改
 
April 2024 - Crypto Market Report's Analysis
April 2024 - Crypto Market Report's AnalysisApril 2024 - Crypto Market Report's Analysis
April 2024 - Crypto Market Report's Analysis
 

Introduction to Deep Learning