SlideShare ist ein Scribd-Unternehmen logo
1 von 64
Downloaden Sie, um offline zu lesen
Deep Neural Networks Tutorial
Andrey Filchenkov
Computer Technology Chair
Computer Technologies Lab
ITMO University
afilchenkov@corp.ifmo.ru
AINL FRUCT’ 6. No , 6, “t. Peters urg, Russia
Tutorial topics
Very brief introduction to AI, ML and ANN
What is ANN and how to learn it
DNN and standard DNN architectures
Beyond discriminative models
2 / 64
Next topic
Very brief introduction to AI, ML and ANN
What is ANN and how to learn it
DNN and standard DNN architectures
Beyond discriminative models
3 / 64
Artificial intelligence
Strong AI (A general I): functionality is similar to the human
brain or better.
Weak AI: good in solving certain well-formulated tasks.
Machine learning is a part of Weak AI
Many people have been thinking that artificial neural
networks are a path to Strong AI
Many people are thinking that deep learning networks are a
path to Strong AI
4 / 64
Neural networks as a machine learning algorithm
Neural paradigm is not only about machine learning
(computer architecture, computations, etc.)
Machine learning is about creating algorithms which can learn
patterns, regularities and rules from given data.
The biggest part of machine learning is supervised learning:
we are given a set of objects, each has a label, and we want to
learn how to find these label for objects we have never seen.
5 / 64
Biological neuron
6 / 64
Brief early history of artificial neural networks
1943 Artificial neuron by McCulloch and Pitts
1949 Neuron learning rule by Hebb
1957 Perceptron by Rosenblatt
1960 Perceptron learning rule by Widrow and Hoff
1968 Group Method of Data Handling to learn multilayered
networks by Ivakhnenko
1969 Perceptrons by Minski and Papert
1974 Back propagation algorithm by Webros and by Galushkin
7 / 64
Brief modern history of ANN
1980 Convolutional NN by Fukushima
1982 Recurrent NN by Hopfield
1991 Va ishi g gradient pro le was identified by Hochreiter
1997 Long short term memory network by Hochreiter and
Schmidhuber
1998 Gradient descent for convolutional NN by LeCun et al.
2006 Deep model by Hinton, Osindero and Teh
2012 DNN started to become mainstream in ML and AI
8 / 64
Next topic
Very brief introduction to AI, ML and ANN
What is ANN and how to learn it
DNN and standard DNN architectures
Beyond discriminative models
9 / 64
Two sources of knowledge
Experts
we need to ask wisely and process
Data
we need to process and apply machine learning algorithms
How do we obtain knowledge?
10 / 64
Algorithms, performance of which grows with experience
The most popular task is prediction
Algorithms require data and labels (for predicting)
Learning of these algorithms is to minimize error rate in
prediction or maximize similar to the known answer
Machine learning
11 / 64
Each object is represented
as a feature vector. Each
object thus is a point in a
multidimensional space.
Vector representation of objects
12 / 64
Linear discriminant surface
Linear model (neuron)
13 / 64
Non-linear case
14 / 64
Multilayer neural network
We just build a composition of
neurons (as functions)
15 / 64
Next topic
Very brief introduction to AI, ML and ANN
What is ANN and how to learn it
DNN and standard DNN architectures
• Deep learning introduction and best practices
• Deep Boltzmann Machines (DBM) and Deep Belief Network
(DBN)
• Convolution Neural Network (CNN)
• Autoencoders
• Recurrental Neural Network (RNN) and Long-Short Term
Memory (LSTM)
Beyond discriminative models
16 / 64
Deep architecture
Definition: Deep architectures are composed of multiple levels of
non-linear operations, such as neural nets with many hidden
layers
Most machine learning algorithms have shallow (1–3 layers)
architecture (SVM, PCA, kNN, Logistic Regression, etc.)
Goal: Deep learning methods aim at:
Learning feature hierarchies, no more feature engineering!
Where features from higher levels of the hierarchy are formed
by lower level features.
17 / 64
Why to go deep?
Some functions cannot be efficiently represented (in terms of
number of tunable elements) by architectures that are too shallow
Functions that can be compactly represented by a depth k
architecture might require an exponential number of computational
elements to be represented by a depth k− ar hite ture
Deep Representations might allow non-local generalization and
comprehensibility
Deep learning gets state of the art results in many fields (vision,
audio, NLP, etc.)!
18 / 64
DNN best practices: ReLU, PReLU
ReLU
19 / 64
PReLU
Sigmoid, hyperbolic tangent activation functions have a problem
with vanishing of gradients and tend to overfitting
DNN best practices: Data augmentation
The easiest and most common method to reduce overfitting on
image data is to artificially enlarge the dataset using label-
preserving transformations.
Types of data augmentation:
Image translation
Horizontal/vertical reflections + cropping
Changing RGB intensities
20 / 64
DNN best practices: Dropout
Dropout: set the output of each hidden neuron to zero w.p. 0.5
The euro s hi h are dropped out i this a do ot o tri ute
to the forward pass and do not participate in backpropagation
So every time an input is presented, the neural network samples a
different architecture, but all these architectures share weights
This technique reduces complex co-adaptations of neurons, since a
neuron cannot rely on the presence of particular other neurons
It is, therefore, forced to learn more robust features that are useful
in conjunction with many different random subsets of the other
neurons
Without dropout, a network exhibits substantial overfitting
Dropout roughly doubles the number of iterations required to
converge
21 / 64
Greedy Layer-Wise Training (1/2)
1. Train first layer using your data without the labels (unsupervised)
• Since there are no targets at this level, labels don't help. Could
also use the more abundant unlabeled data which is not part of
the training set (i.e. self-taught learning).
2. Then freeze the first layer parameters and start training the second
layer using the output of the first layer as the unsupervised input
to the second layer
3. Repeat this for as many layers as desired
• This builds our set of robust features
4. Use the outputs of the final layer as inputs to a supervised
layer/model and train the last supervised layer(s) (leave early
weights frozen)
5. Unfreeze all weights and fine tune the full network by training with
a supervised approach, given the pre-processed weight settings
22 / 64
Greedy Layer-Wise Training (2/2)
Greedy layer-wise training avoids many of the problems of trying to
train a deep net in a supervised fashion:
• Each layer gets full learning focus in its turn since it is the only
current "top" layer
• Can take advantage of unlabeled data
• When you finally tune the entire network with supervised training
the network weights have already been adjusted so that you are in
a good error basin and just need fine tuning. This helps with
problems of:
• Ineffective early layer learning
• Deep network local minima
23 / 64
Restricted Boltzmann machine
Two types of nodes: hidden and visible.
We are minimizing system energy which is not to converge by
updating its weights with propagating new objects. Probability
distribution on visible and hidden layers is Gibbs distribution.
24 / 64
Deep Belief Network
25 / 64
First train unsupervisedly with several
(two) levels of RBM (or autoencoders).
Then train next layers supervisedly and
consecutively.
Hinton, G. E. and Salakhutdinov, R. R. (2006). Reducing the dimensionality of data with neural networks. Science, 313:504-507.
Deep Boltzmann machine
26 / 64
Convolution Neural Network (CNN)
Core concepts:
Local perception – each neuron sees a small part of the
object. Use kernels (filters) to capture 1-D or 2-D structure of
objects. For instance, capture all pixel neighbors for an image.
Weight sharing – use small and the same sets of kernels for all
objects, this leads to reduction of number of adjusting
parameters in comparison with MLP
Subsampling/pooling – use dimensionality reduction for
images in order to provide invariance to scale
27 / 64
Discrete convolution
28 / 64
What can kernels do?
29 / 64
*
*
*
=
=
=
blur
edge
detection
sharpen
Discrete convolution
30 / 64
How do trained kernels look like?
31 / 64
low feature medium feature high feature
Each kernel composes a local patch of lower-level features
into high level representation
Levels of abstraction
Hierarchical Learning:
Natural progression from low
level to high level structure
as seen in natural complexity
Easier to monitor what is
being learnt and to guide the
machine to better subspaces
A good lower level
representation can be used
for many distinct tasks
32 / 64
LeNet
33 / 64
LeCun, Yann, et al. "Gradient-based learning applied to document recognition." Proceedings of the
IEEE 86.11 (1998): 2278-2324.
GoogLeNet
34 / 64
Szegedy, Christian, et al. "Going deeper with convolutions." arXiv preprint
arXiv:1409.4842 (2014).
Application of CNN
Image Recognition
Image Search (enhance search engine)
Visual Question Answering
NLP (Sentence Classification, etc.)
Speech Recognition
…
35 / 64
Visual question answering (1/2)
36 / 64
Visual question answering (2/2)
37 / 64
Application of CNN: Visual question answering
38 / 64
Autoencoders (1/3)
Autoencoder: a feed-forward neural
network trained to reproduce its input
at the output layer
Do non-linear dimensionality
reduction
Train via backpropagation
1-layer autoencoder gets similar
results as PCA
39 / 64
Autoencoders (2/3)
40 / 64
Autoencoders (3/3)
41 / 64
Autoencoders in bioinformatics
42 / 64
Fakoor, Rasool, et al. "Using deep learning to enhance cancer diagnosis and classification." Proceedings of the International
Conference on Machine Learning. 2013.
Deep autoencoders: document processing
We can use an autoencoder to find low-dimensional codes for
documents that allow fast and accurate retrieval of similar
documents from a large set.
We start with o erti g ea h do u e t i to a ag of
ords . This is a 2000 dimensional vector that contains the
counts for each of the 2000 commonest words.
43 / 64
Deep autoencoders: document retrieval
We train the neural network to
reproduce its input vector as
its output
This forces it to compress as
much information as possible
into the 10 numbers in the
central bottleneck.
These 10 numbers are then a
good way to compare
documents.
44 / 64
2000 reconstructed counts
500 neurons
250 neurons
10
250 neurons
500 neurons
2000 word counts
Deep autoencoders: document visualization (1/2)
Instead of using codes to
retrieve documents, we can
use 2-D codes to visualize sets
of documents.
This works much better than
2-D PCA
45 / 64
2000 reconstructed counts
500 neurons
250 neurons
2
250 neurons
500 neurons
2000 word counts
Deep autoencoders: document visualization (2/2)
46 / 64
Recurrent Neural Network (RNN)
RNN: a neural network with recurrent connections
Good for sequence data: time series, text, audio
47 / 64
Recurrent Neural Network (RNN): scheme
48 / 64
Long Short-Term Memory (LSTM)
LSTM: a special case of RNN capable of learning long-term
dependencies
There are four neural network layers in repeating module
49 / 64Hochreiter, Sepp, and Jürgen Schmidhuber. "Long short-term memory." Neural computation 9.8 (1997): 1735-1780.
LSTM: Cell state
Cell state: runs straight down the entire chain, with only
some minor linear interactions
LSTM does have the ability to remove or add information
to the cell state, carefully regulated by structures called
gates. Gates are a way to optionally let information
through.
The sigmoid layer outputs numbers
between zero and one, describing
how much of each component
should be let through. LSTM has
3 gates.
50 / 64
LSTM: Forget gate layer
It looks at ℎt−1 and �t, and outputs a number between 0 and
1 for each number in the cell state �t−1. 1 represents
o pletel keep this hile 0 represe ts o pletel get rid
of this .
51 / 64
LSTM: Input gate layer (1/2)
How to decide what new information to store in the cell state?
First, a sig oid la er alled the i put gate la er de ides
which values should be updated.
Next, a tanh layer creates a vector of new candidate values, �t,
that ould e added to the state. I the e t step, e’ll
combine these two to create an update to the state
52 / 64
LSTM: Input gate layer (2/2)
It’s o ti e to update the old ell state, ��−1, into the new
cell state ��. The previous steps already decided what to do,
we just need to actually do it
We multiply the old state by ��, forgetting the things we
decided to forget earlier. Then we add �� ⋅ ��−1. This is the
new candidate values, scaled by how much we decided to
update each state value
53 / 64
LSTM: Output gate layer
The output will be based on cell state, but will be a filtered
version. First, we run a sigmoid layer which decides what
parts of the ell state e’re goi g to output. The , e put the
cell state through tanh to push the alues to e et ee −
and 1) and multiply it by the output of the sigmoid gate, so
that we only output the parts we decided to.
54 / 64
Visual question answering: LSTM
55 / 64
Deep learning analysis: advantages
Extremely strong model, which can potentially solve any
problem of machine learning.
Already learnt model can be reused: multi-task support
Many good models are already known which are state-of-the-
art for many tasks:
• image recognition;
• speech recognition;
• natural language processing;
• etc.
56 / 64
Deep learning analysis: disadvantages
The deeper the net is
• the more data you need;
• the more time you need;
• the stronger processors you need.
Usually no intuition how it works exactly;
Usually you work with DNN as a black box;
Prone to overfitting: regularization must be used.
57 / 64
Next topic
Very brief introduction to AI, ML and ANN
What is ANN and how to learn it
DNN and standard DNN architectures
Beyond discriminative models
58 / 64
Reverse the network and make it predict images given labels
Image synthesis
Dosovitskiy, A., Tobias Springenberg, J., & Brox, T. (2015). Learning to generate chairs with convolutional neural networks. In Proceedings of the IEEE
Conference on Computer Vision and Pattern Recognition (pp. 1538-1546). 59 / 64
Keep i er represe tatio of a i age
(Gram matrix �� for convolutional layers)
Then create a new random network and
learn it no have similar inner
representation as the one we have kept.
Texture synthesis
Gatys, L., Ecker, A. S., & Bethge, M. (2015). Texture synthesis using convolutional neural networks. In Advances in Neural Information Processing
Systems (pp. 262-270). 60 / 64
Style = texture.
Image = content and
is represented with
the last
convolutional layer.
We will learn an
image that is similar
both to image and
content.
Style transmission
Gatys, L. A., Ecker, A. S., & Bethge, M. (2015). A neural algorithm of artistic style. arXiv preprint arXiv:1508.06576.
61 / 64
DeetArt was created in 2015:
https://deepart.io/
They implemented the algorithm
described before.
DeepArt and Prisma
62 / 64
Prisma was created in June,
2016.
They made it optimized,
mobiles and with preselected
filters (instead of styles)
Materials
Presentation was prepared using:
1. http://avisingh599.github.io/deeplearning/visual-qa/
2. http://colah.github.io/posts/2015-08-Understanding-LSTMs/
3. https://class.coursera.org/ml-003/lecture
4. K. Vorontsov Ma hi e lear i g ourse i Russia
63 / 64
Thanks for attention!
Questions?

Weitere ähnliche Inhalte

Was ist angesagt?

Was ist angesagt? (20)

Deep Learning for NLP Applications
Deep Learning for NLP ApplicationsDeep Learning for NLP Applications
Deep Learning for NLP Applications
 
Anthiil Inside workshop on NLP
Anthiil Inside workshop on NLPAnthiil Inside workshop on NLP
Anthiil Inside workshop on NLP
 
Short story presentation
Short story presentationShort story presentation
Short story presentation
 
Convolutional neural networks for sentiment classification
Convolutional neural networks for sentiment classificationConvolutional neural networks for sentiment classification
Convolutional neural networks for sentiment classification
 
Natural Language Generation / Stanford cs224n 2019w lecture 15 Review
Natural Language Generation / Stanford cs224n 2019w lecture 15 ReviewNatural Language Generation / Stanford cs224n 2019w lecture 15 Review
Natural Language Generation / Stanford cs224n 2019w lecture 15 Review
 
Deep Learning for Natural Language Processing: Word Embeddings
Deep Learning for Natural Language Processing: Word EmbeddingsDeep Learning for Natural Language Processing: Word Embeddings
Deep Learning for Natural Language Processing: Word Embeddings
 
A Brief Introduction on Recurrent Neural Network and Its Application
A Brief Introduction on Recurrent Neural Network and Its ApplicationA Brief Introduction on Recurrent Neural Network and Its Application
A Brief Introduction on Recurrent Neural Network and Its Application
 
Word embeddings, RNN, GRU and LSTM
Word embeddings, RNN, GRU and LSTMWord embeddings, RNN, GRU and LSTM
Word embeddings, RNN, GRU and LSTM
 
Natural language processing techniques transition from machine learning to de...
Natural language processing techniques transition from machine learning to de...Natural language processing techniques transition from machine learning to de...
Natural language processing techniques transition from machine learning to de...
 
NLP Bootcamp
NLP BootcampNLP Bootcamp
NLP Bootcamp
 
Deeplearning NLP
Deeplearning NLPDeeplearning NLP
Deeplearning NLP
 
NLP from scratch
NLP from scratch NLP from scratch
NLP from scratch
 
Sequence to Sequence Learning with Neural Networks
Sequence to Sequence Learning with Neural NetworksSequence to Sequence Learning with Neural Networks
Sequence to Sequence Learning with Neural Networks
 
Recurrent Neural Networks, LSTM and GRU
Recurrent Neural Networks, LSTM and GRURecurrent Neural Networks, LSTM and GRU
Recurrent Neural Networks, LSTM and GRU
 
IRJET- Survey on Generating Suggestions for Erroneous Part in a Sentence
IRJET- Survey on Generating Suggestions for Erroneous Part in a SentenceIRJET- Survey on Generating Suggestions for Erroneous Part in a Sentence
IRJET- Survey on Generating Suggestions for Erroneous Part in a Sentence
 
Deep Learning - A Literature survey
Deep Learning - A Literature surveyDeep Learning - A Literature survey
Deep Learning - A Literature survey
 
Neural Networks and Deep Learning
Neural Networks and Deep LearningNeural Networks and Deep Learning
Neural Networks and Deep Learning
 
Alberto Massidda - Images and words: mechanics of automated captioning with n...
Alberto Massidda - Images and words: mechanics of automated captioning with n...Alberto Massidda - Images and words: mechanics of automated captioning with n...
Alberto Massidda - Images and words: mechanics of automated captioning with n...
 
Learning to understand phrases by embedding the dictionary
Learning to understand phrases by embedding the dictionaryLearning to understand phrases by embedding the dictionary
Learning to understand phrases by embedding the dictionary
 
Talk from NVidia Developer Connect
Talk from NVidia Developer ConnectTalk from NVidia Developer Connect
Talk from NVidia Developer Connect
 

Andere mochten auch

Andere mochten auch (20)

AINL 2016: Fenogenova, Karpov, Kazorin
AINL 2016: Fenogenova, Karpov, KazorinAINL 2016: Fenogenova, Karpov, Kazorin
AINL 2016: Fenogenova, Karpov, Kazorin
 
AINL 2016: Bugaychenko
AINL 2016: BugaychenkoAINL 2016: Bugaychenko
AINL 2016: Bugaychenko
 
AINL 2016: Proncheva
AINL 2016: PronchevaAINL 2016: Proncheva
AINL 2016: Proncheva
 
AINL 2016: Bodrunova, Blekanov, Maksimov
AINL 2016: Bodrunova, Blekanov, MaksimovAINL 2016: Bodrunova, Blekanov, Maksimov
AINL 2016: Bodrunova, Blekanov, Maksimov
 
AINL 2016: Bastrakova, Ledesma, Millan, Zighed
AINL 2016: Bastrakova, Ledesma, Millan, ZighedAINL 2016: Bastrakova, Ledesma, Millan, Zighed
AINL 2016: Bastrakova, Ledesma, Millan, Zighed
 
AINL 2016: Kravchenko
AINL 2016: KravchenkoAINL 2016: Kravchenko
AINL 2016: Kravchenko
 
AINL 2016: Alekseev, Nikolenko
AINL 2016: Alekseev, NikolenkoAINL 2016: Alekseev, Nikolenko
AINL 2016: Alekseev, Nikolenko
 
AINL 2016: Rykov, Nagornyy, Koltsova, Natta, Kremenets, Manovich, Cerrone, Cr...
AINL 2016: Rykov, Nagornyy, Koltsova, Natta, Kremenets, Manovich, Cerrone, Cr...AINL 2016: Rykov, Nagornyy, Koltsova, Natta, Kremenets, Manovich, Cerrone, Cr...
AINL 2016: Rykov, Nagornyy, Koltsova, Natta, Kremenets, Manovich, Cerrone, Cr...
 
AINL 2016: Yagunova
AINL 2016: YagunovaAINL 2016: Yagunova
AINL 2016: Yagunova
 
AINL 2016: Ustalov
AINL 2016: Ustalov AINL 2016: Ustalov
AINL 2016: Ustalov
 
AINL 2016: Galinsky, Alekseev, Nikolenko
AINL 2016: Galinsky, Alekseev, NikolenkoAINL 2016: Galinsky, Alekseev, Nikolenko
AINL 2016: Galinsky, Alekseev, Nikolenko
 
AINL 2016: Boldyreva
AINL 2016: BoldyrevaAINL 2016: Boldyreva
AINL 2016: Boldyreva
 
AINL 2016: Skornyakov
AINL 2016: SkornyakovAINL 2016: Skornyakov
AINL 2016: Skornyakov
 
AINL 2016: Eyecioglu
AINL 2016: EyeciogluAINL 2016: Eyecioglu
AINL 2016: Eyecioglu
 
AINL 2016: Kozerenko
AINL 2016: Kozerenko AINL 2016: Kozerenko
AINL 2016: Kozerenko
 
AINL 2016: Goncharov
AINL 2016: GoncharovAINL 2016: Goncharov
AINL 2016: Goncharov
 
AINL 2016: Muravyov
AINL 2016: MuravyovAINL 2016: Muravyov
AINL 2016: Muravyov
 
AINL 2016: Panicheva, Ledovaya
AINL 2016: Panicheva, LedovayaAINL 2016: Panicheva, Ledovaya
AINL 2016: Panicheva, Ledovaya
 
AINL 2016: Kuznetsova
AINL 2016: KuznetsovaAINL 2016: Kuznetsova
AINL 2016: Kuznetsova
 
AINL 2016: Romanova, Nefedov
AINL 2016: Romanova, NefedovAINL 2016: Romanova, Nefedov
AINL 2016: Romanova, Nefedov
 

Ähnlich wie AINL 2016: Filchenkov

Deep learning - A Visual Introduction
Deep learning - A Visual IntroductionDeep learning - A Visual Introduction
Deep learning - A Visual Introduction
Lukas Masuch
 

Ähnlich wie AINL 2016: Filchenkov (20)

Introduction to deep learning
Introduction to deep learningIntroduction to deep learning
Introduction to deep learning
 
A Survey of Deep Learning Algorithms for Malware Detection
A Survey of Deep Learning Algorithms for Malware DetectionA Survey of Deep Learning Algorithms for Malware Detection
A Survey of Deep Learning Algorithms for Malware Detection
 
Looking into the Black Box - A Theoretical Insight into Deep Learning Networks
Looking into the Black Box - A Theoretical Insight into Deep Learning NetworksLooking into the Black Box - A Theoretical Insight into Deep Learning Networks
Looking into the Black Box - A Theoretical Insight into Deep Learning Networks
 
Deep learning - A Visual Introduction
Deep learning - A Visual IntroductionDeep learning - A Visual Introduction
Deep learning - A Visual Introduction
 
IRJET- Survey on Text Error Detection using Deep Learning
IRJET-  	  Survey on Text Error Detection using Deep LearningIRJET-  	  Survey on Text Error Detection using Deep Learning
IRJET- Survey on Text Error Detection using Deep Learning
 
deeplearning
deeplearningdeeplearning
deeplearning
 
BASIC CONCEPT OF DEEP LEARNING.pptx
BASIC CONCEPT OF DEEP LEARNING.pptxBASIC CONCEPT OF DEEP LEARNING.pptx
BASIC CONCEPT OF DEEP LEARNING.pptx
 
Deep Learning Training at Intel
Deep Learning Training at IntelDeep Learning Training at Intel
Deep Learning Training at Intel
 
Introduction to un supervised learning
Introduction to un supervised learningIntroduction to un supervised learning
Introduction to un supervised learning
 
DSRLab seminar Introduction to deep learning
DSRLab seminar   Introduction to deep learningDSRLab seminar   Introduction to deep learning
DSRLab seminar Introduction to deep learning
 
MDEC Data Matters Series: machine learning and Deep Learning, A Primer
MDEC Data Matters Series: machine learning and Deep Learning, A PrimerMDEC Data Matters Series: machine learning and Deep Learning, A Primer
MDEC Data Matters Series: machine learning and Deep Learning, A Primer
 
Machine learning in science and industry — day 4
Machine learning in science and industry — day 4Machine learning in science and industry — day 4
Machine learning in science and industry — day 4
 
UNSUPERVISED LEARNING MODELS OF INVARIANT FEATURES IN IMAGES: RECENT DEVELOPM...
UNSUPERVISED LEARNING MODELS OF INVARIANT FEATURES IN IMAGES: RECENT DEVELOPM...UNSUPERVISED LEARNING MODELS OF INVARIANT FEATURES IN IMAGES: RECENT DEVELOPM...
UNSUPERVISED LEARNING MODELS OF INVARIANT FEATURES IN IMAGES: RECENT DEVELOPM...
 
UNSUPERVISED LEARNING MODELS OF INVARIANT FEATURES IN IMAGES: RECENT DEVELOPM...
UNSUPERVISED LEARNING MODELS OF INVARIANT FEATURES IN IMAGES: RECENT DEVELOPM...UNSUPERVISED LEARNING MODELS OF INVARIANT FEATURES IN IMAGES: RECENT DEVELOPM...
UNSUPERVISED LEARNING MODELS OF INVARIANT FEATURES IN IMAGES: RECENT DEVELOPM...
 
Unsupervised learning models of invariant features in images: Recent developm...
Unsupervised learning models of invariant features in images: Recent developm...Unsupervised learning models of invariant features in images: Recent developm...
Unsupervised learning models of invariant features in images: Recent developm...
 
Top 10 deep learning algorithms you should know in
Top 10 deep learning algorithms you should know inTop 10 deep learning algorithms you should know in
Top 10 deep learning algorithms you should know in
 
CCS355 Neural Networks & Deep Learning Unit 1 PDF notes with Question bank .pdf
CCS355 Neural Networks & Deep Learning Unit 1 PDF notes with Question bank .pdfCCS355 Neural Networks & Deep Learning Unit 1 PDF notes with Question bank .pdf
CCS355 Neural Networks & Deep Learning Unit 1 PDF notes with Question bank .pdf
 
Basics of Deep learning
Basics of Deep learningBasics of Deep learning
Basics of Deep learning
 
Deep Learning - The Past, Present and Future of Artificial Intelligence
Deep Learning - The Past, Present and Future of Artificial IntelligenceDeep Learning - The Past, Present and Future of Artificial Intelligence
Deep Learning - The Past, Present and Future of Artificial Intelligence
 
Big Data Intelligence: from Correlation Discovery to Causal Reasoning
Big Data Intelligence: from Correlation Discovery to Causal Reasoning Big Data Intelligence: from Correlation Discovery to Causal Reasoning
Big Data Intelligence: from Correlation Discovery to Causal Reasoning
 

Mehr von Lidia Pivovarova

Mehr von Lidia Pivovarova (12)

Classification and clustering in media monitoring: from knowledge engineering...
Classification and clustering in media monitoring: from knowledge engineering...Classification and clustering in media monitoring: from knowledge engineering...
Classification and clustering in media monitoring: from knowledge engineering...
 
Convolutional neural networks for text classification
Convolutional neural networks for text classificationConvolutional neural networks for text classification
Convolutional neural networks for text classification
 
Grouping business news stories based on salience of named entities
Grouping business news stories based on salience of named entitiesGrouping business news stories based on salience of named entities
Grouping business news stories based on salience of named entities
 
Интеллектуальный анализ текста
Интеллектуальный анализ текстаИнтеллектуальный анализ текста
Интеллектуальный анализ текста
 
AINL 2016: Shavrina, Selegey
AINL 2016: Shavrina, SelegeyAINL 2016: Shavrina, Selegey
AINL 2016: Shavrina, Selegey
 
AINL 2016: Khudobakhshov
AINL 2016: KhudobakhshovAINL 2016: Khudobakhshov
AINL 2016: Khudobakhshov
 
AINL 2016:
AINL 2016: AINL 2016:
AINL 2016:
 
AINL 2016: Grigorieva
AINL 2016: GrigorievaAINL 2016: Grigorieva
AINL 2016: Grigorieva
 
AINL 2016: Just AI
AINL 2016: Just AIAINL 2016: Just AI
AINL 2016: Just AI
 
AINL 2016: Moskvichev
AINL 2016: MoskvichevAINL 2016: Moskvichev
AINL 2016: Moskvichev
 
AINL 2016: Malykh
AINL 2016: MalykhAINL 2016: Malykh
AINL 2016: Malykh
 
AINL 2016: Strijov
AINL 2016: StrijovAINL 2016: Strijov
AINL 2016: Strijov
 

Kürzlich hochgeladen

dkNET Webinar "Texera: A Scalable Cloud Computing Platform for Sharing Data a...
dkNET Webinar "Texera: A Scalable Cloud Computing Platform for Sharing Data a...dkNET Webinar "Texera: A Scalable Cloud Computing Platform for Sharing Data a...
dkNET Webinar "Texera: A Scalable Cloud Computing Platform for Sharing Data a...
dkNET
 
Chemical Tests; flame test, positive and negative ions test Edexcel Internati...
Chemical Tests; flame test, positive and negative ions test Edexcel Internati...Chemical Tests; flame test, positive and negative ions test Edexcel Internati...
Chemical Tests; flame test, positive and negative ions test Edexcel Internati...
ssuser79fe74
 
Bacterial Identification and Classifications
Bacterial Identification and ClassificationsBacterial Identification and Classifications
Bacterial Identification and Classifications
Areesha Ahmad
 
Pests of cotton_Borer_Pests_Binomics_Dr.UPR.pdf
Pests of cotton_Borer_Pests_Binomics_Dr.UPR.pdfPests of cotton_Borer_Pests_Binomics_Dr.UPR.pdf
Pests of cotton_Borer_Pests_Binomics_Dr.UPR.pdf
PirithiRaju
 
Conjugation, transduction and transformation
Conjugation, transduction and transformationConjugation, transduction and transformation
Conjugation, transduction and transformation
Areesha Ahmad
 
SCIENCE-4-QUARTER4-WEEK-4-PPT-1 (1).pptx
SCIENCE-4-QUARTER4-WEEK-4-PPT-1 (1).pptxSCIENCE-4-QUARTER4-WEEK-4-PPT-1 (1).pptx
SCIENCE-4-QUARTER4-WEEK-4-PPT-1 (1).pptx
RizalinePalanog2
 
Module for Grade 9 for Asynchronous/Distance learning
Module for Grade 9 for Asynchronous/Distance learningModule for Grade 9 for Asynchronous/Distance learning
Module for Grade 9 for Asynchronous/Distance learning
levieagacer
 
Introduction,importance and scope of horticulture.pptx
Introduction,importance and scope of horticulture.pptxIntroduction,importance and scope of horticulture.pptx
Introduction,importance and scope of horticulture.pptx
Bhagirath Gogikar
 

Kürzlich hochgeladen (20)

Kochi ❤CALL GIRL 84099*07087 ❤CALL GIRLS IN Kochi ESCORT SERVICE❤CALL GIRL
Kochi ❤CALL GIRL 84099*07087 ❤CALL GIRLS IN Kochi ESCORT SERVICE❤CALL GIRLKochi ❤CALL GIRL 84099*07087 ❤CALL GIRLS IN Kochi ESCORT SERVICE❤CALL GIRL
Kochi ❤CALL GIRL 84099*07087 ❤CALL GIRLS IN Kochi ESCORT SERVICE❤CALL GIRL
 
dkNET Webinar "Texera: A Scalable Cloud Computing Platform for Sharing Data a...
dkNET Webinar "Texera: A Scalable Cloud Computing Platform for Sharing Data a...dkNET Webinar "Texera: A Scalable Cloud Computing Platform for Sharing Data a...
dkNET Webinar "Texera: A Scalable Cloud Computing Platform for Sharing Data a...
 
Chemical Tests; flame test, positive and negative ions test Edexcel Internati...
Chemical Tests; flame test, positive and negative ions test Edexcel Internati...Chemical Tests; flame test, positive and negative ions test Edexcel Internati...
Chemical Tests; flame test, positive and negative ions test Edexcel Internati...
 
Bacterial Identification and Classifications
Bacterial Identification and ClassificationsBacterial Identification and Classifications
Bacterial Identification and Classifications
 
pumpkin fruit fly, water melon fruit fly, cucumber fruit fly
pumpkin fruit fly, water melon fruit fly, cucumber fruit flypumpkin fruit fly, water melon fruit fly, cucumber fruit fly
pumpkin fruit fly, water melon fruit fly, cucumber fruit fly
 
CELL -Structural and Functional unit of life.pdf
CELL -Structural and Functional unit of life.pdfCELL -Structural and Functional unit of life.pdf
CELL -Structural and Functional unit of life.pdf
 
Pests of cotton_Borer_Pests_Binomics_Dr.UPR.pdf
Pests of cotton_Borer_Pests_Binomics_Dr.UPR.pdfPests of cotton_Borer_Pests_Binomics_Dr.UPR.pdf
Pests of cotton_Borer_Pests_Binomics_Dr.UPR.pdf
 
Call Girls Alandi Call Me 7737669865 Budget Friendly No Advance Booking
Call Girls Alandi Call Me 7737669865 Budget Friendly No Advance BookingCall Girls Alandi Call Me 7737669865 Budget Friendly No Advance Booking
Call Girls Alandi Call Me 7737669865 Budget Friendly No Advance Booking
 
Connaught Place, Delhi Call girls :8448380779 Model Escorts | 100% verified
Connaught Place, Delhi Call girls :8448380779 Model Escorts | 100% verifiedConnaught Place, Delhi Call girls :8448380779 Model Escorts | 100% verified
Connaught Place, Delhi Call girls :8448380779 Model Escorts | 100% verified
 
Zoology 5th semester notes( Sumit_yadav).pdf
Zoology 5th semester notes( Sumit_yadav).pdfZoology 5th semester notes( Sumit_yadav).pdf
Zoology 5th semester notes( Sumit_yadav).pdf
 
9654467111 Call Girls In Raj Nagar Delhi Short 1500 Night 6000
9654467111 Call Girls In Raj Nagar Delhi Short 1500 Night 60009654467111 Call Girls In Raj Nagar Delhi Short 1500 Night 6000
9654467111 Call Girls In Raj Nagar Delhi Short 1500 Night 6000
 
Conjugation, transduction and transformation
Conjugation, transduction and transformationConjugation, transduction and transformation
Conjugation, transduction and transformation
 
SCIENCE-4-QUARTER4-WEEK-4-PPT-1 (1).pptx
SCIENCE-4-QUARTER4-WEEK-4-PPT-1 (1).pptxSCIENCE-4-QUARTER4-WEEK-4-PPT-1 (1).pptx
SCIENCE-4-QUARTER4-WEEK-4-PPT-1 (1).pptx
 
Module for Grade 9 for Asynchronous/Distance learning
Module for Grade 9 for Asynchronous/Distance learningModule for Grade 9 for Asynchronous/Distance learning
Module for Grade 9 for Asynchronous/Distance learning
 
Clean In Place(CIP).pptx .
Clean In Place(CIP).pptx                 .Clean In Place(CIP).pptx                 .
Clean In Place(CIP).pptx .
 
Introduction,importance and scope of horticulture.pptx
Introduction,importance and scope of horticulture.pptxIntroduction,importance and scope of horticulture.pptx
Introduction,importance and scope of horticulture.pptx
 
COMPUTING ANTI-DERIVATIVES (Integration by SUBSTITUTION)
COMPUTING ANTI-DERIVATIVES(Integration by SUBSTITUTION)COMPUTING ANTI-DERIVATIVES(Integration by SUBSTITUTION)
COMPUTING ANTI-DERIVATIVES (Integration by SUBSTITUTION)
 
Vip profile Call Girls In Lonavala 9748763073 For Genuine Sex Service At Just...
Vip profile Call Girls In Lonavala 9748763073 For Genuine Sex Service At Just...Vip profile Call Girls In Lonavala 9748763073 For Genuine Sex Service At Just...
Vip profile Call Girls In Lonavala 9748763073 For Genuine Sex Service At Just...
 
PSYCHOSOCIAL NEEDS. in nursing II sem pptx
PSYCHOSOCIAL NEEDS. in nursing II sem pptxPSYCHOSOCIAL NEEDS. in nursing II sem pptx
PSYCHOSOCIAL NEEDS. in nursing II sem pptx
 
9999266834 Call Girls In Noida Sector 22 (Delhi) Call Girl Service
9999266834 Call Girls In Noida Sector 22 (Delhi) Call Girl Service9999266834 Call Girls In Noida Sector 22 (Delhi) Call Girl Service
9999266834 Call Girls In Noida Sector 22 (Delhi) Call Girl Service
 

AINL 2016: Filchenkov

  • 1. Deep Neural Networks Tutorial Andrey Filchenkov Computer Technology Chair Computer Technologies Lab ITMO University afilchenkov@corp.ifmo.ru AINL FRUCT’ 6. No , 6, “t. Peters urg, Russia
  • 2. Tutorial topics Very brief introduction to AI, ML and ANN What is ANN and how to learn it DNN and standard DNN architectures Beyond discriminative models 2 / 64
  • 3. Next topic Very brief introduction to AI, ML and ANN What is ANN and how to learn it DNN and standard DNN architectures Beyond discriminative models 3 / 64
  • 4. Artificial intelligence Strong AI (A general I): functionality is similar to the human brain or better. Weak AI: good in solving certain well-formulated tasks. Machine learning is a part of Weak AI Many people have been thinking that artificial neural networks are a path to Strong AI Many people are thinking that deep learning networks are a path to Strong AI 4 / 64
  • 5. Neural networks as a machine learning algorithm Neural paradigm is not only about machine learning (computer architecture, computations, etc.) Machine learning is about creating algorithms which can learn patterns, regularities and rules from given data. The biggest part of machine learning is supervised learning: we are given a set of objects, each has a label, and we want to learn how to find these label for objects we have never seen. 5 / 64
  • 7. Brief early history of artificial neural networks 1943 Artificial neuron by McCulloch and Pitts 1949 Neuron learning rule by Hebb 1957 Perceptron by Rosenblatt 1960 Perceptron learning rule by Widrow and Hoff 1968 Group Method of Data Handling to learn multilayered networks by Ivakhnenko 1969 Perceptrons by Minski and Papert 1974 Back propagation algorithm by Webros and by Galushkin 7 / 64
  • 8. Brief modern history of ANN 1980 Convolutional NN by Fukushima 1982 Recurrent NN by Hopfield 1991 Va ishi g gradient pro le was identified by Hochreiter 1997 Long short term memory network by Hochreiter and Schmidhuber 1998 Gradient descent for convolutional NN by LeCun et al. 2006 Deep model by Hinton, Osindero and Teh 2012 DNN started to become mainstream in ML and AI 8 / 64
  • 9. Next topic Very brief introduction to AI, ML and ANN What is ANN and how to learn it DNN and standard DNN architectures Beyond discriminative models 9 / 64
  • 10. Two sources of knowledge Experts we need to ask wisely and process Data we need to process and apply machine learning algorithms How do we obtain knowledge? 10 / 64
  • 11. Algorithms, performance of which grows with experience The most popular task is prediction Algorithms require data and labels (for predicting) Learning of these algorithms is to minimize error rate in prediction or maximize similar to the known answer Machine learning 11 / 64
  • 12. Each object is represented as a feature vector. Each object thus is a point in a multidimensional space. Vector representation of objects 12 / 64
  • 13. Linear discriminant surface Linear model (neuron) 13 / 64
  • 15. Multilayer neural network We just build a composition of neurons (as functions) 15 / 64
  • 16. Next topic Very brief introduction to AI, ML and ANN What is ANN and how to learn it DNN and standard DNN architectures • Deep learning introduction and best practices • Deep Boltzmann Machines (DBM) and Deep Belief Network (DBN) • Convolution Neural Network (CNN) • Autoencoders • Recurrental Neural Network (RNN) and Long-Short Term Memory (LSTM) Beyond discriminative models 16 / 64
  • 17. Deep architecture Definition: Deep architectures are composed of multiple levels of non-linear operations, such as neural nets with many hidden layers Most machine learning algorithms have shallow (1–3 layers) architecture (SVM, PCA, kNN, Logistic Regression, etc.) Goal: Deep learning methods aim at: Learning feature hierarchies, no more feature engineering! Where features from higher levels of the hierarchy are formed by lower level features. 17 / 64
  • 18. Why to go deep? Some functions cannot be efficiently represented (in terms of number of tunable elements) by architectures that are too shallow Functions that can be compactly represented by a depth k architecture might require an exponential number of computational elements to be represented by a depth k− ar hite ture Deep Representations might allow non-local generalization and comprehensibility Deep learning gets state of the art results in many fields (vision, audio, NLP, etc.)! 18 / 64
  • 19. DNN best practices: ReLU, PReLU ReLU 19 / 64 PReLU Sigmoid, hyperbolic tangent activation functions have a problem with vanishing of gradients and tend to overfitting
  • 20. DNN best practices: Data augmentation The easiest and most common method to reduce overfitting on image data is to artificially enlarge the dataset using label- preserving transformations. Types of data augmentation: Image translation Horizontal/vertical reflections + cropping Changing RGB intensities 20 / 64
  • 21. DNN best practices: Dropout Dropout: set the output of each hidden neuron to zero w.p. 0.5 The euro s hi h are dropped out i this a do ot o tri ute to the forward pass and do not participate in backpropagation So every time an input is presented, the neural network samples a different architecture, but all these architectures share weights This technique reduces complex co-adaptations of neurons, since a neuron cannot rely on the presence of particular other neurons It is, therefore, forced to learn more robust features that are useful in conjunction with many different random subsets of the other neurons Without dropout, a network exhibits substantial overfitting Dropout roughly doubles the number of iterations required to converge 21 / 64
  • 22. Greedy Layer-Wise Training (1/2) 1. Train first layer using your data without the labels (unsupervised) • Since there are no targets at this level, labels don't help. Could also use the more abundant unlabeled data which is not part of the training set (i.e. self-taught learning). 2. Then freeze the first layer parameters and start training the second layer using the output of the first layer as the unsupervised input to the second layer 3. Repeat this for as many layers as desired • This builds our set of robust features 4. Use the outputs of the final layer as inputs to a supervised layer/model and train the last supervised layer(s) (leave early weights frozen) 5. Unfreeze all weights and fine tune the full network by training with a supervised approach, given the pre-processed weight settings 22 / 64
  • 23. Greedy Layer-Wise Training (2/2) Greedy layer-wise training avoids many of the problems of trying to train a deep net in a supervised fashion: • Each layer gets full learning focus in its turn since it is the only current "top" layer • Can take advantage of unlabeled data • When you finally tune the entire network with supervised training the network weights have already been adjusted so that you are in a good error basin and just need fine tuning. This helps with problems of: • Ineffective early layer learning • Deep network local minima 23 / 64
  • 24. Restricted Boltzmann machine Two types of nodes: hidden and visible. We are minimizing system energy which is not to converge by updating its weights with propagating new objects. Probability distribution on visible and hidden layers is Gibbs distribution. 24 / 64
  • 25. Deep Belief Network 25 / 64 First train unsupervisedly with several (two) levels of RBM (or autoencoders). Then train next layers supervisedly and consecutively. Hinton, G. E. and Salakhutdinov, R. R. (2006). Reducing the dimensionality of data with neural networks. Science, 313:504-507.
  • 27. Convolution Neural Network (CNN) Core concepts: Local perception – each neuron sees a small part of the object. Use kernels (filters) to capture 1-D or 2-D structure of objects. For instance, capture all pixel neighbors for an image. Weight sharing – use small and the same sets of kernels for all objects, this leads to reduction of number of adjusting parameters in comparison with MLP Subsampling/pooling – use dimensionality reduction for images in order to provide invariance to scale 27 / 64
  • 29. What can kernels do? 29 / 64 * * * = = = blur edge detection sharpen
  • 31. How do trained kernels look like? 31 / 64 low feature medium feature high feature Each kernel composes a local patch of lower-level features into high level representation
  • 32. Levels of abstraction Hierarchical Learning: Natural progression from low level to high level structure as seen in natural complexity Easier to monitor what is being learnt and to guide the machine to better subspaces A good lower level representation can be used for many distinct tasks 32 / 64
  • 33. LeNet 33 / 64 LeCun, Yann, et al. "Gradient-based learning applied to document recognition." Proceedings of the IEEE 86.11 (1998): 2278-2324.
  • 34. GoogLeNet 34 / 64 Szegedy, Christian, et al. "Going deeper with convolutions." arXiv preprint arXiv:1409.4842 (2014).
  • 35. Application of CNN Image Recognition Image Search (enhance search engine) Visual Question Answering NLP (Sentence Classification, etc.) Speech Recognition … 35 / 64
  • 36. Visual question answering (1/2) 36 / 64
  • 37. Visual question answering (2/2) 37 / 64
  • 38. Application of CNN: Visual question answering 38 / 64
  • 39. Autoencoders (1/3) Autoencoder: a feed-forward neural network trained to reproduce its input at the output layer Do non-linear dimensionality reduction Train via backpropagation 1-layer autoencoder gets similar results as PCA 39 / 64
  • 42. Autoencoders in bioinformatics 42 / 64 Fakoor, Rasool, et al. "Using deep learning to enhance cancer diagnosis and classification." Proceedings of the International Conference on Machine Learning. 2013.
  • 43. Deep autoencoders: document processing We can use an autoencoder to find low-dimensional codes for documents that allow fast and accurate retrieval of similar documents from a large set. We start with o erti g ea h do u e t i to a ag of ords . This is a 2000 dimensional vector that contains the counts for each of the 2000 commonest words. 43 / 64
  • 44. Deep autoencoders: document retrieval We train the neural network to reproduce its input vector as its output This forces it to compress as much information as possible into the 10 numbers in the central bottleneck. These 10 numbers are then a good way to compare documents. 44 / 64 2000 reconstructed counts 500 neurons 250 neurons 10 250 neurons 500 neurons 2000 word counts
  • 45. Deep autoencoders: document visualization (1/2) Instead of using codes to retrieve documents, we can use 2-D codes to visualize sets of documents. This works much better than 2-D PCA 45 / 64 2000 reconstructed counts 500 neurons 250 neurons 2 250 neurons 500 neurons 2000 word counts
  • 46. Deep autoencoders: document visualization (2/2) 46 / 64
  • 47. Recurrent Neural Network (RNN) RNN: a neural network with recurrent connections Good for sequence data: time series, text, audio 47 / 64
  • 48. Recurrent Neural Network (RNN): scheme 48 / 64
  • 49. Long Short-Term Memory (LSTM) LSTM: a special case of RNN capable of learning long-term dependencies There are four neural network layers in repeating module 49 / 64Hochreiter, Sepp, and Jürgen Schmidhuber. "Long short-term memory." Neural computation 9.8 (1997): 1735-1780.
  • 50. LSTM: Cell state Cell state: runs straight down the entire chain, with only some minor linear interactions LSTM does have the ability to remove or add information to the cell state, carefully regulated by structures called gates. Gates are a way to optionally let information through. The sigmoid layer outputs numbers between zero and one, describing how much of each component should be let through. LSTM has 3 gates. 50 / 64
  • 51. LSTM: Forget gate layer It looks at ℎt−1 and �t, and outputs a number between 0 and 1 for each number in the cell state �t−1. 1 represents o pletel keep this hile 0 represe ts o pletel get rid of this . 51 / 64
  • 52. LSTM: Input gate layer (1/2) How to decide what new information to store in the cell state? First, a sig oid la er alled the i put gate la er de ides which values should be updated. Next, a tanh layer creates a vector of new candidate values, �t, that ould e added to the state. I the e t step, e’ll combine these two to create an update to the state 52 / 64
  • 53. LSTM: Input gate layer (2/2) It’s o ti e to update the old ell state, ��−1, into the new cell state ��. The previous steps already decided what to do, we just need to actually do it We multiply the old state by ��, forgetting the things we decided to forget earlier. Then we add �� ⋅ ��−1. This is the new candidate values, scaled by how much we decided to update each state value 53 / 64
  • 54. LSTM: Output gate layer The output will be based on cell state, but will be a filtered version. First, we run a sigmoid layer which decides what parts of the ell state e’re goi g to output. The , e put the cell state through tanh to push the alues to e et ee − and 1) and multiply it by the output of the sigmoid gate, so that we only output the parts we decided to. 54 / 64
  • 56. Deep learning analysis: advantages Extremely strong model, which can potentially solve any problem of machine learning. Already learnt model can be reused: multi-task support Many good models are already known which are state-of-the- art for many tasks: • image recognition; • speech recognition; • natural language processing; • etc. 56 / 64
  • 57. Deep learning analysis: disadvantages The deeper the net is • the more data you need; • the more time you need; • the stronger processors you need. Usually no intuition how it works exactly; Usually you work with DNN as a black box; Prone to overfitting: regularization must be used. 57 / 64
  • 58. Next topic Very brief introduction to AI, ML and ANN What is ANN and how to learn it DNN and standard DNN architectures Beyond discriminative models 58 / 64
  • 59. Reverse the network and make it predict images given labels Image synthesis Dosovitskiy, A., Tobias Springenberg, J., & Brox, T. (2015). Learning to generate chairs with convolutional neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (pp. 1538-1546). 59 / 64
  • 60. Keep i er represe tatio of a i age (Gram matrix �� for convolutional layers) Then create a new random network and learn it no have similar inner representation as the one we have kept. Texture synthesis Gatys, L., Ecker, A. S., & Bethge, M. (2015). Texture synthesis using convolutional neural networks. In Advances in Neural Information Processing Systems (pp. 262-270). 60 / 64
  • 61. Style = texture. Image = content and is represented with the last convolutional layer. We will learn an image that is similar both to image and content. Style transmission Gatys, L. A., Ecker, A. S., & Bethge, M. (2015). A neural algorithm of artistic style. arXiv preprint arXiv:1508.06576. 61 / 64
  • 62. DeetArt was created in 2015: https://deepart.io/ They implemented the algorithm described before. DeepArt and Prisma 62 / 64 Prisma was created in June, 2016. They made it optimized, mobiles and with preselected filters (instead of styles)
  • 63. Materials Presentation was prepared using: 1. http://avisingh599.github.io/deeplearning/visual-qa/ 2. http://colah.github.io/posts/2015-08-Understanding-LSTMs/ 3. https://class.coursera.org/ml-003/lecture 4. K. Vorontsov Ma hi e lear i g ourse i Russia 63 / 64