SlideShare ist ein Scribd-Unternehmen logo
1 von 21
Deep Learning Basics
• Artificial Neural Networks (ANN): Deep
learning is a subset of machine learning that
uses artificial neural networks to learn from
data.
• Layers: ANNs are composed of layers of
interconnected nodes or neurons. The input
layer receives the data, the output layer
produces the predictions, and the hidden
layers perform the computations in between.
• Activation functions: Activation functions are applied
to the outputs of each layer to introduce non-linearity
and increase the model's expressiveness. Common
activation functions include ReLU, sigmoid, and tanh.
• Backpropagation: This is a method for training neural
networks by iteratively adjusting the weights in each
layer to minimize the difference between the predicted
outputs and the actual outputs. It works by
propagating the error backwards from the output layer
to the input layer and adjusting the weights accordingly
• Loss function: The loss function measures the
difference between the predicted outputs and the
actual outputs. The goal of training is to minimize the
loss function by adjusting the weights in each layer.
• Optimization algorithms: Optimization algorithms,
such as stochastic gradient descent (SGD) and Adam,
are used to adjust the weights in each layer during
training to minimize the loss function.
• Overfitting: Overfitting occurs when a model is too
complex and starts to memorize the training data
instead of learning the underlying patterns. This can be
prevented by using techniques
Activation Functions in Artificial Neural
Networks (ANNs):
• Activation functions are mathematical functions
used in ANNs to introduce non-linearity into the
output of a neuron or a layer.
• They are typically applied to the weighted sum of
inputs and biases before being passed through to
the next layer of the network.
• Without activation functions, ANNs would simply
be a linear regression model, which can only
model linear relationships between input and
output.
• Common activation functions include Sigmoid, Tanh, ReLU, and
Softmax.
• Sigmoid and Tanh functions are sigmoidal, meaning they produce
an S-shaped curve. They are used to squash the output of a neuron
to a range between 0 and 1 or -1 and 1, respectively.
• ReLU (Rectified Linear Unit) function is non-sigmoidal and is defined
as f(x) = max(0, x). It is one of the most commonly used activation
functions due to its simplicity and effectiveness.
• Softmax function is used in the output layer of a network to
produce a probability distribution over multiple classes.
• Choosing the right activation function can have a significant impact
on the performance of a neural network, and it is often an area of
active research.
Non Linearity
• In the context of machine learning, nonlinearity is
an important property of neural networks. Neural
networks are composed of many interconnected
processing units (neurons), which apply a
nonlinear activation function to their inputs
before passing them to the next layer of the
network. This nonlinearity allows neural
networks to model complex relationships
between inputs and outputs, and to learn
representations that are not directly observable
in the input data.
Simple Linear Regression Model
• For example, a linear model can only learn a linear
decision boundary between two classes, which may
not be sufficient to accurately classify complex data. In
contrast, a nonlinear model such as a neural network
can learn more complex decision boundaries that can
better separate the classes.
• Some common nonlinear activation functions used in
neural networks include the Rectified Linear Unit
(ReLU), sigmoid, tanh, and others. These functions
introduce nonlinearity into the network, allowing it to
learn more complex representations of the input data.
Rectified Linear Unit (ReLU) activation
function:
• ReLU is a non-linear activation function
commonly used in neural networks.
• It takes an input value x and returns the
maximum of 0 and x as the output value.
• The formula for ReLU is f(x) = max(0, x).
• ReLU is computationally efficient, since it
requires only simple thresholding of the input
value, compared to other activation functions
like sigmoid or tanh.
Sigmoid
• Sigmoid is a non-linear activation function commonly used in neural networks.
• It takes an input value x and maps it to a range between 0 and 1 using the formula
f(x) = 1 / (1 + e^-x).
• The output of the sigmoid function can be interpreted as a probability or
likelihood, since it always produces a value between 0 and 1.
• Sigmoid was one of the earliest activation functions used in neural networks, and
it is still used in some applications, such as logistic regression.
• Sigmoid is smooth and differentiable, which makes it useful for backpropagation
and gradient descent optimization.
• One limitation of sigmoid is that it suffers from the "vanishing gradient" problem,
where the gradient becomes very small as the input value becomes very large or
very small, making it difficult for the network to learn.
• Another limitation of sigmoid is that it is not zero-centered, which can slow down
the convergence of gradient descent.
• Due to these limitations, sigmoid is not as commonly used as other activation
functions like ReLU or its variants.
•
Tanh
• Tanh is a non-linear activation function commonly used in neural networks.
• It takes an input value x and maps it to a range between -1 and 1 using the formula f(x) = (e^x - e^-
x) / (e^x + e^-x).
• Tanh is a shifted and rescaled version of the sigmoid function, with the output value zero-centered.
• Like sigmoid, tanh is smooth and differentiable, which makes it useful for backpropagation and
gradient descent optimization.
• Tanh is often used in the hidden layers of neural networks, especially in recurrent neural networks
(RNNs) and long short-term memory (LSTM) networks.
• One limitation of tanh is that it also suffers from the "vanishing gradient" problem, where the
gradient becomes very small as the input value becomes very large or very small, making it difficult
for the network to learn.
• Another limitation of tanh is that it is more computationally expensive than ReLU or its variants,
since it involves exponentials.
• Despite its limitations, tanh can be useful in certain situations, such as when the input data is
standardized and zero-centered, or when the network needs to model both positive and negative
values.
Difference between Activation
Function and ML Algorithm
• An activation function is a mathematical function used in artificial neural networks
to introduce non-linearity into the output of a neuron.
• Activation functions are used to decide whether the neuron should be activated or
not based on the input it receives.
• Common activation functions include sigmoid, ReLU, tanh, and softmax.
• On the other hand, a machine learning algorithm is a method or set of methods
used to learn patterns and relationships
• in data in order to make predictions or decisions. Machine learning algorithms can
be supervised, unsupervised,
• or semi-supervised, and can be used for a wide range of tasks, such as regression,
classification, clustering, and reinforcement learning.
• While activation functions are used in neural networks to introduce non-linearity
and make them more expressive, machine learning algorithms are used to learn
patterns and relationships in data and make predictions based on that learning.
Why Accuracy of ML is better than DL
• Complexity: ML models are often simpler and more
interpretable than ANNs, which can make them easier to
train and optimize. In some cases, a simpler model may be
sufficient to achieve good performance on a given task,
without the need for a complex neural network.
• Data size: ANNs require large amounts of data to train
effectively, and may not perform well on small datasets. In
contrast, some ML models, such as decision trees or logistic
regression, can perform well even on smaller datasets.
• Feature engineering: ANNs often require extensive feature
engineering and preprocessing of input data, which can be
time-consuming and require domain expertise. In contrast,
some ML models, such as decision trees or Naive Bayes,
can perform well with minimal feature engineering.
• Model selection: Choosing the right ANN architecture and
hyperparameters can be a challenging task, and may require
extensive experimentation and tuning. In contrast, some ML
models, such as decision trees or Naive Bayes, have fewer
hyperparameters to tune and may be easier to select and optimize.
• Overfitting: ANNs are prone to overfitting, where the model
becomes too complex and performs well on the training data but
poorly on new, unseen data. ML models may be less prone to
overfitting, especially when regularized or constrained in some way.
• Overall, the choice of model depends on the specific task and
dataset at hand, and there is no one-size-fits-all solution. In some
cases, an ML model may be more suitable, while in other cases, an
ANN may be necessary to achieve the desired performance

Weitere ähnliche Inhalte

Ähnlich wie Deep Learning Basics.pptx

What are activation functions and why do we need those.pdf
What are activation functions and why do we need those.pdfWhat are activation functions and why do we need those.pdf
What are activation functions and why do we need those.pdfseo18
 
Facial Emotion Detection on Children's Emotional Face
Facial Emotion Detection on Children's Emotional FaceFacial Emotion Detection on Children's Emotional Face
Facial Emotion Detection on Children's Emotional FaceTakrim Ul Islam Laskar
 
Neural Network_basic_Reza_Lecture_3.pptx
Neural Network_basic_Reza_Lecture_3.pptxNeural Network_basic_Reza_Lecture_3.pptx
Neural Network_basic_Reza_Lecture_3.pptxshamimreza94
 
Unit one ppt of deeep learning which includes Ann cnn
Unit one ppt of  deeep learning which includes Ann cnnUnit one ppt of  deeep learning which includes Ann cnn
Unit one ppt of deeep learning which includes Ann cnnkartikaursang53
 
A Survey of Convolutional Neural Networks
A Survey of Convolutional Neural NetworksA Survey of Convolutional Neural Networks
A Survey of Convolutional Neural NetworksRimzim Thube
 
Deep learning concepts
Deep learning conceptsDeep learning concepts
Deep learning conceptsJoe li
 
08 neural networks
08 neural networks08 neural networks
08 neural networksankit_ppt
 
Visualization of Deep Learning
Visualization of Deep LearningVisualization of Deep Learning
Visualization of Deep LearningYaminiAlapati1
 
Machine learning Module-2, 6th Semester Elective
Machine learning Module-2, 6th Semester ElectiveMachine learning Module-2, 6th Semester Elective
Machine learning Module-2, 6th Semester ElectiveMayuraD1
 
Hands on machine learning with scikit-learn and tensor flow by ahmed yousry
Hands on machine learning with scikit-learn and tensor flow by ahmed yousryHands on machine learning with scikit-learn and tensor flow by ahmed yousry
Hands on machine learning with scikit-learn and tensor flow by ahmed yousryAhmed Yousry
 
Deep learning crash course
Deep learning crash courseDeep learning crash course
Deep learning crash courseVishwas N
 
Introduction to Deep learning and H2O for beginner's
Introduction to Deep learning and H2O for beginner'sIntroduction to Deep learning and H2O for beginner's
Introduction to Deep learning and H2O for beginner'sVidyasagar Bhargava
 
ML Module 3 Non Linear Learning.pptx
ML Module 3 Non Linear Learning.pptxML Module 3 Non Linear Learning.pptx
ML Module 3 Non Linear Learning.pptxDebabrataPain1
 
33.-Multi-Layer-Perceptron.pdf
33.-Multi-Layer-Perceptron.pdf33.-Multi-Layer-Perceptron.pdf
33.-Multi-Layer-Perceptron.pdfgnans Kgnanshek
 
Designing your neural networks – a step by step walkthrough
Designing your neural networks – a step by step walkthroughDesigning your neural networks – a step by step walkthrough
Designing your neural networks – a step by step walkthroughLavanya Shukla
 
nil-100128213838-phpapp02.pdf
nil-100128213838-phpapp02.pdfnil-100128213838-phpapp02.pdf
nil-100128213838-phpapp02.pdfdlakmlkfma
 

Ähnlich wie Deep Learning Basics.pptx (20)

What are activation functions and why do we need those.pdf
What are activation functions and why do we need those.pdfWhat are activation functions and why do we need those.pdf
What are activation functions and why do we need those.pdf
 
Facial Emotion Detection on Children's Emotional Face
Facial Emotion Detection on Children's Emotional FaceFacial Emotion Detection on Children's Emotional Face
Facial Emotion Detection on Children's Emotional Face
 
Neural Network_basic_Reza_Lecture_3.pptx
Neural Network_basic_Reza_Lecture_3.pptxNeural Network_basic_Reza_Lecture_3.pptx
Neural Network_basic_Reza_Lecture_3.pptx
 
Unit one ppt of deeep learning which includes Ann cnn
Unit one ppt of  deeep learning which includes Ann cnnUnit one ppt of  deeep learning which includes Ann cnn
Unit one ppt of deeep learning which includes Ann cnn
 
Deep learning
Deep learningDeep learning
Deep learning
 
A Survey of Convolutional Neural Networks
A Survey of Convolutional Neural NetworksA Survey of Convolutional Neural Networks
A Survey of Convolutional Neural Networks
 
Deep learning concepts
Deep learning conceptsDeep learning concepts
Deep learning concepts
 
08 neural networks
08 neural networks08 neural networks
08 neural networks
 
Visualization of Deep Learning
Visualization of Deep LearningVisualization of Deep Learning
Visualization of Deep Learning
 
Machine learning Module-2, 6th Semester Elective
Machine learning Module-2, 6th Semester ElectiveMachine learning Module-2, 6th Semester Elective
Machine learning Module-2, 6th Semester Elective
 
Presentationnnnn
PresentationnnnnPresentationnnnn
Presentationnnnn
 
Hands on machine learning with scikit-learn and tensor flow by ahmed yousry
Hands on machine learning with scikit-learn and tensor flow by ahmed yousryHands on machine learning with scikit-learn and tensor flow by ahmed yousry
Hands on machine learning with scikit-learn and tensor flow by ahmed yousry
 
Deep learning crash course
Deep learning crash courseDeep learning crash course
Deep learning crash course
 
Unit 2 ml.pptx
Unit 2 ml.pptxUnit 2 ml.pptx
Unit 2 ml.pptx
 
Introduction to Deep learning and H2O for beginner's
Introduction to Deep learning and H2O for beginner'sIntroduction to Deep learning and H2O for beginner's
Introduction to Deep learning and H2O for beginner's
 
ML Module 3 Non Linear Learning.pptx
ML Module 3 Non Linear Learning.pptxML Module 3 Non Linear Learning.pptx
ML Module 3 Non Linear Learning.pptx
 
33.-Multi-Layer-Perceptron.pdf
33.-Multi-Layer-Perceptron.pdf33.-Multi-Layer-Perceptron.pdf
33.-Multi-Layer-Perceptron.pdf
 
Designing your neural networks – a step by step walkthrough
Designing your neural networks – a step by step walkthroughDesigning your neural networks – a step by step walkthrough
Designing your neural networks – a step by step walkthrough
 
nil-100128213838-phpapp02.pdf
nil-100128213838-phpapp02.pdfnil-100128213838-phpapp02.pdf
nil-100128213838-phpapp02.pdf
 
Deep learning - a primer
Deep learning - a primerDeep learning - a primer
Deep learning - a primer
 

Mehr von CallplanetsDeveloper

Data Warehousing , Data Mining and BI.pptx
Data Warehousing , Data Mining and BI.pptxData Warehousing , Data Mining and BI.pptx
Data Warehousing , Data Mining and BI.pptxCallplanetsDeveloper
 
Topic 12 Miscellaneous Concepts in IT.pptx
Topic 12 Miscellaneous Concepts in IT.pptxTopic 12 Miscellaneous Concepts in IT.pptx
Topic 12 Miscellaneous Concepts in IT.pptxCallplanetsDeveloper
 
Unit 11 AI , ML , DL and Expert Systems.pptx
Unit 11 AI , ML , DL and Expert Systems.pptxUnit 11 AI , ML , DL and Expert Systems.pptx
Unit 11 AI , ML , DL and Expert Systems.pptxCallplanetsDeveloper
 
Unit 10 Business Intelligence.pptx
Unit 10 Business Intelligence.pptxUnit 10 Business Intelligence.pptx
Unit 10 Business Intelligence.pptxCallplanetsDeveloper
 
Class 12 Probability Distributions.pptx
Class 12 Probability Distributions.pptxClass 12 Probability Distributions.pptx
Class 12 Probability Distributions.pptxCallplanetsDeveloper
 
Class 9 Covariance & Correlation Concepts.pptx
Class 9 Covariance & Correlation Concepts.pptxClass 9 Covariance & Correlation Concepts.pptx
Class 9 Covariance & Correlation Concepts.pptxCallplanetsDeveloper
 
Unit 7 Management Challenges in IT.pptx
Unit 7 Management Challenges in IT.pptxUnit 7 Management Challenges in IT.pptx
Unit 7 Management Challenges in IT.pptxCallplanetsDeveloper
 

Mehr von CallplanetsDeveloper (20)

Basics of Machine Learning.pptx
Basics of Machine Learning.pptxBasics of Machine Learning.pptx
Basics of Machine Learning.pptx
 
Artificial-Neural-Networks.pptx
Artificial-Neural-Networks.pptxArtificial-Neural-Networks.pptx
Artificial-Neural-Networks.pptx
 
Data Warehousing , Data Mining and BI.pptx
Data Warehousing , Data Mining and BI.pptxData Warehousing , Data Mining and BI.pptx
Data Warehousing , Data Mining and BI.pptx
 
Database Management Systems.pptx
Database Management Systems.pptxDatabase Management Systems.pptx
Database Management Systems.pptx
 
Regression.pptx
Regression.pptxRegression.pptx
Regression.pptx
 
Basics of Machine Learning.pptx
Basics of Machine Learning.pptxBasics of Machine Learning.pptx
Basics of Machine Learning.pptx
 
MIS.pptx
MIS.pptxMIS.pptx
MIS.pptx
 
Topic 13 Business Analytics.pptx
Topic 13 Business Analytics.pptxTopic 13 Business Analytics.pptx
Topic 13 Business Analytics.pptx
 
Topic 12 Miscellaneous Concepts in IT.pptx
Topic 12 Miscellaneous Concepts in IT.pptxTopic 12 Miscellaneous Concepts in IT.pptx
Topic 12 Miscellaneous Concepts in IT.pptx
 
Unit 11 AI , ML , DL and Expert Systems.pptx
Unit 11 AI , ML , DL and Expert Systems.pptxUnit 11 AI , ML , DL and Expert Systems.pptx
Unit 11 AI , ML , DL and Expert Systems.pptx
 
Unit 10 Business Intelligence.pptx
Unit 10 Business Intelligence.pptxUnit 10 Business Intelligence.pptx
Unit 10 Business Intelligence.pptx
 
9. Data Warehousing & Mining.pptx
9. Data Warehousing & Mining.pptx9. Data Warehousing & Mining.pptx
9. Data Warehousing & Mining.pptx
 
Class 12 Probability Distributions.pptx
Class 12 Probability Distributions.pptxClass 12 Probability Distributions.pptx
Class 12 Probability Distributions.pptx
 
Class 11 Basic Probability.pptx
Class 11 Basic Probability.pptxClass 11 Basic Probability.pptx
Class 11 Basic Probability.pptx
 
8.DBMS.pptx
8.DBMS.pptx8.DBMS.pptx
8.DBMS.pptx
 
Claas 11 Basic Probability.pptx
Claas 11 Basic Probability.pptxClaas 11 Basic Probability.pptx
Claas 11 Basic Probability.pptx
 
5.Developing IT Solution.pptx
5.Developing IT Solution.pptx5.Developing IT Solution.pptx
5.Developing IT Solution.pptx
 
4. E Commerce Types.pptx
4. E Commerce Types.pptx4. E Commerce Types.pptx
4. E Commerce Types.pptx
 
Class 9 Covariance & Correlation Concepts.pptx
Class 9 Covariance & Correlation Concepts.pptxClass 9 Covariance & Correlation Concepts.pptx
Class 9 Covariance & Correlation Concepts.pptx
 
Unit 7 Management Challenges in IT.pptx
Unit 7 Management Challenges in IT.pptxUnit 7 Management Challenges in IT.pptx
Unit 7 Management Challenges in IT.pptx
 

Kürzlich hochgeladen

Sociology 101 Demonstration of Learning Exhibit
Sociology 101 Demonstration of Learning ExhibitSociology 101 Demonstration of Learning Exhibit
Sociology 101 Demonstration of Learning Exhibitjbellavia9
 
Unit 3 Emotional Intelligence and Spiritual Intelligence.pdf
Unit 3 Emotional Intelligence and Spiritual Intelligence.pdfUnit 3 Emotional Intelligence and Spiritual Intelligence.pdf
Unit 3 Emotional Intelligence and Spiritual Intelligence.pdfDr Vijay Vishwakarma
 
TỔNG ÔN TẬP THI VÀO LỚP 10 MÔN TIẾNG ANH NĂM HỌC 2023 - 2024 CÓ ĐÁP ÁN (NGỮ Â...
TỔNG ÔN TẬP THI VÀO LỚP 10 MÔN TIẾNG ANH NĂM HỌC 2023 - 2024 CÓ ĐÁP ÁN (NGỮ Â...TỔNG ÔN TẬP THI VÀO LỚP 10 MÔN TIẾNG ANH NĂM HỌC 2023 - 2024 CÓ ĐÁP ÁN (NGỮ Â...
TỔNG ÔN TẬP THI VÀO LỚP 10 MÔN TIẾNG ANH NĂM HỌC 2023 - 2024 CÓ ĐÁP ÁN (NGỮ Â...Nguyen Thanh Tu Collection
 
Sensory_Experience_and_Emotional_Resonance_in_Gabriel_Okaras_The_Piano_and_Th...
Sensory_Experience_and_Emotional_Resonance_in_Gabriel_Okaras_The_Piano_and_Th...Sensory_Experience_and_Emotional_Resonance_in_Gabriel_Okaras_The_Piano_and_Th...
Sensory_Experience_and_Emotional_Resonance_in_Gabriel_Okaras_The_Piano_and_Th...Pooja Bhuva
 
Making communications land - Are they received and understood as intended? we...
Making communications land - Are they received and understood as intended? we...Making communications land - Are they received and understood as intended? we...
Making communications land - Are they received and understood as intended? we...Association for Project Management
 
UGC NET Paper 1 Mathematical Reasoning & Aptitude.pdf
UGC NET Paper 1 Mathematical Reasoning & Aptitude.pdfUGC NET Paper 1 Mathematical Reasoning & Aptitude.pdf
UGC NET Paper 1 Mathematical Reasoning & Aptitude.pdfNirmal Dwivedi
 
ICT role in 21st century education and it's challenges.
ICT role in 21st century education and it's challenges.ICT role in 21st century education and it's challenges.
ICT role in 21st century education and it's challenges.MaryamAhmad92
 
ICT Role in 21st Century Education & its Challenges.pptx
ICT Role in 21st Century Education & its Challenges.pptxICT Role in 21st Century Education & its Challenges.pptx
ICT Role in 21st Century Education & its Challenges.pptxAreebaZafar22
 
Basic Civil Engineering first year Notes- Chapter 4 Building.pptx
Basic Civil Engineering first year Notes- Chapter 4 Building.pptxBasic Civil Engineering first year Notes- Chapter 4 Building.pptx
Basic Civil Engineering first year Notes- Chapter 4 Building.pptxDenish Jangid
 
HMCS Max Bernays Pre-Deployment Brief (May 2024).pptx
HMCS Max Bernays Pre-Deployment Brief (May 2024).pptxHMCS Max Bernays Pre-Deployment Brief (May 2024).pptx
HMCS Max Bernays Pre-Deployment Brief (May 2024).pptxEsquimalt MFRC
 
The basics of sentences session 3pptx.pptx
The basics of sentences session 3pptx.pptxThe basics of sentences session 3pptx.pptx
The basics of sentences session 3pptx.pptxheathfieldcps1
 
How to Create and Manage Wizard in Odoo 17
How to Create and Manage Wizard in Odoo 17How to Create and Manage Wizard in Odoo 17
How to Create and Manage Wizard in Odoo 17Celine George
 
Salient Features of India constitution especially power and functions
Salient Features of India constitution especially power and functionsSalient Features of India constitution especially power and functions
Salient Features of India constitution especially power and functionsKarakKing
 
How to Manage Global Discount in Odoo 17 POS
How to Manage Global Discount in Odoo 17 POSHow to Manage Global Discount in Odoo 17 POS
How to Manage Global Discount in Odoo 17 POSCeline George
 
Python Notes for mca i year students osmania university.docx
Python Notes for mca i year students osmania university.docxPython Notes for mca i year students osmania university.docx
Python Notes for mca i year students osmania university.docxRamakrishna Reddy Bijjam
 
Micro-Scholarship, What it is, How can it help me.pdf
Micro-Scholarship, What it is, How can it help me.pdfMicro-Scholarship, What it is, How can it help me.pdf
Micro-Scholarship, What it is, How can it help me.pdfPoh-Sun Goh
 
Towards a code of practice for AI in AT.pptx
Towards a code of practice for AI in AT.pptxTowards a code of practice for AI in AT.pptx
Towards a code of practice for AI in AT.pptxJisc
 
Kodo Millet PPT made by Ghanshyam bairwa college of Agriculture kumher bhara...
Kodo Millet  PPT made by Ghanshyam bairwa college of Agriculture kumher bhara...Kodo Millet  PPT made by Ghanshyam bairwa college of Agriculture kumher bhara...
Kodo Millet PPT made by Ghanshyam bairwa college of Agriculture kumher bhara...pradhanghanshyam7136
 
HMCS Vancouver Pre-Deployment Brief - May 2024 (Web Version).pptx
HMCS Vancouver Pre-Deployment Brief - May 2024 (Web Version).pptxHMCS Vancouver Pre-Deployment Brief - May 2024 (Web Version).pptx
HMCS Vancouver Pre-Deployment Brief - May 2024 (Web Version).pptxmarlenawright1
 
On National Teacher Day, meet the 2024-25 Kenan Fellows
On National Teacher Day, meet the 2024-25 Kenan FellowsOn National Teacher Day, meet the 2024-25 Kenan Fellows
On National Teacher Day, meet the 2024-25 Kenan FellowsMebane Rash
 

Kürzlich hochgeladen (20)

Sociology 101 Demonstration of Learning Exhibit
Sociology 101 Demonstration of Learning ExhibitSociology 101 Demonstration of Learning Exhibit
Sociology 101 Demonstration of Learning Exhibit
 
Unit 3 Emotional Intelligence and Spiritual Intelligence.pdf
Unit 3 Emotional Intelligence and Spiritual Intelligence.pdfUnit 3 Emotional Intelligence and Spiritual Intelligence.pdf
Unit 3 Emotional Intelligence and Spiritual Intelligence.pdf
 
TỔNG ÔN TẬP THI VÀO LỚP 10 MÔN TIẾNG ANH NĂM HỌC 2023 - 2024 CÓ ĐÁP ÁN (NGỮ Â...
TỔNG ÔN TẬP THI VÀO LỚP 10 MÔN TIẾNG ANH NĂM HỌC 2023 - 2024 CÓ ĐÁP ÁN (NGỮ Â...TỔNG ÔN TẬP THI VÀO LỚP 10 MÔN TIẾNG ANH NĂM HỌC 2023 - 2024 CÓ ĐÁP ÁN (NGỮ Â...
TỔNG ÔN TẬP THI VÀO LỚP 10 MÔN TIẾNG ANH NĂM HỌC 2023 - 2024 CÓ ĐÁP ÁN (NGỮ Â...
 
Sensory_Experience_and_Emotional_Resonance_in_Gabriel_Okaras_The_Piano_and_Th...
Sensory_Experience_and_Emotional_Resonance_in_Gabriel_Okaras_The_Piano_and_Th...Sensory_Experience_and_Emotional_Resonance_in_Gabriel_Okaras_The_Piano_and_Th...
Sensory_Experience_and_Emotional_Resonance_in_Gabriel_Okaras_The_Piano_and_Th...
 
Making communications land - Are they received and understood as intended? we...
Making communications land - Are they received and understood as intended? we...Making communications land - Are they received and understood as intended? we...
Making communications land - Are they received and understood as intended? we...
 
UGC NET Paper 1 Mathematical Reasoning & Aptitude.pdf
UGC NET Paper 1 Mathematical Reasoning & Aptitude.pdfUGC NET Paper 1 Mathematical Reasoning & Aptitude.pdf
UGC NET Paper 1 Mathematical Reasoning & Aptitude.pdf
 
ICT role in 21st century education and it's challenges.
ICT role in 21st century education and it's challenges.ICT role in 21st century education and it's challenges.
ICT role in 21st century education and it's challenges.
 
ICT Role in 21st Century Education & its Challenges.pptx
ICT Role in 21st Century Education & its Challenges.pptxICT Role in 21st Century Education & its Challenges.pptx
ICT Role in 21st Century Education & its Challenges.pptx
 
Basic Civil Engineering first year Notes- Chapter 4 Building.pptx
Basic Civil Engineering first year Notes- Chapter 4 Building.pptxBasic Civil Engineering first year Notes- Chapter 4 Building.pptx
Basic Civil Engineering first year Notes- Chapter 4 Building.pptx
 
HMCS Max Bernays Pre-Deployment Brief (May 2024).pptx
HMCS Max Bernays Pre-Deployment Brief (May 2024).pptxHMCS Max Bernays Pre-Deployment Brief (May 2024).pptx
HMCS Max Bernays Pre-Deployment Brief (May 2024).pptx
 
The basics of sentences session 3pptx.pptx
The basics of sentences session 3pptx.pptxThe basics of sentences session 3pptx.pptx
The basics of sentences session 3pptx.pptx
 
How to Create and Manage Wizard in Odoo 17
How to Create and Manage Wizard in Odoo 17How to Create and Manage Wizard in Odoo 17
How to Create and Manage Wizard in Odoo 17
 
Salient Features of India constitution especially power and functions
Salient Features of India constitution especially power and functionsSalient Features of India constitution especially power and functions
Salient Features of India constitution especially power and functions
 
How to Manage Global Discount in Odoo 17 POS
How to Manage Global Discount in Odoo 17 POSHow to Manage Global Discount in Odoo 17 POS
How to Manage Global Discount in Odoo 17 POS
 
Python Notes for mca i year students osmania university.docx
Python Notes for mca i year students osmania university.docxPython Notes for mca i year students osmania university.docx
Python Notes for mca i year students osmania university.docx
 
Micro-Scholarship, What it is, How can it help me.pdf
Micro-Scholarship, What it is, How can it help me.pdfMicro-Scholarship, What it is, How can it help me.pdf
Micro-Scholarship, What it is, How can it help me.pdf
 
Towards a code of practice for AI in AT.pptx
Towards a code of practice for AI in AT.pptxTowards a code of practice for AI in AT.pptx
Towards a code of practice for AI in AT.pptx
 
Kodo Millet PPT made by Ghanshyam bairwa college of Agriculture kumher bhara...
Kodo Millet  PPT made by Ghanshyam bairwa college of Agriculture kumher bhara...Kodo Millet  PPT made by Ghanshyam bairwa college of Agriculture kumher bhara...
Kodo Millet PPT made by Ghanshyam bairwa college of Agriculture kumher bhara...
 
HMCS Vancouver Pre-Deployment Brief - May 2024 (Web Version).pptx
HMCS Vancouver Pre-Deployment Brief - May 2024 (Web Version).pptxHMCS Vancouver Pre-Deployment Brief - May 2024 (Web Version).pptx
HMCS Vancouver Pre-Deployment Brief - May 2024 (Web Version).pptx
 
On National Teacher Day, meet the 2024-25 Kenan Fellows
On National Teacher Day, meet the 2024-25 Kenan FellowsOn National Teacher Day, meet the 2024-25 Kenan Fellows
On National Teacher Day, meet the 2024-25 Kenan Fellows
 

Deep Learning Basics.pptx

  • 2. • Artificial Neural Networks (ANN): Deep learning is a subset of machine learning that uses artificial neural networks to learn from data. • Layers: ANNs are composed of layers of interconnected nodes or neurons. The input layer receives the data, the output layer produces the predictions, and the hidden layers perform the computations in between.
  • 3. • Activation functions: Activation functions are applied to the outputs of each layer to introduce non-linearity and increase the model's expressiveness. Common activation functions include ReLU, sigmoid, and tanh. • Backpropagation: This is a method for training neural networks by iteratively adjusting the weights in each layer to minimize the difference between the predicted outputs and the actual outputs. It works by propagating the error backwards from the output layer to the input layer and adjusting the weights accordingly
  • 4. • Loss function: The loss function measures the difference between the predicted outputs and the actual outputs. The goal of training is to minimize the loss function by adjusting the weights in each layer. • Optimization algorithms: Optimization algorithms, such as stochastic gradient descent (SGD) and Adam, are used to adjust the weights in each layer during training to minimize the loss function. • Overfitting: Overfitting occurs when a model is too complex and starts to memorize the training data instead of learning the underlying patterns. This can be prevented by using techniques
  • 5. Activation Functions in Artificial Neural Networks (ANNs): • Activation functions are mathematical functions used in ANNs to introduce non-linearity into the output of a neuron or a layer. • They are typically applied to the weighted sum of inputs and biases before being passed through to the next layer of the network. • Without activation functions, ANNs would simply be a linear regression model, which can only model linear relationships between input and output.
  • 6. • Common activation functions include Sigmoid, Tanh, ReLU, and Softmax. • Sigmoid and Tanh functions are sigmoidal, meaning they produce an S-shaped curve. They are used to squash the output of a neuron to a range between 0 and 1 or -1 and 1, respectively. • ReLU (Rectified Linear Unit) function is non-sigmoidal and is defined as f(x) = max(0, x). It is one of the most commonly used activation functions due to its simplicity and effectiveness. • Softmax function is used in the output layer of a network to produce a probability distribution over multiple classes. • Choosing the right activation function can have a significant impact on the performance of a neural network, and it is often an area of active research.
  • 7. Non Linearity • In the context of machine learning, nonlinearity is an important property of neural networks. Neural networks are composed of many interconnected processing units (neurons), which apply a nonlinear activation function to their inputs before passing them to the next layer of the network. This nonlinearity allows neural networks to model complex relationships between inputs and outputs, and to learn representations that are not directly observable in the input data.
  • 9.
  • 10.
  • 11.
  • 12.
  • 13. • For example, a linear model can only learn a linear decision boundary between two classes, which may not be sufficient to accurately classify complex data. In contrast, a nonlinear model such as a neural network can learn more complex decision boundaries that can better separate the classes. • Some common nonlinear activation functions used in neural networks include the Rectified Linear Unit (ReLU), sigmoid, tanh, and others. These functions introduce nonlinearity into the network, allowing it to learn more complex representations of the input data.
  • 14. Rectified Linear Unit (ReLU) activation function: • ReLU is a non-linear activation function commonly used in neural networks. • It takes an input value x and returns the maximum of 0 and x as the output value. • The formula for ReLU is f(x) = max(0, x). • ReLU is computationally efficient, since it requires only simple thresholding of the input value, compared to other activation functions like sigmoid or tanh.
  • 15. Sigmoid • Sigmoid is a non-linear activation function commonly used in neural networks. • It takes an input value x and maps it to a range between 0 and 1 using the formula f(x) = 1 / (1 + e^-x). • The output of the sigmoid function can be interpreted as a probability or likelihood, since it always produces a value between 0 and 1. • Sigmoid was one of the earliest activation functions used in neural networks, and it is still used in some applications, such as logistic regression. • Sigmoid is smooth and differentiable, which makes it useful for backpropagation and gradient descent optimization. • One limitation of sigmoid is that it suffers from the "vanishing gradient" problem, where the gradient becomes very small as the input value becomes very large or very small, making it difficult for the network to learn. • Another limitation of sigmoid is that it is not zero-centered, which can slow down the convergence of gradient descent. • Due to these limitations, sigmoid is not as commonly used as other activation functions like ReLU or its variants. •
  • 16.
  • 17.
  • 18. Tanh • Tanh is a non-linear activation function commonly used in neural networks. • It takes an input value x and maps it to a range between -1 and 1 using the formula f(x) = (e^x - e^- x) / (e^x + e^-x). • Tanh is a shifted and rescaled version of the sigmoid function, with the output value zero-centered. • Like sigmoid, tanh is smooth and differentiable, which makes it useful for backpropagation and gradient descent optimization. • Tanh is often used in the hidden layers of neural networks, especially in recurrent neural networks (RNNs) and long short-term memory (LSTM) networks. • One limitation of tanh is that it also suffers from the "vanishing gradient" problem, where the gradient becomes very small as the input value becomes very large or very small, making it difficult for the network to learn. • Another limitation of tanh is that it is more computationally expensive than ReLU or its variants, since it involves exponentials. • Despite its limitations, tanh can be useful in certain situations, such as when the input data is standardized and zero-centered, or when the network needs to model both positive and negative values.
  • 19. Difference between Activation Function and ML Algorithm • An activation function is a mathematical function used in artificial neural networks to introduce non-linearity into the output of a neuron. • Activation functions are used to decide whether the neuron should be activated or not based on the input it receives. • Common activation functions include sigmoid, ReLU, tanh, and softmax. • On the other hand, a machine learning algorithm is a method or set of methods used to learn patterns and relationships • in data in order to make predictions or decisions. Machine learning algorithms can be supervised, unsupervised, • or semi-supervised, and can be used for a wide range of tasks, such as regression, classification, clustering, and reinforcement learning. • While activation functions are used in neural networks to introduce non-linearity and make them more expressive, machine learning algorithms are used to learn patterns and relationships in data and make predictions based on that learning.
  • 20. Why Accuracy of ML is better than DL • Complexity: ML models are often simpler and more interpretable than ANNs, which can make them easier to train and optimize. In some cases, a simpler model may be sufficient to achieve good performance on a given task, without the need for a complex neural network. • Data size: ANNs require large amounts of data to train effectively, and may not perform well on small datasets. In contrast, some ML models, such as decision trees or logistic regression, can perform well even on smaller datasets. • Feature engineering: ANNs often require extensive feature engineering and preprocessing of input data, which can be time-consuming and require domain expertise. In contrast, some ML models, such as decision trees or Naive Bayes, can perform well with minimal feature engineering.
  • 21. • Model selection: Choosing the right ANN architecture and hyperparameters can be a challenging task, and may require extensive experimentation and tuning. In contrast, some ML models, such as decision trees or Naive Bayes, have fewer hyperparameters to tune and may be easier to select and optimize. • Overfitting: ANNs are prone to overfitting, where the model becomes too complex and performs well on the training data but poorly on new, unseen data. ML models may be less prone to overfitting, especially when regularized or constrained in some way. • Overall, the choice of model depends on the specific task and dataset at hand, and there is no one-size-fits-all solution. In some cases, an ML model may be more suitable, while in other cases, an ANN may be necessary to achieve the desired performance