SlideShare ist ein Scribd-Unternehmen logo
1 von 16
Downloaden Sie, um offline zu lesen
1
                     FEEDFORWARD
                 NEURAL NETWORKS:
                  AN INTRODUCTION
                                                   Simon Haykin




A neural network is a massively parallel distributed processor that
has a natural propensity for storing experiential knowledge and
making it available for use. It resembles the brain in two respects
(Haykin 1998):
  1. Knowledge is acquired by the network through a learning
     process.
  2. Interconnection strengths known as synaptic weights are used
     to store the knowledge.
Basically, learning is a process by which the free parameters (i.e.,
synaptic weights and bias levels) of a neural network are adapted
through a continuing process of stimulation by the environment in
which the network is embedded. The type of learning is determined
by the manner in which the parameter changes take place. In a general
sense, the learning process may be classified as follows:
                                                                   1
2     FEEDFORWARD NEURAL NETWORKS: AN INTRODUCTION


    • Learning with a teacher, also referred to as supervised learning
    • Learning without a teacher, also referred to as unsupervised
      learning


1.1 SUPERVISED LEARNING

This form of learning assumes the availability of a labeled (i.e.,
ground-truthed) set of training data made up of N input—output
examples:

                                                   N
                            T = {( x i , di )}i =1                 (1.1)

where xi = input vector of ith example
      di = desired (target) response of ith example, assumed to be
           scalar for convenience of presentation
      N = sample size

Given the training sample T, the requirement is to compute the free
parameters of the neural network so that the actual output yi of the
neural network due to xi is close enough to di for all i in a statistical
sense. For example, we may use the mean-square error
                                        N
                                    1                       2
                         E ( n) =
                                    N
                                        Â (d   i   - yi )          (1.2)
                                        i =1


as the index of performance to be minimized.

1.1.1 Multilayer Perceptrons and
Back-Propagation Learning
The back-propagation algorithm has emerged as the workhorse for
the design of a special class of layered feedforward networks known
as multilayer perceptrons (MLP). As shown in Fig. 1.1, a multilayer
perceptron has an input layer of source nodes and an output layer
of neurons (i.e., computation nodes); these two layers connect the
network to the outside world. In addition to these two layers,
the multilayer perceptron usually has one or more layers of hidden
neurons, which are so called because these neurons are not directly
accessible. The hidden neurons extract important features contained
in the input data.
SUPERVISED LEARNING      3




Figure 1.1 Fully connected feedforward with one hidden layer and one
output layer.

  The training of an MLP is usually accomplished by using a back-
propagation (BP) algorithm that involves two phases (Werbos 1974;
Rumelhart et al. 1986):
  • Forward Phase. During this phase the free parameters of the
    network are fixed, and the input signal is propagated through
    the network of Fig. 1.1 layer by layer. The forward phase fin-
    ishes with the computation of an error signal
                             ei = di - yi                           (1.3)
    where di is the desired response and yi is the actual output pro-
    duced by the network in response to the input xi.
  • Backward Phase. During this second phase, the error signal ei is
    propagated through the network of Fig. 1.1 in the backward
    direction, hence the name of the algorithm. It is during this phase
    that adjustments are applied to the free parameters of the
    network so as to minimize the error ei in a statistical sense.
  Back-propagation learning may be implemented in one of two
basic ways, as summarized here:
4     FEEDFORWARD NEURAL NETWORKS: AN INTRODUCTION


    1. Sequential mode (also referred to as the on-line mode or sto-
       chastic mode): In this mode of BP learning, adjustments are
       made to the free parameters of the network on an example-by-
       example basis. The sequential mode is best suited for pattern
       classification.
    2. Batch mode: In this second mode of BP learning, adjustments
       are made to the free parameters of the network on an epoch-
       by-epoch basis, where each epoch consists of the entire set of
       training examples. The batch mode is best suited for nonlinear
       regression.

The back-propagation learning algorithm is simple to implement and
computationally efficient in that its complexity is linear in the synap-
tic weights of the network. However, a major limitation of the algo-
rithm is that it does not always converge and can be excruciatingly
slow, particularly when we have to deal with a difficult learning task
that requires the use of a large network.
   We may try to make back-propagation learning perform better by
invoking the following list of heuristics:

    • Use neurons with antisymmetric activation functions (e.g.,
      hyperbolic tangent function) in preference to nonsymmetric
      activation functions (e.g., logistic function). Figure 1.2 shows
      examples of these two forms of activation functions.
    • Shuffle the training examples after the presentation of each
      epoch; an epoch involves the presentation of the entire set of
      training examples to the network.
    • Follow an easy-to-learn example with a difficult one.
    • Preprocess the input data so as to remove the mean and decor-
      relate the data.
    • Arrange for the neurons in the different layers to learn at essen-
      tially the same rate. This may be attained by assigning a learn-
      ing rate parameter to neurons in the last layers that is smaller
      than those at the front end.
    • Incorporate prior information into the network design when-
      ever it is available.

   One other heuristic that deserves to be mentioned relates to the
size of the training set, N, for a pattern classification task. Given a mul-
tilayer perceptron with a total number of synaptic weights including
bias levels, denoted by W, a rule of thumb for selecting N is
SUPERVISED LEARNING   5




Figure 1.2 (a) Antisymmetric activation function. (b) Nonsymmetric
activation function.
6     FEEDFORWARD NEURAL NETWORKS: AN INTRODUCTION


                                    W
                              N = OÊ ˆ                           (1.4)
                                   Ë e¯


where O denotes “the order of,” and e denotes the fraction of clas-
sification errors permitted on test data. For example, with an error
of 10% the number of training examples needed should be about 10
times the number of synaptic weights in the network.
   Supposing that we have chosen a multilayer perceptron to be
trained with the back-propagation algorithm, how do we determine
when it is “best” to stop the training session? How do we select the
size of individual hidden layers of the MLP? The answers to these
important questions may be gotten though the use of a statistical
technique known as cross-validation, which proceeds as follows
(Haykin 1999):

    • The set of training examples is split into two parts:
      • Estimation subset used for training of the model
      • Validation subset used for evaluating the model performance
    • The network is finally tuned by using the entire set of training
      examples and then tested on test data not seen before.


1.1.2 Radial-Basis Function Networks
Another popular layered feedforward network is the radial-basis
function (RBF) network which has important universal approxima-
tion properties (Park and Sandberg 1993), and whose structure is
shown in Fig. 13. RBF networks use memory-based learning for their
design. Specifically, learning is viewed as a curve-fitting problem in
high-dimensional space (Broomhead and Lowe 1989; Poggio and
Girosi 1990):

    1. Learning is equivalent to finding a surface in a multidimen-
       sional space that provides a best fit to the training data.
    2. Generalization (i.e., response of the network to input data not
       seen before) is equivalent to the use of this multidimensional
       surface to interpolate the test data.

  RBF networks differ from multilayer perceptrons in some funda-
mental respects:
SUPERVISED LEARNING   7




              Figure 1.3 Radial-basis function network.



  • RBF networks are local approximators, whereas multilayer per-
    ceptrons are global approximators.
  • RBF networks have a single hidden layer, whereas multilayer
    perceptrons can have any number of hidden layers.
  • The output layer of a RBF network is always linear, whereas in
    a multilayer perceptron it can be linear or nonlinear.
  • The activation function of the hidden layer in an RBF network
    computes the Euclidean distance between the input signal
    vector and parameter vector of the network, whereas the acti-
    vation function of a multilayer perceptron computes the inner
    product between the input signal vector and the pertinent
    synaptic weight vector.

  The use of a linear output layer in an RBF network may be justi-
fied in light of Cover’s theorem on the separability of patterns.
According to this theorem, provided that the transformation from
the input space to the feature (hidden) space is nonlinear and the
8     FEEDFORWARD NEURAL NETWORKS: AN INTRODUCTION


dimensionality of the feature space is high compared to that of the
input (data) space, then there is a high likelihood that a nonsepara-
ble pattern classification task in the input space is transformed into
a linearly separable one in the feature space.Another analytical basis
for the use of RBF networks (and multilayer perceptrons) in classi-
fication problems is provided by the results in Chapter 2, where (as
a special case) it is shown that a large family of classification prob-
lems in n can be solved using nonlinear static networks.
   Design methods for RBF networks include the following:

    1. Random selection of fixed centers (Broomhead and Lowe
       1998)
    2. Self-organized selection of centers (Moody and Darken 1989)
    3. Supervised selection of centers (Poggio and Girosi 1990)
    4. Regularized interpolation exploiting the connection between
       an RBF network and the Watson–Nadaraya regression kernel
       (Yee 1998).


1.2 UNSUPERVISED LEARNING

Turning next to unsupervised learning, adjustment of synaptic
weights may be carried through the use of neurobiological principles
such as Hebbian learning and competitive learning. In this section
we will describe specific applications of these two approaches.

1.2.1    Principal Components Analysis
According to Hebb’s postulate of learning, the change in synaptic
weight Dwji of a neural network is defined by

                             Dwji = hxiyj                        (1.5)

where h = learning-rate parameter
      xi = input (presynaptic) signal
      yi = output (postsynaptic) signal

Principal component analysis (PCA) networks use a modified form
of this self-organized learning rule. To begin with, consider a linear
neuron designed to operate as a maximum eigenfilter; such a
neuron is referred to as Oja’s neuron (Oja 1982). It is characterized
as follow:
UNSUPERVISED LEARNING      9

                         Dw ji = hy j ( xi - y j w ji )                     (1.6)

where the term -hy2wji is added to stabilize the learning process. As
                   j
the number of iterations approaches infinity, we find the following:

  1. The synaptic weight vector of neuron j approaches the eigen-
     vector associated with the largest eigenvalue l max of the corre-
     lation matrix of the input vector (assumed to be of zero mean).
  2. The variance of the output of neuron j approaches the largest
     eigenvalue l max.

   The generalized Hebbian algorithm (GHA), due to Sanger (1989),
is a straightforward generalization of Oja’s neuron for the extraction
of any desired number of principal components.


1.2.2   Self-Organizing Maps
In a self-organizing map (SOM), due to Kohonen (1997), the neurons
are placed at the nodes of a lattice, and they become selectively
tuned to various input patterns (vectors) in the course of a compet-
itive learning process. The process is characterized by the formation
of a topographic map in which the spatial locations (i.e., coordinates)
of the neurons in the lattice correspond to intrinsic features of the
input patterns. Figure 1.4 illustrates the basic idea of a self-organiz-
ing map, assuming the use of a two-dimensional lattice of neurons as
the network structure.
   In reality, the SOM belongs to the class of vector-coding algo-
rithms (Luttrell, 1989). That is, a fixed number of codewords are
placed into a higher-dimensional input space, thereby facilitating
data compression.
   An integral feature of the SOM algorithm is the neighborhood
function centered around a neuron that wins the competitive
process. The neighborhood function starts by enclosing the entire
lattice initially and is then allowed to shrink gradually until it encom-
passes the winning neuron.
   The algorithm exhibits two distinct phases in its operation:

  1. Ordering phase, during which the topological ordering of the
     weight vectors takes place
  2. Convergence phase, during which the computational map is fine
     tuned
10      FEEDFORWARD NEURAL NETWORKS: AN INTRODUCTION




Figure 1.4 Illustration of relationship between feature map f and weight
vector wi of winning neuron i.


The SOM algorithm exhibits the following properties:
     1. Approximation of the continuous input space by the weight
        vectors of the discrete lattice.
     2. Topological ordering exemplified by the fact that the spatial
        location of a neuron in the lattice corresponds to a particular
        feature of the input pattern.
     3. The feature map computed by the algorithm reflects variations
        in the statistics of the input distribution.
     4. SOM may be viewed as a nonlinear form of principal compo-
        nents analysis.
Refinements to the SOM algorithm are discussed in Van Hulle (2000).

1.3 TEMPORAL PROCESSING USING
FEEDFORWARD NETWORKS

The material just described is concerned basically with the approxi-
mation of systems without dynamics (i.e., with static systems). At
TEMPORAL PROCESSING USING FEEDFORWARD NETWORKS                 11

about the same time as the appearance of the early universal-
approximation theorems for static neural networks there began
(Sandberg 1991a) a corresponding study (see Chapter 2) of the neural
network approximation of approximately-finite-memory maps and
myopic maps. It was found that large classes of these maps can be uni-
formly approximated arbitrarily well by the maps of certain simple
nonlinear structures using, for example, sigmoidal nonlinearities or
radial basis functions. The approximating networks are two-stage
structures comprising a linear preprocessing stage followed by a
memoryless nonlinear network. Much is now known about the prop-
erties of these networks, and examples of these properties are given in
the following.
   From another perspective, time is an essential dimension of
learning. We may incorporate time into the design of a neural
network implicitly or explicitly. A straightforward method of implicit
representation of time1 is to add a short-term memory structure in
the input layer of a static neural network (e.g., multilayer percep-
tron). The resulting configuration is sometimes called a focused time-
lagged feedforward network (TLFN).
   The short-term memory structure may be implemented in one of
two forms, as described here:
    1. Tapped-Delay-Line (TDL) Memory. This is the most com-
       monly used form of short-term memory. It consists of p unit
       delays with ( p + 1) terminals, as shown in Fig. 1.5, which may
       be viewed as a single input–multiple output network. Figure 1.6
       shows a focused TLFN network using the combination of a
       TDL memory and multilayer perceptron. In Figs. 1.5 and 1.6,
       the unit-delay is denoted by z-1.
          The memory depth of a TDL memory is fixed at p, and its
       memory resolution is fixed at unity, giving a depth resolution
       constant of p.
    2. Gamma Memory. We may exercise control over the memory
       depth by building a feedback loop around each unit delay, as
       illustrated in Fig. 1.7 (deVries and Principe 1992). In effect, the
       unit delay z-1 of the standard TDL memory is replaced by the
       transfer function
1
  Another practical way of accounting for time in a neural network is to employ
feedback at the local or global level. Neural networks so configured are referred to
as recurrent networks. For detailed treatment of recurrent networks, see Haykin
(1999).
12     FEEDFORWARD NEURAL NETWORKS: AN INTRODUCTION




       Figure 1.5 Ordinary tapped-delay line memory of order p.




Figure 1.6 Focused time-lagged feedforward network (TLFN); the bias
levels have been omitted for convenience of presentation.




     Figure 1.7 Signal-flow graph for one section of gamma memory.
TEMPORAL PROCESSING USING FEEDFORWARD NETWORKS         13

                                           mz-1
                           G (z) =
                                      1 - (1 - m )z-1
                                           m
                                  =
                                      z - (1 - m )

     where m is an adjustable parameter. For stability, the only pole
     of G(z) at z = (1 - m) must lie inside the unit circle in the z
     plane. This, in turn, requires that we restrict the choice of m to
     the following range of values:

                                  0<m<2

     The overall impulse response of the gamma memory, consist-
     ing of p sections, is the inverse z transform of the overall trans-
     fer function
                                                        p
                                          m
                          Gp (z) = Ê              ˆ
                                   Ë z - (1 - m ) ¯

     Denoting the impulse response by gp(n), we have

                             n - 1ˆ p
                 g p (z) = Ê
                                               n- p
                                    m (1 - m )              n≥ p
                           Ë p - 1¯

     where (:) is a binomial coefficient.The overall impulse response
     gp(n) for varying p represents a discrete version of the inte-
     grand of the gamma function (deVries and Principe 1992);
     hence the name “gamma memory.”
        The depth of the gamma memory is p/m and its resolution is
     m, for a depth resolution product of p. Accordingly, by choos-
     ing m to be less than unity, the gamma memory provides
     improvement in depth over the TDL memory, but at the
     expense of memory resolution.

   With regard to the utility of gamma networks—which are partic-
ular cases of the family of two-stage structures comprising a linear
preprocessing stage followed by a memoryless nonlinear network—
experimental results have been reported which indicate that the
structure is useful. In fact, it is known (Sandberg and Xu 1997) that
14    FEEDFORWARD NEURAL NETWORKS: AN INTRODUCTION


for a large class of discrete-time dynamic system maps H, and for
any choice of m in the interval (0, 1), there is a focused gamma
network that approximates H uniformly arbitrarily well. It is known
that tapped-delay-line networks (i.e., networks with m = 1) also have
the universal approximation property (Sandberg 1991b).
   Focused TLFNs using ordinary tapped-delay memory or gamma
memory are limited to stationary environments. To deal with
nonstationary dynamical processes, we may use distributed TLFNs
where the effect of time is distributed at the synaptic level through-
out the network. One way in which this may be accomplished is to
use finite-duration impulse response (FIR) filters to implement the
synaptic connections of an MLP; Fig. 1.8 shows an FIR model of a
synapse. The training of a distributed TLFN is naturally a more dif-
ficult proposition than the training of a focused TLFN. Whereas we
may use the ordinary back-propagation algorithm to train a focused
TLFN, we have to extend the back-propagation algorithm to cope
with the replacement of a synaptic weight in the ordinary MLP by a
synaptic weight vector. This extension is referred to as the temporal
back-propagation algorithm due to Wan (1994).


1.4   CONCLUDING REMARKS

In this chapter, we briefly reviewed the feedforward type of neural
networks, which are exemplified by multilayer perceptrons (MLPs),
radial-basis function (RBF) networks, principal component analysis
(PCA) networks, and self-organizing maps (SOMs). The training of
MLPs and RBF networks proceeds in a supervised manner, whereas
the training of PCA networks and SOMs proceeds in an unsuper-
vised manner.
   Feedforward networks by themselves are nonlinear static net-
works. They can be made to operate as nonlinear dynamical systems




         Figure 1.8 Finite-duration impulse response (FIR) filter.
BIBLIOGRAPHY      15

by incorporating short-term memory into their input layer. Two
important examples of short-term memory are the standard tapped-
delay-line and the gamma memory that provides control over attain-
able memory depth. The attractive feature of nonlinear dynamical
systems built in this way is that they are inherently stable.


BIBLIOGRAPHY

Anderson, J. A., 1995, Introduction to Neural Networks (Cambridge, MA:
  MIT Press).
Barlow, H. B., 1989, “Unsupervised learning,” Neural Computation, vol. 1,
  pp. 295–311.
Becker, S., and G. E. Hinton, 1982, “A self-organizing neural network that
  discovers surfaces in random-dot stereograms,” Nature (London), vol.
  355, pp. 161–163.
Broomhead, D. S., and D. Lowe, 1988, “Multivariable functional inter-
  polation and adaptive networks,” Complex Systems, vol. 2, pp. 321–
  355.
Comon, P., 1994,“Independent component analysis:A new concept?” Signal
  Processing, vol. 36, pp. 287–314.
deVries, B., and J. C. Principe, 1992, “The gamma model—A new
  neural model for temporal processing,” Neural Networks, vol. 4, pp.
  565–576.
Haykin, S., 1999, Neural Networks: A Comprehensive Foundation, 2nd ed.
  (Englewood Cliffs, NJ: Prentice-Hall).
Moody and Darken, 1989, “Fast learning in networks of locally-tuned pro-
  cessing unites,” Neural Computation, vol. 1, pp. 281–294.
Oja, E., 1982, “A simplified neuron model as a principal component ana-
  lyzer,” J. Math. Biol., vol. 15, pp. 267–273.
Park, J., and Sandberg, I. W., 1993, “Approximation and radial-basis func-
  tion networks,” Neural computation, vol. 5, pp. 305–316.
Poggio, T., and F. Girosi, 1990, “Networks for approximation and learning,”
  Proc. IEEE, vol. 78, pp. 1481–1497.
Rumelhart, D. E., G. E. Hinton, and R. J. Williams, 1986, “Learning internal
  representations by error propagation,” in D. E. Rumelhart and J. L.
  McCleland, eds. (Cambridge, MA: MIT Press), vol. 1, Chapter 8.
Sandberg, I. W., 1991a, “Structure theorems for nonlinear systems,” Multi-
  dimensional Sys. Sig. Process. vol. 2, pp. 267–286. (Errata in 1992, vol. 3,
  p. 101.)
Sandberg, I. W., 1991b, “Approximation theorems for discrete-time
  systems,” IEEE Trans. Circuits Sys. vol. 38, no. 5, pp. 564–566, May 1991.
16    FEEDFORWARD NEURAL NETWORKS: AN INTRODUCTION


Sandberg, I. W., and Xu, L., 1997, “Uniform approximation and gamma net-
  works,” Neural Networks, vol. 10, pp. 781–784.
Van Hulle, M. M., 2000, Faithful Representations and Topographic Maps:
  From Distortion-to-Information-Based Self Organization (New York:
  Wiley).
Wan, E. A., 1994, “Time series prediction by using a connectionist network
  with internal delay lines,” in A. S. Weigend and N. A. Gershenfield, eds.,
  Time Series Prediction: Forecasting the Future and Understanding the Past
  (Reading, MA: Addison-Wesley), pp. 195–217.
Werbos, P. J., 1974, “Beyond regression: New tools for prediction and analy-
  sis in the behavioral sciences,” Ph.D. Thesis, Harvard University,
  Cambridge, MA.
Werbos, P. J., 1990, “Backpropagation through time: What it does and how
  to do it,” Proc. IEEE, vol. 78, pp. 1550–1560.
Yee, P. V., 1998, “Regularized radial basis function networks: Theory and
  applications to probability estimation, classification, and time series pre-
  diction,” Ph.D. Thesis, McMaster University, Hamilton, Ontario.

Weitere ähnliche Inhalte

Was ist angesagt?

Neural network in matlab
Neural network in matlab Neural network in matlab
Neural network in matlab Fahim Khan
 
nural network ER. Abhishek k. upadhyay
nural network ER. Abhishek  k. upadhyaynural network ER. Abhishek  k. upadhyay
nural network ER. Abhishek k. upadhyayabhishek upadhyay
 
Lecture artificial neural networks and pattern recognition
Lecture   artificial neural networks and pattern recognitionLecture   artificial neural networks and pattern recognition
Lecture artificial neural networks and pattern recognitionHưng Đặng
 
Learning in Networks: were Pavlov and Hebb right?
Learning in Networks: were Pavlov and Hebb right?Learning in Networks: were Pavlov and Hebb right?
Learning in Networks: were Pavlov and Hebb right?Victor Miagkikh
 
Regularization in deep learning
Regularization in deep learningRegularization in deep learning
Regularization in deep learningKien Le
 
The Art Of Backpropagation
The Art Of BackpropagationThe Art Of Backpropagation
The Art Of BackpropagationJennifer Prendki
 
An introduction to machine learning for particle physics
An introduction to machine learning for particle physicsAn introduction to machine learning for particle physics
An introduction to machine learning for particle physicsAndrew Lowe
 
Fundamental, An Introduction to Neural Networks
Fundamental, An Introduction to Neural NetworksFundamental, An Introduction to Neural Networks
Fundamental, An Introduction to Neural NetworksNelson Piedra
 
Artificial Neural Network
Artificial Neural NetworkArtificial Neural Network
Artificial Neural NetworkDessy Amirudin
 
Machine Learning: Introduction to Neural Networks
Machine Learning: Introduction to Neural NetworksMachine Learning: Introduction to Neural Networks
Machine Learning: Introduction to Neural NetworksFrancesco Collova'
 
Web spam classification using supervised artificial neural network algorithms
Web spam classification using supervised artificial neural network algorithmsWeb spam classification using supervised artificial neural network algorithms
Web spam classification using supervised artificial neural network algorithmsaciijournal
 
lecture07.ppt
lecture07.pptlecture07.ppt
lecture07.pptbutest
 
A NEAT Way for Evolving Echo State Networks
A NEAT Way for Evolving Echo State NetworksA NEAT Way for Evolving Echo State Networks
A NEAT Way for Evolving Echo State NetworksKyriakos Chatzidimitriou
 
ARTIFICIAL NEURAL NETWORKS
ARTIFICIAL NEURAL NETWORKSARTIFICIAL NEURAL NETWORKS
ARTIFICIAL NEURAL NETWORKSAIMS Education
 
Echo state networks and locomotion patterns
Echo state networks and locomotion patternsEcho state networks and locomotion patterns
Echo state networks and locomotion patternsVito Strano
 
Neural networks1
Neural networks1Neural networks1
Neural networks1Mohan Raj
 
Introduction to Neural networks (under graduate course) Lecture 9 of 9
Introduction to Neural networks (under graduate course) Lecture 9 of 9Introduction to Neural networks (under graduate course) Lecture 9 of 9
Introduction to Neural networks (under graduate course) Lecture 9 of 9Randa Elanwar
 
2.5 backpropagation
2.5 backpropagation2.5 backpropagation
2.5 backpropagationKrish_ver2
 
Artificial Neural Networks Lect3: Neural Network Learning rules
Artificial Neural Networks Lect3: Neural Network Learning rulesArtificial Neural Networks Lect3: Neural Network Learning rules
Artificial Neural Networks Lect3: Neural Network Learning rulesMohammed Bennamoun
 

Was ist angesagt? (20)

Neural network in matlab
Neural network in matlab Neural network in matlab
Neural network in matlab
 
nural network ER. Abhishek k. upadhyay
nural network ER. Abhishek  k. upadhyaynural network ER. Abhishek  k. upadhyay
nural network ER. Abhishek k. upadhyay
 
Lecture artificial neural networks and pattern recognition
Lecture   artificial neural networks and pattern recognitionLecture   artificial neural networks and pattern recognition
Lecture artificial neural networks and pattern recognition
 
Learning in Networks: were Pavlov and Hebb right?
Learning in Networks: were Pavlov and Hebb right?Learning in Networks: were Pavlov and Hebb right?
Learning in Networks: were Pavlov and Hebb right?
 
Regularization in deep learning
Regularization in deep learningRegularization in deep learning
Regularization in deep learning
 
The Art Of Backpropagation
The Art Of BackpropagationThe Art Of Backpropagation
The Art Of Backpropagation
 
An introduction to machine learning for particle physics
An introduction to machine learning for particle physicsAn introduction to machine learning for particle physics
An introduction to machine learning for particle physics
 
Fundamental, An Introduction to Neural Networks
Fundamental, An Introduction to Neural NetworksFundamental, An Introduction to Neural Networks
Fundamental, An Introduction to Neural Networks
 
Artificial Neural Network
Artificial Neural NetworkArtificial Neural Network
Artificial Neural Network
 
Ann
Ann Ann
Ann
 
Machine Learning: Introduction to Neural Networks
Machine Learning: Introduction to Neural NetworksMachine Learning: Introduction to Neural Networks
Machine Learning: Introduction to Neural Networks
 
Web spam classification using supervised artificial neural network algorithms
Web spam classification using supervised artificial neural network algorithmsWeb spam classification using supervised artificial neural network algorithms
Web spam classification using supervised artificial neural network algorithms
 
lecture07.ppt
lecture07.pptlecture07.ppt
lecture07.ppt
 
A NEAT Way for Evolving Echo State Networks
A NEAT Way for Evolving Echo State NetworksA NEAT Way for Evolving Echo State Networks
A NEAT Way for Evolving Echo State Networks
 
ARTIFICIAL NEURAL NETWORKS
ARTIFICIAL NEURAL NETWORKSARTIFICIAL NEURAL NETWORKS
ARTIFICIAL NEURAL NETWORKS
 
Echo state networks and locomotion patterns
Echo state networks and locomotion patternsEcho state networks and locomotion patterns
Echo state networks and locomotion patterns
 
Neural networks1
Neural networks1Neural networks1
Neural networks1
 
Introduction to Neural networks (under graduate course) Lecture 9 of 9
Introduction to Neural networks (under graduate course) Lecture 9 of 9Introduction to Neural networks (under graduate course) Lecture 9 of 9
Introduction to Neural networks (under graduate course) Lecture 9 of 9
 
2.5 backpropagation
2.5 backpropagation2.5 backpropagation
2.5 backpropagation
 
Artificial Neural Networks Lect3: Neural Network Learning rules
Artificial Neural Networks Lect3: Neural Network Learning rulesArtificial Neural Networks Lect3: Neural Network Learning rules
Artificial Neural Networks Lect3: Neural Network Learning rules
 

Andere mochten auch

REAP Assessment/Feedback Principles and Examples
REAP Assessment/Feedback Principles and ExamplesREAP Assessment/Feedback Principles and Examples
REAP Assessment/Feedback Principles and ExamplesMartin Hawksey
 
Feed-forward approaches for enhancing assessment and feedback
Feed-forward approaches for enhancing assessment and feedbackFeed-forward approaches for enhancing assessment and feedback
Feed-forward approaches for enhancing assessment and feedbackjisc-elearning
 
Radial Basis Function Network (RBFN)
Radial Basis Function Network (RBFN)Radial Basis Function Network (RBFN)
Radial Basis Function Network (RBFN)ahmad haidaroh
 
Radial basis function network ppt bySheetal,Samreen and Dhanashri
Radial basis function network ppt bySheetal,Samreen and DhanashriRadial basis function network ppt bySheetal,Samreen and Dhanashri
Radial basis function network ppt bySheetal,Samreen and Dhanashrisheetal katkar
 
Neural Networks: Radial Bases Functions (RBF)
Neural Networks: Radial Bases Functions (RBF)Neural Networks: Radial Bases Functions (RBF)
Neural Networks: Radial Bases Functions (RBF)Mostafa G. M. Mostafa
 
Does your feedback feed forward?
Does your feedback feed forward?Does your feedback feed forward?
Does your feedback feed forward?Eddy White, Ph.D.
 
An introduction to Machine Learning
An introduction to Machine LearningAn introduction to Machine Learning
An introduction to Machine Learningbutest
 

Andere mochten auch (9)

REAP Assessment/Feedback Principles and Examples
REAP Assessment/Feedback Principles and ExamplesREAP Assessment/Feedback Principles and Examples
REAP Assessment/Feedback Principles and Examples
 
Feed-forward approaches for enhancing assessment and feedback
Feed-forward approaches for enhancing assessment and feedbackFeed-forward approaches for enhancing assessment and feedback
Feed-forward approaches for enhancing assessment and feedback
 
Feedback feedforward coaching_130722 - alternatives
Feedback feedforward coaching_130722 - alternativesFeedback feedforward coaching_130722 - alternatives
Feedback feedforward coaching_130722 - alternatives
 
Radial Basis Function Network (RBFN)
Radial Basis Function Network (RBFN)Radial Basis Function Network (RBFN)
Radial Basis Function Network (RBFN)
 
Radial Basis Function
Radial Basis FunctionRadial Basis Function
Radial Basis Function
 
Radial basis function network ppt bySheetal,Samreen and Dhanashri
Radial basis function network ppt bySheetal,Samreen and DhanashriRadial basis function network ppt bySheetal,Samreen and Dhanashri
Radial basis function network ppt bySheetal,Samreen and Dhanashri
 
Neural Networks: Radial Bases Functions (RBF)
Neural Networks: Radial Bases Functions (RBF)Neural Networks: Radial Bases Functions (RBF)
Neural Networks: Radial Bases Functions (RBF)
 
Does your feedback feed forward?
Does your feedback feed forward?Does your feedback feed forward?
Does your feedback feed forward?
 
An introduction to Machine Learning
An introduction to Machine LearningAn introduction to Machine Learning
An introduction to Machine Learning
 

Ähnlich wie Ffnn

Intro to Deep learning - Autoencoders
Intro to Deep learning - Autoencoders Intro to Deep learning - Autoencoders
Intro to Deep learning - Autoencoders Akash Goel
 
NEURALNETWORKS_DM_SOWMYAJYOTHI.pdf
NEURALNETWORKS_DM_SOWMYAJYOTHI.pdfNEURALNETWORKS_DM_SOWMYAJYOTHI.pdf
NEURALNETWORKS_DM_SOWMYAJYOTHI.pdfSowmyaJyothi3
 
Artificial neural networks seminar presentation using MSWord.
Artificial neural networks seminar presentation using MSWord.Artificial neural networks seminar presentation using MSWord.
Artificial neural networks seminar presentation using MSWord.Mohd Faiz
 
Web Spam Classification Using Supervised Artificial Neural Network Algorithms
Web Spam Classification Using Supervised Artificial Neural Network AlgorithmsWeb Spam Classification Using Supervised Artificial Neural Network Algorithms
Web Spam Classification Using Supervised Artificial Neural Network Algorithmsaciijournal
 
CCS355 Neural Networks & Deep Learning Unit 1 PDF notes with Question bank .pdf
CCS355 Neural Networks & Deep Learning Unit 1 PDF notes with Question bank .pdfCCS355 Neural Networks & Deep Learning Unit 1 PDF notes with Question bank .pdf
CCS355 Neural Networks & Deep Learning Unit 1 PDF notes with Question bank .pdfAsst.prof M.Gokilavani
 
Artificial neural networks and its application
Artificial neural networks and its applicationArtificial neural networks and its application
Artificial neural networks and its applicationHưng Đặng
 
Artificial neural networks and its application
Artificial neural networks and its applicationArtificial neural networks and its application
Artificial neural networks and its applicationHưng Đặng
 
Classification by back propagation, multi layered feed forward neural network...
Classification by back propagation, multi layered feed forward neural network...Classification by back propagation, multi layered feed forward neural network...
Classification by back propagation, multi layered feed forward neural network...bihira aggrey
 
DESIGN AND IMPLEMENTATION OF BINARY NEURAL NETWORK LEARNING WITH FUZZY CLUSTE...
DESIGN AND IMPLEMENTATION OF BINARY NEURAL NETWORK LEARNING WITH FUZZY CLUSTE...DESIGN AND IMPLEMENTATION OF BINARY NEURAL NETWORK LEARNING WITH FUZZY CLUSTE...
DESIGN AND IMPLEMENTATION OF BINARY NEURAL NETWORK LEARNING WITH FUZZY CLUSTE...cscpconf
 
ACUMENS ON NEURAL NET AKG 20 7 23.pptx
ACUMENS ON NEURAL NET AKG 20 7 23.pptxACUMENS ON NEURAL NET AKG 20 7 23.pptx
ACUMENS ON NEURAL NET AKG 20 7 23.pptxgnans Kgnanshek
 
Electricity Demand Forecasting Using Fuzzy-Neural Network
Electricity Demand Forecasting Using Fuzzy-Neural NetworkElectricity Demand Forecasting Using Fuzzy-Neural Network
Electricity Demand Forecasting Using Fuzzy-Neural NetworkNaren Chandra Kattla
 
Levenberg marquardt-algorithm-for-karachi-stock-exchange-share-rates-forecast...
Levenberg marquardt-algorithm-for-karachi-stock-exchange-share-rates-forecast...Levenberg marquardt-algorithm-for-karachi-stock-exchange-share-rates-forecast...
Levenberg marquardt-algorithm-for-karachi-stock-exchange-share-rates-forecast...Cemal Ardil
 
A survey research summary on neural networks
A survey research summary on neural networksA survey research summary on neural networks
A survey research summary on neural networkseSAT Publishing House
 
ARTIFICIAL NEURAL NETWORK APPROACH TO MODELING OF POLYPROPYLENE REACTOR
ARTIFICIAL NEURAL NETWORK APPROACH TO MODELING OF POLYPROPYLENE REACTORARTIFICIAL NEURAL NETWORK APPROACH TO MODELING OF POLYPROPYLENE REACTOR
ARTIFICIAL NEURAL NETWORK APPROACH TO MODELING OF POLYPROPYLENE REACTORijac123
 
Artificial neural network
Artificial neural networkArtificial neural network
Artificial neural networkmustafa aadel
 

Ähnlich wie Ffnn (20)

Intro to Deep learning - Autoencoders
Intro to Deep learning - Autoencoders Intro to Deep learning - Autoencoders
Intro to Deep learning - Autoencoders
 
NEURALNETWORKS_DM_SOWMYAJYOTHI.pdf
NEURALNETWORKS_DM_SOWMYAJYOTHI.pdfNEURALNETWORKS_DM_SOWMYAJYOTHI.pdf
NEURALNETWORKS_DM_SOWMYAJYOTHI.pdf
 
Artificial neural networks seminar presentation using MSWord.
Artificial neural networks seminar presentation using MSWord.Artificial neural networks seminar presentation using MSWord.
Artificial neural networks seminar presentation using MSWord.
 
Web Spam Classification Using Supervised Artificial Neural Network Algorithms
Web Spam Classification Using Supervised Artificial Neural Network AlgorithmsWeb Spam Classification Using Supervised Artificial Neural Network Algorithms
Web Spam Classification Using Supervised Artificial Neural Network Algorithms
 
N ns 1
N ns 1N ns 1
N ns 1
 
CCS355 Neural Networks & Deep Learning Unit 1 PDF notes with Question bank .pdf
CCS355 Neural Networks & Deep Learning Unit 1 PDF notes with Question bank .pdfCCS355 Neural Networks & Deep Learning Unit 1 PDF notes with Question bank .pdf
CCS355 Neural Networks & Deep Learning Unit 1 PDF notes with Question bank .pdf
 
Artificial neural networks and its application
Artificial neural networks and its applicationArtificial neural networks and its application
Artificial neural networks and its application
 
Artificial neural networks and its application
Artificial neural networks and its applicationArtificial neural networks and its application
Artificial neural networks and its application
 
02 Fundamental Concepts of ANN
02 Fundamental Concepts of ANN02 Fundamental Concepts of ANN
02 Fundamental Concepts of ANN
 
Classification by back propagation, multi layered feed forward neural network...
Classification by back propagation, multi layered feed forward neural network...Classification by back propagation, multi layered feed forward neural network...
Classification by back propagation, multi layered feed forward neural network...
 
DESIGN AND IMPLEMENTATION OF BINARY NEURAL NETWORK LEARNING WITH FUZZY CLUSTE...
DESIGN AND IMPLEMENTATION OF BINARY NEURAL NETWORK LEARNING WITH FUZZY CLUSTE...DESIGN AND IMPLEMENTATION OF BINARY NEURAL NETWORK LEARNING WITH FUZZY CLUSTE...
DESIGN AND IMPLEMENTATION OF BINARY NEURAL NETWORK LEARNING WITH FUZZY CLUSTE...
 
ACUMENS ON NEURAL NET AKG 20 7 23.pptx
ACUMENS ON NEURAL NET AKG 20 7 23.pptxACUMENS ON NEURAL NET AKG 20 7 23.pptx
ACUMENS ON NEURAL NET AKG 20 7 23.pptx
 
Electricity Demand Forecasting Using Fuzzy-Neural Network
Electricity Demand Forecasting Using Fuzzy-Neural NetworkElectricity Demand Forecasting Using Fuzzy-Neural Network
Electricity Demand Forecasting Using Fuzzy-Neural Network
 
Levenberg marquardt-algorithm-for-karachi-stock-exchange-share-rates-forecast...
Levenberg marquardt-algorithm-for-karachi-stock-exchange-share-rates-forecast...Levenberg marquardt-algorithm-for-karachi-stock-exchange-share-rates-forecast...
Levenberg marquardt-algorithm-for-karachi-stock-exchange-share-rates-forecast...
 
A survey research summary on neural networks
A survey research summary on neural networksA survey research summary on neural networks
A survey research summary on neural networks
 
Neural network and fuzzy logic
Neural network and fuzzy logicNeural network and fuzzy logic
Neural network and fuzzy logic
 
ARTIFICIAL NEURAL NETWORK APPROACH TO MODELING OF POLYPROPYLENE REACTOR
ARTIFICIAL NEURAL NETWORK APPROACH TO MODELING OF POLYPROPYLENE REACTORARTIFICIAL NEURAL NETWORK APPROACH TO MODELING OF POLYPROPYLENE REACTOR
ARTIFICIAL NEURAL NETWORK APPROACH TO MODELING OF POLYPROPYLENE REACTOR
 
Neural network
Neural networkNeural network
Neural network
 
Artificial neural network
Artificial neural networkArtificial neural network
Artificial neural network
 
Jack
JackJack
Jack
 

Kürzlich hochgeladen

[2024]Digital Global Overview Report 2024 Meltwater.pdf
[2024]Digital Global Overview Report 2024 Meltwater.pdf[2024]Digital Global Overview Report 2024 Meltwater.pdf
[2024]Digital Global Overview Report 2024 Meltwater.pdfhans926745
 
IAC 2024 - IA Fast Track to Search Focused AI Solutions
IAC 2024 - IA Fast Track to Search Focused AI SolutionsIAC 2024 - IA Fast Track to Search Focused AI Solutions
IAC 2024 - IA Fast Track to Search Focused AI SolutionsEnterprise Knowledge
 
A Domino Admins Adventures (Engage 2024)
A Domino Admins Adventures (Engage 2024)A Domino Admins Adventures (Engage 2024)
A Domino Admins Adventures (Engage 2024)Gabriella Davis
 
Artificial Intelligence: Facts and Myths
Artificial Intelligence: Facts and MythsArtificial Intelligence: Facts and Myths
Artificial Intelligence: Facts and MythsJoaquim Jorge
 
Scaling API-first – The story of a global engineering organization
Scaling API-first – The story of a global engineering organizationScaling API-first – The story of a global engineering organization
Scaling API-first – The story of a global engineering organizationRadu Cotescu
 
From Event to Action: Accelerate Your Decision Making with Real-Time Automation
From Event to Action: Accelerate Your Decision Making with Real-Time AutomationFrom Event to Action: Accelerate Your Decision Making with Real-Time Automation
From Event to Action: Accelerate Your Decision Making with Real-Time AutomationSafe Software
 
Advantages of Hiring UIUX Design Service Providers for Your Business
Advantages of Hiring UIUX Design Service Providers for Your BusinessAdvantages of Hiring UIUX Design Service Providers for Your Business
Advantages of Hiring UIUX Design Service Providers for Your BusinessPixlogix Infotech
 
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...Miguel Araújo
 
Exploring the Future Potential of AI-Enabled Smartphone Processors
Exploring the Future Potential of AI-Enabled Smartphone ProcessorsExploring the Future Potential of AI-Enabled Smartphone Processors
Exploring the Future Potential of AI-Enabled Smartphone Processorsdebabhi2
 
EIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptx
EIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptxEIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptx
EIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptxEarley Information Science
 
The Codex of Business Writing Software for Real-World Solutions 2.pptx
The Codex of Business Writing Software for Real-World Solutions 2.pptxThe Codex of Business Writing Software for Real-World Solutions 2.pptx
The Codex of Business Writing Software for Real-World Solutions 2.pptxMalak Abu Hammad
 
04-2024-HHUG-Sales-and-Marketing-Alignment.pptx
04-2024-HHUG-Sales-and-Marketing-Alignment.pptx04-2024-HHUG-Sales-and-Marketing-Alignment.pptx
04-2024-HHUG-Sales-and-Marketing-Alignment.pptxHampshireHUG
 
Real Time Object Detection Using Open CV
Real Time Object Detection Using Open CVReal Time Object Detection Using Open CV
Real Time Object Detection Using Open CVKhem
 
Breaking the Kubernetes Kill Chain: Host Path Mount
Breaking the Kubernetes Kill Chain: Host Path MountBreaking the Kubernetes Kill Chain: Host Path Mount
Breaking the Kubernetes Kill Chain: Host Path MountPuma Security, LLC
 
Handwritten Text Recognition for manuscripts and early printed texts
Handwritten Text Recognition for manuscripts and early printed textsHandwritten Text Recognition for manuscripts and early printed texts
Handwritten Text Recognition for manuscripts and early printed textsMaria Levchenko
 
A Year of the Servo Reboot: Where Are We Now?
A Year of the Servo Reboot: Where Are We Now?A Year of the Servo Reboot: Where Are We Now?
A Year of the Servo Reboot: Where Are We Now?Igalia
 
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024The Digital Insurer
 
2024: Domino Containers - The Next Step. News from the Domino Container commu...
2024: Domino Containers - The Next Step. News from the Domino Container commu...2024: Domino Containers - The Next Step. News from the Domino Container commu...
2024: Domino Containers - The Next Step. News from the Domino Container commu...Martijn de Jong
 
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...Drew Madelung
 
The Role of Taxonomy and Ontology in Semantic Layers - Heather Hedden.pdf
The Role of Taxonomy and Ontology in Semantic Layers - Heather Hedden.pdfThe Role of Taxonomy and Ontology in Semantic Layers - Heather Hedden.pdf
The Role of Taxonomy and Ontology in Semantic Layers - Heather Hedden.pdfEnterprise Knowledge
 

Kürzlich hochgeladen (20)

[2024]Digital Global Overview Report 2024 Meltwater.pdf
[2024]Digital Global Overview Report 2024 Meltwater.pdf[2024]Digital Global Overview Report 2024 Meltwater.pdf
[2024]Digital Global Overview Report 2024 Meltwater.pdf
 
IAC 2024 - IA Fast Track to Search Focused AI Solutions
IAC 2024 - IA Fast Track to Search Focused AI SolutionsIAC 2024 - IA Fast Track to Search Focused AI Solutions
IAC 2024 - IA Fast Track to Search Focused AI Solutions
 
A Domino Admins Adventures (Engage 2024)
A Domino Admins Adventures (Engage 2024)A Domino Admins Adventures (Engage 2024)
A Domino Admins Adventures (Engage 2024)
 
Artificial Intelligence: Facts and Myths
Artificial Intelligence: Facts and MythsArtificial Intelligence: Facts and Myths
Artificial Intelligence: Facts and Myths
 
Scaling API-first – The story of a global engineering organization
Scaling API-first – The story of a global engineering organizationScaling API-first – The story of a global engineering organization
Scaling API-first – The story of a global engineering organization
 
From Event to Action: Accelerate Your Decision Making with Real-Time Automation
From Event to Action: Accelerate Your Decision Making with Real-Time AutomationFrom Event to Action: Accelerate Your Decision Making with Real-Time Automation
From Event to Action: Accelerate Your Decision Making with Real-Time Automation
 
Advantages of Hiring UIUX Design Service Providers for Your Business
Advantages of Hiring UIUX Design Service Providers for Your BusinessAdvantages of Hiring UIUX Design Service Providers for Your Business
Advantages of Hiring UIUX Design Service Providers for Your Business
 
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...
 
Exploring the Future Potential of AI-Enabled Smartphone Processors
Exploring the Future Potential of AI-Enabled Smartphone ProcessorsExploring the Future Potential of AI-Enabled Smartphone Processors
Exploring the Future Potential of AI-Enabled Smartphone Processors
 
EIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptx
EIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptxEIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptx
EIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptx
 
The Codex of Business Writing Software for Real-World Solutions 2.pptx
The Codex of Business Writing Software for Real-World Solutions 2.pptxThe Codex of Business Writing Software for Real-World Solutions 2.pptx
The Codex of Business Writing Software for Real-World Solutions 2.pptx
 
04-2024-HHUG-Sales-and-Marketing-Alignment.pptx
04-2024-HHUG-Sales-and-Marketing-Alignment.pptx04-2024-HHUG-Sales-and-Marketing-Alignment.pptx
04-2024-HHUG-Sales-and-Marketing-Alignment.pptx
 
Real Time Object Detection Using Open CV
Real Time Object Detection Using Open CVReal Time Object Detection Using Open CV
Real Time Object Detection Using Open CV
 
Breaking the Kubernetes Kill Chain: Host Path Mount
Breaking the Kubernetes Kill Chain: Host Path MountBreaking the Kubernetes Kill Chain: Host Path Mount
Breaking the Kubernetes Kill Chain: Host Path Mount
 
Handwritten Text Recognition for manuscripts and early printed texts
Handwritten Text Recognition for manuscripts and early printed textsHandwritten Text Recognition for manuscripts and early printed texts
Handwritten Text Recognition for manuscripts and early printed texts
 
A Year of the Servo Reboot: Where Are We Now?
A Year of the Servo Reboot: Where Are We Now?A Year of the Servo Reboot: Where Are We Now?
A Year of the Servo Reboot: Where Are We Now?
 
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
 
2024: Domino Containers - The Next Step. News from the Domino Container commu...
2024: Domino Containers - The Next Step. News from the Domino Container commu...2024: Domino Containers - The Next Step. News from the Domino Container commu...
2024: Domino Containers - The Next Step. News from the Domino Container commu...
 
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
 
The Role of Taxonomy and Ontology in Semantic Layers - Heather Hedden.pdf
The Role of Taxonomy and Ontology in Semantic Layers - Heather Hedden.pdfThe Role of Taxonomy and Ontology in Semantic Layers - Heather Hedden.pdf
The Role of Taxonomy and Ontology in Semantic Layers - Heather Hedden.pdf
 

Ffnn

  • 1. 1 FEEDFORWARD NEURAL NETWORKS: AN INTRODUCTION Simon Haykin A neural network is a massively parallel distributed processor that has a natural propensity for storing experiential knowledge and making it available for use. It resembles the brain in two respects (Haykin 1998): 1. Knowledge is acquired by the network through a learning process. 2. Interconnection strengths known as synaptic weights are used to store the knowledge. Basically, learning is a process by which the free parameters (i.e., synaptic weights and bias levels) of a neural network are adapted through a continuing process of stimulation by the environment in which the network is embedded. The type of learning is determined by the manner in which the parameter changes take place. In a general sense, the learning process may be classified as follows: 1
  • 2. 2 FEEDFORWARD NEURAL NETWORKS: AN INTRODUCTION • Learning with a teacher, also referred to as supervised learning • Learning without a teacher, also referred to as unsupervised learning 1.1 SUPERVISED LEARNING This form of learning assumes the availability of a labeled (i.e., ground-truthed) set of training data made up of N input—output examples: N T = {( x i , di )}i =1 (1.1) where xi = input vector of ith example di = desired (target) response of ith example, assumed to be scalar for convenience of presentation N = sample size Given the training sample T, the requirement is to compute the free parameters of the neural network so that the actual output yi of the neural network due to xi is close enough to di for all i in a statistical sense. For example, we may use the mean-square error N 1 2 E ( n) = N Â (d i - yi ) (1.2) i =1 as the index of performance to be minimized. 1.1.1 Multilayer Perceptrons and Back-Propagation Learning The back-propagation algorithm has emerged as the workhorse for the design of a special class of layered feedforward networks known as multilayer perceptrons (MLP). As shown in Fig. 1.1, a multilayer perceptron has an input layer of source nodes and an output layer of neurons (i.e., computation nodes); these two layers connect the network to the outside world. In addition to these two layers, the multilayer perceptron usually has one or more layers of hidden neurons, which are so called because these neurons are not directly accessible. The hidden neurons extract important features contained in the input data.
  • 3. SUPERVISED LEARNING 3 Figure 1.1 Fully connected feedforward with one hidden layer and one output layer. The training of an MLP is usually accomplished by using a back- propagation (BP) algorithm that involves two phases (Werbos 1974; Rumelhart et al. 1986): • Forward Phase. During this phase the free parameters of the network are fixed, and the input signal is propagated through the network of Fig. 1.1 layer by layer. The forward phase fin- ishes with the computation of an error signal ei = di - yi (1.3) where di is the desired response and yi is the actual output pro- duced by the network in response to the input xi. • Backward Phase. During this second phase, the error signal ei is propagated through the network of Fig. 1.1 in the backward direction, hence the name of the algorithm. It is during this phase that adjustments are applied to the free parameters of the network so as to minimize the error ei in a statistical sense. Back-propagation learning may be implemented in one of two basic ways, as summarized here:
  • 4. 4 FEEDFORWARD NEURAL NETWORKS: AN INTRODUCTION 1. Sequential mode (also referred to as the on-line mode or sto- chastic mode): In this mode of BP learning, adjustments are made to the free parameters of the network on an example-by- example basis. The sequential mode is best suited for pattern classification. 2. Batch mode: In this second mode of BP learning, adjustments are made to the free parameters of the network on an epoch- by-epoch basis, where each epoch consists of the entire set of training examples. The batch mode is best suited for nonlinear regression. The back-propagation learning algorithm is simple to implement and computationally efficient in that its complexity is linear in the synap- tic weights of the network. However, a major limitation of the algo- rithm is that it does not always converge and can be excruciatingly slow, particularly when we have to deal with a difficult learning task that requires the use of a large network. We may try to make back-propagation learning perform better by invoking the following list of heuristics: • Use neurons with antisymmetric activation functions (e.g., hyperbolic tangent function) in preference to nonsymmetric activation functions (e.g., logistic function). Figure 1.2 shows examples of these two forms of activation functions. • Shuffle the training examples after the presentation of each epoch; an epoch involves the presentation of the entire set of training examples to the network. • Follow an easy-to-learn example with a difficult one. • Preprocess the input data so as to remove the mean and decor- relate the data. • Arrange for the neurons in the different layers to learn at essen- tially the same rate. This may be attained by assigning a learn- ing rate parameter to neurons in the last layers that is smaller than those at the front end. • Incorporate prior information into the network design when- ever it is available. One other heuristic that deserves to be mentioned relates to the size of the training set, N, for a pattern classification task. Given a mul- tilayer perceptron with a total number of synaptic weights including bias levels, denoted by W, a rule of thumb for selecting N is
  • 5. SUPERVISED LEARNING 5 Figure 1.2 (a) Antisymmetric activation function. (b) Nonsymmetric activation function.
  • 6. 6 FEEDFORWARD NEURAL NETWORKS: AN INTRODUCTION W N = OÊ ˆ (1.4) Ë e¯ where O denotes “the order of,” and e denotes the fraction of clas- sification errors permitted on test data. For example, with an error of 10% the number of training examples needed should be about 10 times the number of synaptic weights in the network. Supposing that we have chosen a multilayer perceptron to be trained with the back-propagation algorithm, how do we determine when it is “best” to stop the training session? How do we select the size of individual hidden layers of the MLP? The answers to these important questions may be gotten though the use of a statistical technique known as cross-validation, which proceeds as follows (Haykin 1999): • The set of training examples is split into two parts: • Estimation subset used for training of the model • Validation subset used for evaluating the model performance • The network is finally tuned by using the entire set of training examples and then tested on test data not seen before. 1.1.2 Radial-Basis Function Networks Another popular layered feedforward network is the radial-basis function (RBF) network which has important universal approxima- tion properties (Park and Sandberg 1993), and whose structure is shown in Fig. 13. RBF networks use memory-based learning for their design. Specifically, learning is viewed as a curve-fitting problem in high-dimensional space (Broomhead and Lowe 1989; Poggio and Girosi 1990): 1. Learning is equivalent to finding a surface in a multidimen- sional space that provides a best fit to the training data. 2. Generalization (i.e., response of the network to input data not seen before) is equivalent to the use of this multidimensional surface to interpolate the test data. RBF networks differ from multilayer perceptrons in some funda- mental respects:
  • 7. SUPERVISED LEARNING 7 Figure 1.3 Radial-basis function network. • RBF networks are local approximators, whereas multilayer per- ceptrons are global approximators. • RBF networks have a single hidden layer, whereas multilayer perceptrons can have any number of hidden layers. • The output layer of a RBF network is always linear, whereas in a multilayer perceptron it can be linear or nonlinear. • The activation function of the hidden layer in an RBF network computes the Euclidean distance between the input signal vector and parameter vector of the network, whereas the acti- vation function of a multilayer perceptron computes the inner product between the input signal vector and the pertinent synaptic weight vector. The use of a linear output layer in an RBF network may be justi- fied in light of Cover’s theorem on the separability of patterns. According to this theorem, provided that the transformation from the input space to the feature (hidden) space is nonlinear and the
  • 8. 8 FEEDFORWARD NEURAL NETWORKS: AN INTRODUCTION dimensionality of the feature space is high compared to that of the input (data) space, then there is a high likelihood that a nonsepara- ble pattern classification task in the input space is transformed into a linearly separable one in the feature space.Another analytical basis for the use of RBF networks (and multilayer perceptrons) in classi- fication problems is provided by the results in Chapter 2, where (as a special case) it is shown that a large family of classification prob- lems in n can be solved using nonlinear static networks. Design methods for RBF networks include the following: 1. Random selection of fixed centers (Broomhead and Lowe 1998) 2. Self-organized selection of centers (Moody and Darken 1989) 3. Supervised selection of centers (Poggio and Girosi 1990) 4. Regularized interpolation exploiting the connection between an RBF network and the Watson–Nadaraya regression kernel (Yee 1998). 1.2 UNSUPERVISED LEARNING Turning next to unsupervised learning, adjustment of synaptic weights may be carried through the use of neurobiological principles such as Hebbian learning and competitive learning. In this section we will describe specific applications of these two approaches. 1.2.1 Principal Components Analysis According to Hebb’s postulate of learning, the change in synaptic weight Dwji of a neural network is defined by Dwji = hxiyj (1.5) where h = learning-rate parameter xi = input (presynaptic) signal yi = output (postsynaptic) signal Principal component analysis (PCA) networks use a modified form of this self-organized learning rule. To begin with, consider a linear neuron designed to operate as a maximum eigenfilter; such a neuron is referred to as Oja’s neuron (Oja 1982). It is characterized as follow:
  • 9. UNSUPERVISED LEARNING 9 Dw ji = hy j ( xi - y j w ji ) (1.6) where the term -hy2wji is added to stabilize the learning process. As j the number of iterations approaches infinity, we find the following: 1. The synaptic weight vector of neuron j approaches the eigen- vector associated with the largest eigenvalue l max of the corre- lation matrix of the input vector (assumed to be of zero mean). 2. The variance of the output of neuron j approaches the largest eigenvalue l max. The generalized Hebbian algorithm (GHA), due to Sanger (1989), is a straightforward generalization of Oja’s neuron for the extraction of any desired number of principal components. 1.2.2 Self-Organizing Maps In a self-organizing map (SOM), due to Kohonen (1997), the neurons are placed at the nodes of a lattice, and they become selectively tuned to various input patterns (vectors) in the course of a compet- itive learning process. The process is characterized by the formation of a topographic map in which the spatial locations (i.e., coordinates) of the neurons in the lattice correspond to intrinsic features of the input patterns. Figure 1.4 illustrates the basic idea of a self-organiz- ing map, assuming the use of a two-dimensional lattice of neurons as the network structure. In reality, the SOM belongs to the class of vector-coding algo- rithms (Luttrell, 1989). That is, a fixed number of codewords are placed into a higher-dimensional input space, thereby facilitating data compression. An integral feature of the SOM algorithm is the neighborhood function centered around a neuron that wins the competitive process. The neighborhood function starts by enclosing the entire lattice initially and is then allowed to shrink gradually until it encom- passes the winning neuron. The algorithm exhibits two distinct phases in its operation: 1. Ordering phase, during which the topological ordering of the weight vectors takes place 2. Convergence phase, during which the computational map is fine tuned
  • 10. 10 FEEDFORWARD NEURAL NETWORKS: AN INTRODUCTION Figure 1.4 Illustration of relationship between feature map f and weight vector wi of winning neuron i. The SOM algorithm exhibits the following properties: 1. Approximation of the continuous input space by the weight vectors of the discrete lattice. 2. Topological ordering exemplified by the fact that the spatial location of a neuron in the lattice corresponds to a particular feature of the input pattern. 3. The feature map computed by the algorithm reflects variations in the statistics of the input distribution. 4. SOM may be viewed as a nonlinear form of principal compo- nents analysis. Refinements to the SOM algorithm are discussed in Van Hulle (2000). 1.3 TEMPORAL PROCESSING USING FEEDFORWARD NETWORKS The material just described is concerned basically with the approxi- mation of systems without dynamics (i.e., with static systems). At
  • 11. TEMPORAL PROCESSING USING FEEDFORWARD NETWORKS 11 about the same time as the appearance of the early universal- approximation theorems for static neural networks there began (Sandberg 1991a) a corresponding study (see Chapter 2) of the neural network approximation of approximately-finite-memory maps and myopic maps. It was found that large classes of these maps can be uni- formly approximated arbitrarily well by the maps of certain simple nonlinear structures using, for example, sigmoidal nonlinearities or radial basis functions. The approximating networks are two-stage structures comprising a linear preprocessing stage followed by a memoryless nonlinear network. Much is now known about the prop- erties of these networks, and examples of these properties are given in the following. From another perspective, time is an essential dimension of learning. We may incorporate time into the design of a neural network implicitly or explicitly. A straightforward method of implicit representation of time1 is to add a short-term memory structure in the input layer of a static neural network (e.g., multilayer percep- tron). The resulting configuration is sometimes called a focused time- lagged feedforward network (TLFN). The short-term memory structure may be implemented in one of two forms, as described here: 1. Tapped-Delay-Line (TDL) Memory. This is the most com- monly used form of short-term memory. It consists of p unit delays with ( p + 1) terminals, as shown in Fig. 1.5, which may be viewed as a single input–multiple output network. Figure 1.6 shows a focused TLFN network using the combination of a TDL memory and multilayer perceptron. In Figs. 1.5 and 1.6, the unit-delay is denoted by z-1. The memory depth of a TDL memory is fixed at p, and its memory resolution is fixed at unity, giving a depth resolution constant of p. 2. Gamma Memory. We may exercise control over the memory depth by building a feedback loop around each unit delay, as illustrated in Fig. 1.7 (deVries and Principe 1992). In effect, the unit delay z-1 of the standard TDL memory is replaced by the transfer function 1 Another practical way of accounting for time in a neural network is to employ feedback at the local or global level. Neural networks so configured are referred to as recurrent networks. For detailed treatment of recurrent networks, see Haykin (1999).
  • 12. 12 FEEDFORWARD NEURAL NETWORKS: AN INTRODUCTION Figure 1.5 Ordinary tapped-delay line memory of order p. Figure 1.6 Focused time-lagged feedforward network (TLFN); the bias levels have been omitted for convenience of presentation. Figure 1.7 Signal-flow graph for one section of gamma memory.
  • 13. TEMPORAL PROCESSING USING FEEDFORWARD NETWORKS 13 mz-1 G (z) = 1 - (1 - m )z-1 m = z - (1 - m ) where m is an adjustable parameter. For stability, the only pole of G(z) at z = (1 - m) must lie inside the unit circle in the z plane. This, in turn, requires that we restrict the choice of m to the following range of values: 0<m<2 The overall impulse response of the gamma memory, consist- ing of p sections, is the inverse z transform of the overall trans- fer function p m Gp (z) = Ê ˆ Ë z - (1 - m ) ¯ Denoting the impulse response by gp(n), we have n - 1ˆ p g p (z) = Ê n- p m (1 - m ) n≥ p Ë p - 1¯ where (:) is a binomial coefficient.The overall impulse response gp(n) for varying p represents a discrete version of the inte- grand of the gamma function (deVries and Principe 1992); hence the name “gamma memory.” The depth of the gamma memory is p/m and its resolution is m, for a depth resolution product of p. Accordingly, by choos- ing m to be less than unity, the gamma memory provides improvement in depth over the TDL memory, but at the expense of memory resolution. With regard to the utility of gamma networks—which are partic- ular cases of the family of two-stage structures comprising a linear preprocessing stage followed by a memoryless nonlinear network— experimental results have been reported which indicate that the structure is useful. In fact, it is known (Sandberg and Xu 1997) that
  • 14. 14 FEEDFORWARD NEURAL NETWORKS: AN INTRODUCTION for a large class of discrete-time dynamic system maps H, and for any choice of m in the interval (0, 1), there is a focused gamma network that approximates H uniformly arbitrarily well. It is known that tapped-delay-line networks (i.e., networks with m = 1) also have the universal approximation property (Sandberg 1991b). Focused TLFNs using ordinary tapped-delay memory or gamma memory are limited to stationary environments. To deal with nonstationary dynamical processes, we may use distributed TLFNs where the effect of time is distributed at the synaptic level through- out the network. One way in which this may be accomplished is to use finite-duration impulse response (FIR) filters to implement the synaptic connections of an MLP; Fig. 1.8 shows an FIR model of a synapse. The training of a distributed TLFN is naturally a more dif- ficult proposition than the training of a focused TLFN. Whereas we may use the ordinary back-propagation algorithm to train a focused TLFN, we have to extend the back-propagation algorithm to cope with the replacement of a synaptic weight in the ordinary MLP by a synaptic weight vector. This extension is referred to as the temporal back-propagation algorithm due to Wan (1994). 1.4 CONCLUDING REMARKS In this chapter, we briefly reviewed the feedforward type of neural networks, which are exemplified by multilayer perceptrons (MLPs), radial-basis function (RBF) networks, principal component analysis (PCA) networks, and self-organizing maps (SOMs). The training of MLPs and RBF networks proceeds in a supervised manner, whereas the training of PCA networks and SOMs proceeds in an unsuper- vised manner. Feedforward networks by themselves are nonlinear static net- works. They can be made to operate as nonlinear dynamical systems Figure 1.8 Finite-duration impulse response (FIR) filter.
  • 15. BIBLIOGRAPHY 15 by incorporating short-term memory into their input layer. Two important examples of short-term memory are the standard tapped- delay-line and the gamma memory that provides control over attain- able memory depth. The attractive feature of nonlinear dynamical systems built in this way is that they are inherently stable. BIBLIOGRAPHY Anderson, J. A., 1995, Introduction to Neural Networks (Cambridge, MA: MIT Press). Barlow, H. B., 1989, “Unsupervised learning,” Neural Computation, vol. 1, pp. 295–311. Becker, S., and G. E. Hinton, 1982, “A self-organizing neural network that discovers surfaces in random-dot stereograms,” Nature (London), vol. 355, pp. 161–163. Broomhead, D. S., and D. Lowe, 1988, “Multivariable functional inter- polation and adaptive networks,” Complex Systems, vol. 2, pp. 321– 355. Comon, P., 1994,“Independent component analysis:A new concept?” Signal Processing, vol. 36, pp. 287–314. deVries, B., and J. C. Principe, 1992, “The gamma model—A new neural model for temporal processing,” Neural Networks, vol. 4, pp. 565–576. Haykin, S., 1999, Neural Networks: A Comprehensive Foundation, 2nd ed. (Englewood Cliffs, NJ: Prentice-Hall). Moody and Darken, 1989, “Fast learning in networks of locally-tuned pro- cessing unites,” Neural Computation, vol. 1, pp. 281–294. Oja, E., 1982, “A simplified neuron model as a principal component ana- lyzer,” J. Math. Biol., vol. 15, pp. 267–273. Park, J., and Sandberg, I. W., 1993, “Approximation and radial-basis func- tion networks,” Neural computation, vol. 5, pp. 305–316. Poggio, T., and F. Girosi, 1990, “Networks for approximation and learning,” Proc. IEEE, vol. 78, pp. 1481–1497. Rumelhart, D. E., G. E. Hinton, and R. J. Williams, 1986, “Learning internal representations by error propagation,” in D. E. Rumelhart and J. L. McCleland, eds. (Cambridge, MA: MIT Press), vol. 1, Chapter 8. Sandberg, I. W., 1991a, “Structure theorems for nonlinear systems,” Multi- dimensional Sys. Sig. Process. vol. 2, pp. 267–286. (Errata in 1992, vol. 3, p. 101.) Sandberg, I. W., 1991b, “Approximation theorems for discrete-time systems,” IEEE Trans. Circuits Sys. vol. 38, no. 5, pp. 564–566, May 1991.
  • 16. 16 FEEDFORWARD NEURAL NETWORKS: AN INTRODUCTION Sandberg, I. W., and Xu, L., 1997, “Uniform approximation and gamma net- works,” Neural Networks, vol. 10, pp. 781–784. Van Hulle, M. M., 2000, Faithful Representations and Topographic Maps: From Distortion-to-Information-Based Self Organization (New York: Wiley). Wan, E. A., 1994, “Time series prediction by using a connectionist network with internal delay lines,” in A. S. Weigend and N. A. Gershenfield, eds., Time Series Prediction: Forecasting the Future and Understanding the Past (Reading, MA: Addison-Wesley), pp. 195–217. Werbos, P. J., 1974, “Beyond regression: New tools for prediction and analy- sis in the behavioral sciences,” Ph.D. Thesis, Harvard University, Cambridge, MA. Werbos, P. J., 1990, “Backpropagation through time: What it does and how to do it,” Proc. IEEE, vol. 78, pp. 1550–1560. Yee, P. V., 1998, “Regularized radial basis function networks: Theory and applications to probability estimation, classification, and time series pre- diction,” Ph.D. Thesis, McMaster University, Hamilton, Ontario.