SlideShare ist ein Scribd-Unternehmen logo
1 von 14
Downloaden Sie, um offline zu lesen
The Multivariate Gaussian Probability
Distribution
Peter Ahrendt
IMM, Technical University of Denmark
mail : pa@imm.dtu.dk, web : www.imm.dtu.dk/∼pa
January 7, 2005
Contents
1 Definition 2
2 Functions of Gaussian Variables 4
3 Characteristic function and Moments 6
4 Marginalization and Conditional Distribution 9
4.1 Marginalization . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9
4.2 Conditional distribution . . . . . . . . . . . . . . . . . . . . . . . 10
5 Tips and Tricks 11
5.1 Products . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11
5.2 Gaussian Integrals . . . . . . . . . . . . . . . . . . . . . . . . . . 12
5.3 Useful integrals . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12
1
Chapter 1
Definition
The definition of a multivariate gaussian probability distribution can be stated
in several equivalent ways. A random vector X = [X1X2 . . . XN ] can be said to
belong to a multivariate gaussian distribution if one of the following statements
is true.
• Any linear combination Y = a1X1 + a2X2 + . . . + aN XN , ai ∈ R is a
(univariate) gaussian distribution.
• There exists a random vector Z = [Z1, . . . , ZM ] with components that are
independent and standard normal distributed, a vector µ = [µ1, . . . , µN ]
and an N-by-M matrix A such that X = AZ + µ.
• There exists a vector µ and a symmetric, positive semi-definite matrix Γ
such that the characteristic function of X can be written φx(t) ≡ heitT
X
i =
eiµT
t− 1
2 tT
Γt
.
Under the assumption that the covariance matrix Σ is non-singular, the prob-
ability density function (pdf) can be written as :
Nx(µ, Σ) =
1
p
(2π)d|Σ|
exp
µ
−
1
2
(x − µ)T
Σ−1
(x − µ)
¶
= |2πΣ|− 1
2 exp
µ
−
1
2
(x − µ)T
Σ−1
(x − µ)
¶ (1.1)
Then µ is the mean value, Σ is the covariance matrix and | · | denote the
determinant. Note that it is possible to have multivariate gaussian distributions
with singular covariance matrix and then the above expression cannot be used
for the pdf. In the following, however, non-singular covariance matrices will be
assumed.
In the limit of one dimension, the familiar expression of the univariate gaussian
pdf is found.
2
Nx(µ, σ2
) =
1
√
2πσ2
exp
µ
−
(x − µ)2
2σ2
¶
=
1
√
2πσ2
exp
µ
−
1
2
(x − µ)σ−2
(x − µ)
¶
(1.2)
Neither of them have a closed-form expression for the cumulative density func-
tion.
Symmetries
It is noted that in the one-dimensional case there is a symmetry in the pdf.
Nx(µ, σ2
) which is centered on µ. This can be seen by looking at ”contour
lines”, i.e. setting the exponent −(x−µ)2
2σ2 = c. It is seen that σ determines the
width of the distribution.
In the multivariate case, it is similarly useful to look at −1
2 (x − µ)T
Σ−1
(x − µ) = c.
This is a quadratic form and geometrically the contour curves (for fixed c) are hy-
perellipsoids. In 2D, this is normal ellipsoids with the form (x−x0
a )2
+(y−y0
b )2
=
r2
, which gives symmetries along the principal axes. Similarly, the hyperellip-
soids show symmetries along their principal axes.
Notation: If a random variable X has a gaussian distribution, it is written as
X∼ N(µ, Σ). The probability density function of this variable is then given by
Nx(µ, Σ).
3
Chapter 2
Functions of Gaussian
Variables
Linear transformation and addition of variables
Let A, B ∈ Mc·d
and c ∈ Rd
. Let X ∼ N(µx, Σx) and Y ∼ N(µy, Σy) be
independent variables. Then
Z = AX + BY + c ∼ N(Aµx + Bµy + c, AΣxAT
+ BΣyBT
) (2.1)
Transform to standard normal variables
Let X ∼ N(µ, Σ). Then
Z = Σ− 1
2 (X − µ) ∼ N(0, 1) (2.2)
Note, that by Σ− 1
2 is actually meant a unique matrix, although in general
matrices with fractional exponents are not. The matrix that is meant can be
found from the diagonalisation into Σ = UΛUT
= (UΛ
1
2 )(UΛ
1
2 )T
where Λ
is the diagonal matrix with the eigenvalues of Σ and U is the matrix with the
eigenvectors. Then Σ− 1
2 = (UΛ
1
2 )−1
= Λ− 1
2 U−1
.
In the one-dimensional case, this corresponds to the transformation of X ∼
N(µ, σ2
) into Y = σ−1
(X − µ) ∼ N(0, 1).
Addition
Let Xi ∼ N(µi, Σi), i ∈ 1, ..., N be independent variables. Then
N
X
i
Xi ∼ N(
N
X
i
µi,
N
X
i
Σi) (2.3)
Note: This is a direct implication of equation (2.1).
4
Quadratic
Let Xi ∼ N(0, 1), i ∈ 1, ..., N be independent variables. Then
N
X
i
X2
i ∼ χ2
n (2.4)
Alternatively let X ∼ N(µx, Σx). Then
Z = (X − µ)T
Σ−1
(X − µ) ∼ χ2
n (2.5)
This is, however, the same thing since Z= X̃
T
X̃ =
PN
i X̃i
2
, where X̃i are the
decorrelated components (see eqn. (2.2)).
5
Chapter 3
Characteristic function and
Moments
The characteristic function of the univariate gaussian distribution is given by
φx(t) ≡ heitX
i = eitµ−σ2
t2
/2
. The generalization to multivariate gaussian distri-
butions is
φx(t) ≡ heitT
X
i = eiµT
t− 1
2 tT
Σt
(3.1)
The pdf p(x) is related to the characteristic function.
p(x) =
1
(2π)d
Z
Rd
φx(t) e−itT
x
dt (3.2)
It is seen that the characteristic function is the inverse Fourier transform of the
pdf.
Moments of a pdf are generally defined as :
hXk1
1 Xk2
2 · · · XkN
N i ≡
Z
Rd
xk1
1 xk2
2 · · · xkN
N p(x) dx (3.3)
where hXk1
1 Xk2
2 · · · XkN
N i is the k’th order moment, k = [k1, k2, . . . , kN ] (ki ∈ N)
and k = k1 + k2 + . . . + kN . A well-known example is the first order moment,
called the mean value µi (of variable Xi) - or the mean µ ≡ [µ1µ2 . . . µN ] of the
whole random vector X.
The k’th order central moment is defined as above, but with Xi replaced by
Xi − µi in equation (3.3). An example is the second order central moment,
called the variance, which is given by h(Xi − µi)2
i.
Any moment (that exists) can be found from the characteristic function [8]:
6
hXk1
1 Xk2
2 · · · XkN
N i = (−j)k ∂k
φx(t)
∂tk1
1 . . . ∂tkN
N
¯
¯
¯
¯
t=0
(3.4)
where k = k1 + k2 + . . . + kN .
1. Order Moments
Mean µ ≡ hXi (3.5)
2. Order Moments
Variance cii ≡ h(Xi − µi)2
i = hX2
i i − µ2
i (3.6)
Covariance cij ≡ h(Xi − µi)(Xj − µj)i (3.7)
Covariance matrix Σ ≡ h(X − µ)(X − µ)T
i ≡ [cij] (3.8)
3. Order Moments
Often the skewness is used.
Skew(X) ≡
h(Xi − hXii)3
i
h(Xi − hXii)2i
3
2
=
h(Xi − µi)3
i
h(Xi − µi)2i
3
2
(3.9)
All 3. order central moments are zero for gaussian distributions and thus also
the skewness.
4. Order Moments
The kurtosis is (in newer literature) given as
Kurt(X) ≡
h(Xi − µi)4
i
h(Xi − µi)2i2
− 3 (3.10)
Let X ∼ N(µ, Σ). Then
h(Xi − µi)(Xj − µj)(Xk − µk)(Xl − µl)i = cijckl + cilcjk + cikclj (3.11)
and
Kurt(X) = 0 (3.12)
N. Order Moments
Any central moment of a gaussian distribution can (fairly easily) be calculated
with the following method [3] (sometimes known as Wicks theorem).
Let X ∼ N(µ, Σ). Then
7
• Assume k is odd. Then the central k’th order moments are all zero.
• Assume k is even. Then the central k’th order moments are equal to
P
(cijckl . . . cxz). The sum is taken over all different permutations of the k
indices, where it is noted that cij = cji. This gives (k − 1)!/(2k/2−1
(k/2 −
1)!) terms which each is the product of k/2 covariances.
An example is illustrative. The different 4. order central moments of X are found
with the above method to give
h(Xi − µi)4
i = 3c2
ii
h(Xi − µi)3
(Xj − µj)i = 3ciicij
h(Xi − µi)2
(Xj − µj)2
i = ciicjj + 2c2
ij
h(Xi − µi)2
(Xj − µj)(Xk − µk)i = ciicjk + 2cijcik
h(Xi − µi)(Xj − µj)(Xk − µk)(Xl − µl)i = cijckl + cilcjk + cikclj
(3.13)
The above results were found by seeing that the different permutations of the
k=4 indices are (12)(34), (13)(24) and (14)(23). Other permutations are equiv-
alents, such as for instance (32)(14) which is equivalent to (14)(23). When cal-
culating e.g. h(Xi − µi)2
(Xj − µj)(Xk − µk)i, the assignment (1 → i, 2 → i, 3 →
j, 4 → k) gives the terms ciicjk, cijcik and cijcik in the sum.
Calculations with moments
Let b ∈ Rc
, A, B ∈ Mc·d
. Let X and Y be random vectors and f and g vector
functions. Then
hAf(X) + Bg(X) + bi = Ahf(X)i + Bhg(X)i + b (3.14)
hAX + bi = AhXi + b (3.15)
hhY|Xii ≡ E(E(Y|X)) = hYi (3.16)
If Xi and Xj are independent then
hXiXji = hXiihXji (3.17)
8
Chapter 4
Marginalization and
Conditional Distribution
4.1 Marginalization
Marginalization is the operation of integrating out variables of the pdf of a
random vector X. Assume that X is split into two parts (since the order-
ing of the Xi is arbitrary, this corresponds to any division of the variables),
X = [XT
1:cXT
c+1:N ]T
= [X1X2 . . . XcXc+1 . . . XN ]T
. Let the pdf of X be p(x) =
p(x1, . . . , xN ), then :
p(x1, . . . , xc) =
Z
· · ·
Z
Rc+1:N
p(x1, . . . , xN ) dxc+1 . . . xN (4.1)
The nice part about gaussian distributions is that every marginal distribution
of a gaussian distribution is itself a gaussian. More specifically, let X be split
into two parts as above and X ∼ N(µ, Σ), then :
p(x1, . . . , xc) = p(x1:c) = N1:c(µ1:c, Σ1:c) (4.2)
where µ1:c = [µ1, µ2, . . . , µc] and
Σ1:c =






c11 c21 . . . cc1
c12 c22
.
.
.
.
.
.
...
.
.
.
cc1 . . . . . . ccc






In words, the mean and covariance matrix of the marginal distribution is the
same as the corresponding elements of the joint distribution.
9
4.2. CONDITIONAL DISTRIBUTION
4.2 Conditional distribution
As in the previous, let X = [XT
1:cXT
c+1:N ]T
= [X1X2 . . . XcXc+1 . . . XN ]T
be a
division of the variables into two parts. Let X ∼ N(µ, Σ) and use the notation
X = [XT
1:cXT
c+1:N ]T
= [XT
(1)XT
(2)]T
and
µ =
µ
µ(1)
µ(2)
¶
and
Σ =
µ
Σ11 Σ12
Σ21 Σ22
¶
It is found that the conditional distribution p(x(1)|x(2)) is in fact again a gaus-
sian distribution and
X(1)|X(2) ∼ N(µ(1) + Σ12Σ−1
22 (x2 − µ(2)), Σ11 − Σ12Σ−1
22 ΣT
12) (4.3)
10
Chapter 5
Tips and Tricks
5.1 Products
Consider the product Nx(µa, Σa) · Nx(µb, Σb) and note that they both have x
as their ”random variable”. Then
Nx(µa, Σa) · Nx(µb, Σb) = zcNx(µc, Σc) (5.1)
where Σc = (Σ−1
a + Σ−1
b )−1
and µc = Σc(Σ−1
a µa + Σ−1
b µb) and
zc = |2πΣaΣbΣ−1
c |− 1
2 exp
µ
−
1
2
(µa − µb)T
Σ−1
a ΣcΣ−1
b (µa − µb)
¶
= |2π(Σa + Σb)|− 1
2 exp
µ
−
1
2
(µa − µb)T
(Σa + Σb)−1
(µa − µb)
¶ (5.2)
In words, the product of two gaussians is another gaussian (unnormalized). This
can be generalised to a product of K gaussians with distributions Xk ∼ N(µk, Σk).
K
Y
k=1
Nx(µk, Σk) = z̃ · Nx(µ̃, Σ̃) (5.3)
where Σ̃ =
³PK
k=1 Σ−1
k
´−1
and µ̃ = Σ̃
³PK
k=1 Σ−1
k µk
´
=
³PK
k=1 Σ−1
k
´−1 ³PK
k=1 Σ−1
k µk
´
and
z̃ =
|2πΣd|
1
2
QK
k=1 |2πΣk|
1
2
Y
i<j
exp
µ
−
1
2
(µi − µj)T
Bij(µi − µj)
¶
(5.4)
where
11
5.2. GAUSSIAN INTEGRALS
Bij = Σ−1
i
à K
X
k=1
Σ−1
k
!−1
Σ−1
j (5.5)
5.2 Gaussian Integrals
A nice thing about the fact that products of gaussian functions are again a
gaussian function, is that it makes gaussian integrals easier to calculate since
R
Nx(µ̃, Σ̃)dx = 1. Using this with the equations (5.1) and (5.3) of the previous
section gives the following.
Z
Rd
Nx(µa, Σa) · Nx(µb, Σb) dx =
Z
Rd
zcNx(µc, Σc) dx
= zc
(5.6)
Similarly,
Z
Rd
K
Y
k=1
Nx(µk, Σk) dx = z̃ (5.7)
Equation (5.3) can also be used to calculate integrals such as
R
|x|q
(
Q
k Nx(µk, Σk)) dx
or similar by using the same technique as above.
5.3 Useful integrals
Let X ∼ N(µ, Σ) and a ∈ Rd
an arbitrary vector. Then
heaT
x
i ≡
Z
Nx(µ, Σ)eaT
x
dx = eaT
µ+ 1
2 aT
Σa
(5.8)
From this expression, it is possible to find integrals such as
R
exp(xT
Ax +
aT
x) dx. Another useful integral is
hexT
Ax
i ≡
Z
Nx(µ, Σ)exT
Ax
dx = |I − 2ΣA|− 1
2 e− 1
2 µT
(Σ−2A−1
)−1
µ
(5.9)
where A ∈ Md·d
is a non-singular matrix.
12
Bibliography
[1] T.W. Anderson, An introduction to multivariate statistical analysis. Wiley,
1984.
[2] C. M. Bishop, Neural Networks for Pattern Recognition. Oxford University
Press, 1995.
[3] K. Triantafyllopoulos, “On the central moments of the multidimensional
Gaussian distribution,” The Mathematical Scientist, vol. 28, pp. 125–128,
2003.
[4] S. Roweis, “Gaussian Identities,” http://www.cs.toronto.edu/ roweis/notes.html.
[5] J. Larsen, “Gaussian Integrals,” Tech. Rep., http://isp.imm.dtu.dk/staff-
/jlarsen/pubs/frame.htm
[6] www.Wikipedia.org.
[7] www.MathWorld.wolfram.com
[8] P. Kidmose, “Blind Separation of Heavy Tail Signals,” Ph.d. Thesis,
http://isp.imm.dtu.dk/staff/kidmose/pk publications.html
13

Weitere ähnliche Inhalte

Was ist angesagt?

APPLICATION OF DSP IN BIOMEDICAL ENGINEERING
APPLICATION OF DSP IN BIOMEDICAL ENGINEERINGAPPLICATION OF DSP IN BIOMEDICAL ENGINEERING
APPLICATION OF DSP IN BIOMEDICAL ENGINEERINGpirh khan
 
DSP_2018_FOEHU - Lec 07 - IIR Filter Design
DSP_2018_FOEHU - Lec 07 - IIR Filter DesignDSP_2018_FOEHU - Lec 07 - IIR Filter Design
DSP_2018_FOEHU - Lec 07 - IIR Filter DesignAmr E. Mohamed
 
Digital speech processing lecture1
Digital speech processing lecture1Digital speech processing lecture1
Digital speech processing lecture1Samiul Parag
 
Hard versus Soft real time system
Hard versus Soft real time systemHard versus Soft real time system
Hard versus Soft real time systemKamal Acharya
 
DSP_2018_FOEHU - Lec 0 - Course Outlines
DSP_2018_FOEHU - Lec 0 - Course OutlinesDSP_2018_FOEHU - Lec 0 - Course Outlines
DSP_2018_FOEHU - Lec 0 - Course OutlinesAmr E. Mohamed
 
Basics of digital filters
Basics of digital filtersBasics of digital filters
Basics of digital filtersSmile Hossain
 
Hilbert huang transform(hht)
Hilbert huang transform(hht)Hilbert huang transform(hht)
Hilbert huang transform(hht)Puneet Gupta
 
Integration of Sensors & Actuators With Arduino.pptx
Integration of Sensors & Actuators With Arduino.pptxIntegration of Sensors & Actuators With Arduino.pptx
Integration of Sensors & Actuators With Arduino.pptxNShravani1
 
Basics of Digital Filters
Basics of Digital FiltersBasics of Digital Filters
Basics of Digital Filtersop205
 
Speech emotion recognition from audio converted
Speech emotion recognition from audio convertedSpeech emotion recognition from audio converted
Speech emotion recognition from audio converteddataalcott
 
Hand gesture recognition system(FYP REPORT)
Hand gesture recognition system(FYP REPORT)Hand gesture recognition system(FYP REPORT)
Hand gesture recognition system(FYP REPORT)Afnan Rehman
 
Fingerprint Recognition Technique(PPT)
Fingerprint Recognition Technique(PPT)Fingerprint Recognition Technique(PPT)
Fingerprint Recognition Technique(PPT)Sandeep Kumar Panda
 
DSP_FOEHU - Lec 07 - Digital Filters
DSP_FOEHU - Lec 07 - Digital FiltersDSP_FOEHU - Lec 07 - Digital Filters
DSP_FOEHU - Lec 07 - Digital FiltersAmr E. Mohamed
 
Stationary & Non stationary signal.pptx
Stationary & Non stationary signal.pptxStationary & Non stationary signal.pptx
Stationary & Non stationary signal.pptxRCCIITAEIE
 

Was ist angesagt? (20)

Digital filter structures
Digital filter structuresDigital filter structures
Digital filter structures
 
APPLICATION OF DSP IN BIOMEDICAL ENGINEERING
APPLICATION OF DSP IN BIOMEDICAL ENGINEERINGAPPLICATION OF DSP IN BIOMEDICAL ENGINEERING
APPLICATION OF DSP IN BIOMEDICAL ENGINEERING
 
DSP_2018_FOEHU - Lec 07 - IIR Filter Design
DSP_2018_FOEHU - Lec 07 - IIR Filter DesignDSP_2018_FOEHU - Lec 07 - IIR Filter Design
DSP_2018_FOEHU - Lec 07 - IIR Filter Design
 
Adaptive filter
Adaptive filterAdaptive filter
Adaptive filter
 
Digital speech processing lecture1
Digital speech processing lecture1Digital speech processing lecture1
Digital speech processing lecture1
 
Hard versus Soft real time system
Hard versus Soft real time systemHard versus Soft real time system
Hard versus Soft real time system
 
DSP_2018_FOEHU - Lec 0 - Course Outlines
DSP_2018_FOEHU - Lec 0 - Course OutlinesDSP_2018_FOEHU - Lec 0 - Course Outlines
DSP_2018_FOEHU - Lec 0 - Course Outlines
 
Basics of digital filters
Basics of digital filtersBasics of digital filters
Basics of digital filters
 
Hilbert huang transform(hht)
Hilbert huang transform(hht)Hilbert huang transform(hht)
Hilbert huang transform(hht)
 
Dsp lecture vol 7 adaptive filter
Dsp lecture vol 7 adaptive filterDsp lecture vol 7 adaptive filter
Dsp lecture vol 7 adaptive filter
 
Integration of Sensors & Actuators With Arduino.pptx
Integration of Sensors & Actuators With Arduino.pptxIntegration of Sensors & Actuators With Arduino.pptx
Integration of Sensors & Actuators With Arduino.pptx
 
Sampling theorem
Sampling theoremSampling theorem
Sampling theorem
 
Digital Filters Part 2
Digital Filters Part 2Digital Filters Part 2
Digital Filters Part 2
 
Basics of Digital Filters
Basics of Digital FiltersBasics of Digital Filters
Basics of Digital Filters
 
Speech emotion recognition from audio converted
Speech emotion recognition from audio convertedSpeech emotion recognition from audio converted
Speech emotion recognition from audio converted
 
Hand gesture recognition system(FYP REPORT)
Hand gesture recognition system(FYP REPORT)Hand gesture recognition system(FYP REPORT)
Hand gesture recognition system(FYP REPORT)
 
Fingerprint Recognition Technique(PPT)
Fingerprint Recognition Technique(PPT)Fingerprint Recognition Technique(PPT)
Fingerprint Recognition Technique(PPT)
 
DSP_FOEHU - Lec 07 - Digital Filters
DSP_FOEHU - Lec 07 - Digital FiltersDSP_FOEHU - Lec 07 - Digital Filters
DSP_FOEHU - Lec 07 - Digital Filters
 
Adaptive filter
Adaptive filterAdaptive filter
Adaptive filter
 
Stationary & Non stationary signal.pptx
Stationary & Non stationary signal.pptxStationary & Non stationary signal.pptx
Stationary & Non stationary signal.pptx
 

Ähnlich wie The Multivariate Gaussian Probability Distribution

A Note on “   Geraghty contraction type mappings”
A Note on “   Geraghty contraction type mappings”A Note on “   Geraghty contraction type mappings”
A Note on “   Geraghty contraction type mappings”IOSRJM
 
Hierarchical matrices for approximating large covariance matries and computin...
Hierarchical matrices for approximating large covariance matries and computin...Hierarchical matrices for approximating large covariance matries and computin...
Hierarchical matrices for approximating large covariance matries and computin...Alexander Litvinenko
 
A new implementation of k-MLE for mixture modelling of Wishart distributions
A new implementation of k-MLE for mixture modelling of Wishart distributionsA new implementation of k-MLE for mixture modelling of Wishart distributions
A new implementation of k-MLE for mixture modelling of Wishart distributionsFrank Nielsen
 
Applications of Differential Calculus in real life
Applications of Differential Calculus in real life Applications of Differential Calculus in real life
Applications of Differential Calculus in real life OlooPundit
 
Multivariate Gaussin, Rayleigh & Rician distributions
Multivariate Gaussin, Rayleigh & Rician distributionsMultivariate Gaussin, Rayleigh & Rician distributions
Multivariate Gaussin, Rayleigh & Rician distributionsHAmindavarLectures
 
Principal Component Analysis
Principal Component AnalysisPrincipal Component Analysis
Principal Component AnalysisSumit Singh
 
Computing f-Divergences and Distances of\\ High-Dimensional Probability Densi...
Computing f-Divergences and Distances of\\ High-Dimensional Probability Densi...Computing f-Divergences and Distances of\\ High-Dimensional Probability Densi...
Computing f-Divergences and Distances of\\ High-Dimensional Probability Densi...Alexander Litvinenko
 
Moment-Generating Functions and Reproductive Properties of Distributions
Moment-Generating Functions and Reproductive Properties of DistributionsMoment-Generating Functions and Reproductive Properties of Distributions
Moment-Generating Functions and Reproductive Properties of DistributionsIJSRED
 
IJSRED-V2I5P56
IJSRED-V2I5P56IJSRED-V2I5P56
IJSRED-V2I5P56IJSRED
 
Linear models for classification
Linear models for classificationLinear models for classification
Linear models for classificationSung Yub Kim
 
multivariate normal distribution.pdf
multivariate normal distribution.pdfmultivariate normal distribution.pdf
multivariate normal distribution.pdfrishumaurya10
 
Fixed Point Theorem in Fuzzy Metric Space Using (CLRg) Property
Fixed Point Theorem in Fuzzy Metric Space Using (CLRg) PropertyFixed Point Theorem in Fuzzy Metric Space Using (CLRg) Property
Fixed Point Theorem in Fuzzy Metric Space Using (CLRg) Propertyinventionjournals
 
Fixed points theorem on a pair of random generalized non linear contractions
Fixed points theorem on a pair of random generalized non linear contractionsFixed points theorem on a pair of random generalized non linear contractions
Fixed points theorem on a pair of random generalized non linear contractionsAlexander Decker
 
Multivriada ppt ms
Multivriada   ppt msMultivriada   ppt ms
Multivriada ppt msFaeco Bot
 
Extension of Some Common Fixed Point Theorems using Compatible Mappings in Fu...
Extension of Some Common Fixed Point Theorems using Compatible Mappings in Fu...Extension of Some Common Fixed Point Theorems using Compatible Mappings in Fu...
Extension of Some Common Fixed Point Theorems using Compatible Mappings in Fu...IJRES Journal
 
Clustering Random Walk Time Series
Clustering Random Walk Time SeriesClustering Random Walk Time Series
Clustering Random Walk Time SeriesGautier Marti
 

Ähnlich wie The Multivariate Gaussian Probability Distribution (20)

A Note on “   Geraghty contraction type mappings”
A Note on “   Geraghty contraction type mappings”A Note on “   Geraghty contraction type mappings”
A Note on “   Geraghty contraction type mappings”
 
Section3 stochastic
Section3 stochasticSection3 stochastic
Section3 stochastic
 
Bayes gauss
Bayes gaussBayes gauss
Bayes gauss
 
Section5 stochastic
Section5 stochasticSection5 stochastic
Section5 stochastic
 
Hierarchical matrices for approximating large covariance matries and computin...
Hierarchical matrices for approximating large covariance matries and computin...Hierarchical matrices for approximating large covariance matries and computin...
Hierarchical matrices for approximating large covariance matries and computin...
 
A new implementation of k-MLE for mixture modelling of Wishart distributions
A new implementation of k-MLE for mixture modelling of Wishart distributionsA new implementation of k-MLE for mixture modelling of Wishart distributions
A new implementation of k-MLE for mixture modelling of Wishart distributions
 
Applications of Differential Calculus in real life
Applications of Differential Calculus in real life Applications of Differential Calculus in real life
Applications of Differential Calculus in real life
 
Metodo gauss_newton.pdf
Metodo gauss_newton.pdfMetodo gauss_newton.pdf
Metodo gauss_newton.pdf
 
Multivariate Gaussin, Rayleigh & Rician distributions
Multivariate Gaussin, Rayleigh & Rician distributionsMultivariate Gaussin, Rayleigh & Rician distributions
Multivariate Gaussin, Rayleigh & Rician distributions
 
Principal Component Analysis
Principal Component AnalysisPrincipal Component Analysis
Principal Component Analysis
 
Computing f-Divergences and Distances of\\ High-Dimensional Probability Densi...
Computing f-Divergences and Distances of\\ High-Dimensional Probability Densi...Computing f-Divergences and Distances of\\ High-Dimensional Probability Densi...
Computing f-Divergences and Distances of\\ High-Dimensional Probability Densi...
 
Moment-Generating Functions and Reproductive Properties of Distributions
Moment-Generating Functions and Reproductive Properties of DistributionsMoment-Generating Functions and Reproductive Properties of Distributions
Moment-Generating Functions and Reproductive Properties of Distributions
 
IJSRED-V2I5P56
IJSRED-V2I5P56IJSRED-V2I5P56
IJSRED-V2I5P56
 
Linear models for classification
Linear models for classificationLinear models for classification
Linear models for classification
 
multivariate normal distribution.pdf
multivariate normal distribution.pdfmultivariate normal distribution.pdf
multivariate normal distribution.pdf
 
Fixed Point Theorem in Fuzzy Metric Space Using (CLRg) Property
Fixed Point Theorem in Fuzzy Metric Space Using (CLRg) PropertyFixed Point Theorem in Fuzzy Metric Space Using (CLRg) Property
Fixed Point Theorem in Fuzzy Metric Space Using (CLRg) Property
 
Fixed points theorem on a pair of random generalized non linear contractions
Fixed points theorem on a pair of random generalized non linear contractionsFixed points theorem on a pair of random generalized non linear contractions
Fixed points theorem on a pair of random generalized non linear contractions
 
Multivriada ppt ms
Multivriada   ppt msMultivriada   ppt ms
Multivriada ppt ms
 
Extension of Some Common Fixed Point Theorems using Compatible Mappings in Fu...
Extension of Some Common Fixed Point Theorems using Compatible Mappings in Fu...Extension of Some Common Fixed Point Theorems using Compatible Mappings in Fu...
Extension of Some Common Fixed Point Theorems using Compatible Mappings in Fu...
 
Clustering Random Walk Time Series
Clustering Random Walk Time SeriesClustering Random Walk Time Series
Clustering Random Walk Time Series
 

Kürzlich hochgeladen

Davis plaque method.pptx recombinant DNA technology
Davis plaque method.pptx recombinant DNA technologyDavis plaque method.pptx recombinant DNA technology
Davis plaque method.pptx recombinant DNA technologycaarthichand2003
 
(9818099198) Call Girls In Noida Sector 14 (NOIDA ESCORTS)
(9818099198) Call Girls In Noida Sector 14 (NOIDA ESCORTS)(9818099198) Call Girls In Noida Sector 14 (NOIDA ESCORTS)
(9818099198) Call Girls In Noida Sector 14 (NOIDA ESCORTS)riyaescorts54
 
Neurodevelopmental disorders according to the dsm 5 tr
Neurodevelopmental disorders according to the dsm 5 trNeurodevelopmental disorders according to the dsm 5 tr
Neurodevelopmental disorders according to the dsm 5 trssuser06f238
 
preservation, maintanence and improvement of industrial organism.pptx
preservation, maintanence and improvement of industrial organism.pptxpreservation, maintanence and improvement of industrial organism.pptx
preservation, maintanence and improvement of industrial organism.pptxnoordubaliya2003
 
STOPPED FLOW METHOD & APPLICATION MURUGAVENI B.pptx
STOPPED FLOW METHOD & APPLICATION MURUGAVENI B.pptxSTOPPED FLOW METHOD & APPLICATION MURUGAVENI B.pptx
STOPPED FLOW METHOD & APPLICATION MURUGAVENI B.pptxMurugaveni B
 
Microteaching on terms used in filtration .Pharmaceutical Engineering
Microteaching on terms used in filtration .Pharmaceutical EngineeringMicroteaching on terms used in filtration .Pharmaceutical Engineering
Microteaching on terms used in filtration .Pharmaceutical EngineeringPrajakta Shinde
 
Speech, hearing, noise, intelligibility.pptx
Speech, hearing, noise, intelligibility.pptxSpeech, hearing, noise, intelligibility.pptx
Speech, hearing, noise, intelligibility.pptxpriyankatabhane
 
Radiation physics in Dental Radiology...
Radiation physics in Dental Radiology...Radiation physics in Dental Radiology...
Radiation physics in Dental Radiology...navyadasi1992
 
REVISTA DE BIOLOGIA E CIÊNCIAS DA TERRA ISSN 1519-5228 - Artigo_Bioterra_V24_...
REVISTA DE BIOLOGIA E CIÊNCIAS DA TERRA ISSN 1519-5228 - Artigo_Bioterra_V24_...REVISTA DE BIOLOGIA E CIÊNCIAS DA TERRA ISSN 1519-5228 - Artigo_Bioterra_V24_...
REVISTA DE BIOLOGIA E CIÊNCIAS DA TERRA ISSN 1519-5228 - Artigo_Bioterra_V24_...Universidade Federal de Sergipe - UFS
 
Call Girls In Nihal Vihar Delhi ❤️8860477959 Looking Escorts In 24/7 Delhi NCR
Call Girls In Nihal Vihar Delhi ❤️8860477959 Looking Escorts In 24/7 Delhi NCRCall Girls In Nihal Vihar Delhi ❤️8860477959 Looking Escorts In 24/7 Delhi NCR
Call Girls In Nihal Vihar Delhi ❤️8860477959 Looking Escorts In 24/7 Delhi NCRlizamodels9
 
User Guide: Pulsar™ Weather Station (Columbia Weather Systems)
User Guide: Pulsar™ Weather Station (Columbia Weather Systems)User Guide: Pulsar™ Weather Station (Columbia Weather Systems)
User Guide: Pulsar™ Weather Station (Columbia Weather Systems)Columbia Weather Systems
 
Bioteknologi kelas 10 kumer smapsa .pptx
Bioteknologi kelas 10 kumer smapsa .pptxBioteknologi kelas 10 kumer smapsa .pptx
Bioteknologi kelas 10 kumer smapsa .pptx023NiWayanAnggiSriWa
 
User Guide: Capricorn FLX™ Weather Station
User Guide: Capricorn FLX™ Weather StationUser Guide: Capricorn FLX™ Weather Station
User Guide: Capricorn FLX™ Weather StationColumbia Weather Systems
 
ALL ABOUT MIXTURES IN GRADE 7 CLASS PPTX
ALL ABOUT MIXTURES IN GRADE 7 CLASS PPTXALL ABOUT MIXTURES IN GRADE 7 CLASS PPTX
ALL ABOUT MIXTURES IN GRADE 7 CLASS PPTXDole Philippines School
 
Pests of soyabean_Binomics_IdentificationDr.UPR.pdf
Pests of soyabean_Binomics_IdentificationDr.UPR.pdfPests of soyabean_Binomics_IdentificationDr.UPR.pdf
Pests of soyabean_Binomics_IdentificationDr.UPR.pdfPirithiRaju
 
The dark energy paradox leads to a new structure of spacetime.pptx
The dark energy paradox leads to a new structure of spacetime.pptxThe dark energy paradox leads to a new structure of spacetime.pptx
The dark energy paradox leads to a new structure of spacetime.pptxEran Akiva Sinbar
 
basic entomology with insect anatomy and taxonomy
basic entomology with insect anatomy and taxonomybasic entomology with insect anatomy and taxonomy
basic entomology with insect anatomy and taxonomyDrAnita Sharma
 
Carbon Dioxide Capture and Storage (CSS)
Carbon Dioxide Capture and Storage (CSS)Carbon Dioxide Capture and Storage (CSS)
Carbon Dioxide Capture and Storage (CSS)Tamer Koksalan, PhD
 
Pests of Blackgram, greengram, cowpea_Dr.UPR.pdf
Pests of Blackgram, greengram, cowpea_Dr.UPR.pdfPests of Blackgram, greengram, cowpea_Dr.UPR.pdf
Pests of Blackgram, greengram, cowpea_Dr.UPR.pdfPirithiRaju
 
GenBio2 - Lesson 1 - Introduction to Genetics.pptx
GenBio2 - Lesson 1 - Introduction to Genetics.pptxGenBio2 - Lesson 1 - Introduction to Genetics.pptx
GenBio2 - Lesson 1 - Introduction to Genetics.pptxBerniceCayabyab1
 

Kürzlich hochgeladen (20)

Davis plaque method.pptx recombinant DNA technology
Davis plaque method.pptx recombinant DNA technologyDavis plaque method.pptx recombinant DNA technology
Davis plaque method.pptx recombinant DNA technology
 
(9818099198) Call Girls In Noida Sector 14 (NOIDA ESCORTS)
(9818099198) Call Girls In Noida Sector 14 (NOIDA ESCORTS)(9818099198) Call Girls In Noida Sector 14 (NOIDA ESCORTS)
(9818099198) Call Girls In Noida Sector 14 (NOIDA ESCORTS)
 
Neurodevelopmental disorders according to the dsm 5 tr
Neurodevelopmental disorders according to the dsm 5 trNeurodevelopmental disorders according to the dsm 5 tr
Neurodevelopmental disorders according to the dsm 5 tr
 
preservation, maintanence and improvement of industrial organism.pptx
preservation, maintanence and improvement of industrial organism.pptxpreservation, maintanence and improvement of industrial organism.pptx
preservation, maintanence and improvement of industrial organism.pptx
 
STOPPED FLOW METHOD & APPLICATION MURUGAVENI B.pptx
STOPPED FLOW METHOD & APPLICATION MURUGAVENI B.pptxSTOPPED FLOW METHOD & APPLICATION MURUGAVENI B.pptx
STOPPED FLOW METHOD & APPLICATION MURUGAVENI B.pptx
 
Microteaching on terms used in filtration .Pharmaceutical Engineering
Microteaching on terms used in filtration .Pharmaceutical EngineeringMicroteaching on terms used in filtration .Pharmaceutical Engineering
Microteaching on terms used in filtration .Pharmaceutical Engineering
 
Speech, hearing, noise, intelligibility.pptx
Speech, hearing, noise, intelligibility.pptxSpeech, hearing, noise, intelligibility.pptx
Speech, hearing, noise, intelligibility.pptx
 
Radiation physics in Dental Radiology...
Radiation physics in Dental Radiology...Radiation physics in Dental Radiology...
Radiation physics in Dental Radiology...
 
REVISTA DE BIOLOGIA E CIÊNCIAS DA TERRA ISSN 1519-5228 - Artigo_Bioterra_V24_...
REVISTA DE BIOLOGIA E CIÊNCIAS DA TERRA ISSN 1519-5228 - Artigo_Bioterra_V24_...REVISTA DE BIOLOGIA E CIÊNCIAS DA TERRA ISSN 1519-5228 - Artigo_Bioterra_V24_...
REVISTA DE BIOLOGIA E CIÊNCIAS DA TERRA ISSN 1519-5228 - Artigo_Bioterra_V24_...
 
Call Girls In Nihal Vihar Delhi ❤️8860477959 Looking Escorts In 24/7 Delhi NCR
Call Girls In Nihal Vihar Delhi ❤️8860477959 Looking Escorts In 24/7 Delhi NCRCall Girls In Nihal Vihar Delhi ❤️8860477959 Looking Escorts In 24/7 Delhi NCR
Call Girls In Nihal Vihar Delhi ❤️8860477959 Looking Escorts In 24/7 Delhi NCR
 
User Guide: Pulsar™ Weather Station (Columbia Weather Systems)
User Guide: Pulsar™ Weather Station (Columbia Weather Systems)User Guide: Pulsar™ Weather Station (Columbia Weather Systems)
User Guide: Pulsar™ Weather Station (Columbia Weather Systems)
 
Bioteknologi kelas 10 kumer smapsa .pptx
Bioteknologi kelas 10 kumer smapsa .pptxBioteknologi kelas 10 kumer smapsa .pptx
Bioteknologi kelas 10 kumer smapsa .pptx
 
User Guide: Capricorn FLX™ Weather Station
User Guide: Capricorn FLX™ Weather StationUser Guide: Capricorn FLX™ Weather Station
User Guide: Capricorn FLX™ Weather Station
 
ALL ABOUT MIXTURES IN GRADE 7 CLASS PPTX
ALL ABOUT MIXTURES IN GRADE 7 CLASS PPTXALL ABOUT MIXTURES IN GRADE 7 CLASS PPTX
ALL ABOUT MIXTURES IN GRADE 7 CLASS PPTX
 
Pests of soyabean_Binomics_IdentificationDr.UPR.pdf
Pests of soyabean_Binomics_IdentificationDr.UPR.pdfPests of soyabean_Binomics_IdentificationDr.UPR.pdf
Pests of soyabean_Binomics_IdentificationDr.UPR.pdf
 
The dark energy paradox leads to a new structure of spacetime.pptx
The dark energy paradox leads to a new structure of spacetime.pptxThe dark energy paradox leads to a new structure of spacetime.pptx
The dark energy paradox leads to a new structure of spacetime.pptx
 
basic entomology with insect anatomy and taxonomy
basic entomology with insect anatomy and taxonomybasic entomology with insect anatomy and taxonomy
basic entomology with insect anatomy and taxonomy
 
Carbon Dioxide Capture and Storage (CSS)
Carbon Dioxide Capture and Storage (CSS)Carbon Dioxide Capture and Storage (CSS)
Carbon Dioxide Capture and Storage (CSS)
 
Pests of Blackgram, greengram, cowpea_Dr.UPR.pdf
Pests of Blackgram, greengram, cowpea_Dr.UPR.pdfPests of Blackgram, greengram, cowpea_Dr.UPR.pdf
Pests of Blackgram, greengram, cowpea_Dr.UPR.pdf
 
GenBio2 - Lesson 1 - Introduction to Genetics.pptx
GenBio2 - Lesson 1 - Introduction to Genetics.pptxGenBio2 - Lesson 1 - Introduction to Genetics.pptx
GenBio2 - Lesson 1 - Introduction to Genetics.pptx
 

The Multivariate Gaussian Probability Distribution

  • 1. The Multivariate Gaussian Probability Distribution Peter Ahrendt IMM, Technical University of Denmark mail : pa@imm.dtu.dk, web : www.imm.dtu.dk/∼pa January 7, 2005
  • 2. Contents 1 Definition 2 2 Functions of Gaussian Variables 4 3 Characteristic function and Moments 6 4 Marginalization and Conditional Distribution 9 4.1 Marginalization . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9 4.2 Conditional distribution . . . . . . . . . . . . . . . . . . . . . . . 10 5 Tips and Tricks 11 5.1 Products . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11 5.2 Gaussian Integrals . . . . . . . . . . . . . . . . . . . . . . . . . . 12 5.3 Useful integrals . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12 1
  • 3. Chapter 1 Definition The definition of a multivariate gaussian probability distribution can be stated in several equivalent ways. A random vector X = [X1X2 . . . XN ] can be said to belong to a multivariate gaussian distribution if one of the following statements is true. • Any linear combination Y = a1X1 + a2X2 + . . . + aN XN , ai ∈ R is a (univariate) gaussian distribution. • There exists a random vector Z = [Z1, . . . , ZM ] with components that are independent and standard normal distributed, a vector µ = [µ1, . . . , µN ] and an N-by-M matrix A such that X = AZ + µ. • There exists a vector µ and a symmetric, positive semi-definite matrix Γ such that the characteristic function of X can be written φx(t) ≡ heitT X i = eiµT t− 1 2 tT Γt . Under the assumption that the covariance matrix Σ is non-singular, the prob- ability density function (pdf) can be written as : Nx(µ, Σ) = 1 p (2π)d|Σ| exp µ − 1 2 (x − µ)T Σ−1 (x − µ) ¶ = |2πΣ|− 1 2 exp µ − 1 2 (x − µ)T Σ−1 (x − µ) ¶ (1.1) Then µ is the mean value, Σ is the covariance matrix and | · | denote the determinant. Note that it is possible to have multivariate gaussian distributions with singular covariance matrix and then the above expression cannot be used for the pdf. In the following, however, non-singular covariance matrices will be assumed. In the limit of one dimension, the familiar expression of the univariate gaussian pdf is found. 2
  • 4. Nx(µ, σ2 ) = 1 √ 2πσ2 exp µ − (x − µ)2 2σ2 ¶ = 1 √ 2πσ2 exp µ − 1 2 (x − µ)σ−2 (x − µ) ¶ (1.2) Neither of them have a closed-form expression for the cumulative density func- tion. Symmetries It is noted that in the one-dimensional case there is a symmetry in the pdf. Nx(µ, σ2 ) which is centered on µ. This can be seen by looking at ”contour lines”, i.e. setting the exponent −(x−µ)2 2σ2 = c. It is seen that σ determines the width of the distribution. In the multivariate case, it is similarly useful to look at −1 2 (x − µ)T Σ−1 (x − µ) = c. This is a quadratic form and geometrically the contour curves (for fixed c) are hy- perellipsoids. In 2D, this is normal ellipsoids with the form (x−x0 a )2 +(y−y0 b )2 = r2 , which gives symmetries along the principal axes. Similarly, the hyperellip- soids show symmetries along their principal axes. Notation: If a random variable X has a gaussian distribution, it is written as X∼ N(µ, Σ). The probability density function of this variable is then given by Nx(µ, Σ). 3
  • 5. Chapter 2 Functions of Gaussian Variables Linear transformation and addition of variables Let A, B ∈ Mc·d and c ∈ Rd . Let X ∼ N(µx, Σx) and Y ∼ N(µy, Σy) be independent variables. Then Z = AX + BY + c ∼ N(Aµx + Bµy + c, AΣxAT + BΣyBT ) (2.1) Transform to standard normal variables Let X ∼ N(µ, Σ). Then Z = Σ− 1 2 (X − µ) ∼ N(0, 1) (2.2) Note, that by Σ− 1 2 is actually meant a unique matrix, although in general matrices with fractional exponents are not. The matrix that is meant can be found from the diagonalisation into Σ = UΛUT = (UΛ 1 2 )(UΛ 1 2 )T where Λ is the diagonal matrix with the eigenvalues of Σ and U is the matrix with the eigenvectors. Then Σ− 1 2 = (UΛ 1 2 )−1 = Λ− 1 2 U−1 . In the one-dimensional case, this corresponds to the transformation of X ∼ N(µ, σ2 ) into Y = σ−1 (X − µ) ∼ N(0, 1). Addition Let Xi ∼ N(µi, Σi), i ∈ 1, ..., N be independent variables. Then N X i Xi ∼ N( N X i µi, N X i Σi) (2.3) Note: This is a direct implication of equation (2.1). 4
  • 6. Quadratic Let Xi ∼ N(0, 1), i ∈ 1, ..., N be independent variables. Then N X i X2 i ∼ χ2 n (2.4) Alternatively let X ∼ N(µx, Σx). Then Z = (X − µ)T Σ−1 (X − µ) ∼ χ2 n (2.5) This is, however, the same thing since Z= X̃ T X̃ = PN i X̃i 2 , where X̃i are the decorrelated components (see eqn. (2.2)). 5
  • 7. Chapter 3 Characteristic function and Moments The characteristic function of the univariate gaussian distribution is given by φx(t) ≡ heitX i = eitµ−σ2 t2 /2 . The generalization to multivariate gaussian distri- butions is φx(t) ≡ heitT X i = eiµT t− 1 2 tT Σt (3.1) The pdf p(x) is related to the characteristic function. p(x) = 1 (2π)d Z Rd φx(t) e−itT x dt (3.2) It is seen that the characteristic function is the inverse Fourier transform of the pdf. Moments of a pdf are generally defined as : hXk1 1 Xk2 2 · · · XkN N i ≡ Z Rd xk1 1 xk2 2 · · · xkN N p(x) dx (3.3) where hXk1 1 Xk2 2 · · · XkN N i is the k’th order moment, k = [k1, k2, . . . , kN ] (ki ∈ N) and k = k1 + k2 + . . . + kN . A well-known example is the first order moment, called the mean value µi (of variable Xi) - or the mean µ ≡ [µ1µ2 . . . µN ] of the whole random vector X. The k’th order central moment is defined as above, but with Xi replaced by Xi − µi in equation (3.3). An example is the second order central moment, called the variance, which is given by h(Xi − µi)2 i. Any moment (that exists) can be found from the characteristic function [8]: 6
  • 8. hXk1 1 Xk2 2 · · · XkN N i = (−j)k ∂k φx(t) ∂tk1 1 . . . ∂tkN N ¯ ¯ ¯ ¯ t=0 (3.4) where k = k1 + k2 + . . . + kN . 1. Order Moments Mean µ ≡ hXi (3.5) 2. Order Moments Variance cii ≡ h(Xi − µi)2 i = hX2 i i − µ2 i (3.6) Covariance cij ≡ h(Xi − µi)(Xj − µj)i (3.7) Covariance matrix Σ ≡ h(X − µ)(X − µ)T i ≡ [cij] (3.8) 3. Order Moments Often the skewness is used. Skew(X) ≡ h(Xi − hXii)3 i h(Xi − hXii)2i 3 2 = h(Xi − µi)3 i h(Xi − µi)2i 3 2 (3.9) All 3. order central moments are zero for gaussian distributions and thus also the skewness. 4. Order Moments The kurtosis is (in newer literature) given as Kurt(X) ≡ h(Xi − µi)4 i h(Xi − µi)2i2 − 3 (3.10) Let X ∼ N(µ, Σ). Then h(Xi − µi)(Xj − µj)(Xk − µk)(Xl − µl)i = cijckl + cilcjk + cikclj (3.11) and Kurt(X) = 0 (3.12) N. Order Moments Any central moment of a gaussian distribution can (fairly easily) be calculated with the following method [3] (sometimes known as Wicks theorem). Let X ∼ N(µ, Σ). Then 7
  • 9. • Assume k is odd. Then the central k’th order moments are all zero. • Assume k is even. Then the central k’th order moments are equal to P (cijckl . . . cxz). The sum is taken over all different permutations of the k indices, where it is noted that cij = cji. This gives (k − 1)!/(2k/2−1 (k/2 − 1)!) terms which each is the product of k/2 covariances. An example is illustrative. The different 4. order central moments of X are found with the above method to give h(Xi − µi)4 i = 3c2 ii h(Xi − µi)3 (Xj − µj)i = 3ciicij h(Xi − µi)2 (Xj − µj)2 i = ciicjj + 2c2 ij h(Xi − µi)2 (Xj − µj)(Xk − µk)i = ciicjk + 2cijcik h(Xi − µi)(Xj − µj)(Xk − µk)(Xl − µl)i = cijckl + cilcjk + cikclj (3.13) The above results were found by seeing that the different permutations of the k=4 indices are (12)(34), (13)(24) and (14)(23). Other permutations are equiv- alents, such as for instance (32)(14) which is equivalent to (14)(23). When cal- culating e.g. h(Xi − µi)2 (Xj − µj)(Xk − µk)i, the assignment (1 → i, 2 → i, 3 → j, 4 → k) gives the terms ciicjk, cijcik and cijcik in the sum. Calculations with moments Let b ∈ Rc , A, B ∈ Mc·d . Let X and Y be random vectors and f and g vector functions. Then hAf(X) + Bg(X) + bi = Ahf(X)i + Bhg(X)i + b (3.14) hAX + bi = AhXi + b (3.15) hhY|Xii ≡ E(E(Y|X)) = hYi (3.16) If Xi and Xj are independent then hXiXji = hXiihXji (3.17) 8
  • 10. Chapter 4 Marginalization and Conditional Distribution 4.1 Marginalization Marginalization is the operation of integrating out variables of the pdf of a random vector X. Assume that X is split into two parts (since the order- ing of the Xi is arbitrary, this corresponds to any division of the variables), X = [XT 1:cXT c+1:N ]T = [X1X2 . . . XcXc+1 . . . XN ]T . Let the pdf of X be p(x) = p(x1, . . . , xN ), then : p(x1, . . . , xc) = Z · · · Z Rc+1:N p(x1, . . . , xN ) dxc+1 . . . xN (4.1) The nice part about gaussian distributions is that every marginal distribution of a gaussian distribution is itself a gaussian. More specifically, let X be split into two parts as above and X ∼ N(µ, Σ), then : p(x1, . . . , xc) = p(x1:c) = N1:c(µ1:c, Σ1:c) (4.2) where µ1:c = [µ1, µ2, . . . , µc] and Σ1:c =       c11 c21 . . . cc1 c12 c22 . . . . . . ... . . . cc1 . . . . . . ccc       In words, the mean and covariance matrix of the marginal distribution is the same as the corresponding elements of the joint distribution. 9
  • 11. 4.2. CONDITIONAL DISTRIBUTION 4.2 Conditional distribution As in the previous, let X = [XT 1:cXT c+1:N ]T = [X1X2 . . . XcXc+1 . . . XN ]T be a division of the variables into two parts. Let X ∼ N(µ, Σ) and use the notation X = [XT 1:cXT c+1:N ]T = [XT (1)XT (2)]T and µ = µ µ(1) µ(2) ¶ and Σ = µ Σ11 Σ12 Σ21 Σ22 ¶ It is found that the conditional distribution p(x(1)|x(2)) is in fact again a gaus- sian distribution and X(1)|X(2) ∼ N(µ(1) + Σ12Σ−1 22 (x2 − µ(2)), Σ11 − Σ12Σ−1 22 ΣT 12) (4.3) 10
  • 12. Chapter 5 Tips and Tricks 5.1 Products Consider the product Nx(µa, Σa) · Nx(µb, Σb) and note that they both have x as their ”random variable”. Then Nx(µa, Σa) · Nx(µb, Σb) = zcNx(µc, Σc) (5.1) where Σc = (Σ−1 a + Σ−1 b )−1 and µc = Σc(Σ−1 a µa + Σ−1 b µb) and zc = |2πΣaΣbΣ−1 c |− 1 2 exp µ − 1 2 (µa − µb)T Σ−1 a ΣcΣ−1 b (µa − µb) ¶ = |2π(Σa + Σb)|− 1 2 exp µ − 1 2 (µa − µb)T (Σa + Σb)−1 (µa − µb) ¶ (5.2) In words, the product of two gaussians is another gaussian (unnormalized). This can be generalised to a product of K gaussians with distributions Xk ∼ N(µk, Σk). K Y k=1 Nx(µk, Σk) = z̃ · Nx(µ̃, Σ̃) (5.3) where Σ̃ = ³PK k=1 Σ−1 k ´−1 and µ̃ = Σ̃ ³PK k=1 Σ−1 k µk ´ = ³PK k=1 Σ−1 k ´−1 ³PK k=1 Σ−1 k µk ´ and z̃ = |2πΣd| 1 2 QK k=1 |2πΣk| 1 2 Y i<j exp µ − 1 2 (µi − µj)T Bij(µi − µj) ¶ (5.4) where 11
  • 13. 5.2. GAUSSIAN INTEGRALS Bij = Σ−1 i à K X k=1 Σ−1 k !−1 Σ−1 j (5.5) 5.2 Gaussian Integrals A nice thing about the fact that products of gaussian functions are again a gaussian function, is that it makes gaussian integrals easier to calculate since R Nx(µ̃, Σ̃)dx = 1. Using this with the equations (5.1) and (5.3) of the previous section gives the following. Z Rd Nx(µa, Σa) · Nx(µb, Σb) dx = Z Rd zcNx(µc, Σc) dx = zc (5.6) Similarly, Z Rd K Y k=1 Nx(µk, Σk) dx = z̃ (5.7) Equation (5.3) can also be used to calculate integrals such as R |x|q ( Q k Nx(µk, Σk)) dx or similar by using the same technique as above. 5.3 Useful integrals Let X ∼ N(µ, Σ) and a ∈ Rd an arbitrary vector. Then heaT x i ≡ Z Nx(µ, Σ)eaT x dx = eaT µ+ 1 2 aT Σa (5.8) From this expression, it is possible to find integrals such as R exp(xT Ax + aT x) dx. Another useful integral is hexT Ax i ≡ Z Nx(µ, Σ)exT Ax dx = |I − 2ΣA|− 1 2 e− 1 2 µT (Σ−2A−1 )−1 µ (5.9) where A ∈ Md·d is a non-singular matrix. 12
  • 14. Bibliography [1] T.W. Anderson, An introduction to multivariate statistical analysis. Wiley, 1984. [2] C. M. Bishop, Neural Networks for Pattern Recognition. Oxford University Press, 1995. [3] K. Triantafyllopoulos, “On the central moments of the multidimensional Gaussian distribution,” The Mathematical Scientist, vol. 28, pp. 125–128, 2003. [4] S. Roweis, “Gaussian Identities,” http://www.cs.toronto.edu/ roweis/notes.html. [5] J. Larsen, “Gaussian Integrals,” Tech. Rep., http://isp.imm.dtu.dk/staff- /jlarsen/pubs/frame.htm [6] www.Wikipedia.org. [7] www.MathWorld.wolfram.com [8] P. Kidmose, “Blind Separation of Heavy Tail Signals,” Ph.d. Thesis, http://isp.imm.dtu.dk/staff/kidmose/pk publications.html 13