SlideShare ist ein Scribd-Unternehmen logo
1 von 22
Downloaden Sie, um offline zu lesen
An Introduction to the
 Bayesian Approach

     J Guzmán, PhD
     15 August 2011
Bayesian Evolution
Bayesian: one who asks you
what you think before a study in
order to tell you what you think
afterwards

             Adapted from:
    S Senn (1997). Statistical Issues in
        Drug Development. Wiley
Rev. Thomas Bayes


English Theologian and
     Mathematician

   ca. 1700 – 1761
Bayesian Methods
•  1763 – Bayes’ article on inverse probability
•  Laplace extended Bayesian ideas in different
   scientific areas in Théorie Analytique des
   Probabilités [1812]
•  Both Laplace & Gauss used the inverse method
•  1st three quarters of 20th Century dominated by
   frequentist methods
•  Last quarter of 20th Century – resurgence of
   Bayesian methods [computational advances]
•  21st Century – Bayesian Century [Lindley]
Pierre-Simon Laplace


French Mathematician

    1749 – 1827
Karl Friedrich Gauss


“Prince of Mathematics”

     1777 – 1855

Used inverse probability
Bayesian Methods
•  Key components: prior, likelihood function,
   posterior, and predictive distribution
•  Suppose a study is carried out to compare new
   and standard teaching methods
•  Ho: Methods are equally effective
•  HA: New method increases grades by 20%
•  A Bayesian presents the probability that new &
   standard methods are equally effective, given
   the results of the experiment at hand:
             P(Ho | data)
Bayesian Methods
•  Data – observed data from experiment

•  Find the probability that the new method is at least 20%
   more effective than the standard, given the results of the
   experiment [Posterior Probability]

•  Another conclusion could be the probability distribution
   for the outcome of interest for the next student

•  Predictive Probabilities – refer to future observations on
   individuals or on set of individuals
Bayes’ Theorem
•  Basic tool of Bayesian analysis
•  Provide the means by which we learn from
   data
•  Given prior state of knowledge, it tells how
   to update belief based upon observations:
                       ∝



    P(H | data) = P(H) · P(data | H) / P(data)
                   €



                α P(H) · P(data | H)
               ∝

   α means “is proportonal to”
•  Bayes’ theorem can be re-expressed in
           €



   odds terms: let data ≡ y
Bayes’ Theorem
Bayes’ Theorem
•  Can also consider posterior probability of
   any measure θ:
       P(θ | data) α P(θ) · P( data | θ)
•  Bayes’ theorem states that the posterior
   probability of any measure θ, is
   proportional to the information on θ
   external to the experiment times the
   likelihood function evaluated at θ:
       Prior · likelihood → posterior
Prior
•  Prior information about θ assessed as a
   probability distribution on θ
•  Distribution on θ depends on the assessor: it is
   subjective
•  A subjective probability can be calculated any
   time a person has an opinion
•  Diffuse prior - when a person’ s opinion on θ
   includes a broad range of possibilities & all
   values are thought to be roughly equally
   probable
Prior
•     Conjugate prior – if the posterior distribution
      has same shape as the prior distribution,
      regardless of the observed sample values
•     Examples:
     1.  Beta prior & binomial likelihood yield a beta posterior
     2.  Normal prior & normal likelihood yield a normal
         posterior
     3.  Gamma prior & Poisson likelihood yield a gamma
         posterior
Community of Priors
•  Expressing a range of reasonable opinions
•  Reference – represents minimal prior
   information
•  Expertise – formalizes opinion of well-
   informed experts
•  Skeptical – downgrades superiority of new
   method
•  Enthusiastic – counterbalance of skeptical
Likelihood Function
                      P(data | θ)
•    Represents the weighting of evidence from the
     experiment about θ
•    It states what the experiment says about the
     measure of interest [Savage, 1962]
•    It is the probability of getting certain result,
     conditioning on the model
•    As the amount of data increases, Prior is
     dominated by the Likelihood :
     –  Two investigators with different prior opinions
        could reach a consensus after the results of an
        experiment
Likelihood Principle
•  States that the likelihood function contains
   all relevant information from the data
•  Two samples have equivalent information
   if their likelihoods are proportional
•  Adherence to the Likelihood Principle
   means that inference are conditional on
   the observed data
•  Bayesian analysts base all inferences
   about θ solely on its posterior distribution
Likelihood Principle
•  Two experiments: one yields data y1 and
   the other yields data y2
•  If the likelihoods: P(y1 | θ) & P(y2 | θ) are
   identical up to multiplication by arbitrary
   functions of y1 & y2 then they contain
   identical information about θ and lead to
   identical posterior distributions
•  Therefore, to equivalent inferences
Example
•  EXP 1: In a study of a   •  EXP 2: Students are
   fixed sample of 20          entered into a study
   students, 12 of them        until 12 of them
   respond positively to       respond positively to
   the method [Binomial        the method [Negative-
   distribution]               binomial distribution]

•  Likelihood is            •  Likelihood at n = 20 is
   proportional to             proportional to
       θ12 (1 – θ)8                θ12 (1 – θ)8
Exchangeability
•  Key idea in statistical inference in general
•  Two observations are exchangeable if they provide
   equivalent statistical information
•  Two students randomly selected from a particular
   population of students can be considered
   exchangeable
•  If the students in a study are exchangeable with the
   students in the population for which the method is
   intended, then the study can be used to make
   inferences about the entire population
•  Exchangeability in terms of experiments: Two studies
   are exchangeable if they provide equivalent statistical
   information about some super-population of
   experiments
Laplace on Probability
It is remarkable that a science, which
commenced with the consideration of
games of chance, should be elevated to
the rank of the most important subjects
of human knowledge.
A Philosophical Essay on Probabilities.
John Wiley & Sons, 1902, page 195.
Original French edition 1814.
References
•  Computation:
   OpenBUGS http://mathstat.helsinki.fi/openbugs/
   R packages: BRugs, bayesm, R2WinBUGS from CRAN: http://
   cran.r-project.org/
•  Gelman, A, Carlin, JB, Stern, HS, & Rubin, DB (2004). Bayesian
   Data Analysis. Second Ed.. Chapman and Hall
•  Gilks, WR, Richardson, S, & Spiegelhalter, DJ (1996). Markov Chain
   Monte Carlo in Practice. Chapman & Hall

•  More Advanced:
   Bernardo, J & Smith, AFM (1994). Bayesian Theory. Wiley
   O'Hagan, A & Forster, JJ (2004). Bayesian Inference, 2nd Edition.
   Vol. 2B of "Kendall's Advanced Theory of Statistics". Arnold

Weitere ähnliche Inhalte

Was ist angesagt?

Basis of statistical inference
Basis of statistical inferenceBasis of statistical inference
Basis of statistical inference
zahidacademy
 
Logistic regression
Logistic regressionLogistic regression
Logistic regression
saba khan
 
Estimation and hypothesis testing 1 (graduate statistics2)
Estimation and hypothesis testing 1 (graduate statistics2)Estimation and hypothesis testing 1 (graduate statistics2)
Estimation and hypothesis testing 1 (graduate statistics2)
Harve Abella
 

Was ist angesagt? (20)

Bayes' theorem
Bayes' theoremBayes' theorem
Bayes' theorem
 
Belief Networks & Bayesian Classification
Belief Networks & Bayesian ClassificationBelief Networks & Bayesian Classification
Belief Networks & Bayesian Classification
 
Bernoulli distribution
Bernoulli distributionBernoulli distribution
Bernoulli distribution
 
Bayesian Networks - A Brief Introduction
Bayesian Networks - A Brief IntroductionBayesian Networks - A Brief Introduction
Bayesian Networks - A Brief Introduction
 
Mle
MleMle
Mle
 
Basis of statistical inference
Basis of statistical inferenceBasis of statistical inference
Basis of statistical inference
 
NAIVE BAYES CLASSIFIER
NAIVE BAYES CLASSIFIERNAIVE BAYES CLASSIFIER
NAIVE BAYES CLASSIFIER
 
Bayes Theorem
Bayes TheoremBayes Theorem
Bayes Theorem
 
Bayes rule (Bayes Law)
Bayes rule (Bayes Law)Bayes rule (Bayes Law)
Bayes rule (Bayes Law)
 
Sample Space and Event,Probability,The Axioms of Probability,Bayes Theorem
Sample Space and Event,Probability,The Axioms of Probability,Bayes TheoremSample Space and Event,Probability,The Axioms of Probability,Bayes Theorem
Sample Space and Event,Probability,The Axioms of Probability,Bayes Theorem
 
Logistic regression
Logistic regressionLogistic regression
Logistic regression
 
Linear regression
Linear regressionLinear regression
Linear regression
 
Logistic regression
Logistic regressionLogistic regression
Logistic regression
 
Machine learning session4(linear regression)
Machine learning   session4(linear regression)Machine learning   session4(linear regression)
Machine learning session4(linear regression)
 
Linear regression
Linear regressionLinear regression
Linear regression
 
Estimation and hypothesis testing 1 (graduate statistics2)
Estimation and hypothesis testing 1 (graduate statistics2)Estimation and hypothesis testing 1 (graduate statistics2)
Estimation and hypothesis testing 1 (graduate statistics2)
 
Logistic regression with SPSS examples
Logistic regression with SPSS examplesLogistic regression with SPSS examples
Logistic regression with SPSS examples
 
Principal Component Analysis (PCA) and LDA PPT Slides
Principal Component Analysis (PCA) and LDA PPT SlidesPrincipal Component Analysis (PCA) and LDA PPT Slides
Principal Component Analysis (PCA) and LDA PPT Slides
 
Bayes network
Bayes networkBayes network
Bayes network
 
Chapter 12 outlier
Chapter 12 outlierChapter 12 outlier
Chapter 12 outlier
 

Andere mochten auch

Bayes theorem explained
Bayes theorem explainedBayes theorem explained
Bayes theorem explained
Daniel Ross
 
Chapter 3 maximum likelihood and bayesian estimation-fix
Chapter 3   maximum likelihood and bayesian estimation-fixChapter 3   maximum likelihood and bayesian estimation-fix
Chapter 3 maximum likelihood and bayesian estimation-fix
jelli123
 

Andere mochten auch (19)

Bayesian statistics using r intro
Bayesian statistics using r   introBayesian statistics using r   intro
Bayesian statistics using r intro
 
An introduction to Bayesian Statistics using Python
An introduction to Bayesian Statistics using PythonAn introduction to Bayesian Statistics using Python
An introduction to Bayesian Statistics using Python
 
An introduction to bayesian statistics
An introduction to bayesian statisticsAn introduction to bayesian statistics
An introduction to bayesian statistics
 
What is bayesian statistics and how is it different?
What is bayesian statistics and how is it different?What is bayesian statistics and how is it different?
What is bayesian statistics and how is it different?
 
Bayesian statistical concepts
Bayesian statistical conceptsBayesian statistical concepts
Bayesian statistical concepts
 
Bayes theorem explained
Bayes theorem explainedBayes theorem explained
Bayes theorem explained
 
On being Bayesian
On being BayesianOn being Bayesian
On being Bayesian
 
Modeling and Mining Sequential Data
Modeling and Mining Sequential DataModeling and Mining Sequential Data
Modeling and Mining Sequential Data
 
Lecture 2
Lecture 2Lecture 2
Lecture 2
 
ma12012id536
ma12012id536ma12012id536
ma12012id536
 
Bayesian statistics
Bayesian statisticsBayesian statistics
Bayesian statistics
 
Monte carlo tree search
Monte carlo tree searchMonte carlo tree search
Monte carlo tree search
 
A Markov Chain Monte Carlo approach to the Steiner Tree Problem in water netw...
A Markov Chain Monte Carlo approach to the Steiner Tree Problem in water netw...A Markov Chain Monte Carlo approach to the Steiner Tree Problem in water netw...
A Markov Chain Monte Carlo approach to the Steiner Tree Problem in water netw...
 
Application of Monte Carlo Tree Search in a Fighting Game AI (GCCE 2016)
Application of Monte Carlo Tree Search in a Fighting Game AI (GCCE 2016)Application of Monte Carlo Tree Search in a Fighting Game AI (GCCE 2016)
Application of Monte Carlo Tree Search in a Fighting Game AI (GCCE 2016)
 
Chapter 3 maximum likelihood and bayesian estimation-fix
Chapter 3   maximum likelihood and bayesian estimation-fixChapter 3   maximum likelihood and bayesian estimation-fix
Chapter 3 maximum likelihood and bayesian estimation-fix
 
"Monte-Carlo Tree Search for the game of Go"
"Monte-Carlo Tree Search for the game of Go""Monte-Carlo Tree Search for the game of Go"
"Monte-Carlo Tree Search for the game of Go"
 
3 recursive bayesian estimation
3  recursive bayesian estimation3  recursive bayesian estimation
3 recursive bayesian estimation
 
Contingency approach
Contingency approachContingency approach
Contingency approach
 
Slideshare ppt
Slideshare pptSlideshare ppt
Slideshare ppt
 

Ähnlich wie Bayesian intro

D. Mayo: Philosophical Interventions in the Statistics Wars
D. Mayo: Philosophical Interventions in the Statistics WarsD. Mayo: Philosophical Interventions in the Statistics Wars
D. Mayo: Philosophical Interventions in the Statistics Wars
jemille6
 
Quantitative and Qualitative Research Methods.pptx
Quantitative and Qualitative Research Methods.pptxQuantitative and Qualitative Research Methods.pptx
Quantitative and Qualitative Research Methods.pptx
kiran513883
 

Ähnlich wie Bayesian intro (20)

Bayesian statistics intro using r
Bayesian statistics intro using rBayesian statistics intro using r
Bayesian statistics intro using r
 
Philosophy of Science and Philosophy of Statistics
Philosophy of Science and Philosophy of StatisticsPhilosophy of Science and Philosophy of Statistics
Philosophy of Science and Philosophy of Statistics
 
The Statistics Wars: Errors and Casualties
The Statistics Wars: Errors and CasualtiesThe Statistics Wars: Errors and Casualties
The Statistics Wars: Errors and Casualties
 
Meeting #1 Slides Phil 6334/Econ 6614 SP2019
Meeting #1 Slides Phil 6334/Econ 6614 SP2019Meeting #1 Slides Phil 6334/Econ 6614 SP2019
Meeting #1 Slides Phil 6334/Econ 6614 SP2019
 
Mayod@psa 21(na)
Mayod@psa 21(na)Mayod@psa 21(na)
Mayod@psa 21(na)
 
D. Mayo: Philosophy of Statistics & the Replication Crisis in Science
D. Mayo: Philosophy of Statistics & the Replication Crisis in ScienceD. Mayo: Philosophy of Statistics & the Replication Crisis in Science
D. Mayo: Philosophy of Statistics & the Replication Crisis in Science
 
D. Mayo: Philosophical Interventions in the Statistics Wars
D. Mayo: Philosophical Interventions in the Statistics WarsD. Mayo: Philosophical Interventions in the Statistics Wars
D. Mayo: Philosophical Interventions in the Statistics Wars
 
sigir2017bayesian
sigir2017bayesiansigir2017bayesian
sigir2017bayesian
 
Ds 2251 -_hypothesis test
Ds 2251 -_hypothesis testDs 2251 -_hypothesis test
Ds 2251 -_hypothesis test
 
Stat 1163 -statistics in environmental science
Stat 1163 -statistics in environmental scienceStat 1163 -statistics in environmental science
Stat 1163 -statistics in environmental science
 
Test of significance in Statistics
Test of significance in StatisticsTest of significance in Statistics
Test of significance in Statistics
 
2주차
2주차2주차
2주차
 
BBA 020
BBA 020BBA 020
BBA 020
 
D. G. Mayo Columbia slides for Workshop on Probability &Learning
D. G. Mayo Columbia slides for Workshop on Probability &LearningD. G. Mayo Columbia slides for Workshop on Probability &Learning
D. G. Mayo Columbia slides for Workshop on Probability &Learning
 
Mayo & parker spsp 2016 june 16
Mayo & parker   spsp 2016 june 16Mayo & parker   spsp 2016 june 16
Mayo & parker spsp 2016 june 16
 
Bayesian data analysis1
Bayesian data analysis1Bayesian data analysis1
Bayesian data analysis1
 
Quantitative and Qualitative Research Methods.pptx
Quantitative and Qualitative Research Methods.pptxQuantitative and Qualitative Research Methods.pptx
Quantitative and Qualitative Research Methods.pptx
 
Bayesian learning
Bayesian learningBayesian learning
Bayesian learning
 
Statistical Inference as Severe Testing: Beyond Performance and Probabilism
Statistical Inference as Severe Testing: Beyond Performance and ProbabilismStatistical Inference as Severe Testing: Beyond Performance and Probabilism
Statistical Inference as Severe Testing: Beyond Performance and Probabilism
 
Non-parametric.pptx qualitative and quantity data
Non-parametric.pptx  qualitative  and quantity  dataNon-parametric.pptx  qualitative  and quantity  data
Non-parametric.pptx qualitative and quantity data
 

Bayesian intro

  • 1. An Introduction to the Bayesian Approach J Guzmán, PhD 15 August 2011
  • 3. Bayesian: one who asks you what you think before a study in order to tell you what you think afterwards Adapted from: S Senn (1997). Statistical Issues in Drug Development. Wiley
  • 4. Rev. Thomas Bayes English Theologian and Mathematician ca. 1700 – 1761
  • 5. Bayesian Methods •  1763 – Bayes’ article on inverse probability •  Laplace extended Bayesian ideas in different scientific areas in Théorie Analytique des Probabilités [1812] •  Both Laplace & Gauss used the inverse method •  1st three quarters of 20th Century dominated by frequentist methods •  Last quarter of 20th Century – resurgence of Bayesian methods [computational advances] •  21st Century – Bayesian Century [Lindley]
  • 7. Karl Friedrich Gauss “Prince of Mathematics” 1777 – 1855 Used inverse probability
  • 8. Bayesian Methods •  Key components: prior, likelihood function, posterior, and predictive distribution •  Suppose a study is carried out to compare new and standard teaching methods •  Ho: Methods are equally effective •  HA: New method increases grades by 20% •  A Bayesian presents the probability that new & standard methods are equally effective, given the results of the experiment at hand: P(Ho | data)
  • 9. Bayesian Methods •  Data – observed data from experiment •  Find the probability that the new method is at least 20% more effective than the standard, given the results of the experiment [Posterior Probability] •  Another conclusion could be the probability distribution for the outcome of interest for the next student •  Predictive Probabilities – refer to future observations on individuals or on set of individuals
  • 10. Bayes’ Theorem •  Basic tool of Bayesian analysis •  Provide the means by which we learn from data •  Given prior state of knowledge, it tells how to update belief based upon observations: ∝ P(H | data) = P(H) · P(data | H) / P(data) € α P(H) · P(data | H) ∝ α means “is proportonal to” •  Bayes’ theorem can be re-expressed in € odds terms: let data ≡ y
  • 12. Bayes’ Theorem •  Can also consider posterior probability of any measure θ: P(θ | data) α P(θ) · P( data | θ) •  Bayes’ theorem states that the posterior probability of any measure θ, is proportional to the information on θ external to the experiment times the likelihood function evaluated at θ: Prior · likelihood → posterior
  • 13. Prior •  Prior information about θ assessed as a probability distribution on θ •  Distribution on θ depends on the assessor: it is subjective •  A subjective probability can be calculated any time a person has an opinion •  Diffuse prior - when a person’ s opinion on θ includes a broad range of possibilities & all values are thought to be roughly equally probable
  • 14. Prior •  Conjugate prior – if the posterior distribution has same shape as the prior distribution, regardless of the observed sample values •  Examples: 1.  Beta prior & binomial likelihood yield a beta posterior 2.  Normal prior & normal likelihood yield a normal posterior 3.  Gamma prior & Poisson likelihood yield a gamma posterior
  • 15. Community of Priors •  Expressing a range of reasonable opinions •  Reference – represents minimal prior information •  Expertise – formalizes opinion of well- informed experts •  Skeptical – downgrades superiority of new method •  Enthusiastic – counterbalance of skeptical
  • 16. Likelihood Function P(data | θ) •  Represents the weighting of evidence from the experiment about θ •  It states what the experiment says about the measure of interest [Savage, 1962] •  It is the probability of getting certain result, conditioning on the model •  As the amount of data increases, Prior is dominated by the Likelihood : –  Two investigators with different prior opinions could reach a consensus after the results of an experiment
  • 17. Likelihood Principle •  States that the likelihood function contains all relevant information from the data •  Two samples have equivalent information if their likelihoods are proportional •  Adherence to the Likelihood Principle means that inference are conditional on the observed data •  Bayesian analysts base all inferences about θ solely on its posterior distribution
  • 18. Likelihood Principle •  Two experiments: one yields data y1 and the other yields data y2 •  If the likelihoods: P(y1 | θ) & P(y2 | θ) are identical up to multiplication by arbitrary functions of y1 & y2 then they contain identical information about θ and lead to identical posterior distributions •  Therefore, to equivalent inferences
  • 19. Example •  EXP 1: In a study of a •  EXP 2: Students are fixed sample of 20 entered into a study students, 12 of them until 12 of them respond positively to respond positively to the method [Binomial the method [Negative- distribution] binomial distribution] •  Likelihood is •  Likelihood at n = 20 is proportional to proportional to θ12 (1 – θ)8 θ12 (1 – θ)8
  • 20. Exchangeability •  Key idea in statistical inference in general •  Two observations are exchangeable if they provide equivalent statistical information •  Two students randomly selected from a particular population of students can be considered exchangeable •  If the students in a study are exchangeable with the students in the population for which the method is intended, then the study can be used to make inferences about the entire population •  Exchangeability in terms of experiments: Two studies are exchangeable if they provide equivalent statistical information about some super-population of experiments
  • 21. Laplace on Probability It is remarkable that a science, which commenced with the consideration of games of chance, should be elevated to the rank of the most important subjects of human knowledge. A Philosophical Essay on Probabilities. John Wiley & Sons, 1902, page 195. Original French edition 1814.
  • 22. References •  Computation: OpenBUGS http://mathstat.helsinki.fi/openbugs/ R packages: BRugs, bayesm, R2WinBUGS from CRAN: http:// cran.r-project.org/ •  Gelman, A, Carlin, JB, Stern, HS, & Rubin, DB (2004). Bayesian Data Analysis. Second Ed.. Chapman and Hall •  Gilks, WR, Richardson, S, & Spiegelhalter, DJ (1996). Markov Chain Monte Carlo in Practice. Chapman & Hall •  More Advanced: Bernardo, J & Smith, AFM (1994). Bayesian Theory. Wiley O'Hagan, A & Forster, JJ (2004). Bayesian Inference, 2nd Edition. Vol. 2B of "Kendall's Advanced Theory of Statistics". Arnold