SlideShare ist ein Scribd-Unternehmen logo
1 von 56
Downloaden Sie, um offline zu lesen
Introduction to Bayesian Methods
   Theory, Computation, Inference and Prediction




Corey Chivers
PhD Candidate
Department of Biology
McGill University
Script to run examples in
these slides can be found
here:

bit.ly/Wnmb2W

These slides are here:

bit.ly/P9Xa9G
Corey Chivers, 2012
The Likelihood Principle
      ●   All information contained in data x, with
          respect to inference about the value of θ, is
          contained in the likelihood function:


                      L | x ∝ P X= x |  


Corey Chivers, 2012
The Likelihood Principle




                      L.J. Savage    R.A. Fisher



Corey Chivers, 2012
The Likelihood Function
                      L | x ∝ P X= x |  
                      L | x =f  | x 
        Where θ is(are) our parameter(s) of interest
        ex:
            Attack rate
            Fitness
            Mean body mass
            Mortality
            etc...

Corey Chivers, 2012
The Ecologist's Quarter
       Lands tails (caribou up) 60% of the time




Corey Chivers, 2012
The Ecologist's Quarter
       Lands tails (caribou up) 60% of the time
      ●    1) What is the probability that I will flip tails, given that
           I am flipping an ecologist's quarter (p(tail=0.6))?

                         P x | =0.6 
      ●    2) What is the likelihood that I am flipping an
           ecologist's quarter, given the flip(s) that I have
           observed?
                         L=0.6 | x 

Corey Chivers, 2012
The Ecologist's Quarter
                                T       H
                      L | x = ∏  ∏ 1−
                                t=1     h=1

                        L=0.6 | x=H T T H T 
                          3         2
                        = ∏ 0.6 ∏ 0.4
                         t =1   h=1
                        = 0.03456
Corey Chivers, 2012
The Ecologist's Quarter
                                T       H
                      L | x = ∏  ∏ 1−
                                t=1     h=1

                        L=0.6 | x=H T T H T 
                          3         2         But what does this
                        = ∏ 0.6 ∏ 0.4         mean?
                                              0.03456 ≠ P(θ|x) !!!!
                         t =1   h=1
                        = 0.03456
Corey Chivers, 2012
How do we ask Statistical Questions?
    A Frequentist asks: What is the probability of
      having observed data at least as extreme as my
      data if the null hypothesis is true?
           P(data | H0) ? ← note: P=1 does not mean P(H0)=1
    A Bayesian asks: What is the probability of
      hypotheses given that I have observed my data?
           P(H | data) ? ← note: here H denotes the space of all
                         possible hypotheses




Corey Chivers, 2012
P(data | H0)                            P(H | data)

                                     But we both want to make
                                     inferences about our hypotheses,
                                     not the data.




Corey Chivers, 2012
Bayes Theorem
   ●    The posterior probability of θ, given
        our observation (x) is proportional to the
        likelihood times the prior probability of θ.

                                P  x |   P 
                      P | x=
                                     P  x


Corey Chivers, 2012
The Ecologist's Quarter
                     Redux
       Lands tails (caribou up) 60% of the time




Corey Chivers, 2012
The Ecologist's Quarter
                                T       H
                      L | x = ∏  ∏ 1−
                                t=1     h=1

                        L=0.6 | x=H T T H T 
                          3         2
                        = ∏ 0.6 ∏ 0.4
                         t =1   h=1
                        = 0.03456
Corey Chivers, 2012
Likelihood of data
    given hypothesis


  P( x | θ)

  But we want to know

    P(θ | x )



Corey Chivers, 2012
●   How can we make inferences about our
          ecologist's quarter using Bayes?



                                 P( x | θ) P(θ)
                      P(θ | x )=
                                     P( x )



Corey Chivers, 2012
●   How can we make inferences about our
          ecologist's quarter using Bayes?
                                  Likelihood


                                P  x |   P 
                      P | x=
                                     P  x



Corey Chivers, 2012
●   How can we make inferences about our
          ecologist's quarter using Bayes?
                                 Likelihood   Prior


                                 P( x | θ) P(θ)
                      P(θ | x )=
                                     P( x )



Corey Chivers, 2012
●   How can we make inferences about our
          ecologist's quarter using Bayes?
                                  Likelihood   Prior


                                P  x |   P 
                      P | x=
                      Posterior
                                     P  x



Corey Chivers, 2012
●   How can we make inferences about our
          ecologist's quarter using Bayes?
                                    Likelihood   Prior


                                P  x |   P 
                      P | x=
                      Posterior
                                     P  x

                        P x =∫ P  x |  P   d 
               Not always a closed form solution possible!!

Corey Chivers, 2012
Randomization to Solve Difficult
                       Problems

                                              `


                                  Feynman, Ulam &
                                  Von Neumann




          ∫ f  d 
Corey Chivers, 2012
Monte Carlo
        Throw darts at random               Feynman, Ulam &
                                            Von Neumann


                                    (0,1)
        P(blue) = ?
        P(blue) = 1/2
        P(blue) ~ 7/15 ~ 1/2


                                               (0.5,0)        (1,0)




Corey Chivers, 2012
Your turn...
Let's use Monte Carlo to estimate π

- Generate random x and y values using the number sheet

- Plot those points on your graph

How many of the points fall
within the circle?
                                               y=17




                                                          x=4
Your turn...

Estimate π using the formula:




             ≈4 # in circle / total
Now using a more powerful
 computer!
Posterior Integration via Markov
                   Chain Monte Carlo
         A Markov Chain is a mathematical construct
           where given the present, the past and the
           future are independent.
                “Where I decide to go next depends not
                  on where I have been, or where I may
                  go in the future – but only on where I
                  am right now.”
                              -Andrey Markov (maybe)




Corey Chivers, 2012
Corey Chivers, 2012
Metropolis-Hastings Algorithm
                             1. Pick a starting location at
   The Markovian Explorer!
                             random.

                             2. Choose a new location in
                             your vicinity.

                             3. Go to the new location with
                             probability:
                                  p=min 1,  x proposal 
                                             x current     
                             4. Otherwise stay where you
                             are.

                             5. Repeat.
Corey Chivers, 2012
MCMC in Action!




Corey Chivers, 2012
●   We've solved our integration problem!




                                P  x |   P 
                      P | x=
                                     P  x
                      P | x∝ P x |  P 

Corey Chivers, 2012
Ex: Bayesian Regression
      ●   Regression coefficients are traditionally
          estimated via maximum likelihood.

      ●   To obtain full posterior distributions, we can
          view the regression problem from a Bayesian
          perspective.




Corey Chivers, 2012
##@ 2.1 @##




Corey Chivers, 2012
Example: Salmon Regression
                      Model               Priors

     Y =a+ bX +ϵ                   a ~ Normal (0,100)
       ϵ ~ Normal( 0, σ)           b ~ Normal (0,100)
                                   σ ~ gamma (1,1/ 100)



          P( a , b , σ | X , Y )∝ P( X ,Y | a , b , σ)
                                     P( a) P(b) P( σ)
Corey Chivers, 2012
Example: Salmon Regression

                Likelihood of the data (x,y), given
                the parameters (a,b,σ):
                              n
   P( X ,Y | a , b , σ)= ∏ N ( y i ,μ=a+ b x i , sd=σ)
                             i=1




Corey Chivers, 2012
Corey Chivers, 2012
Corey Chivers, 2012
Corey Chivers, 2012
##@ 2.5 @##
      >## Print the Bayesian Credible Intervals
      > BCI(mcmc_salmon)

                      0.025       0.975           post_mean
      a               -13.16485   14.84092        0.9762583
      b               0.127730    0.455046        0.2911597
      Sigma           1.736082    3.186122        2.3303188
                                         Inference:

                                         Does body length have
      EM =ab BL                     an effect on egg mass?



Corey Chivers, 2012
The Prior revisited
    ●   What if we do have prior information?

    ●   You have done a literature search and find that a
        previous study on the same salmon population
        found a slope of 0.6mg/cm (SE=0.1), and an
        intercept of -3.1mg (SE=1.2).
        How does this prior information change your
        analysis?



Corey Chivers, 2012
Corey Chivers, 2012
Example: Salmon Regression
                                   Informative
                      Model           Priors
      EM =ab BL          a ~ Normal (−3.1,1 .2)
       ~ Normal 0,        b ~ Normal (0.6,0 .1)
                               ~ gamma1,1 /100 




Corey Chivers, 2012
If you can formulate the likelihood function, you
          can estimate the posterior, and we have a
          coherent way to incorporate prior information.




                      Most experiments do happen in a vacuum.



Corey Chivers, 2012
Making predictions using point estimates can
          be a dangerous endeavor – using the posterior
          (aka predictive) distribution allows us to take
          full account of uncertainty.




                       How sure are we about our predictions?


Corey Chivers, 2012
Aleatory
 Stochasticity, randomness
Epistemic
 Incomplete knowledge
##@ 3.1 @##
      ●   Suppose you have a 90cm long individual
          salmon, what do you predict to be the egg
          mass produced by this individual?

      ●   What is the posterior probability that the egg
          mass produced will be greater than 35mg?




Corey Chivers, 2012
Corey Chivers, 2012
P(EM>35mg | θ)




Corey Chivers, 2012
Extensions:




              Clark (2005)
Extensions:
      ●    By quantifying our uncertainty through
           integration of the posterior distribution, we can
           make better informed decisions.

      ●    Bayesian analysis provides the basis for
           decision theory.

      ●    Bayesian analysis allows us to construct
           hierarchical models of arbitrary complexity.


Corey Chivers, 2012
Summary
      ●   The output of a Bayesian analysis is not a single estimate of
          θ, but rather the entire posterior distribution., which
          represents our degree of belief about the value of θ.

      ●   To get a posterior distribution, we need to specify our prior
          belief about θ.

      ●   Complex Bayesian models can be estimated using MCMC.

      ●   The posterior can be used to make both inference about θ,
          and quantitative predictions with proper accounting of
          uncertainty.



Corey Chivers, 2012
Questions for Corey
●   You can email me!
    Corey.chivers@mail.mcgill.ca

●   I blog about statistics:
    bayesianbiologist.com

●   I tweet about statistics:
    @cjbayesian
Resources
      ●    Bayesian Updating using Gibbs Sampling
           http://www.mrc-bsu.cam.ac.uk/bugs/winbugs/

      ●    Just Another Gibbs Sampler
           http://www-ice.iarc.fr/~martyn/software/jags/

      ●    Chi-squared example, done Bayesian:
           http://madere.biol.mcgill.ca/cchivers/biol373/chi-
           squared_done_bayesian.pdf


Corey Chivers, 2012

Weitere ähnliche Inhalte

Was ist angesagt? (20)

Bayesian inference
Bayesian inferenceBayesian inference
Bayesian inference
 
Categorical data analysis
Categorical data analysisCategorical data analysis
Categorical data analysis
 
Logistic Regression Analysis
Logistic Regression AnalysisLogistic Regression Analysis
Logistic Regression Analysis
 
Statistics
StatisticsStatistics
Statistics
 
Regression analysis
Regression analysisRegression analysis
Regression analysis
 
F-Distribution
F-DistributionF-Distribution
F-Distribution
 
Bayesian statistics
Bayesian statisticsBayesian statistics
Bayesian statistics
 
Probability Theory for Data Scientists
Probability Theory for Data ScientistsProbability Theory for Data Scientists
Probability Theory for Data Scientists
 
Logistic regression
Logistic regressionLogistic regression
Logistic regression
 
Different types of distributions
Different types of distributionsDifferent types of distributions
Different types of distributions
 
Mle
MleMle
Mle
 
Pca ppt
Pca pptPca ppt
Pca ppt
 
The Method Of Maximum Likelihood
The Method Of Maximum LikelihoodThe Method Of Maximum Likelihood
The Method Of Maximum Likelihood
 
Statistical distributions
Statistical distributionsStatistical distributions
Statistical distributions
 
Logistic regression
Logistic regressionLogistic regression
Logistic regression
 
Bayesian statistics
Bayesian statisticsBayesian statistics
Bayesian statistics
 
Probability And Probability Distributions
Probability And Probability Distributions Probability And Probability Distributions
Probability And Probability Distributions
 
Simple linear regression
Simple linear regressionSimple linear regression
Simple linear regression
 
Bayes rule (Bayes Law)
Bayes rule (Bayes Law)Bayes rule (Bayes Law)
Bayes rule (Bayes Law)
 
Simple linear regression
Simple linear regressionSimple linear regression
Simple linear regression
 

Ähnlich wie Introduction to Bayesian Methods

Statistical learning
Statistical learningStatistical learning
Statistical learningSlideshare
 
Bayes Classification
Bayes ClassificationBayes Classification
Bayes Classificationsathish sak
 
Statistics in Astronomy
Statistics in AstronomyStatistics in Astronomy
Statistics in AstronomyPeter Coles
 
NICE Implementations of Variational Inference
NICE Implementations of Variational Inference NICE Implementations of Variational Inference
NICE Implementations of Variational Inference Natan Katz
 
NICE Research -Variational inference project
NICE Research -Variational inference projectNICE Research -Variational inference project
NICE Research -Variational inference projectNatan Katz
 
Chapter 3 projection
Chapter 3 projectionChapter 3 projection
Chapter 3 projectionNBER
 
14th Athens Colloquium on Algorithms and Complexity (ACAC19)
14th Athens Colloquium on Algorithms and Complexity (ACAC19)14th Athens Colloquium on Algorithms and Complexity (ACAC19)
14th Athens Colloquium on Algorithms and Complexity (ACAC19)Apostolos Chalkis
 
The slides of my Ph.D. defense
The slides of my Ph.D. defenseThe slides of my Ph.D. defense
The slides of my Ph.D. defenseApostolos Chalkis
 
Multiple estimators for Monte Carlo approximations
Multiple estimators for Monte Carlo approximationsMultiple estimators for Monte Carlo approximations
Multiple estimators for Monte Carlo approximationsChristian Robert
 
Mit18 05 s14_class4slides
Mit18 05 s14_class4slidesMit18 05 s14_class4slides
Mit18 05 s14_class4slidesAnimesh Kumar
 

Ähnlich wie Introduction to Bayesian Methods (13)

Statistical learning
Statistical learningStatistical learning
Statistical learning
 
Bayes Classification
Bayes ClassificationBayes Classification
Bayes Classification
 
Statistics in Astronomy
Statistics in AstronomyStatistics in Astronomy
Statistics in Astronomy
 
NICE Implementations of Variational Inference
NICE Implementations of Variational Inference NICE Implementations of Variational Inference
NICE Implementations of Variational Inference
 
NICE Research -Variational inference project
NICE Research -Variational inference projectNICE Research -Variational inference project
NICE Research -Variational inference project
 
Linear models2
Linear models2Linear models2
Linear models2
 
Uncertainity
Uncertainity Uncertainity
Uncertainity
 
Chapter 3 projection
Chapter 3 projectionChapter 3 projection
Chapter 3 projection
 
14th Athens Colloquium on Algorithms and Complexity (ACAC19)
14th Athens Colloquium on Algorithms and Complexity (ACAC19)14th Athens Colloquium on Algorithms and Complexity (ACAC19)
14th Athens Colloquium on Algorithms and Complexity (ACAC19)
 
The slides of my Ph.D. defense
The slides of my Ph.D. defenseThe slides of my Ph.D. defense
The slides of my Ph.D. defense
 
Points of view in conceptual space
Points of view in conceptual spacePoints of view in conceptual space
Points of view in conceptual space
 
Multiple estimators for Monte Carlo approximations
Multiple estimators for Monte Carlo approximationsMultiple estimators for Monte Carlo approximations
Multiple estimators for Monte Carlo approximations
 
Mit18 05 s14_class4slides
Mit18 05 s14_class4slidesMit18 05 s14_class4slides
Mit18 05 s14_class4slides
 

Kürzlich hochgeladen

SOCIAL AND HISTORICAL CONTEXT - LFTVD.pptx
SOCIAL AND HISTORICAL CONTEXT - LFTVD.pptxSOCIAL AND HISTORICAL CONTEXT - LFTVD.pptx
SOCIAL AND HISTORICAL CONTEXT - LFTVD.pptxiammrhaywood
 
Measures of Dispersion and Variability: Range, QD, AD and SD
Measures of Dispersion and Variability: Range, QD, AD and SDMeasures of Dispersion and Variability: Range, QD, AD and SD
Measures of Dispersion and Variability: Range, QD, AD and SDThiyagu K
 
General AI for Medical Educators April 2024
General AI for Medical Educators April 2024General AI for Medical Educators April 2024
General AI for Medical Educators April 2024Janet Corral
 
Introduction to Nonprofit Accounting: The Basics
Introduction to Nonprofit Accounting: The BasicsIntroduction to Nonprofit Accounting: The Basics
Introduction to Nonprofit Accounting: The BasicsTechSoup
 
Interactive Powerpoint_How to Master effective communication
Interactive Powerpoint_How to Master effective communicationInteractive Powerpoint_How to Master effective communication
Interactive Powerpoint_How to Master effective communicationnomboosow
 
Advanced Views - Calendar View in Odoo 17
Advanced Views - Calendar View in Odoo 17Advanced Views - Calendar View in Odoo 17
Advanced Views - Calendar View in Odoo 17Celine George
 
BAG TECHNIQUE Bag technique-a tool making use of public health bag through wh...
BAG TECHNIQUE Bag technique-a tool making use of public health bag through wh...BAG TECHNIQUE Bag technique-a tool making use of public health bag through wh...
BAG TECHNIQUE Bag technique-a tool making use of public health bag through wh...Sapna Thakur
 
A Critique of the Proposed National Education Policy Reform
A Critique of the Proposed National Education Policy ReformA Critique of the Proposed National Education Policy Reform
A Critique of the Proposed National Education Policy ReformChameera Dedduwage
 
Key note speaker Neum_Admir Softic_ENG.pdf
Key note speaker Neum_Admir Softic_ENG.pdfKey note speaker Neum_Admir Softic_ENG.pdf
Key note speaker Neum_Admir Softic_ENG.pdfAdmir Softic
 
social pharmacy d-pharm 1st year by Pragati K. Mahajan
social pharmacy d-pharm 1st year by Pragati K. Mahajansocial pharmacy d-pharm 1st year by Pragati K. Mahajan
social pharmacy d-pharm 1st year by Pragati K. Mahajanpragatimahajan3
 
Q4-W6-Restating Informational Text Grade 3
Q4-W6-Restating Informational Text Grade 3Q4-W6-Restating Informational Text Grade 3
Q4-W6-Restating Informational Text Grade 3JemimahLaneBuaron
 
fourth grading exam for kindergarten in writing
fourth grading exam for kindergarten in writingfourth grading exam for kindergarten in writing
fourth grading exam for kindergarten in writingTeacherCyreneCayanan
 
9548086042 for call girls in Indira Nagar with room service
9548086042  for call girls in Indira Nagar  with room service9548086042  for call girls in Indira Nagar  with room service
9548086042 for call girls in Indira Nagar with room servicediscovermytutordmt
 
Arihant handbook biology for class 11 .pdf
Arihant handbook biology for class 11 .pdfArihant handbook biology for class 11 .pdf
Arihant handbook biology for class 11 .pdfchloefrazer622
 
1029 - Danh muc Sach Giao Khoa 10 . pdf
1029 -  Danh muc Sach Giao Khoa 10 . pdf1029 -  Danh muc Sach Giao Khoa 10 . pdf
1029 - Danh muc Sach Giao Khoa 10 . pdfQucHHunhnh
 
Call Girls in Dwarka Mor Delhi Contact Us 9654467111
Call Girls in Dwarka Mor Delhi Contact Us 9654467111Call Girls in Dwarka Mor Delhi Contact Us 9654467111
Call Girls in Dwarka Mor Delhi Contact Us 9654467111Sapana Sha
 
Class 11th Physics NEET formula sheet pdf
Class 11th Physics NEET formula sheet pdfClass 11th Physics NEET formula sheet pdf
Class 11th Physics NEET formula sheet pdfAyushMahapatra5
 

Kürzlich hochgeladen (20)

SOCIAL AND HISTORICAL CONTEXT - LFTVD.pptx
SOCIAL AND HISTORICAL CONTEXT - LFTVD.pptxSOCIAL AND HISTORICAL CONTEXT - LFTVD.pptx
SOCIAL AND HISTORICAL CONTEXT - LFTVD.pptx
 
Measures of Dispersion and Variability: Range, QD, AD and SD
Measures of Dispersion and Variability: Range, QD, AD and SDMeasures of Dispersion and Variability: Range, QD, AD and SD
Measures of Dispersion and Variability: Range, QD, AD and SD
 
General AI for Medical Educators April 2024
General AI for Medical Educators April 2024General AI for Medical Educators April 2024
General AI for Medical Educators April 2024
 
Introduction to Nonprofit Accounting: The Basics
Introduction to Nonprofit Accounting: The BasicsIntroduction to Nonprofit Accounting: The Basics
Introduction to Nonprofit Accounting: The Basics
 
Interactive Powerpoint_How to Master effective communication
Interactive Powerpoint_How to Master effective communicationInteractive Powerpoint_How to Master effective communication
Interactive Powerpoint_How to Master effective communication
 
Código Creativo y Arte de Software | Unidad 1
Código Creativo y Arte de Software | Unidad 1Código Creativo y Arte de Software | Unidad 1
Código Creativo y Arte de Software | Unidad 1
 
Mattingly "AI & Prompt Design: Structured Data, Assistants, & RAG"
Mattingly "AI & Prompt Design: Structured Data, Assistants, & RAG"Mattingly "AI & Prompt Design: Structured Data, Assistants, & RAG"
Mattingly "AI & Prompt Design: Structured Data, Assistants, & RAG"
 
Advanced Views - Calendar View in Odoo 17
Advanced Views - Calendar View in Odoo 17Advanced Views - Calendar View in Odoo 17
Advanced Views - Calendar View in Odoo 17
 
BAG TECHNIQUE Bag technique-a tool making use of public health bag through wh...
BAG TECHNIQUE Bag technique-a tool making use of public health bag through wh...BAG TECHNIQUE Bag technique-a tool making use of public health bag through wh...
BAG TECHNIQUE Bag technique-a tool making use of public health bag through wh...
 
A Critique of the Proposed National Education Policy Reform
A Critique of the Proposed National Education Policy ReformA Critique of the Proposed National Education Policy Reform
A Critique of the Proposed National Education Policy Reform
 
Key note speaker Neum_Admir Softic_ENG.pdf
Key note speaker Neum_Admir Softic_ENG.pdfKey note speaker Neum_Admir Softic_ENG.pdf
Key note speaker Neum_Admir Softic_ENG.pdf
 
INDIA QUIZ 2024 RLAC DELHI UNIVERSITY.pptx
INDIA QUIZ 2024 RLAC DELHI UNIVERSITY.pptxINDIA QUIZ 2024 RLAC DELHI UNIVERSITY.pptx
INDIA QUIZ 2024 RLAC DELHI UNIVERSITY.pptx
 
social pharmacy d-pharm 1st year by Pragati K. Mahajan
social pharmacy d-pharm 1st year by Pragati K. Mahajansocial pharmacy d-pharm 1st year by Pragati K. Mahajan
social pharmacy d-pharm 1st year by Pragati K. Mahajan
 
Q4-W6-Restating Informational Text Grade 3
Q4-W6-Restating Informational Text Grade 3Q4-W6-Restating Informational Text Grade 3
Q4-W6-Restating Informational Text Grade 3
 
fourth grading exam for kindergarten in writing
fourth grading exam for kindergarten in writingfourth grading exam for kindergarten in writing
fourth grading exam for kindergarten in writing
 
9548086042 for call girls in Indira Nagar with room service
9548086042  for call girls in Indira Nagar  with room service9548086042  for call girls in Indira Nagar  with room service
9548086042 for call girls in Indira Nagar with room service
 
Arihant handbook biology for class 11 .pdf
Arihant handbook biology for class 11 .pdfArihant handbook biology for class 11 .pdf
Arihant handbook biology for class 11 .pdf
 
1029 - Danh muc Sach Giao Khoa 10 . pdf
1029 -  Danh muc Sach Giao Khoa 10 . pdf1029 -  Danh muc Sach Giao Khoa 10 . pdf
1029 - Danh muc Sach Giao Khoa 10 . pdf
 
Call Girls in Dwarka Mor Delhi Contact Us 9654467111
Call Girls in Dwarka Mor Delhi Contact Us 9654467111Call Girls in Dwarka Mor Delhi Contact Us 9654467111
Call Girls in Dwarka Mor Delhi Contact Us 9654467111
 
Class 11th Physics NEET formula sheet pdf
Class 11th Physics NEET formula sheet pdfClass 11th Physics NEET formula sheet pdf
Class 11th Physics NEET formula sheet pdf
 

Introduction to Bayesian Methods

  • 1. Introduction to Bayesian Methods Theory, Computation, Inference and Prediction Corey Chivers PhD Candidate Department of Biology McGill University
  • 2. Script to run examples in these slides can be found here: bit.ly/Wnmb2W These slides are here: bit.ly/P9Xa9G
  • 3.
  • 5. The Likelihood Principle ● All information contained in data x, with respect to inference about the value of θ, is contained in the likelihood function: L | x ∝ P X= x |   Corey Chivers, 2012
  • 6. The Likelihood Principle L.J. Savage R.A. Fisher Corey Chivers, 2012
  • 7. The Likelihood Function L | x ∝ P X= x |   L | x =f  | x  Where θ is(are) our parameter(s) of interest ex: Attack rate Fitness Mean body mass Mortality etc... Corey Chivers, 2012
  • 8. The Ecologist's Quarter Lands tails (caribou up) 60% of the time Corey Chivers, 2012
  • 9. The Ecologist's Quarter Lands tails (caribou up) 60% of the time ● 1) What is the probability that I will flip tails, given that I am flipping an ecologist's quarter (p(tail=0.6))? P x | =0.6  ● 2) What is the likelihood that I am flipping an ecologist's quarter, given the flip(s) that I have observed? L=0.6 | x  Corey Chivers, 2012
  • 10. The Ecologist's Quarter T H L | x = ∏  ∏ 1− t=1 h=1 L=0.6 | x=H T T H T  3 2 = ∏ 0.6 ∏ 0.4 t =1 h=1 = 0.03456 Corey Chivers, 2012
  • 11. The Ecologist's Quarter T H L | x = ∏  ∏ 1− t=1 h=1 L=0.6 | x=H T T H T  3 2 But what does this = ∏ 0.6 ∏ 0.4 mean? 0.03456 ≠ P(θ|x) !!!! t =1 h=1 = 0.03456 Corey Chivers, 2012
  • 12. How do we ask Statistical Questions? A Frequentist asks: What is the probability of having observed data at least as extreme as my data if the null hypothesis is true? P(data | H0) ? ← note: P=1 does not mean P(H0)=1 A Bayesian asks: What is the probability of hypotheses given that I have observed my data? P(H | data) ? ← note: here H denotes the space of all possible hypotheses Corey Chivers, 2012
  • 13. P(data | H0) P(H | data) But we both want to make inferences about our hypotheses, not the data. Corey Chivers, 2012
  • 14. Bayes Theorem ● The posterior probability of θ, given our observation (x) is proportional to the likelihood times the prior probability of θ. P  x |   P  P | x= P  x Corey Chivers, 2012
  • 15. The Ecologist's Quarter Redux Lands tails (caribou up) 60% of the time Corey Chivers, 2012
  • 16. The Ecologist's Quarter T H L | x = ∏  ∏ 1− t=1 h=1 L=0.6 | x=H T T H T  3 2 = ∏ 0.6 ∏ 0.4 t =1 h=1 = 0.03456 Corey Chivers, 2012
  • 17. Likelihood of data given hypothesis P( x | θ) But we want to know P(θ | x ) Corey Chivers, 2012
  • 18. How can we make inferences about our ecologist's quarter using Bayes? P( x | θ) P(θ) P(θ | x )= P( x ) Corey Chivers, 2012
  • 19. How can we make inferences about our ecologist's quarter using Bayes? Likelihood P  x |   P  P | x= P  x Corey Chivers, 2012
  • 20. How can we make inferences about our ecologist's quarter using Bayes? Likelihood Prior P( x | θ) P(θ) P(θ | x )= P( x ) Corey Chivers, 2012
  • 21. How can we make inferences about our ecologist's quarter using Bayes? Likelihood Prior P  x |   P  P | x= Posterior P  x Corey Chivers, 2012
  • 22. How can we make inferences about our ecologist's quarter using Bayes? Likelihood Prior P  x |   P  P | x= Posterior P  x P x =∫ P  x |  P   d  Not always a closed form solution possible!! Corey Chivers, 2012
  • 23.
  • 24. Randomization to Solve Difficult Problems ` Feynman, Ulam & Von Neumann ∫ f  d  Corey Chivers, 2012
  • 25. Monte Carlo Throw darts at random Feynman, Ulam & Von Neumann (0,1) P(blue) = ? P(blue) = 1/2 P(blue) ~ 7/15 ~ 1/2 (0.5,0) (1,0) Corey Chivers, 2012
  • 26. Your turn... Let's use Monte Carlo to estimate π - Generate random x and y values using the number sheet - Plot those points on your graph How many of the points fall within the circle? y=17 x=4
  • 27. Your turn... Estimate π using the formula: ≈4 # in circle / total
  • 28. Now using a more powerful computer!
  • 29. Posterior Integration via Markov Chain Monte Carlo A Markov Chain is a mathematical construct where given the present, the past and the future are independent. “Where I decide to go next depends not on where I have been, or where I may go in the future – but only on where I am right now.” -Andrey Markov (maybe) Corey Chivers, 2012
  • 30.
  • 32. Metropolis-Hastings Algorithm 1. Pick a starting location at The Markovian Explorer! random. 2. Choose a new location in your vicinity. 3. Go to the new location with probability: p=min 1,  x proposal    x current   4. Otherwise stay where you are. 5. Repeat. Corey Chivers, 2012
  • 33. MCMC in Action! Corey Chivers, 2012
  • 34. We've solved our integration problem! P  x |   P  P | x= P  x P | x∝ P x |  P  Corey Chivers, 2012
  • 35. Ex: Bayesian Regression ● Regression coefficients are traditionally estimated via maximum likelihood. ● To obtain full posterior distributions, we can view the regression problem from a Bayesian perspective. Corey Chivers, 2012
  • 36. ##@ 2.1 @## Corey Chivers, 2012
  • 37. Example: Salmon Regression Model Priors Y =a+ bX +ϵ a ~ Normal (0,100) ϵ ~ Normal( 0, σ) b ~ Normal (0,100) σ ~ gamma (1,1/ 100) P( a , b , σ | X , Y )∝ P( X ,Y | a , b , σ) P( a) P(b) P( σ) Corey Chivers, 2012
  • 38. Example: Salmon Regression Likelihood of the data (x,y), given the parameters (a,b,σ): n P( X ,Y | a , b , σ)= ∏ N ( y i ,μ=a+ b x i , sd=σ) i=1 Corey Chivers, 2012
  • 42. ##@ 2.5 @## >## Print the Bayesian Credible Intervals > BCI(mcmc_salmon) 0.025 0.975 post_mean a -13.16485 14.84092 0.9762583 b 0.127730 0.455046 0.2911597 Sigma 1.736082 3.186122 2.3303188 Inference: Does body length have EM =ab BL an effect on egg mass? Corey Chivers, 2012
  • 43. The Prior revisited ● What if we do have prior information? ● You have done a literature search and find that a previous study on the same salmon population found a slope of 0.6mg/cm (SE=0.1), and an intercept of -3.1mg (SE=1.2). How does this prior information change your analysis? Corey Chivers, 2012
  • 45. Example: Salmon Regression Informative Model Priors EM =ab BL a ~ Normal (−3.1,1 .2)  ~ Normal 0,  b ~ Normal (0.6,0 .1)  ~ gamma1,1 /100  Corey Chivers, 2012
  • 46. If you can formulate the likelihood function, you can estimate the posterior, and we have a coherent way to incorporate prior information. Most experiments do happen in a vacuum. Corey Chivers, 2012
  • 47. Making predictions using point estimates can be a dangerous endeavor – using the posterior (aka predictive) distribution allows us to take full account of uncertainty. How sure are we about our predictions? Corey Chivers, 2012
  • 49. ##@ 3.1 @## ● Suppose you have a 90cm long individual salmon, what do you predict to be the egg mass produced by this individual? ● What is the posterior probability that the egg mass produced will be greater than 35mg? Corey Chivers, 2012
  • 51. P(EM>35mg | θ) Corey Chivers, 2012
  • 52. Extensions: Clark (2005)
  • 53. Extensions: ● By quantifying our uncertainty through integration of the posterior distribution, we can make better informed decisions. ● Bayesian analysis provides the basis for decision theory. ● Bayesian analysis allows us to construct hierarchical models of arbitrary complexity. Corey Chivers, 2012
  • 54. Summary ● The output of a Bayesian analysis is not a single estimate of θ, but rather the entire posterior distribution., which represents our degree of belief about the value of θ. ● To get a posterior distribution, we need to specify our prior belief about θ. ● Complex Bayesian models can be estimated using MCMC. ● The posterior can be used to make both inference about θ, and quantitative predictions with proper accounting of uncertainty. Corey Chivers, 2012
  • 55. Questions for Corey ● You can email me! Corey.chivers@mail.mcgill.ca ● I blog about statistics: bayesianbiologist.com ● I tweet about statistics: @cjbayesian
  • 56. Resources ● Bayesian Updating using Gibbs Sampling http://www.mrc-bsu.cam.ac.uk/bugs/winbugs/ ● Just Another Gibbs Sampler http://www-ice.iarc.fr/~martyn/software/jags/ ● Chi-squared example, done Bayesian: http://madere.biol.mcgill.ca/cchivers/biol373/chi- squared_done_bayesian.pdf Corey Chivers, 2012