SlideShare ist ein Scribd-Unternehmen logo
1 von 58
Downloaden Sie, um offline zu lesen
Testing as estimation: the demise of the Bayes factors
Testing as estimation: the demise of the Bayes
factors
Christian P. Robert
Universit´e Paris-Dauphine and University of Warwick
arXiv:1412.2044
with K. Kamary, K. Mengersen, and J. Rousseau
Testing as estimation: the demise of the Bayes factors
Outline
Introduction
Testing problems as estimating mixture
models
Illustrations
Asymptotic consistency
Conclusion
Testing as estimation: the demise of the Bayes factors
Introduction
Testing hypotheses
Hypothesis testing
central problem of statistical inference
dramatically differentiating feature between classical and
Bayesian paradigms
wide open to controversy and divergent opinions, includ.
within the Bayesian community
non-informative Bayesian testing case mostly unresolved,
witness the Jeffreys–Lindley paradox
[Berger (2003), Mayo & Cox (2006), Gelman (2008)]
Testing as estimation: the demise of the Bayes factors
Introduction
Besting hypotheses
Bayesian model selection as comparison of k potential
statistical models towards the selection of model that fits the
data “best”
mostly accepted perspective: it does not primarily seek to
identify which model is “true”, but compares fits
tools like Bayes factor naturally include a penalisation
addressing model complexity, mimicked by Bayes Information
(BIC) and Deviance Information (DIC) criteria
posterior predictive tools successfully advocated in Gelman et
al. (2013) even though they involve double use of data
Testing as estimation: the demise of the Bayes factors
Introduction
Besting hypotheses
Bayesian model selection as comparison of k potential
statistical models towards the selection of model that fits the
data “best”
mostly accepted perspective: it does not primarily seek to
identify which model is “true”, but compares fits
tools like Bayes factor naturally include a penalisation
addressing model complexity, mimicked by Bayes Information
(BIC) and Deviance Information (DIC) criteria
posterior predictive tools successfully advocated in Gelman et
al. (2013) even though they involve double use of data
Testing as estimation: the demise of the Bayes factors
Introduction
Bayesian modelling
Standard Bayesian approach to testing: consider two families of
models, one for each of the hypotheses under comparison,
M1 : x ∼ f1(x|θ1) , θ1 ∈ Θ1 and M2 : x ∼ f2(x|θ2) , θ2 ∈ Θ2 ,
and associate with each model a prior distribution,
θ1 ∼ π1(θ1) and θ2 ∼ π2(θ2) ,
[Jeffreys, 1939]
Testing as estimation: the demise of the Bayes factors
Introduction
Bayesian modelling
Standard Bayesian approach to testing: consider two families of
models, one for each of the hypotheses under comparison,
M1 : x ∼ f1(x|θ1) , θ1 ∈ Θ1 and M2 : x ∼ f2(x|θ2) , θ2 ∈ Θ2 ,
in order to compare the marginal likelihoods
m1(x) =
Θ1
f1(x|θ1) π1(θ1) dθ1 and m2(x) =
Θ2
f2(x|θ2) π1(θ2) dθ2
[Jeffreys, 1939]
Testing as estimation: the demise of the Bayes factors
Introduction
Bayesian modelling
Standard Bayesian approach to testing: consider two families of
models, one for each of the hypotheses under comparison,
M1 : x ∼ f1(x|θ1) , θ1 ∈ Θ1 and M2 : x ∼ f2(x|θ2) , θ2 ∈ Θ2 ,
either through Bayes factor or posterior probability,
B12 =
m1(x)
m2(x)
, P(M1|x) =
ω1m1(x)
ω1m1(x) + ω2m2(x)
;
the latter depends on the prior weights ωi
[Jeffreys, 1939]
Testing as estimation: the demise of the Bayes factors
Introduction
Bayesian modelling
Standard Bayesian approach to testing: consider two families of
models, one for each of the hypotheses under comparison,
M1 : x ∼ f1(x|θ1) , θ1 ∈ Θ1 and M2 : x ∼ f2(x|θ2) , θ2 ∈ Θ2 ,
Bayesian decision step
comparing Bayes factor B12 with threshold value of one or
comparing posterior probability P(M1|x) with bound α
[Jeffreys, 1939]
Testing as estimation: the demise of the Bayes factors
Introduction
Some difficulties
tension between (i) posterior probabilities justified by binary
loss but depending on unnatural prior weights, and (ii) Bayes
factors that eliminate dependence but lose direct connection
with posterior, unless prior weights are integrated within loss
delicate interpretation (or calibration) of the strength of the
Bayes factor towards supporting a given hypothesis or model:
not a Bayesian decision rule!
difficulty with posterior probabilities: tendency to interpret
them as p-values while they only report respective strengths
of fitting to both models
Testing as estimation: the demise of the Bayes factors
Introduction
Some further difficulties
long-lasting impact of the prior modeling, i.e., choice of prior
distributions on both parameter spaces under comparison,
despite overall consistency for Bayes factor
major discontinuity in use of improper priors, not justified in
most testing situations, leading to ad hoc solutions (zoo),
where data is either used twice or split artificially
binary (accept vs. reject) outcome more suited for immediate
decision (if any) than for model evaluation, connected with
rudimentary loss function
Testing as estimation: the demise of the Bayes factors
Introduction
Lindley’s paradox
In a normal mean testing problem,
¯xn ∼ N(θ, σ2
/n) , H0 : θ = θ0 ,
under Jeffreys prior, θ ∼ N(θ0, σ2), the Bayes factor
B01(tn) = (1 + n)1/2
exp −nt2
n/2[1 + n] ,
where tn =
√
n|¯xn − θ0|/σ, satisfies
B01(tn)
n−→∞
−→ ∞
[assuming a fixed tn]
[Lindley, 1957]
Testing as estimation: the demise of the Bayes factors
Introduction
A strong impropriety
Improper priors not allowed in Bayes factors:
If
Θ1
π1(dθ1) = ∞ or
Θ2
π2(dθ2) = ∞
then π1 or π2 cannot be coherently normalised while the
normalisation matters in the Bayes factor B12
Lack of mathematical justification for “common nuisance
parameter” [and prior of]
[Berger, Pericchi, and Varshavsky, 1998; Marin and Robert, 2013]
Testing as estimation: the demise of the Bayes factors
Introduction
A strong impropriety
Improper priors not allowed in Bayes factors:
If
Θ1
π1(dθ1) = ∞ or
Θ2
π2(dθ2) = ∞
then π1 or π2 cannot be coherently normalised while the
normalisation matters in the Bayes factor B12
Lack of mathematical justification for “common nuisance
parameter” [and prior of]
[Berger, Pericchi, and Varshavsky, 1998; Marin and Robert, 2013]
Testing as estimation: the demise of the Bayes factors
Testing problems as estimating mixture models
Paradigm shift
New proposal as paradigm shift in Bayesian processing of
hypothesis testing and of model selection
convergent and naturally interpretable solution
extended use of improper priors
abandonment of the Neyman-Pearson decision framework
natural strenght of evidence
Simple representation of the testing problem as a
two-component mixture estimation problem where the
weights are formally equal to 0 or 1
Testing as estimation: the demise of the Bayes factors
Testing problems as estimating mixture models
Paradigm shift
New proposal as paradigm shift in Bayesian processing of
hypothesis testing and of model selection
convergent and naturally interpretable solution
extended use of improper priors
abandonment of the Neyman-Pearson decision framework
natural strenght of evidence
Simple representation of the testing problem as a
two-component mixture estimation problem where the
weights are formally equal to 0 or 1
Testing as estimation: the demise of the Bayes factors
Testing problems as estimating mixture models
Paradigm shift
Simple representation of the testing problem as a
two-component mixture estimation problem where the
weights are formally equal to 0 or 1
Approach inspired from consistency result of Rousseau and
Mengersen (2011) on estimated overfitting mixtures
Mixture representation not equivalent to use of a posterior
probability
More natural approach to testing, while sparse in parameters
Calibration of the posterior distribution of mixture weight,
while moving away from artificial notion of the posterior
probability of a model
Testing as estimation: the demise of the Bayes factors
Testing problems as estimating mixture models
Encompassing mixture model
Idea: Given two statistical models,
M1 : x ∼ f1(x|θ1) , θ1 ∈ Θ1 and M2 : x ∼ f2(x|θ2) , θ2 ∈ Θ2 ,
embed both within an encompassing mixture
Mα : x ∼ αf1(x|θ1) + (1 − α)f2(x|θ2) , 0 α 1 (1)
Note: Both models correspond to special cases of (1), one for
α = 1 and one for α = 0
Draw inference on mixture representation (1), as if each
observation was individually and independently produced by the
mixture model
Testing as estimation: the demise of the Bayes factors
Testing problems as estimating mixture models
Encompassing mixture model
Idea: Given two statistical models,
M1 : x ∼ f1(x|θ1) , θ1 ∈ Θ1 and M2 : x ∼ f2(x|θ2) , θ2 ∈ Θ2 ,
embed both within an encompassing mixture
Mα : x ∼ αf1(x|θ1) + (1 − α)f2(x|θ2) , 0 α 1 (1)
Note: Both models correspond to special cases of (1), one for
α = 1 and one for α = 0
Draw inference on mixture representation (1), as if each
observation was individually and independently produced by the
mixture model
Testing as estimation: the demise of the Bayes factors
Testing problems as estimating mixture models
Encompassing mixture model
Idea: Given two statistical models,
M1 : x ∼ f1(x|θ1) , θ1 ∈ Θ1 and M2 : x ∼ f2(x|θ2) , θ2 ∈ Θ2 ,
embed both within an encompassing mixture
Mα : x ∼ αf1(x|θ1) + (1 − α)f2(x|θ2) , 0 α 1 (1)
Note: Both models correspond to special cases of (1), one for
α = 1 and one for α = 0
Draw inference on mixture representation (1), as if each
observation was individually and independently produced by the
mixture model
Testing as estimation: the demise of the Bayes factors
Testing problems as estimating mixture models
Inferential motivations
Sounds like approximation to the real problem, but definitive
advantages to shift:
Bayes estimate of the weight α replaces posterior probability
of model M1, equally convergent indicator of which model is
“true”, while avoiding artificial prior probabilities on model
indices, ω1 and ω2, and 0 − 1 loss setting
posterior on α provides measure of proximity to models, while
being interpretable as data propensity to stand within one
model
further allows for alternative perspectives on testing and
model choice, like predictive tools, cross-validation, and
information indices like WAIC
Testing as estimation: the demise of the Bayes factors
Testing problems as estimating mixture models
Computational motivations
avoids problematic computations of marginal likelihoods, since
standard algorithms are available for Bayesian mixture
estimation
straightforward extension to finite collection of models, which
considers all models at once and eliminates least likely models
by simulation
eliminates famous difficulty of label switching that plagues
both Bayes estimation and computation: components are no
longer exchangeable
posterior distribution on α evaluates more thoroughly strength
of support for a given model than the single figure posterior
probability
variability of posterior distribution on α allows for a more
thorough assessment of the strength of this support
Testing as estimation: the demise of the Bayes factors
Testing problems as estimating mixture models
Noninformative motivations
novel Bayesian feature: a mixture model acknowledges
possibility that, for a finite dataset, both models or none
could be acceptable
standard (proper and informative) prior modeling can be
processed in this setting, but non-informative (improper)
priors also are manageable, provided both models first
reparameterised into shared parameters, e.g. location and
scale parameters
in special case when all parameters are common
Mα : x ∼ αf1(x|θ) + (1 − α)f2(x|θ) , 0 α 1
if θ is a location parameter, a flat prior π(θ) ∝ 1 is available
Testing as estimation: the demise of the Bayes factors
Testing problems as estimating mixture models
Weakly informative motivations
using the same parameters or some identical parameters on
both components highlights that opposition between the two
components is not an issue of enjoying different parameters
common parameters are nuisance parameters, easily integrated
prior model weights ωi rarely discussed in classical Bayesian
approach, with linear impact on posterior probabilities
prior modeling only involves selecting a prior on α, e.g.,
α ∼ B(a0, a0)
while a0 impacts posterior on α, it always leads to mass
accumulation near 1 or 0, i.e. favours most likely model
sensitivity analysis straightforward to carry
approach easily calibrated by parametric boostrap providing
reference posterior of α under each model
natural Metropolis–Hastings alternative
Testing as estimation: the demise of the Bayes factors
Illustrations
Poisson/Geometric example
choice betwen Poisson P(λ) and Geometric Geo(p)
distribution
mixture with common parameter λ
Mα : αP(λ) + (1 − α)Geo(1/1+λ)
Allows for Jeffreys prior since resulting posterior is proper
independent Metropolis–within–Gibbs with proposal
distribution on λ equal to Poisson posterior (with acceptance
rate larger than 75%)
Testing as estimation: the demise of the Bayes factors
Illustrations
Poisson/Geometric example
choice betwen Poisson P(λ) and Geometric Geo(p)
distribution
mixture with common parameter λ
Mα : αP(λ) + (1 − α)Geo(1/1+λ)
Allows for Jeffreys prior since resulting posterior is proper
independent Metropolis–within–Gibbs with proposal
distribution on λ equal to Poisson posterior (with acceptance
rate larger than 75%)
Testing as estimation: the demise of the Bayes factors
Illustrations
Poisson/Geometric example
choice betwen Poisson P(λ) and Geometric Geo(p)
distribution
mixture with common parameter λ
Mα : αP(λ) + (1 − α)Geo(1/1+λ)
Allows for Jeffreys prior since resulting posterior is proper
independent Metropolis–within–Gibbs with proposal
distribution on λ equal to Poisson posterior (with acceptance
rate larger than 75%)
Testing as estimation: the demise of the Bayes factors
Illustrations
Beta prior
When α ∼ Be(a0, a0) prior, full conditional posterior
α ∼ Be(n1(ζ) + a0, n2(ζ) + a0)
Exact Bayes factor opposing Poisson and Geometric
B12 = nn¯xn
n
i=1
xi ! Γ n + 2 +
n
i=1
xi Γ(n + 2)
although undefined from a purely mathematical viewpoint
Testing as estimation: the demise of the Bayes factors
Illustrations
Weight estimation
1e-04 0.001 0.01 0.1 0.2 0.3 0.4 0.5
0.9900.9920.9940.9960.9981.000
Posterior medians of α for 100 Poisson P(4) datasets of size n = 1000, for
a0 = .0001, .001, .01, .1, .2, .3, .4, .5. Each posterior approximation is based on
104
Metropolis-Hastings iterations.
Testing as estimation: the demise of the Bayes factors
Illustrations
Consistency
0 1 2 3 4 5 6 7
0.00.20.40.60.81.0
a0=.1
log(sample size)
0 1 2 3 4 5 6 7
0.00.20.40.60.81.0
a0=.3
log(sample size)
0 1 2 3 4 5 6 7
0.00.20.40.60.81.0
a0=.5
log(sample size)
Posterior means (sky-blue) and medians (grey-dotted) of α, over 100 Poisson
P(4) datasets for sample sizes from 1 to 1000.
Testing as estimation: the demise of the Bayes factors
Illustrations
Behaviour of Bayes factor
0 1 2 3 4 5 6 7
0.00.20.40.60.81.0
log(sample size)
0 1 2 3 4 5 6 7
0.00.20.40.60.81.0
log(sample size)
0 1 2 3 4 5 6 7
0.00.20.40.60.81.0
log(sample size)
Comparison between P(M1|x) (red dotted area) and posterior medians of α
(grey zone) for 100 Poisson P(4) datasets with sample sizes n between 1 and
1000, for a0 = .001, .1, .5
Testing as estimation: the demise of the Bayes factors
Illustrations
Normal-normal comparison
comparison of a normal N(θ1, 1) with a normal N(θ2, 2)
distribution
mixture with identical location parameter θ
αN(θ, 1) + (1 − α)N(θ, 2)
Jeffreys prior π(θ) = 1 can be used, since posterior is proper
Reference (improper) Bayes factor
B12 = 2
n−1/2
exp 1/4
n
i=1
(xi − ¯x)2
,
Testing as estimation: the demise of the Bayes factors
Illustrations
Normal-normal comparison
comparison of a normal N(θ1, 1) with a normal N(θ2, 2)
distribution
mixture with identical location parameter θ
αN(θ, 1) + (1 − α)N(θ, 2)
Jeffreys prior π(θ) = 1 can be used, since posterior is proper
Reference (improper) Bayes factor
B12 = 2
n−1/2
exp 1/4
n
i=1
(xi − ¯x)2
,
Testing as estimation: the demise of the Bayes factors
Illustrations
Normal-normal comparison
comparison of a normal N(θ1, 1) with a normal N(θ2, 2)
distribution
mixture with identical location parameter θ
αN(θ, 1) + (1 − α)N(θ, 2)
Jeffreys prior π(θ) = 1 can be used, since posterior is proper
Reference (improper) Bayes factor
B12 = 2
n−1/2
exp 1/4
n
i=1
(xi − ¯x)2
,
Testing as estimation: the demise of the Bayes factors
Illustrations
Comparison with posterior probability
0 100 200 300 400 500
-50-40-30-20-100
a0=.1
sample size
0 100 200 300 400 500
-50-40-30-20-100
a0=.3
sample size
0 100 200 300 400 500
-50-40-30-20-100
a0=.4
sample size
0 100 200 300 400 500
-50-40-30-20-100
a0=.5
sample size
Plots of ranges of log(n) log(1 − E[α|x]) (gray color) and log(1 − p(M1|x)) (red
dotted) over 100 N(0, 1) samples as sample size n grows from 1 to 500. and α
is the weight of N(0, 1) in the mixture model. The shaded areas indicate the
range of the estimations and each plot is based on a Beta prior with
a0 = .1, .2, .3, .4, .5, 1 and each posterior approximation is based on 104
iterations.
Testing as estimation: the demise of the Bayes factors
Illustrations
Comments
convergence to one boundary value as sample size n grows
impact of hyperarameter a0 slowly vanishes as n increases, but
present for moderate sample sizes
when simulated sample is neither from N(θ1, 1) nor from
N(θ2, 2), behaviour of posterior varies, depending on which
distribution is closest
Testing as estimation: the demise of the Bayes factors
Illustrations
Logit or Probit?
binary dataset, R dataset about diabetes in 200 Pima Indian
women with body mass index as explanatory variable
comparison of logit and probit fits could be suitable. We are
thus comparing both fits via our method
M1 : yi | xi
, θ1 ∼ B(1, pi ) where pi =
exp(xi θ1)
1 + exp(xi θ1)
M2 : yi | xi
, θ2 ∼ B(1, qi ) where qi = Φ(xi
θ2)
Testing as estimation: the demise of the Bayes factors
Illustrations
Common parameterisation
Local reparameterisation strategy that rescales parameters of the
probit model M2 so that the MLE’s of both models coincide.
[Choudhuty et al., 2007]
Φ(xi
θ2) ≈
exp(kxi θ2)
1 + exp(kxi θ2)
and use best estimate of k to bring both parameters into coherency
(k0, k1) = (θ01/θ02, θ11/θ12) ,
reparameterise M1 and M2 as
M1 :yi | xi
, θ ∼ B(1, pi ) where pi =
exp(xi θ)
1 + exp(xi θ)
M2 :yi | xi
, θ ∼ B(1, qi ) where qi = Φ(xi
(κ−1
θ)) ,
with κ−1θ = (θ0/k0, θ1/k1).
Testing as estimation: the demise of the Bayes factors
Illustrations
Prior modelling
Under default g-prior
θ ∼ N2(0, n(XT
X)−1
)
full conditional posterior distributions given allocations
π(θ | y, X, ζ) ∝
exp i Iζi =1yi xi θ
i;ζi =1[1 + exp(xi θ)]
exp −θT
(XT
X)θ 2n
×
i;ζi =2
Φ(xi
(κ−1
θ))yi
(1 − Φ(xi
(κ−1
θ)))(1−yi )
hence posterior distribution clearly defined
Testing as estimation: the demise of the Bayes factors
Illustrations
Results
Logistic Probit
a0 α θ0 θ1
θ0
k0
θ1
k1
.1 .352 -4.06 .103 -2.51 .064
.2 .427 -4.03 .103 -2.49 .064
.3 .440 -4.02 .102 -2.49 .063
.4 .456 -4.01 .102 -2.48 .063
.5 .449 -4.05 .103 -2.51 .064
Histograms of posteriors of α in favour of logistic model where a0 = .1, .2, .3,
.4, .5 for (a) Pima dataset, (b) Data from logistic model, (c) Data from probit
Testing as estimation: the demise of the Bayes factors
Illustrations
Survival analysis models
Testing hypothesis that data comes from a
1. log-Normal(φ, κ2),
2. Weibull(α, λ), or
3. log-Logistic(γ, δ)
distribution
Corresponding mixture given by the density
α1 exp{−(log x − φ)2
/2κ2
}/
√
2πxκ+
α2
α
λ
exp{−(x/λ)α
}((x/λ)α−1
+
α3(δ/γ)(x/γ)δ−1
/(1 + (x/γ)δ
)2
where α1 + α2 + α3 = 1
Testing as estimation: the demise of the Bayes factors
Illustrations
Survival analysis models
Testing hypothesis that data comes from a
1. log-Normal(φ, κ2),
2. Weibull(α, λ), or
3. log-Logistic(γ, δ)
distribution
Corresponding mixture given by the density
α1 exp{−(log x − φ)2
/2κ2
}/
√
2πxκ+
α2
α
λ
exp{−(x/λ)α
}((x/λ)α−1
+
α3(δ/γ)(x/γ)δ−1
/(1 + (x/γ)δ
)2
where α1 + α2 + α3 = 1
Testing as estimation: the demise of the Bayes factors
Illustrations
Reparameterisation
Looking for common parameter(s):
φ = µ + γβ = ξ
σ2
= π2
β2
/6 = ζ2
π2
/3
where γ ≈ 0.5772 is Euler-Mascheroni constant.
Allows for a noninformative prior on the common location scale
parameter,
π(φ, σ2
) = 1/σ2
Testing as estimation: the demise of the Bayes factors
Illustrations
Reparameterisation
Looking for common parameter(s):
φ = µ + γβ = ξ
σ2
= π2
β2
/6 = ζ2
π2
/3
where γ ≈ 0.5772 is Euler-Mascheroni constant.
Allows for a noninformative prior on the common location scale
parameter,
π(φ, σ2
) = 1/σ2
Testing as estimation: the demise of the Bayes factors
Illustrations
Recovery
.01 0.1 1.0 10.0
0.860.880.900.920.940.960.981.00
.01 0.1 1.0 10.0
0.000.020.040.060.08
Boxplots of the posterior distributions of the Normal weight α1 under the two
scenarii: truth = Normal (left panel), truth = Gumbel (right panel), a0=0.01,
0.1, 1.0, 10.0 (from left to right in each panel) and n = 10, 000 simulated
observations.
Testing as estimation: the demise of the Bayes factors
Asymptotic consistency
Asymptotic consistency
Posterior consistency holds for mixture testing procedure [under
minor conditions]
Two different cases
the two models, M1 and M2, are well separated
model M1 is a submodel of M2.
Testing as estimation: the demise of the Bayes factors
Asymptotic consistency
Asymptotic consistency
Posterior consistency holds for mixture testing procedure [under
minor conditions]
Two different cases
the two models, M1 and M2, are well separated
model M1 is a submodel of M2.
Testing as estimation: the demise of the Bayes factors
Asymptotic consistency
Separated models
Assumption: Models are separated, i.e. identifiability holds:
∀α, α ∈ [0, 1], ∀θj , θj , j = 1, 2 Pθ,α = Pθ ,α ⇒ α = α , θ = θ
theorem
Under above assumptions, then for all > 0,
π [|α − α∗
| > |xn
] = op(1)
Testing as estimation: the demise of the Bayes factors
Asymptotic consistency
Separated models
Assumption: Models are separated, i.e. identifiability holds:
∀α, α ∈ [0, 1], ∀θj , θj , j = 1, 2 Pθ,α = Pθ ,α ⇒ α = α , θ = θ
theorem
If
θj → fj,θj
is C2 around θ∗
j , j = 1, 2,
f1,θ∗
1
− f2,θ∗
2
, f1,θ∗
1
, f2,θ∗
2
are linearly independent in y and
there exists δ > 0 such that
f1,θ∗
1
, f2,θ∗
2
, sup
|θ1−θ∗
1 |<δ
|D2
f1,θ1
|, sup
|θ2−θ∗
2 |<δ
|D2
f2,θ2
| ∈ L1
then
π |α − α∗
| > M log n/n xn
= op(1).
Testing as estimation: the demise of the Bayes factors
Asymptotic consistency
Separated models
Assumption: Models are separated, i.e. identifiability holds:
∀α, α ∈ [0, 1], ∀θj , θj , j = 1, 2 Pθ,α = Pθ ,α ⇒ α = α , θ = θ
theorem allows for interpretation of α under the posterior: If data
xn is generated from model M1 then posterior on α concentrates
around α = 1
Testing as estimation: the demise of the Bayes factors
Asymptotic consistency
Embedded case
Here M1 is a submodel of M2, i.e.
θ2 = (θ1, ψ) and θ2 = (θ1, ψ0 = 0)
corresponds to f2,θ2 ∈ M1
Same posterior concentration rate
log n/n
for estimating α when α∗ ∈ (0, 1) and ψ∗ = 0.
Testing as estimation: the demise of the Bayes factors
Asymptotic consistency
Null case
Case where ψ∗ = 0, i.e., f ∗ is in model M1
Two possible paths to approximate f ∗: either α goes to 1
(path 1) or ψ goes to 0 (path 2)
New identifiability condition: Pθ,α = P∗ only if
α = 1, θ1 = θ∗
1, θ2 = (θ∗
1, ψ) or α 1, θ1 = θ∗
1, θ2 = (θ∗
1, 0)
Prior
π(α, θ) = πα(α)π1(θ1)πψ(ψ), θ2 = (θ1, ψ)
with common (prior on) θ1
Testing as estimation: the demise of the Bayes factors
Asymptotic consistency
Null case
Case where ψ∗ = 0, i.e., f ∗ is in model M1
Two possible paths to approximate f ∗: either α goes to 1
(path 1) or ψ goes to 0 (path 2)
New identifiability condition: Pθ,α = P∗ only if
α = 1, θ1 = θ∗
1, θ2 = (θ∗
1, ψ) or α 1, θ1 = θ∗
1, θ2 = (θ∗
1, 0)
Prior
π(α, θ) = πα(α)π1(θ1)πψ(ψ), θ2 = (θ1, ψ)
with common (prior on) θ1
Testing as estimation: the demise of the Bayes factors
Asymptotic consistency
Consistency
theorem
Given the mixture fθ1,ψ,α = αf1,θ1 + (1 − α)f2,θ1,ψ and a sample
xn = (x1, · · · , xn) issued from f1,θ∗
1
, under regularity assumptions,
and an M > 0 such that
π (α, θ); fθ,α − f ∗
1 > M log n/n|xn
= op(1).
If α ∼ B(a1, a2), with a2 < d2, and if the prior πθ1,ψ is absolutely
continuous with positive and continuous density at (θ∗
1, 0), then for
Mn −→ ∞
π |α − α∗
| > Mn(log n)γ
/
√
n|xn
= op(1), γ = max((d1 + a2)/(d2 − a2), 1)/2,
Testing as estimation: the demise of the Bayes factors
Conclusion
Conclusion
many applications of the Bayesian paradigm concentrate on
the comparison of scientific theories and on testing of null
hypotheses
natural tendency to default to Bayes factors
poorly understood sensitivity to prior modeling and posterior
calibration
Time is ripe for a paradigm shift
Testing as estimation: the demise of the Bayes factors
Conclusion
Conclusion
Time is ripe for a paradigm shift
original testing problem replaced with a better controlled
estimation target
allow for posterior variability over the component frequency as
opposed to deterministic Bayes factors
range of acceptance, rejection and indecision conclusions
easily calibrated by simulation
posterior medians quickly settling near the boundary values of
0 and 1
potential derivation of a Bayesian b-value by looking at the
posterior area under the tail of the distribution of the weight
Testing as estimation: the demise of the Bayes factors
Conclusion
Prior modelling
Time is ripe for a paradigm shift
Partly common parameterisation always feasible and hence
allows for reference priors
removal of the absolute prohibition of improper priors in
hypothesis testing
prior on the weight α shows sensitivity that naturally vanishes
as the sample size increases
default value of a0 = 0.5 in the Beta prior
Testing as estimation: the demise of the Bayes factors
Conclusion
Computing aspects
Time is ripe for a paradigm shift
proposal that does not induce additional computational strain
when algorithmic solutions exist for both models, they can be
recycled towards estimating the encompassing mixture
easier than in standard mixture problems due to common
parameters that allow for original MCMC samplers to be
turned into proposals
Gibbs sampling completions useful for assessing potential
outliers but not essential to achieve a conclusion about the
overall problem

Weitere ähnliche Inhalte

Was ist angesagt?

"reflections on the probability space induced by moment conditions with impli...
"reflections on the probability space induced by moment conditions with impli..."reflections on the probability space induced by moment conditions with impli...
"reflections on the probability space induced by moment conditions with impli...Christian Robert
 
random forests for ABC model choice and parameter estimation
random forests for ABC model choice and parameter estimationrandom forests for ABC model choice and parameter estimation
random forests for ABC model choice and parameter estimationChristian Robert
 
Can we estimate a constant?
Can we estimate a constant?Can we estimate a constant?
Can we estimate a constant?Christian Robert
 
Approximate Bayesian model choice via random forests
Approximate Bayesian model choice via random forestsApproximate Bayesian model choice via random forests
Approximate Bayesian model choice via random forestsChristian Robert
 
better together? statistical learning in models made of modules
better together? statistical learning in models made of modulesbetter together? statistical learning in models made of modules
better together? statistical learning in models made of modulesChristian Robert
 
Probability/Statistics Lecture Notes 4: Hypothesis Testing
Probability/Statistics Lecture Notes 4: Hypothesis TestingProbability/Statistics Lecture Notes 4: Hypothesis Testing
Probability/Statistics Lecture Notes 4: Hypothesis Testingjemille6
 
An overview of Bayesian testing
An overview of Bayesian testingAn overview of Bayesian testing
An overview of Bayesian testingChristian Robert
 
NBBC15, Reyjavik, June 08, 2015
NBBC15, Reyjavik, June 08, 2015NBBC15, Reyjavik, June 08, 2015
NBBC15, Reyjavik, June 08, 2015Christian Robert
 
accurate ABC Oliver Ratmann
accurate ABC Oliver Ratmannaccurate ABC Oliver Ratmann
accurate ABC Oliver Ratmannolli0601
 
Big Data Analysis
Big Data AnalysisBig Data Analysis
Big Data AnalysisNBER
 
Likelihood-free Design: a discussion
Likelihood-free Design: a discussionLikelihood-free Design: a discussion
Likelihood-free Design: a discussionChristian Robert
 
NCE, GANs & VAEs (and maybe BAC)
NCE, GANs & VAEs (and maybe BAC)NCE, GANs & VAEs (and maybe BAC)
NCE, GANs & VAEs (and maybe BAC)Christian Robert
 
Laplace's Demon: seminar #1
Laplace's Demon: seminar #1Laplace's Demon: seminar #1
Laplace's Demon: seminar #1Christian Robert
 
Approximating Bayes Factors
Approximating Bayes FactorsApproximating Bayes Factors
Approximating Bayes FactorsChristian Robert
 
Convergence of ABC methods
Convergence of ABC methodsConvergence of ABC methods
Convergence of ABC methodsChristian Robert
 

Was ist angesagt? (20)

2018 MUMS Fall Course - Introduction to statistical and mathematical model un...
2018 MUMS Fall Course - Introduction to statistical and mathematical model un...2018 MUMS Fall Course - Introduction to statistical and mathematical model un...
2018 MUMS Fall Course - Introduction to statistical and mathematical model un...
 
"reflections on the probability space induced by moment conditions with impli...
"reflections on the probability space induced by moment conditions with impli..."reflections on the probability space induced by moment conditions with impli...
"reflections on the probability space induced by moment conditions with impli...
 
2018 MUMS Fall Course - Issue Arising in Several Working Groups: Probabilisti...
2018 MUMS Fall Course - Issue Arising in Several Working Groups: Probabilisti...2018 MUMS Fall Course - Issue Arising in Several Working Groups: Probabilisti...
2018 MUMS Fall Course - Issue Arising in Several Working Groups: Probabilisti...
 
random forests for ABC model choice and parameter estimation
random forests for ABC model choice and parameter estimationrandom forests for ABC model choice and parameter estimation
random forests for ABC model choice and parameter estimation
 
Can we estimate a constant?
Can we estimate a constant?Can we estimate a constant?
Can we estimate a constant?
 
Intractable likelihoods
Intractable likelihoodsIntractable likelihoods
Intractable likelihoods
 
Approximate Bayesian model choice via random forests
Approximate Bayesian model choice via random forestsApproximate Bayesian model choice via random forests
Approximate Bayesian model choice via random forests
 
better together? statistical learning in models made of modules
better together? statistical learning in models made of modulesbetter together? statistical learning in models made of modules
better together? statistical learning in models made of modules
 
Probability/Statistics Lecture Notes 4: Hypothesis Testing
Probability/Statistics Lecture Notes 4: Hypothesis TestingProbability/Statistics Lecture Notes 4: Hypothesis Testing
Probability/Statistics Lecture Notes 4: Hypothesis Testing
 
An overview of Bayesian testing
An overview of Bayesian testingAn overview of Bayesian testing
An overview of Bayesian testing
 
NBBC15, Reyjavik, June 08, 2015
NBBC15, Reyjavik, June 08, 2015NBBC15, Reyjavik, June 08, 2015
NBBC15, Reyjavik, June 08, 2015
 
accurate ABC Oliver Ratmann
accurate ABC Oliver Ratmannaccurate ABC Oliver Ratmann
accurate ABC Oliver Ratmann
 
Big model, big data
Big model, big dataBig model, big data
Big model, big data
 
ABC workshop: 17w5025
ABC workshop: 17w5025ABC workshop: 17w5025
ABC workshop: 17w5025
 
Big Data Analysis
Big Data AnalysisBig Data Analysis
Big Data Analysis
 
Likelihood-free Design: a discussion
Likelihood-free Design: a discussionLikelihood-free Design: a discussion
Likelihood-free Design: a discussion
 
NCE, GANs & VAEs (and maybe BAC)
NCE, GANs & VAEs (and maybe BAC)NCE, GANs & VAEs (and maybe BAC)
NCE, GANs & VAEs (and maybe BAC)
 
Laplace's Demon: seminar #1
Laplace's Demon: seminar #1Laplace's Demon: seminar #1
Laplace's Demon: seminar #1
 
Approximating Bayes Factors
Approximating Bayes FactorsApproximating Bayes Factors
Approximating Bayes Factors
 
Convergence of ABC methods
Convergence of ABC methodsConvergence of ABC methods
Convergence of ABC methods
 

Ähnlich wie Testing as estimation: the demise of the Bayes factor

ABC short course: model choice chapter
ABC short course: model choice chapterABC short course: model choice chapter
ABC short course: model choice chapterChristian Robert
 
Cointegration analysis: Modelling the complex interdependencies between finan...
Cointegration analysis: Modelling the complex interdependencies between finan...Cointegration analysis: Modelling the complex interdependencies between finan...
Cointegration analysis: Modelling the complex interdependencies between finan...Edward Thomas Jones
 
Intro to Model Selection
Intro to Model SelectionIntro to Model Selection
Intro to Model Selectionchenhm
 
TTests.ppt
TTests.pptTTests.ppt
TTests.pptMUzair21
 
SimpleLinearRegressionAnalysisWithExamples.ppt
SimpleLinearRegressionAnalysisWithExamples.pptSimpleLinearRegressionAnalysisWithExamples.ppt
SimpleLinearRegressionAnalysisWithExamples.pptAdnanAli861711
 
Linear regression.ppt
Linear regression.pptLinear regression.ppt
Linear regression.pptbranlymbunga1
 
Slideset Simple Linear Regression models.ppt
Slideset Simple Linear Regression models.pptSlideset Simple Linear Regression models.ppt
Slideset Simple Linear Regression models.pptrahulrkmgb09
 
lecture13.ppt
lecture13.pptlecture13.ppt
lecture13.pptarkian3
 
Basics of Structural Equation Modeling
Basics of Structural Equation ModelingBasics of Structural Equation Modeling
Basics of Structural Equation Modelingsmackinnon
 
A. spanos slides ch14-2013 (4)
A. spanos slides ch14-2013 (4)A. spanos slides ch14-2013 (4)
A. spanos slides ch14-2013 (4)jemille6
 
Bayesian Variable Selection in Linear Regression and A Comparison
Bayesian Variable Selection in Linear Regression and A ComparisonBayesian Variable Selection in Linear Regression and A Comparison
Bayesian Variable Selection in Linear Regression and A ComparisonAtilla YARDIMCI
 
Bayesian Structural Equations Modeling (SEM)
Bayesian Structural Equations Modeling (SEM)Bayesian Structural Equations Modeling (SEM)
Bayesian Structural Equations Modeling (SEM)Hamy Temkit
 
Asymptotic properties of bayes factor in one way repeated measurements model
Asymptotic properties of bayes factor in one  way repeated measurements modelAsymptotic properties of bayes factor in one  way repeated measurements model
Asymptotic properties of bayes factor in one way repeated measurements modelAlexander Decker
 
Asymptotic properties of bayes factor in one way repeated measurements model
Asymptotic properties of bayes factor in one  way repeated measurements modelAsymptotic properties of bayes factor in one  way repeated measurements model
Asymptotic properties of bayes factor in one way repeated measurements modelAlexander Decker
 

Ähnlich wie Testing as estimation: the demise of the Bayes factor (20)

ABC short course: model choice chapter
ABC short course: model choice chapterABC short course: model choice chapter
ABC short course: model choice chapter
 
Cointegration analysis: Modelling the complex interdependencies between finan...
Cointegration analysis: Modelling the complex interdependencies between finan...Cointegration analysis: Modelling the complex interdependencies between finan...
Cointegration analysis: Modelling the complex interdependencies between finan...
 
Causal Inference Opening Workshop - Bayesian Nonparametric Models for Treatme...
Causal Inference Opening Workshop - Bayesian Nonparametric Models for Treatme...Causal Inference Opening Workshop - Bayesian Nonparametric Models for Treatme...
Causal Inference Opening Workshop - Bayesian Nonparametric Models for Treatme...
 
Intro to Model Selection
Intro to Model SelectionIntro to Model Selection
Intro to Model Selection
 
TTests.ppt
TTests.pptTTests.ppt
TTests.ppt
 
SimpleLinearRegressionAnalysisWithExamples.ppt
SimpleLinearRegressionAnalysisWithExamples.pptSimpleLinearRegressionAnalysisWithExamples.ppt
SimpleLinearRegressionAnalysisWithExamples.ppt
 
Linear regression.ppt
Linear regression.pptLinear regression.ppt
Linear regression.ppt
 
lecture13.ppt
lecture13.pptlecture13.ppt
lecture13.ppt
 
lecture13.ppt
lecture13.pptlecture13.ppt
lecture13.ppt
 
Slideset Simple Linear Regression models.ppt
Slideset Simple Linear Regression models.pptSlideset Simple Linear Regression models.ppt
Slideset Simple Linear Regression models.ppt
 
lecture13.ppt
lecture13.pptlecture13.ppt
lecture13.ppt
 
lecture13.ppt
lecture13.pptlecture13.ppt
lecture13.ppt
 
HW1 MIT Fall 2005
HW1 MIT Fall 2005HW1 MIT Fall 2005
HW1 MIT Fall 2005
 
Basics of Structural Equation Modeling
Basics of Structural Equation ModelingBasics of Structural Equation Modeling
Basics of Structural Equation Modeling
 
A. spanos slides ch14-2013 (4)
A. spanos slides ch14-2013 (4)A. spanos slides ch14-2013 (4)
A. spanos slides ch14-2013 (4)
 
Bayesian Variable Selection in Linear Regression and A Comparison
Bayesian Variable Selection in Linear Regression and A ComparisonBayesian Variable Selection in Linear Regression and A Comparison
Bayesian Variable Selection in Linear Regression and A Comparison
 
Bayesian Structural Equations Modeling (SEM)
Bayesian Structural Equations Modeling (SEM)Bayesian Structural Equations Modeling (SEM)
Bayesian Structural Equations Modeling (SEM)
 
Bayesian_SEM_HT
Bayesian_SEM_HTBayesian_SEM_HT
Bayesian_SEM_HT
 
Asymptotic properties of bayes factor in one way repeated measurements model
Asymptotic properties of bayes factor in one  way repeated measurements modelAsymptotic properties of bayes factor in one  way repeated measurements model
Asymptotic properties of bayes factor in one way repeated measurements model
 
Asymptotic properties of bayes factor in one way repeated measurements model
Asymptotic properties of bayes factor in one  way repeated measurements modelAsymptotic properties of bayes factor in one  way repeated measurements model
Asymptotic properties of bayes factor in one way repeated measurements model
 

Mehr von Christian Robert

Asymptotics of ABC, lecture, Collège de France
Asymptotics of ABC, lecture, Collège de FranceAsymptotics of ABC, lecture, Collège de France
Asymptotics of ABC, lecture, Collège de FranceChristian Robert
 
Workshop in honour of Don Poskitt and Gael Martin
Workshop in honour of Don Poskitt and Gael MartinWorkshop in honour of Don Poskitt and Gael Martin
Workshop in honour of Don Poskitt and Gael MartinChristian Robert
 
How many components in a mixture?
How many components in a mixture?How many components in a mixture?
How many components in a mixture?Christian Robert
 
Testing for mixtures at BNP 13
Testing for mixtures at BNP 13Testing for mixtures at BNP 13
Testing for mixtures at BNP 13Christian Robert
 
Inferring the number of components: dream or reality?
Inferring the number of components: dream or reality?Inferring the number of components: dream or reality?
Inferring the number of components: dream or reality?Christian Robert
 
Testing for mixtures by seeking components
Testing for mixtures by seeking componentsTesting for mixtures by seeking components
Testing for mixtures by seeking componentsChristian Robert
 
discussion on Bayesian restricted likelihood
discussion on Bayesian restricted likelihooddiscussion on Bayesian restricted likelihood
discussion on Bayesian restricted likelihoodChristian Robert
 
Coordinate sampler : A non-reversible Gibbs-like sampler
Coordinate sampler : A non-reversible Gibbs-like samplerCoordinate sampler : A non-reversible Gibbs-like sampler
Coordinate sampler : A non-reversible Gibbs-like samplerChristian Robert
 
CISEA 2019: ABC consistency and convergence
CISEA 2019: ABC consistency and convergenceCISEA 2019: ABC consistency and convergence
CISEA 2019: ABC consistency and convergenceChristian Robert
 
a discussion of Chib, Shin, and Simoni (2017-8) Bayesian moment models
a discussion of Chib, Shin, and Simoni (2017-8) Bayesian moment modelsa discussion of Chib, Shin, and Simoni (2017-8) Bayesian moment models
a discussion of Chib, Shin, and Simoni (2017-8) Bayesian moment modelsChristian Robert
 
ABC based on Wasserstein distances
ABC based on Wasserstein distancesABC based on Wasserstein distances
ABC based on Wasserstein distancesChristian Robert
 

Mehr von Christian Robert (20)

Asymptotics of ABC, lecture, Collège de France
Asymptotics of ABC, lecture, Collège de FranceAsymptotics of ABC, lecture, Collège de France
Asymptotics of ABC, lecture, Collège de France
 
Workshop in honour of Don Poskitt and Gael Martin
Workshop in honour of Don Poskitt and Gael MartinWorkshop in honour of Don Poskitt and Gael Martin
Workshop in honour of Don Poskitt and Gael Martin
 
discussion of ICML23.pdf
discussion of ICML23.pdfdiscussion of ICML23.pdf
discussion of ICML23.pdf
 
How many components in a mixture?
How many components in a mixture?How many components in a mixture?
How many components in a mixture?
 
restore.pdf
restore.pdfrestore.pdf
restore.pdf
 
Testing for mixtures at BNP 13
Testing for mixtures at BNP 13Testing for mixtures at BNP 13
Testing for mixtures at BNP 13
 
Inferring the number of components: dream or reality?
Inferring the number of components: dream or reality?Inferring the number of components: dream or reality?
Inferring the number of components: dream or reality?
 
CDT 22 slides.pdf
CDT 22 slides.pdfCDT 22 slides.pdf
CDT 22 slides.pdf
 
Testing for mixtures by seeking components
Testing for mixtures by seeking componentsTesting for mixtures by seeking components
Testing for mixtures by seeking components
 
discussion on Bayesian restricted likelihood
discussion on Bayesian restricted likelihooddiscussion on Bayesian restricted likelihood
discussion on Bayesian restricted likelihood
 
ABC-Gibbs
ABC-GibbsABC-Gibbs
ABC-Gibbs
 
Coordinate sampler : A non-reversible Gibbs-like sampler
Coordinate sampler : A non-reversible Gibbs-like samplerCoordinate sampler : A non-reversible Gibbs-like sampler
Coordinate sampler : A non-reversible Gibbs-like sampler
 
eugenics and statistics
eugenics and statisticseugenics and statistics
eugenics and statistics
 
ABC-Gibbs
ABC-GibbsABC-Gibbs
ABC-Gibbs
 
asymptotics of ABC
asymptotics of ABCasymptotics of ABC
asymptotics of ABC
 
ABC-Gibbs
ABC-GibbsABC-Gibbs
ABC-Gibbs
 
the ABC of ABC
the ABC of ABCthe ABC of ABC
the ABC of ABC
 
CISEA 2019: ABC consistency and convergence
CISEA 2019: ABC consistency and convergenceCISEA 2019: ABC consistency and convergence
CISEA 2019: ABC consistency and convergence
 
a discussion of Chib, Shin, and Simoni (2017-8) Bayesian moment models
a discussion of Chib, Shin, and Simoni (2017-8) Bayesian moment modelsa discussion of Chib, Shin, and Simoni (2017-8) Bayesian moment models
a discussion of Chib, Shin, and Simoni (2017-8) Bayesian moment models
 
ABC based on Wasserstein distances
ABC based on Wasserstein distancesABC based on Wasserstein distances
ABC based on Wasserstein distances
 

Kürzlich hochgeladen

9654467111 Call Girls In Raj Nagar Delhi Short 1500 Night 6000
9654467111 Call Girls In Raj Nagar Delhi Short 1500 Night 60009654467111 Call Girls In Raj Nagar Delhi Short 1500 Night 6000
9654467111 Call Girls In Raj Nagar Delhi Short 1500 Night 6000Sapana Sha
 
GBSN - Microbiology (Unit 1)
GBSN - Microbiology (Unit 1)GBSN - Microbiology (Unit 1)
GBSN - Microbiology (Unit 1)Areesha Ahmad
 
Nanoparticles synthesis and characterization​ ​
Nanoparticles synthesis and characterization​  ​Nanoparticles synthesis and characterization​  ​
Nanoparticles synthesis and characterization​ ​kaibalyasahoo82800
 
❤Jammu Kashmir Call Girls 8617697112 Personal Whatsapp Number 💦✅.
❤Jammu Kashmir Call Girls 8617697112 Personal Whatsapp Number 💦✅.❤Jammu Kashmir Call Girls 8617697112 Personal Whatsapp Number 💦✅.
❤Jammu Kashmir Call Girls 8617697112 Personal Whatsapp Number 💦✅.Nitya salvi
 
Biopesticide (2).pptx .This slides helps to know the different types of biop...
Biopesticide (2).pptx  .This slides helps to know the different types of biop...Biopesticide (2).pptx  .This slides helps to know the different types of biop...
Biopesticide (2).pptx .This slides helps to know the different types of biop...RohitNehra6
 
VIRUSES structure and classification ppt by Dr.Prince C P
VIRUSES structure and classification ppt by Dr.Prince C PVIRUSES structure and classification ppt by Dr.Prince C P
VIRUSES structure and classification ppt by Dr.Prince C PPRINCE C P
 
Vip profile Call Girls In Lonavala 9748763073 For Genuine Sex Service At Just...
Vip profile Call Girls In Lonavala 9748763073 For Genuine Sex Service At Just...Vip profile Call Girls In Lonavala 9748763073 For Genuine Sex Service At Just...
Vip profile Call Girls In Lonavala 9748763073 For Genuine Sex Service At Just...Monika Rani
 
Stunning ➥8448380779▻ Call Girls In Panchshil Enclave Delhi NCR
Stunning ➥8448380779▻ Call Girls In Panchshil Enclave Delhi NCRStunning ➥8448380779▻ Call Girls In Panchshil Enclave Delhi NCR
Stunning ➥8448380779▻ Call Girls In Panchshil Enclave Delhi NCRDelhi Call girls
 
Discovery of an Accretion Streamer and a Slow Wide-angle Outflow around FUOri...
Discovery of an Accretion Streamer and a Slow Wide-angle Outflow around FUOri...Discovery of an Accretion Streamer and a Slow Wide-angle Outflow around FUOri...
Discovery of an Accretion Streamer and a Slow Wide-angle Outflow around FUOri...Sérgio Sacani
 
Seismic Method Estimate velocity from seismic data.pptx
Seismic Method Estimate velocity from seismic  data.pptxSeismic Method Estimate velocity from seismic  data.pptx
Seismic Method Estimate velocity from seismic data.pptxAlMamun560346
 
Nightside clouds and disequilibrium chemistry on the hot Jupiter WASP-43b
Nightside clouds and disequilibrium chemistry on the hot Jupiter WASP-43bNightside clouds and disequilibrium chemistry on the hot Jupiter WASP-43b
Nightside clouds and disequilibrium chemistry on the hot Jupiter WASP-43bSérgio Sacani
 
Asymmetry in the atmosphere of the ultra-hot Jupiter WASP-76 b
Asymmetry in the atmosphere of the ultra-hot Jupiter WASP-76 bAsymmetry in the atmosphere of the ultra-hot Jupiter WASP-76 b
Asymmetry in the atmosphere of the ultra-hot Jupiter WASP-76 bSérgio Sacani
 
Zoology 4th semester series (krishna).pdf
Zoology 4th semester series (krishna).pdfZoology 4th semester series (krishna).pdf
Zoology 4th semester series (krishna).pdfSumit Kumar yadav
 
Botany 4th semester series (krishna).pdf
Botany 4th semester series (krishna).pdfBotany 4th semester series (krishna).pdf
Botany 4th semester series (krishna).pdfSumit Kumar yadav
 
Chemistry 4th semester series (krishna).pdf
Chemistry 4th semester series (krishna).pdfChemistry 4th semester series (krishna).pdf
Chemistry 4th semester series (krishna).pdfSumit Kumar yadav
 
Biogenic Sulfur Gases as Biosignatures on Temperate Sub-Neptune Waterworlds
Biogenic Sulfur Gases as Biosignatures on Temperate Sub-Neptune WaterworldsBiogenic Sulfur Gases as Biosignatures on Temperate Sub-Neptune Waterworlds
Biogenic Sulfur Gases as Biosignatures on Temperate Sub-Neptune WaterworldsSérgio Sacani
 
Recombinant DNA technology (Immunological screening)
Recombinant DNA technology (Immunological screening)Recombinant DNA technology (Immunological screening)
Recombinant DNA technology (Immunological screening)PraveenaKalaiselvan1
 
SAMASTIPUR CALL GIRL 7857803690 LOW PRICE ESCORT SERVICE
SAMASTIPUR CALL GIRL 7857803690  LOW PRICE  ESCORT SERVICESAMASTIPUR CALL GIRL 7857803690  LOW PRICE  ESCORT SERVICE
SAMASTIPUR CALL GIRL 7857803690 LOW PRICE ESCORT SERVICEayushi9330
 
Pests of cotton_Borer_Pests_Binomics_Dr.UPR.pdf
Pests of cotton_Borer_Pests_Binomics_Dr.UPR.pdfPests of cotton_Borer_Pests_Binomics_Dr.UPR.pdf
Pests of cotton_Borer_Pests_Binomics_Dr.UPR.pdfPirithiRaju
 
Isotopic evidence of long-lived volcanism on Io
Isotopic evidence of long-lived volcanism on IoIsotopic evidence of long-lived volcanism on Io
Isotopic evidence of long-lived volcanism on IoSérgio Sacani
 

Kürzlich hochgeladen (20)

9654467111 Call Girls In Raj Nagar Delhi Short 1500 Night 6000
9654467111 Call Girls In Raj Nagar Delhi Short 1500 Night 60009654467111 Call Girls In Raj Nagar Delhi Short 1500 Night 6000
9654467111 Call Girls In Raj Nagar Delhi Short 1500 Night 6000
 
GBSN - Microbiology (Unit 1)
GBSN - Microbiology (Unit 1)GBSN - Microbiology (Unit 1)
GBSN - Microbiology (Unit 1)
 
Nanoparticles synthesis and characterization​ ​
Nanoparticles synthesis and characterization​  ​Nanoparticles synthesis and characterization​  ​
Nanoparticles synthesis and characterization​ ​
 
❤Jammu Kashmir Call Girls 8617697112 Personal Whatsapp Number 💦✅.
❤Jammu Kashmir Call Girls 8617697112 Personal Whatsapp Number 💦✅.❤Jammu Kashmir Call Girls 8617697112 Personal Whatsapp Number 💦✅.
❤Jammu Kashmir Call Girls 8617697112 Personal Whatsapp Number 💦✅.
 
Biopesticide (2).pptx .This slides helps to know the different types of biop...
Biopesticide (2).pptx  .This slides helps to know the different types of biop...Biopesticide (2).pptx  .This slides helps to know the different types of biop...
Biopesticide (2).pptx .This slides helps to know the different types of biop...
 
VIRUSES structure and classification ppt by Dr.Prince C P
VIRUSES structure and classification ppt by Dr.Prince C PVIRUSES structure and classification ppt by Dr.Prince C P
VIRUSES structure and classification ppt by Dr.Prince C P
 
Vip profile Call Girls In Lonavala 9748763073 For Genuine Sex Service At Just...
Vip profile Call Girls In Lonavala 9748763073 For Genuine Sex Service At Just...Vip profile Call Girls In Lonavala 9748763073 For Genuine Sex Service At Just...
Vip profile Call Girls In Lonavala 9748763073 For Genuine Sex Service At Just...
 
Stunning ➥8448380779▻ Call Girls In Panchshil Enclave Delhi NCR
Stunning ➥8448380779▻ Call Girls In Panchshil Enclave Delhi NCRStunning ➥8448380779▻ Call Girls In Panchshil Enclave Delhi NCR
Stunning ➥8448380779▻ Call Girls In Panchshil Enclave Delhi NCR
 
Discovery of an Accretion Streamer and a Slow Wide-angle Outflow around FUOri...
Discovery of an Accretion Streamer and a Slow Wide-angle Outflow around FUOri...Discovery of an Accretion Streamer and a Slow Wide-angle Outflow around FUOri...
Discovery of an Accretion Streamer and a Slow Wide-angle Outflow around FUOri...
 
Seismic Method Estimate velocity from seismic data.pptx
Seismic Method Estimate velocity from seismic  data.pptxSeismic Method Estimate velocity from seismic  data.pptx
Seismic Method Estimate velocity from seismic data.pptx
 
Nightside clouds and disequilibrium chemistry on the hot Jupiter WASP-43b
Nightside clouds and disequilibrium chemistry on the hot Jupiter WASP-43bNightside clouds and disequilibrium chemistry on the hot Jupiter WASP-43b
Nightside clouds and disequilibrium chemistry on the hot Jupiter WASP-43b
 
Asymmetry in the atmosphere of the ultra-hot Jupiter WASP-76 b
Asymmetry in the atmosphere of the ultra-hot Jupiter WASP-76 bAsymmetry in the atmosphere of the ultra-hot Jupiter WASP-76 b
Asymmetry in the atmosphere of the ultra-hot Jupiter WASP-76 b
 
Zoology 4th semester series (krishna).pdf
Zoology 4th semester series (krishna).pdfZoology 4th semester series (krishna).pdf
Zoology 4th semester series (krishna).pdf
 
Botany 4th semester series (krishna).pdf
Botany 4th semester series (krishna).pdfBotany 4th semester series (krishna).pdf
Botany 4th semester series (krishna).pdf
 
Chemistry 4th semester series (krishna).pdf
Chemistry 4th semester series (krishna).pdfChemistry 4th semester series (krishna).pdf
Chemistry 4th semester series (krishna).pdf
 
Biogenic Sulfur Gases as Biosignatures on Temperate Sub-Neptune Waterworlds
Biogenic Sulfur Gases as Biosignatures on Temperate Sub-Neptune WaterworldsBiogenic Sulfur Gases as Biosignatures on Temperate Sub-Neptune Waterworlds
Biogenic Sulfur Gases as Biosignatures on Temperate Sub-Neptune Waterworlds
 
Recombinant DNA technology (Immunological screening)
Recombinant DNA technology (Immunological screening)Recombinant DNA technology (Immunological screening)
Recombinant DNA technology (Immunological screening)
 
SAMASTIPUR CALL GIRL 7857803690 LOW PRICE ESCORT SERVICE
SAMASTIPUR CALL GIRL 7857803690  LOW PRICE  ESCORT SERVICESAMASTIPUR CALL GIRL 7857803690  LOW PRICE  ESCORT SERVICE
SAMASTIPUR CALL GIRL 7857803690 LOW PRICE ESCORT SERVICE
 
Pests of cotton_Borer_Pests_Binomics_Dr.UPR.pdf
Pests of cotton_Borer_Pests_Binomics_Dr.UPR.pdfPests of cotton_Borer_Pests_Binomics_Dr.UPR.pdf
Pests of cotton_Borer_Pests_Binomics_Dr.UPR.pdf
 
Isotopic evidence of long-lived volcanism on Io
Isotopic evidence of long-lived volcanism on IoIsotopic evidence of long-lived volcanism on Io
Isotopic evidence of long-lived volcanism on Io
 

Testing as estimation: the demise of the Bayes factor

  • 1. Testing as estimation: the demise of the Bayes factors Testing as estimation: the demise of the Bayes factors Christian P. Robert Universit´e Paris-Dauphine and University of Warwick arXiv:1412.2044 with K. Kamary, K. Mengersen, and J. Rousseau
  • 2. Testing as estimation: the demise of the Bayes factors Outline Introduction Testing problems as estimating mixture models Illustrations Asymptotic consistency Conclusion
  • 3. Testing as estimation: the demise of the Bayes factors Introduction Testing hypotheses Hypothesis testing central problem of statistical inference dramatically differentiating feature between classical and Bayesian paradigms wide open to controversy and divergent opinions, includ. within the Bayesian community non-informative Bayesian testing case mostly unresolved, witness the Jeffreys–Lindley paradox [Berger (2003), Mayo & Cox (2006), Gelman (2008)]
  • 4. Testing as estimation: the demise of the Bayes factors Introduction Besting hypotheses Bayesian model selection as comparison of k potential statistical models towards the selection of model that fits the data “best” mostly accepted perspective: it does not primarily seek to identify which model is “true”, but compares fits tools like Bayes factor naturally include a penalisation addressing model complexity, mimicked by Bayes Information (BIC) and Deviance Information (DIC) criteria posterior predictive tools successfully advocated in Gelman et al. (2013) even though they involve double use of data
  • 5. Testing as estimation: the demise of the Bayes factors Introduction Besting hypotheses Bayesian model selection as comparison of k potential statistical models towards the selection of model that fits the data “best” mostly accepted perspective: it does not primarily seek to identify which model is “true”, but compares fits tools like Bayes factor naturally include a penalisation addressing model complexity, mimicked by Bayes Information (BIC) and Deviance Information (DIC) criteria posterior predictive tools successfully advocated in Gelman et al. (2013) even though they involve double use of data
  • 6. Testing as estimation: the demise of the Bayes factors Introduction Bayesian modelling Standard Bayesian approach to testing: consider two families of models, one for each of the hypotheses under comparison, M1 : x ∼ f1(x|θ1) , θ1 ∈ Θ1 and M2 : x ∼ f2(x|θ2) , θ2 ∈ Θ2 , and associate with each model a prior distribution, θ1 ∼ π1(θ1) and θ2 ∼ π2(θ2) , [Jeffreys, 1939]
  • 7. Testing as estimation: the demise of the Bayes factors Introduction Bayesian modelling Standard Bayesian approach to testing: consider two families of models, one for each of the hypotheses under comparison, M1 : x ∼ f1(x|θ1) , θ1 ∈ Θ1 and M2 : x ∼ f2(x|θ2) , θ2 ∈ Θ2 , in order to compare the marginal likelihoods m1(x) = Θ1 f1(x|θ1) π1(θ1) dθ1 and m2(x) = Θ2 f2(x|θ2) π1(θ2) dθ2 [Jeffreys, 1939]
  • 8. Testing as estimation: the demise of the Bayes factors Introduction Bayesian modelling Standard Bayesian approach to testing: consider two families of models, one for each of the hypotheses under comparison, M1 : x ∼ f1(x|θ1) , θ1 ∈ Θ1 and M2 : x ∼ f2(x|θ2) , θ2 ∈ Θ2 , either through Bayes factor or posterior probability, B12 = m1(x) m2(x) , P(M1|x) = ω1m1(x) ω1m1(x) + ω2m2(x) ; the latter depends on the prior weights ωi [Jeffreys, 1939]
  • 9. Testing as estimation: the demise of the Bayes factors Introduction Bayesian modelling Standard Bayesian approach to testing: consider two families of models, one for each of the hypotheses under comparison, M1 : x ∼ f1(x|θ1) , θ1 ∈ Θ1 and M2 : x ∼ f2(x|θ2) , θ2 ∈ Θ2 , Bayesian decision step comparing Bayes factor B12 with threshold value of one or comparing posterior probability P(M1|x) with bound α [Jeffreys, 1939]
  • 10. Testing as estimation: the demise of the Bayes factors Introduction Some difficulties tension between (i) posterior probabilities justified by binary loss but depending on unnatural prior weights, and (ii) Bayes factors that eliminate dependence but lose direct connection with posterior, unless prior weights are integrated within loss delicate interpretation (or calibration) of the strength of the Bayes factor towards supporting a given hypothesis or model: not a Bayesian decision rule! difficulty with posterior probabilities: tendency to interpret them as p-values while they only report respective strengths of fitting to both models
  • 11. Testing as estimation: the demise of the Bayes factors Introduction Some further difficulties long-lasting impact of the prior modeling, i.e., choice of prior distributions on both parameter spaces under comparison, despite overall consistency for Bayes factor major discontinuity in use of improper priors, not justified in most testing situations, leading to ad hoc solutions (zoo), where data is either used twice or split artificially binary (accept vs. reject) outcome more suited for immediate decision (if any) than for model evaluation, connected with rudimentary loss function
  • 12. Testing as estimation: the demise of the Bayes factors Introduction Lindley’s paradox In a normal mean testing problem, ¯xn ∼ N(θ, σ2 /n) , H0 : θ = θ0 , under Jeffreys prior, θ ∼ N(θ0, σ2), the Bayes factor B01(tn) = (1 + n)1/2 exp −nt2 n/2[1 + n] , where tn = √ n|¯xn − θ0|/σ, satisfies B01(tn) n−→∞ −→ ∞ [assuming a fixed tn] [Lindley, 1957]
  • 13. Testing as estimation: the demise of the Bayes factors Introduction A strong impropriety Improper priors not allowed in Bayes factors: If Θ1 π1(dθ1) = ∞ or Θ2 π2(dθ2) = ∞ then π1 or π2 cannot be coherently normalised while the normalisation matters in the Bayes factor B12 Lack of mathematical justification for “common nuisance parameter” [and prior of] [Berger, Pericchi, and Varshavsky, 1998; Marin and Robert, 2013]
  • 14. Testing as estimation: the demise of the Bayes factors Introduction A strong impropriety Improper priors not allowed in Bayes factors: If Θ1 π1(dθ1) = ∞ or Θ2 π2(dθ2) = ∞ then π1 or π2 cannot be coherently normalised while the normalisation matters in the Bayes factor B12 Lack of mathematical justification for “common nuisance parameter” [and prior of] [Berger, Pericchi, and Varshavsky, 1998; Marin and Robert, 2013]
  • 15. Testing as estimation: the demise of the Bayes factors Testing problems as estimating mixture models Paradigm shift New proposal as paradigm shift in Bayesian processing of hypothesis testing and of model selection convergent and naturally interpretable solution extended use of improper priors abandonment of the Neyman-Pearson decision framework natural strenght of evidence Simple representation of the testing problem as a two-component mixture estimation problem where the weights are formally equal to 0 or 1
  • 16. Testing as estimation: the demise of the Bayes factors Testing problems as estimating mixture models Paradigm shift New proposal as paradigm shift in Bayesian processing of hypothesis testing and of model selection convergent and naturally interpretable solution extended use of improper priors abandonment of the Neyman-Pearson decision framework natural strenght of evidence Simple representation of the testing problem as a two-component mixture estimation problem where the weights are formally equal to 0 or 1
  • 17. Testing as estimation: the demise of the Bayes factors Testing problems as estimating mixture models Paradigm shift Simple representation of the testing problem as a two-component mixture estimation problem where the weights are formally equal to 0 or 1 Approach inspired from consistency result of Rousseau and Mengersen (2011) on estimated overfitting mixtures Mixture representation not equivalent to use of a posterior probability More natural approach to testing, while sparse in parameters Calibration of the posterior distribution of mixture weight, while moving away from artificial notion of the posterior probability of a model
  • 18. Testing as estimation: the demise of the Bayes factors Testing problems as estimating mixture models Encompassing mixture model Idea: Given two statistical models, M1 : x ∼ f1(x|θ1) , θ1 ∈ Θ1 and M2 : x ∼ f2(x|θ2) , θ2 ∈ Θ2 , embed both within an encompassing mixture Mα : x ∼ αf1(x|θ1) + (1 − α)f2(x|θ2) , 0 α 1 (1) Note: Both models correspond to special cases of (1), one for α = 1 and one for α = 0 Draw inference on mixture representation (1), as if each observation was individually and independently produced by the mixture model
  • 19. Testing as estimation: the demise of the Bayes factors Testing problems as estimating mixture models Encompassing mixture model Idea: Given two statistical models, M1 : x ∼ f1(x|θ1) , θ1 ∈ Θ1 and M2 : x ∼ f2(x|θ2) , θ2 ∈ Θ2 , embed both within an encompassing mixture Mα : x ∼ αf1(x|θ1) + (1 − α)f2(x|θ2) , 0 α 1 (1) Note: Both models correspond to special cases of (1), one for α = 1 and one for α = 0 Draw inference on mixture representation (1), as if each observation was individually and independently produced by the mixture model
  • 20. Testing as estimation: the demise of the Bayes factors Testing problems as estimating mixture models Encompassing mixture model Idea: Given two statistical models, M1 : x ∼ f1(x|θ1) , θ1 ∈ Θ1 and M2 : x ∼ f2(x|θ2) , θ2 ∈ Θ2 , embed both within an encompassing mixture Mα : x ∼ αf1(x|θ1) + (1 − α)f2(x|θ2) , 0 α 1 (1) Note: Both models correspond to special cases of (1), one for α = 1 and one for α = 0 Draw inference on mixture representation (1), as if each observation was individually and independently produced by the mixture model
  • 21. Testing as estimation: the demise of the Bayes factors Testing problems as estimating mixture models Inferential motivations Sounds like approximation to the real problem, but definitive advantages to shift: Bayes estimate of the weight α replaces posterior probability of model M1, equally convergent indicator of which model is “true”, while avoiding artificial prior probabilities on model indices, ω1 and ω2, and 0 − 1 loss setting posterior on α provides measure of proximity to models, while being interpretable as data propensity to stand within one model further allows for alternative perspectives on testing and model choice, like predictive tools, cross-validation, and information indices like WAIC
  • 22. Testing as estimation: the demise of the Bayes factors Testing problems as estimating mixture models Computational motivations avoids problematic computations of marginal likelihoods, since standard algorithms are available for Bayesian mixture estimation straightforward extension to finite collection of models, which considers all models at once and eliminates least likely models by simulation eliminates famous difficulty of label switching that plagues both Bayes estimation and computation: components are no longer exchangeable posterior distribution on α evaluates more thoroughly strength of support for a given model than the single figure posterior probability variability of posterior distribution on α allows for a more thorough assessment of the strength of this support
  • 23. Testing as estimation: the demise of the Bayes factors Testing problems as estimating mixture models Noninformative motivations novel Bayesian feature: a mixture model acknowledges possibility that, for a finite dataset, both models or none could be acceptable standard (proper and informative) prior modeling can be processed in this setting, but non-informative (improper) priors also are manageable, provided both models first reparameterised into shared parameters, e.g. location and scale parameters in special case when all parameters are common Mα : x ∼ αf1(x|θ) + (1 − α)f2(x|θ) , 0 α 1 if θ is a location parameter, a flat prior π(θ) ∝ 1 is available
  • 24. Testing as estimation: the demise of the Bayes factors Testing problems as estimating mixture models Weakly informative motivations using the same parameters or some identical parameters on both components highlights that opposition between the two components is not an issue of enjoying different parameters common parameters are nuisance parameters, easily integrated prior model weights ωi rarely discussed in classical Bayesian approach, with linear impact on posterior probabilities prior modeling only involves selecting a prior on α, e.g., α ∼ B(a0, a0) while a0 impacts posterior on α, it always leads to mass accumulation near 1 or 0, i.e. favours most likely model sensitivity analysis straightforward to carry approach easily calibrated by parametric boostrap providing reference posterior of α under each model natural Metropolis–Hastings alternative
  • 25. Testing as estimation: the demise of the Bayes factors Illustrations Poisson/Geometric example choice betwen Poisson P(λ) and Geometric Geo(p) distribution mixture with common parameter λ Mα : αP(λ) + (1 − α)Geo(1/1+λ) Allows for Jeffreys prior since resulting posterior is proper independent Metropolis–within–Gibbs with proposal distribution on λ equal to Poisson posterior (with acceptance rate larger than 75%)
  • 26. Testing as estimation: the demise of the Bayes factors Illustrations Poisson/Geometric example choice betwen Poisson P(λ) and Geometric Geo(p) distribution mixture with common parameter λ Mα : αP(λ) + (1 − α)Geo(1/1+λ) Allows for Jeffreys prior since resulting posterior is proper independent Metropolis–within–Gibbs with proposal distribution on λ equal to Poisson posterior (with acceptance rate larger than 75%)
  • 27. Testing as estimation: the demise of the Bayes factors Illustrations Poisson/Geometric example choice betwen Poisson P(λ) and Geometric Geo(p) distribution mixture with common parameter λ Mα : αP(λ) + (1 − α)Geo(1/1+λ) Allows for Jeffreys prior since resulting posterior is proper independent Metropolis–within–Gibbs with proposal distribution on λ equal to Poisson posterior (with acceptance rate larger than 75%)
  • 28. Testing as estimation: the demise of the Bayes factors Illustrations Beta prior When α ∼ Be(a0, a0) prior, full conditional posterior α ∼ Be(n1(ζ) + a0, n2(ζ) + a0) Exact Bayes factor opposing Poisson and Geometric B12 = nn¯xn n i=1 xi ! Γ n + 2 + n i=1 xi Γ(n + 2) although undefined from a purely mathematical viewpoint
  • 29. Testing as estimation: the demise of the Bayes factors Illustrations Weight estimation 1e-04 0.001 0.01 0.1 0.2 0.3 0.4 0.5 0.9900.9920.9940.9960.9981.000 Posterior medians of α for 100 Poisson P(4) datasets of size n = 1000, for a0 = .0001, .001, .01, .1, .2, .3, .4, .5. Each posterior approximation is based on 104 Metropolis-Hastings iterations.
  • 30. Testing as estimation: the demise of the Bayes factors Illustrations Consistency 0 1 2 3 4 5 6 7 0.00.20.40.60.81.0 a0=.1 log(sample size) 0 1 2 3 4 5 6 7 0.00.20.40.60.81.0 a0=.3 log(sample size) 0 1 2 3 4 5 6 7 0.00.20.40.60.81.0 a0=.5 log(sample size) Posterior means (sky-blue) and medians (grey-dotted) of α, over 100 Poisson P(4) datasets for sample sizes from 1 to 1000.
  • 31. Testing as estimation: the demise of the Bayes factors Illustrations Behaviour of Bayes factor 0 1 2 3 4 5 6 7 0.00.20.40.60.81.0 log(sample size) 0 1 2 3 4 5 6 7 0.00.20.40.60.81.0 log(sample size) 0 1 2 3 4 5 6 7 0.00.20.40.60.81.0 log(sample size) Comparison between P(M1|x) (red dotted area) and posterior medians of α (grey zone) for 100 Poisson P(4) datasets with sample sizes n between 1 and 1000, for a0 = .001, .1, .5
  • 32. Testing as estimation: the demise of the Bayes factors Illustrations Normal-normal comparison comparison of a normal N(θ1, 1) with a normal N(θ2, 2) distribution mixture with identical location parameter θ αN(θ, 1) + (1 − α)N(θ, 2) Jeffreys prior π(θ) = 1 can be used, since posterior is proper Reference (improper) Bayes factor B12 = 2 n−1/2 exp 1/4 n i=1 (xi − ¯x)2 ,
  • 33. Testing as estimation: the demise of the Bayes factors Illustrations Normal-normal comparison comparison of a normal N(θ1, 1) with a normal N(θ2, 2) distribution mixture with identical location parameter θ αN(θ, 1) + (1 − α)N(θ, 2) Jeffreys prior π(θ) = 1 can be used, since posterior is proper Reference (improper) Bayes factor B12 = 2 n−1/2 exp 1/4 n i=1 (xi − ¯x)2 ,
  • 34. Testing as estimation: the demise of the Bayes factors Illustrations Normal-normal comparison comparison of a normal N(θ1, 1) with a normal N(θ2, 2) distribution mixture with identical location parameter θ αN(θ, 1) + (1 − α)N(θ, 2) Jeffreys prior π(θ) = 1 can be used, since posterior is proper Reference (improper) Bayes factor B12 = 2 n−1/2 exp 1/4 n i=1 (xi − ¯x)2 ,
  • 35. Testing as estimation: the demise of the Bayes factors Illustrations Comparison with posterior probability 0 100 200 300 400 500 -50-40-30-20-100 a0=.1 sample size 0 100 200 300 400 500 -50-40-30-20-100 a0=.3 sample size 0 100 200 300 400 500 -50-40-30-20-100 a0=.4 sample size 0 100 200 300 400 500 -50-40-30-20-100 a0=.5 sample size Plots of ranges of log(n) log(1 − E[α|x]) (gray color) and log(1 − p(M1|x)) (red dotted) over 100 N(0, 1) samples as sample size n grows from 1 to 500. and α is the weight of N(0, 1) in the mixture model. The shaded areas indicate the range of the estimations and each plot is based on a Beta prior with a0 = .1, .2, .3, .4, .5, 1 and each posterior approximation is based on 104 iterations.
  • 36. Testing as estimation: the demise of the Bayes factors Illustrations Comments convergence to one boundary value as sample size n grows impact of hyperarameter a0 slowly vanishes as n increases, but present for moderate sample sizes when simulated sample is neither from N(θ1, 1) nor from N(θ2, 2), behaviour of posterior varies, depending on which distribution is closest
  • 37. Testing as estimation: the demise of the Bayes factors Illustrations Logit or Probit? binary dataset, R dataset about diabetes in 200 Pima Indian women with body mass index as explanatory variable comparison of logit and probit fits could be suitable. We are thus comparing both fits via our method M1 : yi | xi , θ1 ∼ B(1, pi ) where pi = exp(xi θ1) 1 + exp(xi θ1) M2 : yi | xi , θ2 ∼ B(1, qi ) where qi = Φ(xi θ2)
  • 38. Testing as estimation: the demise of the Bayes factors Illustrations Common parameterisation Local reparameterisation strategy that rescales parameters of the probit model M2 so that the MLE’s of both models coincide. [Choudhuty et al., 2007] Φ(xi θ2) ≈ exp(kxi θ2) 1 + exp(kxi θ2) and use best estimate of k to bring both parameters into coherency (k0, k1) = (θ01/θ02, θ11/θ12) , reparameterise M1 and M2 as M1 :yi | xi , θ ∼ B(1, pi ) where pi = exp(xi θ) 1 + exp(xi θ) M2 :yi | xi , θ ∼ B(1, qi ) where qi = Φ(xi (κ−1 θ)) , with κ−1θ = (θ0/k0, θ1/k1).
  • 39. Testing as estimation: the demise of the Bayes factors Illustrations Prior modelling Under default g-prior θ ∼ N2(0, n(XT X)−1 ) full conditional posterior distributions given allocations π(θ | y, X, ζ) ∝ exp i Iζi =1yi xi θ i;ζi =1[1 + exp(xi θ)] exp −θT (XT X)θ 2n × i;ζi =2 Φ(xi (κ−1 θ))yi (1 − Φ(xi (κ−1 θ)))(1−yi ) hence posterior distribution clearly defined
  • 40. Testing as estimation: the demise of the Bayes factors Illustrations Results Logistic Probit a0 α θ0 θ1 θ0 k0 θ1 k1 .1 .352 -4.06 .103 -2.51 .064 .2 .427 -4.03 .103 -2.49 .064 .3 .440 -4.02 .102 -2.49 .063 .4 .456 -4.01 .102 -2.48 .063 .5 .449 -4.05 .103 -2.51 .064 Histograms of posteriors of α in favour of logistic model where a0 = .1, .2, .3, .4, .5 for (a) Pima dataset, (b) Data from logistic model, (c) Data from probit
  • 41. Testing as estimation: the demise of the Bayes factors Illustrations Survival analysis models Testing hypothesis that data comes from a 1. log-Normal(φ, κ2), 2. Weibull(α, λ), or 3. log-Logistic(γ, δ) distribution Corresponding mixture given by the density α1 exp{−(log x − φ)2 /2κ2 }/ √ 2πxκ+ α2 α λ exp{−(x/λ)α }((x/λ)α−1 + α3(δ/γ)(x/γ)δ−1 /(1 + (x/γ)δ )2 where α1 + α2 + α3 = 1
  • 42. Testing as estimation: the demise of the Bayes factors Illustrations Survival analysis models Testing hypothesis that data comes from a 1. log-Normal(φ, κ2), 2. Weibull(α, λ), or 3. log-Logistic(γ, δ) distribution Corresponding mixture given by the density α1 exp{−(log x − φ)2 /2κ2 }/ √ 2πxκ+ α2 α λ exp{−(x/λ)α }((x/λ)α−1 + α3(δ/γ)(x/γ)δ−1 /(1 + (x/γ)δ )2 where α1 + α2 + α3 = 1
  • 43. Testing as estimation: the demise of the Bayes factors Illustrations Reparameterisation Looking for common parameter(s): φ = µ + γβ = ξ σ2 = π2 β2 /6 = ζ2 π2 /3 where γ ≈ 0.5772 is Euler-Mascheroni constant. Allows for a noninformative prior on the common location scale parameter, π(φ, σ2 ) = 1/σ2
  • 44. Testing as estimation: the demise of the Bayes factors Illustrations Reparameterisation Looking for common parameter(s): φ = µ + γβ = ξ σ2 = π2 β2 /6 = ζ2 π2 /3 where γ ≈ 0.5772 is Euler-Mascheroni constant. Allows for a noninformative prior on the common location scale parameter, π(φ, σ2 ) = 1/σ2
  • 45. Testing as estimation: the demise of the Bayes factors Illustrations Recovery .01 0.1 1.0 10.0 0.860.880.900.920.940.960.981.00 .01 0.1 1.0 10.0 0.000.020.040.060.08 Boxplots of the posterior distributions of the Normal weight α1 under the two scenarii: truth = Normal (left panel), truth = Gumbel (right panel), a0=0.01, 0.1, 1.0, 10.0 (from left to right in each panel) and n = 10, 000 simulated observations.
  • 46. Testing as estimation: the demise of the Bayes factors Asymptotic consistency Asymptotic consistency Posterior consistency holds for mixture testing procedure [under minor conditions] Two different cases the two models, M1 and M2, are well separated model M1 is a submodel of M2.
  • 47. Testing as estimation: the demise of the Bayes factors Asymptotic consistency Asymptotic consistency Posterior consistency holds for mixture testing procedure [under minor conditions] Two different cases the two models, M1 and M2, are well separated model M1 is a submodel of M2.
  • 48. Testing as estimation: the demise of the Bayes factors Asymptotic consistency Separated models Assumption: Models are separated, i.e. identifiability holds: ∀α, α ∈ [0, 1], ∀θj , θj , j = 1, 2 Pθ,α = Pθ ,α ⇒ α = α , θ = θ theorem Under above assumptions, then for all > 0, π [|α − α∗ | > |xn ] = op(1)
  • 49. Testing as estimation: the demise of the Bayes factors Asymptotic consistency Separated models Assumption: Models are separated, i.e. identifiability holds: ∀α, α ∈ [0, 1], ∀θj , θj , j = 1, 2 Pθ,α = Pθ ,α ⇒ α = α , θ = θ theorem If θj → fj,θj is C2 around θ∗ j , j = 1, 2, f1,θ∗ 1 − f2,θ∗ 2 , f1,θ∗ 1 , f2,θ∗ 2 are linearly independent in y and there exists δ > 0 such that f1,θ∗ 1 , f2,θ∗ 2 , sup |θ1−θ∗ 1 |<δ |D2 f1,θ1 |, sup |θ2−θ∗ 2 |<δ |D2 f2,θ2 | ∈ L1 then π |α − α∗ | > M log n/n xn = op(1).
  • 50. Testing as estimation: the demise of the Bayes factors Asymptotic consistency Separated models Assumption: Models are separated, i.e. identifiability holds: ∀α, α ∈ [0, 1], ∀θj , θj , j = 1, 2 Pθ,α = Pθ ,α ⇒ α = α , θ = θ theorem allows for interpretation of α under the posterior: If data xn is generated from model M1 then posterior on α concentrates around α = 1
  • 51. Testing as estimation: the demise of the Bayes factors Asymptotic consistency Embedded case Here M1 is a submodel of M2, i.e. θ2 = (θ1, ψ) and θ2 = (θ1, ψ0 = 0) corresponds to f2,θ2 ∈ M1 Same posterior concentration rate log n/n for estimating α when α∗ ∈ (0, 1) and ψ∗ = 0.
  • 52. Testing as estimation: the demise of the Bayes factors Asymptotic consistency Null case Case where ψ∗ = 0, i.e., f ∗ is in model M1 Two possible paths to approximate f ∗: either α goes to 1 (path 1) or ψ goes to 0 (path 2) New identifiability condition: Pθ,α = P∗ only if α = 1, θ1 = θ∗ 1, θ2 = (θ∗ 1, ψ) or α 1, θ1 = θ∗ 1, θ2 = (θ∗ 1, 0) Prior π(α, θ) = πα(α)π1(θ1)πψ(ψ), θ2 = (θ1, ψ) with common (prior on) θ1
  • 53. Testing as estimation: the demise of the Bayes factors Asymptotic consistency Null case Case where ψ∗ = 0, i.e., f ∗ is in model M1 Two possible paths to approximate f ∗: either α goes to 1 (path 1) or ψ goes to 0 (path 2) New identifiability condition: Pθ,α = P∗ only if α = 1, θ1 = θ∗ 1, θ2 = (θ∗ 1, ψ) or α 1, θ1 = θ∗ 1, θ2 = (θ∗ 1, 0) Prior π(α, θ) = πα(α)π1(θ1)πψ(ψ), θ2 = (θ1, ψ) with common (prior on) θ1
  • 54. Testing as estimation: the demise of the Bayes factors Asymptotic consistency Consistency theorem Given the mixture fθ1,ψ,α = αf1,θ1 + (1 − α)f2,θ1,ψ and a sample xn = (x1, · · · , xn) issued from f1,θ∗ 1 , under regularity assumptions, and an M > 0 such that π (α, θ); fθ,α − f ∗ 1 > M log n/n|xn = op(1). If α ∼ B(a1, a2), with a2 < d2, and if the prior πθ1,ψ is absolutely continuous with positive and continuous density at (θ∗ 1, 0), then for Mn −→ ∞ π |α − α∗ | > Mn(log n)γ / √ n|xn = op(1), γ = max((d1 + a2)/(d2 − a2), 1)/2,
  • 55. Testing as estimation: the demise of the Bayes factors Conclusion Conclusion many applications of the Bayesian paradigm concentrate on the comparison of scientific theories and on testing of null hypotheses natural tendency to default to Bayes factors poorly understood sensitivity to prior modeling and posterior calibration Time is ripe for a paradigm shift
  • 56. Testing as estimation: the demise of the Bayes factors Conclusion Conclusion Time is ripe for a paradigm shift original testing problem replaced with a better controlled estimation target allow for posterior variability over the component frequency as opposed to deterministic Bayes factors range of acceptance, rejection and indecision conclusions easily calibrated by simulation posterior medians quickly settling near the boundary values of 0 and 1 potential derivation of a Bayesian b-value by looking at the posterior area under the tail of the distribution of the weight
  • 57. Testing as estimation: the demise of the Bayes factors Conclusion Prior modelling Time is ripe for a paradigm shift Partly common parameterisation always feasible and hence allows for reference priors removal of the absolute prohibition of improper priors in hypothesis testing prior on the weight α shows sensitivity that naturally vanishes as the sample size increases default value of a0 = 0.5 in the Beta prior
  • 58. Testing as estimation: the demise of the Bayes factors Conclusion Computing aspects Time is ripe for a paradigm shift proposal that does not induce additional computational strain when algorithmic solutions exist for both models, they can be recycled towards estimating the encompassing mixture easier than in standard mixture problems due to common parameters that allow for original MCMC samplers to be turned into proposals Gibbs sampling completions useful for assessing potential outliers but not essential to achieve a conclusion about the overall problem