A little summary on several types of computer intensive statistical methods developed from the fantastic book by Malcolm Haddon titled "Modeling and Quantitative Methods in Fisheries"
Authors: Daniele Baker and Stephanie Johnson
2. Playing Card Example
• 52 cards are population
▫ 25 card sample…. What is the mean, median, SE
• Randomization
â–« Randomly reallocate cards to groups (i.e. diamonds,
spades..)
â–« 4 people divide up cards
â–« Do 1000+ times randomly allocating to the 4 groups
â–« Graph up distribution of parameter
â–« If your original parameter is outside of the 95% CI then
it is significantly different from random
3. Playing Card Example
• JackKnife
▫ Take 24 cards… (leave 1 out) and complete test
statistics
â–« Redo for all possible combinations (25 x)
• Bootstrapping
â–« Pick out 25 cards, but put them back each time
â–« Calculate new parameters
ď‚– Redo 1000+ times
â–« If your sample parameter falls within the 95% CI of
the distribution, then it isn’t statistically different
from random
4. Playing Card Example
• Monte Carlo
â–« Find a model that would fit the card trend
ď‚– Relative frequency plot, examine shape
â–« Randomly select value for model parameters or data (cards
picked)
ď‚– Complete 1000+ times
ď‚– Analyze your parameters to fit the data vs. random generated
parameters
http://www.vertex42.com/ExcelArticles/mc/MonteCarloSimulation.html
5. COMPARISON
Randomize Jackknife Bootstrap Monte Carlo
With
Replacement
No No Yes Yes
Exact P-values Yes Likely no Yes Yes
Resample a
theoretical PDF
Parametric
Resample an
empirical
distribution
Yes Yes Non-
parametric
Non-
parametric
Good to… Deal with
unknown
distribution
Detect bias,
calc. SE, good
for biases
parameters
Calc. sample
size for exp.
Design, CI, SE
and Test Hypot.
Flexible,
generic, SE,
CI, Test Hypot.
Good for sparse data sets
Limitations Can’t calc SE,
or CI (weak)
Bad CI
6. 4 Methods
• Randomization
â–« Ho: each group of obs. is a random sample of 1 pop.
â–« Must be characterized by a test stat
â–« Combine all groups, then reallocate, and compare
ď‚– Repeat 1000+ times
â–« Compare obs. Test stat with empirical distribution of
that test stat given available data.
ď‚– A sig diff is when obs. test stat is beyond empirical
distribution
7. 4 Methods
• Jackknife
â–« Sample could be from one arm of the distribution
â–« Subset data (for all combinations of all data minus 1
pt) (total = n-1)
â–« Calculate pseudo-values , diff. btw. this and obs =
estimate of bias
â–« Good when estimating something other than mean
â–« Calculate Jackknife SE and parameter of interest
ď‚– CI can be fitted but issues with DFs
8. 4 Methods
• Bootstrapping
â–« Random samples of the observation (with
replacement
ď‚– Each treated as a separate random sample
â–« Should = the distribution if you had repeatedly
sampled the original population
â–« Provides better CI than the Jackknife
ď‚– Can determine SE, CI and test hypotheses
â–« Have been used for multiple regression and stratified
sampling
9. 4 Methods
• Monte Carlo
â–« Mathematical model of the situation + model
parameters
â–« Randomly select variable, parameter or data values
and then use to determine model output
ď‚– Do 1000+ times and use to test hypotheses or
determine Confidence Intervals
â–« *Resampling from a theoretical distribution
â–« Compare observations with data from a model of the
system + for use in risk assessment
10. Chapter 5: Randomization TestsChapter 5: Randomization Tests
• ANOVA (e.g.) vs. Randomization
â–« Assumptions
• Hypothesis testing
â–« Determinations of likelihood that observations in
nature could have arisen by chance
â–« Relativity: group 1 vs. group 2, hypothesis
11. Standard Significance Testing
• Test statistic (e.g. t-test)
• Significant difference
• Statistically 3 things
needed to test
hypothesis:
1. Formally stated hypothesis
(Ho and Ha)
2. Test statistic: t-test, F-ratio,
r correlation, etc.
3. Means of generating PD of
test statistic under
assumption Ho is true
• idrc.ca
idrc.ca
12. • Observed vs. expected values (d.f., etc.)
• Determine how likely observed values,
assuming Ho to be true
• α-value: 0.05, 0.01, 0.001
0
0.2
0.4
0.6
0.8
1
0 5 10 15 20 25
X2
Statistic
Probability
α 0.05 = ~15
13. However…
• Test statistics are not valid if assumptions are
falsely made
• Can, therefore, reject a real difference in data or
accept a non-existent difference
• Problem with theoretically derived PDF:
â–« If test statistic is not significant, cannot tell without
further analyses if the test failed b/c:
1. Samples are not independently distributed (thing being tested)
or,
2. The data failed to conform to the assumptions necessary for the
validity of the test (e.g. samples were not normally distributed)
This is where randomization tests come into play…
14. Significance Testing by Randomization
Independent of any determined parametric
PDF
Generates empirical PDF for the test-statistic
15. Given a null hypothesis, the expected PDF for a
test statistic can be generated by repeatedly
randomizing the data with respect to sample
group membership and recalculating the test
statistic
1. Done many times (min. 1000)
2. Test statistic values tabulated
3. Compared with original value from un-
randomized data
4. If original value is unusual relative to
permutations, Ho can be rejected
17. Key Points
• Essentially, the null-hypothesis is that the
groups being compared are random samples
from the same population
• Thus, a test of significance is an attempt to
determine whether observed samples could
have been randomly drawn from the same
population
• Answer = probability
â–« PROBABILITY = possibly from same pop. (never
claim definitively)
â–« PROBABILITY = not likely from same pop.
18. Ex: Fish length for in-shore fish vs. off-
shore fish
• Ha: In-shore fish are smaller on average than off-
shore fish
0 50 100 150 200 250 300
Fork Length
Off-shore
In-shore
19. Ex. 5.2
• Randomization can be used to test mean
difference
• Original mean difference occurred 25 x’s out of
1000
• What does this tell us about the data?
• Weight of evidence, not significant difference
(e.g. p=0.05)
21. Selection of a Test Statistic
• Chosen based on sensitivity to the hypothetical
property being checked or compared (e.g. mean
vs. median (ex.5.3))
• Precaution should be taken when selecting a
non-standard test statistic
• Multivariate comparisons
• Determine exactly which hypothesis is being
tested by the test statistic
▫ “When in doubt, be conservative”
• Ex 5.3
22. Ideal Test Statistics
• Greatest statistical power
• Significance – probability of making a Type I
error
• Power – probability of making a Type II error
• Unbiased – using a test that is more likely to
reject a false hypothesis than a true one
No difference
Null true
Difference exists
Null False
Null accepted OK TYPE II ERROR
Null rejected TYPE I ERROR OK
23. Randomization of Structured Data
• Restricted to comparison tests (cannot be used
for parameter estimation)
• Differences in variation – randomize residuals
instead of data values
• Basic rule: unbalanced and highly non-normal
data a randomization procedure should be used
• Question: what should be randomized?
â–« Raw data,
â–« Sub-set of raw data, or
â–« residuals from model
24. Take Home Message…
With structured or non-linear data, care needs to
be taken in what components should be
randomized
25. Summary
• Randomization requires less assumptions than
standard parametric stats
• Significance tests test whether the observed
samples could be from the same pop.
• State hypotheses and determine significance
level
• Test statistics that yield the greatest power
should be utilized
Hinweis der Redaktion
FOR THIS EXAMPLE PICK OUT 3 cards at a time
FOR THIS EXAMPLE PICK OUT 3 cards at a time
Less assumptions are needed for randomization tests, giving extra flexibility.