Conditional calibration and the sage statistician
Section 5. The Bayesian posterior
distribution of
The Bayesian approach
differs from the Neymanian approach (and from Fisher’s fiducial approach) by
formulating the problem so that a real conditional probability
distribution for the estimand
can be calculated, using the laws
of probability theory to condition on the fact that the observed data equals
this distribution is called the posterior distribution of
that is, posterior after seeing
To conduct this activity formally,
must be a random variable, and
thus
needs to have a “starting” probability distribution, called its prior distribution,
meaning prior to seeing any data; in the context of our
setup, this prior distribution is a distribution over the possible local estimands, that is,
a set of
probabilities (summing to one),
one probability for each possible Truth. This prior distribution is essentially
a set of
weights
reflecting the prior beliefs of
experts that each of the
possible local estimands is the
correct one. The Neymanian frequentist has no use for such weights
over the set of possible Truths, because the 95% is supposed to hold for any
set of weights, and thus for each possible Truth (i.e., for all
point mass prior distributions).
Now comes the part of the
argument that hints at a departure from Neyman’s 1970’s claim to me that
conditional inference is too difficult. In the context of the simulation just
described, and admitting some Bayesian or
fiducial logic, when confronted with actual observed data set
attention should be focused on
the parts of the simulation where the generated
equals
the other
can be ignored (at least in the context
of the idealized description
here, where
is essentially infinite) because, to be fully
Bayesian, we want to condition on
equaling
In fact, let us use the simulation itself to describe
the Bayesian posterior distribution of
i.e., the distribution of
conditioning on the fact that
Let
be the proportion of
the
values of
that match
for
that is, for truth
is the proportion of
the generated data sets from truth
that match the actual data set
For example, if
is zero, then the a priori possible truth
could not be the actual truth
because it could not have generated observed data
The posterior probability that
the estimand
equals
the local value of
for Truth
is the weighted average of the
proportions,
weighted by
the prior probability that
is the correct truth. Here, this weighted average
of proportions is generally labeled
where
for the observed data
is labelled
and equals
we could call
the estimated
ability of Truth
to match
observed data
This description of the posterior
distribution of
using simulation is from Rubin
(1984); see Figure 5.1.

Description for Figure 5.1
Description of
posterior distribution from Rubin (1984). Suppose we first draw equally likely
values of
from
and label these
The
can be thought of as representing the possible
populations that might have generated the observed
For each
we now draw an
from
label these
The
represent possible values of
that might have been observed under the full
model
Now some of the
will look just like the observed
and many will not; of course, subject to the
degree of rounding and the number of possible values of
might have to be very large in order to find
generated
that agree with observed
but this creates no problem for our conceptual
experiment. Suppose we collect together all
that match the observed
and then all
that correspond to these
This collection of
represents the values of
that could have generated the observed
formally, this collection of
values represents the posterior distribution
of
An interval that includes 95% of these values
of
is a 95% probability interval for
and has the frequency interpretation that
under the model, 95% of populations that could have generated the data are
included within the 95% interval.
There are objections to
this approach. First, where do the prior weights
come from and who are the experts
providing these weights? Perhaps we should find some way to avoid using these
potentially overly subjective prior weights? Second, perhaps the requirement
for exact equality between a generated data set
and the observed data set
should be relaxed in some way so
that a generated
does not have to equal
exactly but only “look like” it
came from the same distribution as did
and so match
in some way?
More on
this second point first, which is clearly important when trying to conduct an
actual simulation like this idealized one with a finite budget. The
approximate equality between generated data
and observed data
can be
achieved in situations with low-dimensional sufficient statistics, because
only those statistics have to match. But this idea of generated data
being “close to” observed data
is the
basis of all work using this description of the
posterior distribution to conduct “ABC”
Approximate Bayesian Computation,
apparently first described in the paragraph in Figure 5.1 (https://en.wikipedia.org/wiki/Approximate_Bayesian_computation,
Tavare, Balding, Griffiths and Donnelly, 1997). We simply assume at this point that we have
chosen some such metric to define the function
and
use it to define the ability of Truth
to generate data sets that match
the observed data,
ISSN : 1492-0921
Editorial policy
Survey Methodology publishes articles dealing with various aspects of statistical development relevant to a statistical agency, such as design issues in the context of practical constraints, use of different data sources and collection techniques, total survey error, survey evaluation, research in survey methodology, time series analysis, seasonal adjustment, demographic studies, data integration, estimation and data analysis methods, and general survey systems development. The emphasis is placed on the development and evaluation of specific methodologies as applied to data collection or the data themselves. All papers will be refereed. However, the authors retain full responsibility for the contents of their papers and opinions expressed are not necessarily those of the Editorial Board or of Statistics Canada.
Submission of Manuscripts
Survey Methodology is published twice a year in electronic format. Authors are invited to submit their articles in English or French in electronic form, preferably in Word to the Editor, (statcan.smj-rte.statcan@canada.ca, Statistics Canada, 150 Tunney’s Pasture Driveway, Ottawa, Ontario, Canada, K1A 0T6). For formatting instructions, please see the guidelines provided in the journal and on the web site (www.statcan.gc.ca/SurveyMethodology).
Note of appreciation
Canada owes the success of its statistical system to a long-standing partnership between Statistics Canada, the citizens of Canada, its businesses, governments and other institutions. Accurate and timely statistical information could not be produced without their continued co-operation and goodwill.
Standards of service to the public
Statistics Canada is committed to serving its clients in a prompt, reliable and courteous manner. To this end, the Agency has developed standards of service which its employees observe in serving its clients.
Copyright
Published by authority of the Minister responsible for Statistics Canada.
© Her Majesty the Queen in Right of Canada as represented by the Minister of Industry, 2019
Use of this publication is governed by the Statistics Canada Open Licence Agreement.
Catalogue No. 12-001-X
Frequency: Semi-annual
Ottawa