 
 
 
 
 
 
 
  
![[*]](file:/usr/lib/latex2html/icons/crossref.png) and
  and ![[*]](file:/usr/lib/latex2html/icons/crossref.png) . We have seen there
that the expected value of a parameter can be considered, somehow,
to be analogous to the estimators8.10 of the frequentistic approach. 
It is well known, from courses on conventional statistics, that 
one of the nice properties an estimator should have is that 
of being free of bias.
. We have seen there
that the expected value of a parameter can be considered, somehow,
to be analogous to the estimators8.10 of the frequentistic approach. 
It is well known, from courses on conventional statistics, that 
one of the nice properties an estimator should have is that 
of being free of bias. 
Let us consider the case of Poisson and  binomial distributed 
observations, exactly as they have been treated in 
Sections ![[*]](file:/usr/lib/latex2html/icons/crossref.png) and
 and ![[*]](file:/usr/lib/latex2html/icons/crossref.png) , i.e. assuming a 
uniform prior. 
Using the typical notation of frequentistic analysis, let us 
indicate with
, i.e. assuming a 
uniform prior. 
Using the typical notation of frequentistic analysis, let us 
indicate with  the parameter to be inferred,  
with
 the parameter to be inferred,  
with 
 its estimator.
 its estimator. 
 ;
;  indicates the possible observation
and
 indicates the possible observation
and 
 is the estimator in the light of
 is the estimator in the light of  :
:  
|  |  | E ![$\displaystyle [\lambda\,\vert\,X] = X + 1 \,,$](img1199.png) | |
| E ![$\displaystyle [\hat{\theta}]$](img1200.png) |  | E ![$\displaystyle [X + 1] = \lambda + 1 {\bf\ne} \lambda \,.$](img1201.png) | (8.3) | 
 is large).
 is large). 
 ; after
; after  trials one may observe
 trials one may observe 
 favourable results, and the estimator of
 favourable results, and the estimator of  is then
 is then
|  |  | E ![$\displaystyle [p\,\vert\,X] = \frac{X+1}{n+2} \,,$](img1203.png) | |
| E ![$\displaystyle [\hat{\theta}]$](img1200.png) |  | E ![$\displaystyle \left[\frac{X+1}{n+2}\right] =
\frac{n\,p+1}{n+2} {\bf\ne} p \,.$](img1204.png) | (8.4) | 
| E ![$\displaystyle [\hat{\theta}\,\vert\,\theta]$](img1205.png) |  |  e.g.    E ![$\displaystyle [\hat{\lambda}\,\vert\,\lambda] = \lambda) \,,$](img1207.png) | |
|  i.e.  |  |  | (8.5) | 
 ? We don't know, otherwise 
we would not be wasting our time trying to estimate it
(always keep real situations in mind!). 
For this reason, our considerations
cannot depend only on the fluctuations of
? We don't know, otherwise 
we would not be wasting our time trying to estimate it
(always keep real situations in mind!). 
For this reason, our considerations
cannot depend only on the fluctuations of 
 around
 around
 , but also on the different degrees of belief of the possible
values of
, but also on the different degrees of belief of the possible
values of  .
Therefore they must depend also on
.
Therefore they must depend also on 
 . 
For this reason, the Bayesian result is that which makes the
best use8.11 of the state of knowledge about
. 
For this reason, the Bayesian result is that which makes the
best use8.11 of the state of knowledge about
 and  of the distribution of
 and  of the distribution of 
 for each possible value
 for each possible value  .
This can be easily understood by going back to the examples of 
Section
.
This can be easily understood by going back to the examples of 
Section ![[*]](file:/usr/lib/latex2html/icons/crossref.png) . 
It is also easy to see that the  freedom from bias 
of the frequentistic approach requires
. 
It is also easy to see that the  freedom from bias 
of the frequentistic approach requires 
 to
be uniformly distributed from
 to
be uniformly distributed from  to
 to  (implicitly, as frequentists refuse
the very concept of probability of
 
(implicitly, as frequentists refuse
the very concept of probability of  ). Essentially, 
whenever a parameter has a  limited range, the frequentistic analysis
decrees that  Bayesian estimators are biased.
). Essentially, 
whenever a parameter has a  limited range, the frequentistic analysis
decrees that  Bayesian estimators are biased.
There is another important and subtle point related to this problem, 
namely that of the
Monte Carlo check of Bayesian methods. Let us consider the case 
depicted in Fig. ![[*]](file:/usr/lib/latex2html/icons/crossref.png) and imagine 
making a simulation, choosing the value
 and imagine 
making a simulation, choosing the value 
 , 
generating many (e.g. 10
, 
generating many (e.g. 10  000) events, and considering 
three different analyses:
000) events, and considering 
three different analyses:
 ;
;
 `of the 
kind'
 `of the 
kind' 
 of Fig.
 of Fig. ![[*]](file:/usr/lib/latex2html/icons/crossref.png) , 
assuming that we have a good idea of the kind of physics
we are doing.
, 
assuming that we have a good idea of the kind of physics
we are doing. 
 ? 
You don't really need to run the Monte Carlo to realize that 
the first two procedures will perform equally well, while 
the third one, advertised as the best in these notes, will 
systematically underestimate
? 
You don't really need to run the Monte Carlo to realize that 
the first two procedures will perform equally well, while 
the third one, advertised as the best in these notes, will 
systematically underestimate  !
!
Now, let us assume we have observed a value of  , for example
, for example  . 
Which analysis would you use to infer the value of
. 
Which analysis would you use to infer the value of  ?  
Considering only the results of the 
Monte Carlo simulation it seems obvious
that one should choose one of the first two, but 
certainly not the third!
?  
Considering only the results of the 
Monte Carlo simulation it seems obvious
that one should choose one of the first two, but 
certainly not the third! 
This way of thinking is wrong, but unfortunately 
it is often used by practitioners who have no time to understand 
what is behind Bayesian reasoning, who perform some Monte Carlo tests,
and decide that the Bayesian theorem does not
 work!8.12 The solution 
to this apparent paradox is simple.
If you
believe that  is distributed like
 is distributed like 
 of Fig.
of Fig. ![[*]](file:/usr/lib/latex2html/icons/crossref.png) , then you should use this
distribution in the analysis and also 
in the generator. Making a simulation 
based only on a single true value, or on a set of 
points with equal weight, is equivalent to assuming 
a flat distribution for
, then you should use this
distribution in the analysis and also 
in the generator. Making a simulation 
based only on a single true value, or on a set of 
points with equal weight, is equivalent to assuming 
a flat distribution for  and, therefore,
 it is not surprising 
that the most grounded  Bayesian analysis is that
which performs worst
in the simple-minded frequentistic checks.
It is also worth remembering that priors are not
just mathematical objects to be plugged into Bayes' theorem, 
but must reflect prior knowledge. Any inconsistent 
use of them leads to paradoxical results.
 and, therefore,
 it is not surprising 
that the most grounded  Bayesian analysis is that
which performs worst
in the simple-minded frequentistic checks.
It is also worth remembering that priors are not
just mathematical objects to be plugged into Bayes' theorem, 
but must reflect prior knowledge. Any inconsistent 
use of them leads to paradoxical results. 
 
 
 
 
 
 
