Is Edwin Jaynes correct about statistics? I've recently been reading Edwin Jaynes's book, Probabili

Defensorentx9 2022-05-07 Answered
Is Edwin Jaynes correct about statistics?
I've recently been reading Edwin Jaynes's book, Probability Theory: The Logic of Science, and I was struck by Jaynes's hostile view of what he dubs "orthodox statistics." He repeatedly claims that much of statistics is a giant mess, and argues that many commonplace statistical techniques are ad hoc devices which lack a strong theoretical footing. He blames historical giants like Karl Pearson and Ronald Fisher for the field's sorry state, and champions Bayesian methods as a healthier alternative.
From my personal perspective, his arguments make a lot of sense. However, there are a number of factors that prevent me from taking his criticism of statistics at face value. Despite the book's being published in 2003, the majority of its contents were written in the mid 20th century, making it a little dated. The field of statistics is reasonably young, and I'm willing to bet it's changed significantly since he levied his critiques.
Furthermore, I'm skeptical of how he paints statistics as having these giant methodological rifts between "frequentists" and "Bayesians." From my experience with other fields, serious methodological disagreement between professionals is almost nonexistent, and where it does exist it is often exaggerated. I'm also always skeptical of anyone who asserts that an entire field is corrupt--scientists and mathematicians are pretty intelligent people, and it's hard to believe that they were as clueless during Jaynes's lifetime as he claims.
1. Can anyone tell me if Jaynes's criticisms of statistics were valid in the mid 20th century, and furthermore whether they are applicable to statistics in the present day? For example, do serious statisticians still refuse to assign probabilities to different hypotheses, merely because those probabilities don't correspond to any actual "frequencies?"
2. Are "frequentists" and "Bayesians" actual factions with strong disagreements about statistics, or is the conflict exaggerated?
You can still ask an expert for help

Expert Community at Your Service

  • Live experts 24/7
  • Questions are typically answered in as fast as 30 minutes
  • Personalized clear answers
Learn more

Solve your problem for the price of one coffee

  • Available 24/7
  • Math expert for every subject
  • Pay only if we can solve it
Ask Question

Answers (1)

Answered 2022-05-08 Author has 11 answers
I can't speak for all statisticians, but here's my view of these two questions:
First, I think that (1) and (2) are mostly the same question, each is asking about the relevance of the Bayesian-Frequentist debate.
In general, I'd say the debate has shifted significantly, notably via input from Deborah Mayo and Andrew Gelman (among many others...but I read these two the most), away from philosophizing about the nature of probability as applied to statistics to the idea of how we judge/check our statistical models. For example, a Bayesian claiming a 95% probability (posterior) of their model being correct is really not saying much, since we can't place that 95% in an objective context. Why should 95% be reassuring? What does 95% mean in terms of observable outcomes?
If you say, "Well, I accept my model as valid when the posterior probability is over 95%", then does that imply 5% of your accepted models are wrong (or something to that effect)? If so, you are implicitly using a frequentist calibration of your model. If you deny the validity of the above statement, then what can you offer to support confidence in your 95%? Why can't we say its 60%? How would you show that I am wrong in a non-circular way?
However, I digress a bit. My main point is not to criticize subjective probability, but to say that it's become largely irrelevant to modern statistical practice. Here's why:
Regardless of your philosophy, you need to validate your models!
No one, no matter how good their reputation or beautifully coherent their logic should get a free pass. You need to show your model is a reasonably good approximation of reality. To do this, you're going to need to compare your output to data, and if your outputs are probabilities, you can't get away from calculating frequency distributions/histograms/density estimates.
As far as I'm concerned, the philosophical part of Bayesian/Frequentist debate is largely over, with frequentism winning out as the preferred method of calibration or checking of models. There simply aren't many alternatives to showing "hey, my model is pretty good!" without appealing to some sense of frequency estimates. Logical coherence alone is not a valid way to empirical truth.
However, I'd also say that Bayesian models are strongly on the ascendance. Modeling your parameters as "random effects" from some meta model basically means you are entertaining a richer space of models, with the added benefit that your estimates may have less variance and lower overall error (if your priors are not completely off the mark).
So, yes, I'd say Jayenes was a product of (a) his ego and (b) his times. Both Bayesian and non-Bayesian (I'm not going to say Frequentist) approaches have advanced markedly, with quite a bit of convergence at the philosophical level, but proliferation at the model/methodological level. For example we have the bootstrap, empirical likelihood, MCMC, EM, AIC/BIC, diffuse priors/informative prior/non-informative priors, many ways to skin the same cat (but at least we all seem to agree on what it means to have skinned it!).
For (2), I actually think I partially answered this in my answer to (1). If you read modern Bayesians like Gelman, you'll see a strong push towards model checking and verification by comparing observed frequencies vs modeled frequencies or similar measures (e.g., he proposes a version of the P-value in his book for predictive model checking). The massive decrease in computing cost means you can also check a lot more measures of "misfit" than when Fisher and Pearson came up with their P-values and significance tests. Gelman and others seem to support the idea of holistic checks, not blind adherence to a single statistics. I agree with this, and I'm happy to see how objective verification has returned to Bayesian modeling.
Not exactly what you’re looking for?
Ask My Question

Expert Community at Your Service

  • Live experts 24/7
  • Questions are typically answered in as fast as 30 minutes
  • Personalized clear answers
Learn more

Relevant Questions

asked 2022-04-17
What does failing to reject a null hypothesis imply?
asked 2022-04-06
Why is standard deviation calculated differently for finding Z scores and confidence intervals?
Suppose that as a personnel director, you want to test the perception of fairness of two methods of performance evaluation. 63 of 78 employees rated Method 1 as fair. 49 of 82 rated Method 2 as fair. A 99% confidence interval for p 1 p 2 (where p 1 = 63/78 and p 2 = 49/82) is as follows:
p 1 p 2 ± 2.58 p 1 ( 1 p 1 ) n 1 + p 2 ( 1 p 2 ) n 2
0.029 p 1 p 2 0.391
At the 0.01 level of significance the Z score is
Z = p ( 1 p ) n 1 + p ( 1 p ) n 2
(where ( x 1 + x 2 ) / ( n 1 + n 2 ) = 0.70 but sometimes a different formula p = ( p 1 + p 2 ) / 2 is also used)
Z = 2.90
Both tests indicate that there is evidence of a difference.
But you could also find the Z score using the standard deviation formula in the first method to be 2.993. Why are the Z scores different? Where do the formulas for finding the standard deviation come from?
asked 2021-12-18
Which of the following are true? If false, explain briefly.
a) If the null hypothesis is true, youll
asked 2022-03-31
Suppose that you want to do statistical significance tests for four correlations and you want your EWa to be .05. What PCa would you use if you apply the Bonferroni procedure?
asked 2020-12-28
A. Look for the definitions of the following terms related to hypothesis testing.
1. Null Hypothesis
2. Level of Significance
3. Type I error
asked 2022-05-02
Show the difference between two approaches to significance testing?
asked 2022-05-08

The US government is interested in understanding what predicts death rates. They have a set of data that includes the number of deaths in each state, the number of deaths resulting from vehicle accidents (VEHICLE), the number of people dying from diabetes (DIABETES), the number of deaths related to the flu (FLU) and the number of homicide deaths (HOMICIDE).
Your run a regression to predict deaths and get the following output: 
At alpha=0.05, what is indicated by the significance F in this problem? 
A. The regression model does not significantly predict deaths. 
B. The regression model significantly predicts deaths. 
C. Two of the four independent variables significantly predict deaths. 
D. Three of the four independent variables significantly predict deaths. 
Multiple R0.874642613R Square0.764999351Adjusted R Square0.744564512Standart Error62.97881926Obsevations51 


 dfSSMSFSignificanceFRegression4593934..928148483.73237.436035146.36772F14Residual46182451.2571?3966ю331676   Total50776386.1851