Author Archives: Mayo

The F.D.A.’s controversial ruling on an Alzheimer’s drug (letter from a reader)(ii)

I was watching Biogen’s stock (BIIB) climb over 100 points yesterday because its Alzheimer’s drug, aducanumab [brand name: Aduhelm], received surprising FDA approval.  I hadn’t been following the drug at all (it’s enough to try and track some Covid treatments/vaccines). I knew only that the FDA panel had unanimously recommended not to approve it last year, and the general sentiment was that it was heading for FDA rejection yesterday. After I received an email from Geoff Stuart[i] asking what I thought, I found out a bit more. He wrote:


Hi Deborah, 

I guess you may have heard about the approval of this Alzheimer’s drug under controversial circumstances: 

I was asked for my statistical opinion about this prior to the decision. I am giving a brief talk later this week. My answer was although non-effectiveness has not been proven, there was not sufficient evidence to show effectiveness, and I thought the FDA would ask for more research. They kind of did this in the form of requiring a Phase IV trial.[ii] 

There were problems I think with the evidence, and with the FDA’s reasoning. 

The problems were (i) stopping the trials on the basis of futility analyses, (ii) then re-analysing using subgroup analyses and secondary endpoints, and (iii) lack of evidence for a relationship between amyloid levels on imaging and cognitive scores. 

I think this decision may set an unfortunate precedent that undermines the principle of pre-registering trials. Secondary analysis is fine for understanding why a trial failed and generating hypotheses for future studies, but it is not usually considered sufficient evidence for drug approval. Interestingly, the FDA’s own expert panel and statistician didn’t think that the evidence justified approval at this point. 

I think there are critical flaws in the FDA’s reasoning. They argued for a lower standard of evidence given the current lack of effective treatments for Alzheimer’s disease. They argued that the trials showed that the drug prevented the accumulation of amyloid protein, and that should improve cognition. It is the improvement in cognition that is the primary endpoint, and the trials were stopped because of the decline of cognition in the treatment group [in one of two trials]. Even in the positive-trending trial, cognition in the treatment group was not sufficiently different from decline in the control group. The initial hope in the field that amyloid was the actual cause of dementia and clearing it would help has been dashed many times. Previous drugs with similar actions have all failed.  These trials were a bit different in targeting the early phase, so they are more about prevention than treatment. That was a reasonable approach, but we need strong evidence about preventing cognitive decline.

This is a fairly unique example of not sticking to pre-registered analyses. I don’t know of any similar example. Other groups who have conducted ostensibly failed trials may conduct secondary analyses and then seek to have treatments approved based on those analyses. That opens up the possibility of data-dredging to achieve the desired outcome, and I think that you will agree that is something that has contributed to the “replicability crisis” in science. Not the use of significance testing! 

Given the small and questionable treatment effect and the huge cost of the drug, the FDA may have some explaining to do in a few years.  The Phase IV trial may not be successful. I think the insurance companies and other countries will be watching with interest before they commit.  

What do you think about this development? 




I agree with Stuart’s worry about the post hoc searching, especially as it appears the FDA worked with Biogen to carve out the subgroup after the company itself stopped the trial for futility.[iii] I agree with those who criticize the FDA approving the drug despite poor evidence. I don’t know enough about the status of the amyloid plaque theory today to weight in much further (see note iv.) But I’m very concerned about the precedent this sets. In the Buzzfeed link: 

“The FDA’s decision to approve aducanumab shows a stunning disregard for science and eviscerates the agency’s standards for approving new drugs,” Michael Carome, who tracks the pharmaceutical industry for the watchdog group Public Citizen, told BuzzFeed News. “Because of this reckless action, the agency’s credibility has been irreparably damaged.”

Granted, it’s hard to see how the FDA can insist on its own strong position against drug approvals based on such ransacking in the future. (See my ‘P-values on Trial’.) Stuart says he doesn’t know of another similar example, but perhaps others do (please inform us in the comments). I would definitely worry about future FDA panels being as prepared to say “no” in the future as this one was–and that’s a real problem. (One panelist just quit, see comment 6/9/21.)

One good thing is that the FDA admitted that “the drug had provided incomplete evidence to demonstrate effectiveness” and so it was requiring Biogen to conduct a new clinical trial. This I believe is a distinct, possibly a newish, category for especially devastating diseases. Some say people won’t volunteer for a clinical trial where they might be assigned to placebo, now that it has been approved. Maybe so, but maybe those who can’t possibly afford the $56,000 cost per year would be willing for a chance of being assigned to it. [iv]

The main thing, it seems to me, is that the FDA clearly made the distinction between having provided sufficient evidence the drug works in improving cognitive function (Biogen did not), and the reasons it was nevertheless approving it–the fact that Alzheimer’s is a devastating disease with no effective treatment, and that this would give hope to Alzheimers’ sufferers after nearly 20 years with no new treatments. FDA Report. The agony this group went through in coping with the Covid pandemic might also have played a role. (There are also allegations of financial ties of various sorts, both with drug companies and advocacy groups.)The most serious problem with the latest calls to “abandon” statistical significance is their recommendation that such extra-evidential considerations go toward the evaluation of the evidence for effectiveness, and that these policy decisions be mixed in with the evidential evaluation. That’s one of the main reasons I find such calls so irresponsible, especially among those in positions of power and influence.

Please share your thoughts in the comments.

As usual, I indicate updated versions with (i), (ii),… in the title.

[i] Stuart is involved in statistical research in psychology at Melbourne, and has attended sessions of my Phil Stat Wars Forum.

[ii] This grants the drug only conditional approval, which can later be rescinded, though it’s very doubtful this would happen. 

[iii] Stuart elaborated further on the subgroup analysis:

My understanding of the post hoc analysis is that it was only of participants who had completed the full course of treatment and were in the high dose group or the placebo group (not the low dose group – but surely they looked). The problem here is that at the point of the futility analysis, progress would have been compared for everyone at the same time from enrolment and it was looking very much like the treatment group were declining in their cognition at a similar rate to the placebo group, so they stopped the trials. Then they found post-hoc that the high dose “completers” showed a bit of an effect (only a 20% reduction in cognitive decline). Now it could be that  the eventual completers were getting better at the mid-point. So, removing all those who had only partly completed the trial could have helped them get a significant result with the subset who completed before the trials were aborted. There’s no statistical control for that, so the chance that the finding was a Type I error goes up by an unknown amount. Sticking to the pre-registered protocol and completing the trials would have controlled for that.

It could be true that for effectiveness a high dose for the pre-specified treatment period is required, so the hypothesis is still alive. But a lot of us, probably including the FDAs own expert panel and statistician, think there should be a new Phase III trial to test that hypothesis, and that the evidence from the aborted trials is compromised by its post-hoc nature.

[iv] The fact that reducing amyloid in the brain, achieved by other experimental drugs, hasn’t shown improved cognitive function had been taken by some as falsifying the amyloid theory altogether. (They’ve also all tended to have the side effect of brain-swelling.) One valuable consequence of this approval could be to finally falsify the amyloid plaque theory, or, alternatively, give it new life.

Added 6/11/21 Remarks from Bio Pharma collected by Endpoints (though without names that I can see):

  • — It’s a horrifying abdication of the agency’s responsibility as a vanguard of scientific truth. It makes a mockery of statistical rigor and their own advisors, and it will do irreparable damage to health care economics, research into Alzheimer’s, and their own reputation.
  • — Totally discredits the FDA and potentially dangerous to patients without any real indication of benefit.
  • — Unproven efficacy, unproven surrogate that can’t be reasonably expected to predict benefit given large # of other failed attempts.
  • — “Borderline fraud,” one reader noted. “I have an ethical obligation to my patients. I will never prescribe.”
  • — This is the first time in 30 years I’ve seen a decision that’s based almost entirely on lack of scientific reason, lack of science biology and medical evidence (they claim that “it’s reasonable to assume” that losing a-beta plaques will help the patient). There was powerful manipulation by in this case a patient group organization (Alz Assn, all-bad) which is ignorant and subservient to some influence group and accompany this and like minded companies. There was also manipulation of trial data; which made everyone believe this was nonsensical.
  • — The sheer ignorance/disregard of proper scientific method that was displayed in not requiring a confirmatory trial prior to approval is disgraceful.
  • — Clinical studies failed to provide convincing results as to its efficacy. I was interviewing at Biogen the day they first announced they are ceasing clinical trials due to failure of the drug. The team I was interviewing with was part of the group that “discovered” the drug and they clearly told me that they were not surprised at all that the clinical trials failed. 2 years down the line and FDA managed to take science back 20 years …
  • — It is at odds with all prior guidance. They treated it like it was an AIDS drug in the late 1980s but the problem is nothing like AIDS. There is no established link between plaque regression and cognitive improvement. They changed their mind on accelerated approval. The AdComm should have had a chance to weigh in on this. It is not like DMD — this disease affects millions of people.
  • — Cherry picked, post-hoc data. There have been 17 other A-beta antibody studies that showed no benefit or worse outcomes. There have been 16 trials (not all anti-A-beta) in which the treatment arm did worse than placebo despite reduction of amyloid.
  • — This sets a terrible incentive for other companies to make me-too drugs rather than focus on alternative pathways that would be more effective, such as taking a gene-based approach. In addition, there is not enough proof of efficacy. They should have made them do a third trial at the higher dosing. Lastly, Biogen will make a boat-load of profit selling an unproven drug at a high price based on the (false?) hopes of patients.
  • — Data presented by Biogen is at the best weak, and at the worst totally irrelevant. This is coupled with high costs to consumers for, at best, modest improvements in therapeutic outcomes. The FDA was founded to stop the sale of “patent” medicines to unwitting citizens, yet here it is a century later endorsing what is their functional equivalent.
  • — The advisory panel overwhelmingly rejected aducanumab and later the FDA changed the criteria to the surrogate endpoint. It exposes patients to an ineffective and costly treatment, gives misguided encouragement to other similar would-be products, throws the reimbursement system into disarray and completely disregards science.
  • — The indirect implications include validating a hypothesis that has never panned out (and I don’t believe it did here) and from which the world has moved on.
  • — Overall, a very sad decision from the FDA, sad for serious and responsible drug development, not the one which is guided solely by the population’s unmet need. Very sad with major implications for the FDA, pharma, drug development and society. It will cost us years to recover from this one bad decision! I do not envy the patients … sad also for them who will take the risk of having unwanted side effects, for very few beneficial effects (if at all). Praying is less risky and maybe more efficacious.
  • — It undermines the credibility of randomized placebo controlled trials. While post-hoc analyses can be useful in designing subsequent trials, they should not be used as the basis for approval. As Alexander et al. put it in their scathing JAMA commentary, “Any treatment will appear to be more effective if individuals in whom it works least are removed from the analysis.”
  • — Did not meet the FDA’s standard of 2 demonstrative clinical trials. Also across disease states, the FDA has set an appropriate standard of requiring positive data on both patient-reported outcomes/symptoms as well as underlying disease process. In this case the drug only hinted at one of these and not the other at all. What gives? Yes we are desperate for ALZ therapies. But even its critics have previously appreciated the FDA’s role as the toughest reviewer of clinical data. Now like so many other areas of American leadership, that is crumbling.
  • — Ignoring a nearly unanimous Ad Board sets a dangerous and confusing precedent.
  • — The approval decision should be guided by science and not outside pressure (patient advocacy and/or business lobby). The adcom unilaterally rejecting the aducanumab on efficacy screams “no” to me.
  • — The FDA has given Biogen free license to charge $56K (an insane price given how poor their data are) for 6+ million people. Even if Biogen is only able to capture 10% of that market, you’re talking about an economic burden on the US for payers of ~$34B a year. How are payers going to focus on any other disease therapies (with far better efficacy) when they’re distracted with this giant line item? Sovaldi’s CURE for hep C was already an enormous burden and now we have a chronic, ineffective therapy. The FDA has no ability to enforce the requirements of the post-marketing study and Biogen will likely be unable to recruit placebo patients anyway. They’ve handed Biogen a gigantic blockbuster without them earning it. Sad day.
  • — This result will embolden the nearly fully denounced amyloid hypothesis supporters in the field and will ensure that other, more promising approaches fail to get funding meaning likely meaning we will not see progress in treating az for another 20+ years.
  • — Can’t say it better than John’s editorial. Not following the science, terrible precedent and counterproductive for finding real therapies. Biogen will do a Sarepta and use the money to buy a real drug. Puts all the Covid goodwill in peril.
  • — Terrible FDA precedent. Makes drug development and regulatory approval less predictable for all sponsors. FDA has stated many times that accelerated approval is NOT meant to be a consolation prize for a failed study – that’s what happened here.
  • — While the 2nd trial may barely meet the requirements for the FDA’s Accelerated Approval Program, the strong disapproval by such a large number of neuroscientists should NOT have been dismissed … I am the daughter of parents who both died from the disease and it is prevalent in both sides of my family so my disagreement did not come without a lot of thought. I am always concerned when one individual in particular (in this case Dr. Dunn) appears to want it approved over all opposition. What is his relationship to Biogen?
  • — This is simply the FDA unable to make a tough call, and passing the buck, hoping that payers will do the right thing. This is bad for patients, bad for the payers, and very, very bad for those manufacturers that are doing the right thing in developing good drugs supported by good data.
  • — To me, this is worse than Shkreli-type price gouging, because it interferes with our ability to actually know if this drug can actually help anyone (besides Biogen).
  • — What is the point of a p value if you play with the data until you get what you want (stats 101!)? — What is the point of aligning with the FDA prior to the studies on endpoints etc if it doesn’t matter? — What is the point of the FDA if their endorsement no longer means that the drug works? — Why have any adcomms in the future if all the experts will be ignored? But most of all: This provides false hope to a lot of desperate people who cannot (and should not!) understand the nuances above, and undermines efforts to develop something that does work.

There were some supportive remarks scattered in the mix here, but even some of those were qualified with a marked lack of enthusiasm for this drug. From readers:

  • — It is a toe hold into neurodegeneration and a starting point upon which to improve. Much like early approvals in MS and ALS starting points are imperfect.
  • — The data, while not up to the FDA’s usual standard of clear efficacy in 2 independent Phase 3 trials, are the most convincing to date of a disease-modifying effect in this desperate disease. Biogen should have to conduct another confirmatory trial, but patients should not have to wait for the outcome in order to be treated.
  • — With each failure there is increased trepidation to invest in Alzheimer’s treatments, diagnostics and infrastructure. A win, although controversial, could reverse this trend and encourage treatments that are much better.


Categories: PhilStat/Med, preregistration | 8 Comments

Bayesian philosophers vs Bayesian statisticians: Remarks on Jon Williamson

While I would agree that there are differences between Bayesian statisticians and Bayesian philosophers, those differences don’t line up with the ones drawn by Jon Williamson in his presentation to our Phil Stat Wars Forum (May 20 slides). I hope Bayesians (statisticians, or more generally, practitioners, and philosophers) will weigh in on this. 

Continue reading
Categories: Phil Stat Forum, stat wars and their casualties | 11 Comments

Mayo Casualties of O-Bayesianism and Williamson response


After Jon Williamson’s talk, Objective Bayesianism from a Philosophical Perspective, at the PhilStat forum on May 22, I raised some general “casualties” encountered by objective, non-subjective or default Bayesian accounts, not necessarily Williamson’s. I am pasting those remarks below, followed by some additional remarks and the video of his responses to my main kvetches. Continue reading

Categories: frequentist/Bayesian, objective Bayesians, Phil Stat Forum | 4 Comments

May 20: “Objective Bayesianism from a Philosophical Perspective” (Jon Williamson)

The ninth meeting of our Phil Stat Forum*:

The Statistics Wars
and Their Casualties

20 May 2021

TIME: 15:00-16:45 (London); 10:00-11:45 (New York, EST)

For information about the Phil Stat Wars forum and how to join, click on this link.

“Objective Bayesianism from a philosophical perspective” 

Jon Williamson Continue reading

Categories: Error Statistics | Tags: | Leave a comment

Tom Sterkenburg Reviews Mayo’s “Statistical Inference as Severe Testing: How to Get Beyond the Statistics Wars” (2018, CUP)

T. Sterkenburg

Tom Sterkenburg, PhD
Postdoctoral Fellow
Munich Center for Mathematical Philosophy
LMU Munich
Munich, German

Deborah G. Mayo: Statistical Inference as Severe Testing: How to Get Beyond the Statistics Wars

The foundations of statistics is not a land of peace and quiet. “Tribal warfare” is perhaps putting it too strong, but it is the case that for decades now various camps and subcamps have been exchanging heated arguments about the right statistical methodology. That these skirmishes are not just an academic exercise is clear from the widespread use of statistical methods, and contemporary challenges that cry for more secure foundations: the rise of big data, the replication crisis.

Continue reading

Categories: SIST, Statistical Inference as Severe Testing–Review, Tom Sterkenburg | 9 Comments

CUNY zoom talk on Wednesday: Evidence as Passing a Severe Test

If interested, write to me for the zoom link (

Categories: Announcement | Leave a comment

April 22 “How an information metric could bring truce to the statistics wars” (Daniele Fanelli)

The eighth meeting of our Phil Stat Forum*:

The Statistics Wars
and Their Casualties

22 April 2021

TIME: 15:00-16:45 (London); 10:00-11:45 (New York, EST)

For information about the Phil Stat Wars forum and how to join, click on this link.

“How an information metric could bring truce to the statistics wars

Daniele Fanelli Continue reading

Categories: Phil Stat Forum, replication crisis, stat wars and their casualties | Leave a comment

A. Spanos: Jerzy Neyman and his Enduring Legacy (guest post)

I am reblogging a guest post that Aris Spanos wrote for this blog on Neyman’s birthday some years ago.   

A. Spanos

A Statistical Model as a Chance Mechanism
Aris Spanos 

Jerzy Neyman (April 16, 1894 – August 5, 1981), was a Polish/American statistician[i] who spent most of his professional career at the University of California, Berkeley. Neyman is best known in statistics for his pioneering contributions in framing the Neyman-Pearson (N-P) optimal theory of hypothesis testing and his theory of Confidence Intervals. (This article was first posted here.) Continue reading

Categories: Neyman, Spanos | Leave a comment

Happy Birthday Neyman: What was Neyman opposing when he opposed the ‘Inferential’ Probabilists?


Today is Jerzy Neyman’s birthday (April 16, 1894 – August 5, 1981). I’m posting a link to a quirky paper of his that explains one of the most misunderstood of his positions–what he was opposed to in opposing the “inferential theory”. The paper is Neyman, J. (1962), ‘Two Breakthroughs in the Theory of Statistical Decision Making‘ [i] It’s chock full of ideas and arguments. “In the present paper” he tells us, “the term ‘inferential theory’…will be used to describe the attempts to solve the Bayes’ problem with a reference to confidence, beliefs, etc., through some supplementation …either a substitute a priori distribution [exemplified by the so called principle of insufficient reason] or a new measure of uncertainty” such as Fisher’s fiducial probability. It arises on p. 391 of Excursion 5 Tour III of Statistical Inference as Severe Testing: How to Get Beyond the Statistics Wars (2018, CUP). Here’s a link to the proofs of that entire tour. If you hear Neyman rejecting “inferential accounts” you have to understand it in this very specific way: he’s rejecting “new measures of confidence or diffidence”. Here he alludes to them as “easy ways out”. He is not rejecting statistical inference in favor of behavioral performance as typically thought. Neyman always distinguished his error statistical performance conception from Bayesian and Fiducial probabilisms [ii]. The surprising twist here is semantical and the culprit is none other than…Allan Birnbaum. Yet Birnbaum gets short shrift, and no mention is made of our favorite “breakthrough” (or did I miss it?). You can find quite a lot on this blog searching Birnbaum. Continue reading

Categories: Bayesian/frequentist, Error Statistics, Neyman | 3 Comments

Intellectual conflicts of interest: Reviewers


Where do journal editors look to find someone to referee your manuscript (in the typical “double blind” review system in academic journals)? One obvious place to look is the reference list in your paper. After all, if you’ve cited them, they must know about the topic of your paper, putting them in a good position to write a useful review. The problem is that if your paper is on a topic of ardent disagreement, and you argue in favor of one side of the debates, then your reference list is likely to include those with actual or perceived conflicts of interest. After all, if someone has a strong standpoint on an issue of some controversy, and a strong interest in persuading others to accept their side, it creates an intellectual conflict of interest, if that person has power to uphold that view. Since your referee is in a position of significant power to do just that, it follows that they have a conflict of interest (COI). A lot of attention is paid to author’s conflicts of interest, but little into intellectual or ideological conflicts of interests of reviewers. At most, the concern is with the reviewer having special reasons to favor the author, usually thought to be indicated by having been a previous co-author. We’ve been talking about journal editors conflicts of interest as of late (e.g., with Mark Burgman’s presentation at the last Phil Stat Forum) and this brings to mind another one. Continue reading

Categories: conflicts of interest, journal referees | 12 Comments

ASA to Release the Recommendations of its Task Force on Statistical Significance and Replication

The American Statistical Association has announced that it has decided to reverse course and share the recommendations developed by the ASA Task Force on Statistical Significance and Replicability in one of its official channels. The ASA Board created this group [1] in November 2019 “with a charge to develop thoughtful principles and practices that the ASA can endorse and share with scientists and journal editors.” (AMSTATNEWS 1 February 2020). Some members of the ASA Board felt that its earlier decision not to make these recommendations public, but instead to leave the group to publish its recommendations on its own, might give the appearance of a conflict of interest between the obligation of the ASA to represent the wide variety of methodologies used by its members in widely diverse fields, and the advocacy by some members who believe practitioners should stop using the term “statistical significance” and end the practice of using p-value thresholds in interpreting data [the Wasserstein et al. (2019) editorial]. I think that deciding to publicly share the new Task Force recommendations is very welcome, given especially that the Task Force was appointed to avoid just such an apparent conflict of interest. Past ASA President, Karen Kafadar noted: Continue reading

Categories: conflicts of interest | 1 Comment

The Stat Wars and Intellectual conflicts of interest: Journal Editors


Like most wars, the Statistics Wars continues to have casualties. Some of the reforms thought to improve reliability and replication may actually create obstacles to methods known to improve on reliability and replication. At each one of our meeting of the Phil Stat Forum: “The Statistics Wars and Their Casualties,” I take 5 -10 minutes to draw out a proper subset of casualties associated with the topic of the presenter for the day. (The associated workshop that I have been organizing with Roman Frigg at the London School of Economics (CPNSS) now has a date for a hoped for in-person meeting in London: 24-25 September 2021.) Of course we’re interested not just in casualties but in positive contributions, though what counts as a casualty and what a contribution is itself a focus of philosophy of statistics battles.

Continue reading
Categories: Error Statistics | Leave a comment

Reminder: March 25 “How Should Applied Science Journal Editors Deal With Statistical Controversies?” (Mark Burgman)

The seventh meeting of our Phil Stat Forum*:

The Statistics Wars
and Their Casualties

25 March, 2021

TIME: 15:00-16:45 (London); 11:00-12:45 (New York, NOTE TIME CHANGE TO MATCH UK TIME**)

For information about the Phil Stat Wars forum and how to join, click on this link.

How should applied science journal editors deal with statistical controversies?

Mark Burgman Continue reading

Categories: ASA Guide to P-values, confidence intervals and tests, P-values, significance tests | Tags: , | 1 Comment

Pandemic Nostalgia: The Corona Princess: Learning from a petri dish cruise (reblog 1yr)


Last week, giving a long postponed talk for the NY/NY Metro Area Philosophers of Science Group (MAPS), I mentioned how my book Statistical Inference as Severe Testing: How to Get Beyond the Statistics Wars (2018, CUP) invites the reader to see themselves on a special interest cruise as we revisit old and new controversies in the philosophy of statistics–noting that I had no idea in writing the book that cruise ships would themselves become controversial in just a few years. The first thing I wrote during early pandemic days last March was this post on the Diamond Princess. The statistics gleaned from the ship remain important resources which haven’t been far off in many ways. I reblog it here. Continue reading

Categories: covid-19, memory lane | Leave a comment

March 25 “How Should Applied Science Journal Editors Deal With Statistical Controversies?” (Mark Burgman)

The seventh meeting of our Phil Stat Forum*:

The Statistics Wars
and Their Casualties

25 March, 2021

TIME: 15:00-16:45 (London); 11:00-12:45 (New York, NOTE TIME CHANGE)

For information about the Phil Stat Wars forum and how to join, click on this link.

How should applied science journal editors deal with statistical controversies?

Mark Burgman Continue reading

Categories: ASA Guide to P-values, confidence intervals and tests, P-values, significance tests | Tags: , | 1 Comment

Falsifying claims of trust in bat coronavirus research: mysteries of the mine (i)-(iv)


Have you ever wondered if people read Master’s (or even Ph.D) theses a decade out? Whether or not you have, I think you will be intrigued to learn the story of why an obscure Master’s thesis from 2012, translated from Chinese in 2020, is now an integral key for unravelling the puzzle of the global controversy about the mechanism and origins of Covid-19. The Master’s thesis by a doctor, Li Xu [1], “The Analysis of 6 Patients with Severe Pneumonia Caused by Unknown Viruses”, describes 6 patients he helped to treat after they entered a hospital in 2012, one after the other, suffering from an atypical pneumonia from cleaning up after bats in an abandoned copper mine in China. Given the keen interest in finding the origin of the 2002–2003 severe acute respiratory syndrome (SARS) outbreak, Li wrote: “This makes the research of the bats in the mine where the six miners worked and later suffered from severe pneumonia caused by unknown virus a significant research topic”. He and the other doctors treating the mine cleaners hypothesized that their diseases were caused by a SARS-like coronavirus from having been in close proximity to the bats in the mine. Continue reading

Categories: covid-19, falsification, science communication | 22 Comments

Aris Spanos: Modeling vs. Inference in Frequentist Statistics (guest post)


Aris Spanos
Wilson Schmidt Professor of Economics
Department of Economics
Virginia Tech

The following guest post (link to updated PDF) was written in response to C. Hennig’s presentation at our Phil Stat Wars Forum on 18 February, 2021: “Testing With Models That Are Not True”. Continue reading

Categories: misspecification testing, Spanos, stat wars and their casualties | 11 Comments

R.A. Fisher: “Statistical methods and Scientific Induction” with replies by Neyman and E.S. Pearson

In Recognition of Fisher’s birthday (Feb 17), I reblog his contribution to the “Triad”–an exchange between  Fisher, Neyman and Pearson 20 years after the Fisher-Neyman break-up. The other two are below. My favorite is the reply by E.S. Pearson, but all are chock full of gems for different reasons. They are each very short and are worth your rereading. Continue reading

Categories: E.S. Pearson, Fisher, Neyman, phil/history of stat | Leave a comment

R. A. Fisher: How an Outsider Revolutionized Statistics (Aris Spanos)



This is a belated birthday post for R.A. Fisher (17 February, 1890-29 July, 1962)–it’s a guest post from earlier on this blog by Aris Spanos that has gotten the highest number of hits over the years. 

Happy belated birthday to R.A. Fisher!

‘R. A. Fisher: How an Outsider Revolutionized Statistics’

by Aris Spanos

Few statisticians will dispute that R. A. Fisher (February 17, 1890 – July 29, 1962) is the father of modern statistics; see Savage (1976), Rao (1992). Inspired by William Gosset’s (1908) paper on the Student’s t finite sampling distribution, he recast statistics into the modern model-based induction in a series of papers in the early 1920s. He put forward a theory of optimal estimation based on the method of maximum likelihood that has changed only marginally over the last century. His significance testing, spearheaded by the p-value, provided the basis for the Neyman-Pearson theory of optimal testing in the early 1930s. According to Hald (1998) Continue reading

Categories: Fisher, phil/history of stat, Spanos | 2 Comments

Reminder: February 18 “Testing with models that are not true” (Christian Hennig)

The sixth meeting of our Phil Stat Forum*:

The Statistics Wars
and Their Casualties

18 February, 2021

TIME: 15:00-16:45 (London); 10-11:45 a.m. (New York, EST)

For information about the Phil Stat Wars forum and how to join, click on this link. 


Testing with Models that Are Not True Continue reading

Categories: Phil Stat Forum | Leave a comment

Blog at