Why P-hacking is so prevalent in research of psychic/psi abilities, NDE, mediumship, etc

yoshy

Senior Member.
tl;dr: They were instructed to-p hack because their instructors didn't know any better. With p-hacking, you can find statistically significant evidence of anything you want, and often, researchers don't even realize they're p-hacking!

Boring background; skip if you already know what p-hacking is

I originally started writing this as a response to another MB post, Mediumship — triple blind study, but I decided to make a separate thread since the information is so widely applicable across the woo and pseudoscience world (as well as the real science world sadly - though to a lesser degree. See Replication crisis).

What is "p-hacking"? Wikipedia has a good overview of course:
External Quote:
Data dredging, also known as data snooping or p-hacking[1][a] is the misuse of data analysis to find patterns in data that can be presented as statistically significant, thus dramatically increasing and understating the risk of false positives. This is done by performing many statistical tests on the data and only reporting those that come back with significant results.[2] Thus data dredging is also often a misused or misapplied form of data mining.
P-hacking can be boiled down to any statistical analysis that increases the likelihood, or even guarantees, finding "statistically significant" results - the coveted p < 0.05. To avoid getting into the weeds, I'll just use the colloquial definition of statistically significant: the results you have found are likely "real" and not just a happy accident due to random variation. A simple example will show how easy it is to commit p-hacking unintentionally: you have a hypothesis (the effect of X on outcome Y is greater than 0); you collect your data, which includes independent variable X, dependent variable Y, as well as another independent variable Z; you perform your statistical test on both X and Z because modern software makes it so easy; you look at X's effect and, heartbreakingly, you get p > .05. But you also performed the test on Z, and hooray, that got p < 0.05! You write your paper with Z's effect as the focus, and all is good. Unfortunately, that's p-hacking because that wasn't your plan and is an example of the multiple comparisons problem.

Now, onto the actual topic:

Why are p-hacking and other bad stats so common in psychic, psi, NDE, mediumship, etc research?

Because researchers were literally instructed to do so! One of the primary culprits propagating this in the psi-related research field is Daryl J. Bem (sometimes referred to as D.J. Bem). Bem is well known and influential in the field, including a very popular publication Feeling the future: Experimental evidence for anomalous retroactive influences on cognition and affect, which found statistically significant evidence for psi abilities. Bem also wrote two popular articles on methodology: Writing an empirical article in 2000 and Writing the empirical article in 2003. Those articles have serious flaws, and they'll be the focus of this post. (2003 is mostly a rehashing of 2000, including copy/pasting entire paragraphs, with some minor additions. Since the important quotes are present in both versions, I will be citing the 2000 version exclusively. 2003 does not fix any of the conceptual mistakes made in 2000.)

Why Psychologists Must Change the Way They Analyze Their Data: The Case of Psi: Comment on Bem (2011) (I'll refer to it with "Wagenmakers et al. (2011)") thoroughly and clearly exposes the flawed recommendations. The authors break the issues down into 3 problems: Exploration Instead of Confirmation, Fallacy of the Transposed Conditional, and p Values Overstate the Evidence Against the Null. The first problem is most relevant to this post.

Wagenmakers et al. (2011) include a quote from Bem (2000):
To compensate for this remoteness from our participants, let us at least become intimately familiar with the record of their behavior: the data. Examine them from every angle. Analyze the sexes separately. Make up new composite indexes. If a datum suggests a new hypothesis, try to find further evidence for it elsewhere in the data. If you see dim traces of interesting patterns, try to reorganize the data to bring them into bolder relief. If there are participants you don't like, or trials, observers, or interviewers who gave you anomalous results, place them aside temporarily and see if any coherent patterns emerge. Go on a fishing expedition for something—anything—interesting. (Bem, 2000, pp. 4 –5)
None of the above recommendations are bad, per se. In fact, it's great practice to slice and dice your data to deduct useful insights. As Wagenmakers et al. (2011) explains, the issues arise with how you use the information, which the authors denote as "exploratory" or "confirmatory". If you use this exploratory data analysis EDA to generate new ideas for future experiments, that's great! However, if you use these findings in an article as if they were your intent all along and present them as confirmatory, you have just done some bad stats. You cannot use the same data to both generate your hypothesis AND test your hypothesis. Your p-values are invalid in that case.

Wagenmakers et al.:
Instead of presenting exploratory findings as confirmatory, one should ideally use a two-step procedure. First, in the absence of strong theory, one can explore the data until one discovers an interesting new hypothesis. But this phase of exploration and discovery needs to be followed by a second phase, one in which the new hypothesis is tested against new data in a confirmatory fashion.
Bem (2011) does actually attempt to warn against this but immediately follows with a fatal mistake:
If you still plan to report the current data, you may wish to mention the new insights tentatively, stating honestly that they remain to be tested adequately. Alternatively, the data may be strong enough and reliable enough to justify recentering your article around the new findings and subordinating or even ignoring your original hypotheses.
The first sentence is good. The second is back to bad stats. There is no way for the data to be "strong enough and reliable enough" to "[recenter] your article around the new findings and subordinating or even ignoring your original hypotheses." Since you have tested a new hypothesis on the same data from which you generated the hypothesis, there is no way to make the p-values valid. The p-values are inherently wrong, and you cannot adjust for it. Further, ignoring your original hypothesis is a terrible recommendation as it can mislead the reader into believing that the presented hypothesis was the original hypothesis all along.

A similar concept is the multiple comparisons problem. You can understand this concept quite easily: imagine a large dataset where you can generate hundreds or thousands of hypotheses (in this example, you have generated your hypotheses without looking at the data, contrary to the previous paragraph). The original effect you test will have a valid p-value where the probability of finding a "fake" effect is .05 as desired. But if you add another test after the first one failed to find a significant result, you now have a 1 - (1 - .05)^2 = 0.0975 probability of finding a false significant result. The error rate grows as 1 - (1 - .05)^n where n is the number of statistical tests performed. The probability of finding a fake effect asymptotically grows to 1 as you add more tests, essentially guaranteeing that you will find some statistically significant effect eventually. There is an important distinction in the case where you have multiple comparisons, but none of them are testing hypotheses that you generated by looking at the data. In the latter case, you can't fix it as I mentioned above. In the former, you can use corrections, such as the Bonferroni correction, to adjust for the multiple comparisons problem and make valid statistical conclusions.

This ended up being a much longer post than anticipated, and there is still more to be said, such as examples of bad things to look out for, good things to look out for (pre-registration and Bayesian analysis), and who knows what else. However, I'm tired, so I'll continue in a reply eventually. I'll also include commentary from another paper that mentions Bem (2011) - The garden of forking paths: Why multiple comparisons can be a problem, even when there is no fishing expedition or p-hacking and the research hypothesis was posited ahead of time by Andrew Gelman and Eric Loken (2013).
 
Last edited:

Significant

significant.png

So, uh, we did the green study again and got no link. It was probably a-- "RESEARCH CONFLICTED ON GREEN JELLY BEAN/ACNE LINK; MORE STUDY RECOMMENDED!"

Source: https://xkcd.com/882/ , or https://explainxkcd.com/882/ , which includes an explanation.
 
We in the United States are currently dealing with having Robert F Kennedy Jr as our new Secretary of Health. He, a vaccine opponent, wants to have new studies that connect vaccines with autism, despite the fact that there have already been many, many studies that disprove a correlation. In other words, every indication is that he wants his study to be confirmatory rather than exploratory, and I am concerned that some p-hacking can "find a connection" if that is what he demands.

The original "culprit" blamed was thimerosal, Now that it is no longer used in childhood vaccines, attention (and suspicion) has turned to aluminum compounds. But the results have recently been announced from a massive data study in Denmark:

External Quote:

A massive, 24-year-long study of more than 1.2 million children provides reassurance to parents around the world.

The research has found no compelling evidence that childhood vaccines lead to autism, asthma, or dozens of other chronic disorders.

Researchers in Denmark examined the safety of a specific vaccine ingredient – aluminum salts – which, despite frequent debunking, remains a common talking point among vaccine skeptics. Clinical trials have tested their safety extensively, and they've been used in non-live vaccines for more than 70 years to boost the immune system's response to lower doses of medicine.
https://www.sciencealert.com/study-...en-finds-no-risk-from-common-vaccine-additive

One would hope that the size and scale of this study might tamp down the tendency of vaccine skeptics to blame them for any adverse condition, but the current political situation that casts disdain on anything to do with science makes me fear that it will be ignored. And turning people away from vaccines can only boost the return of diseases (such as measles) that the vaccines are designed to suppress. It's happening right now, in areas with low vaccination rates.

External Quote:

Already in 2025, the United States has reported more measles cases than in any full year since the early '90s, according to federal data. And that record-setting case count has been driven by a massive outbreak that began in Texas.


The country's 2025 case count has broken record highs going back more than 30 years, according to data updated on July 9 by the U.S. Centers for Disease Control and Prevention.

Well over half of the U.S.'s 1,288 confirmed measles cases this year have stemmed from Texas' outbreak, which, according to state numbers, has so far led to more than 750 cases in the Lone Star State alone. More than 100 related cases have been reported in New Mexico and Oklahoma.

"Overwhelmingly, this is from the epidemic that began in Texas," said vaccine expert Dr. Peter Hotez. "Of the thousand-plus cases, the lion's share ... are because of a single epidemic that began in West Texas."

Hotez, the dean of the National School of Tropical Medicine at the Baylor College of Medicine, has warned about potential measles outbreaks for years. The Texas outbreak was entirely predictable, he said, based on anti-vaccine sentiment in the state and low vaccination rates in particular counties. That includes Gaines County, where the outbreak began.
https://medicalxpress.com/news/2025-07-measles-case-year-high-texas.html

Misuse and manipulation of data is not benign in any way. It can delude people, it can kill people, and it discredits the entire field of science, so we should call it out whenever we see it.
 
We in the United States are currently dealing with having Robert F Kennedy Jr as our new Secretary of Health. He, a vaccine opponent, wants to have new studies that connect vaccines with autism, despite the fact that there have already been many, many studies that disprove a correlation. In other words, every indication is that he wants his study to be confirmatory rather than exploratory, and I am concerned that some p-hacking can "find a connection" if that is what he demands.
I majorly share your concern. With deliberate p-hacking, you force your data to say whatever you want it to say. Or, to make things even easier, you can also just alter data (outright fabrication is next up):
Secret changes to major U.S. health datasets raise alarms
External Quote:
A new study in the medical journal The Lancet reports that more than 100 United States government health datasets were altered this spring without any public notice.
Many of the changes made were relatively low impact, idiotic rewordings of phrases that are too "woke", such as changing "gender diverse" to "include men and women.", but other changes deserve alarm:
External Quote:
In 89 cases, the revision affected text that defines the data itself, such as column names or category labels.
...
When variable labels shift from "gender" to "sex" in these resources, studies that compare answers given under the old wording with figures retrieved after the change are no longer aligning like‑with‑like.
Scientists differentiate between gender and sex so changing those words is actually a big deal. Changing column names is annoying as hell because now everybody has to rewrite their code.

The original "culprit" blamed was thimerosal, Now that it is no longer used in childhood vaccines, attention (and suspicion) has turned to aluminum compounds. But the results have recently been announced from a massive data study in Denmark:

External Quote:

A massive, 24-year-long study of more than 1.2 million children provides reassurance to parents around the world.

The research has found no compelling evidence that childhood vaccines lead to autism, asthma, or dozens of other chronic disorders.

Researchers in Denmark examined the safety of a specific vaccine ingredient – aluminum salts – which, despite frequent debunking, remains a common talking point among vaccine skeptics. Clinical trials have tested their safety extensively, and they've been used in non-live vaccines for more than 70 years to boost the immune system's response to lower doses of medicine.
https://www.sciencealert.com/study-...en-finds-no-risk-from-common-vaccine-additive

One would hope that the size and scale of this study might tamp down the tendency of vaccine skeptics to blame them for any adverse condition, but the current political situation that casts disdain on anything to do with science makes me fear that it will be ignored.
Great study! In fact, a very smart person already made a post for that study on this forum :p
Misuse and manipulation of data is not benign in any way. It can delude people, it can kill people, and it discredits the entire field of science, so we should call it out whenever we see it.
Preach!
 
Misuse and manipulation of data is not benign in any way. It can delude people, it can kill people, and it discredits the entire field of science, so we should call it out whenever we see it.
Absolutely.

External Quote:

A child has died at Liverpool's Alder Hey Children's Hospital after contracting measles, the BBC understands.

The hospital said the highly contagious virus was on the rise among young people in the region and it had seen a surge in "seriously unwell" children being admitted.
... ...
The World Health Organization (WHO) wants 95% of children fully vaccinated by their fifth birthday.
In the north-west of England the figure is 85%, with lower rates of 73% per cent in Liverpool and 75% Manchester, according to NHS data.
... ...
Professor Matthew Ashton, Liverpool director of public health, said one person with measles can infect 15 others.
He said there was "no link whatsoever" between the MMR vaccine and autism and this claim had been disproved several times.
BBC News, Local News, Liverpool; "Child dies at Alder Hey after contracting measles", 13 July 2025 https://www.bbc.co.uk/news/articles/c8j1k3k44e2o

Yet another tragic reason to condemn Andrew Wakefield's fraudulent 1997-1998 "research" into the MMR vaccine (https://en.wikipedia.org/wiki/Lancet_MMR_autism_fraud)
 
Re the thread Mediumship - triple blind study, discussion has recently moved on to a different "triple-blinded" study conducted in 2022 (from @Merle's post #67 onwards).
Merle provided a useful link to some supporting information from that paper's authors, https://figshare.com/articles/dataset/Mediumship/13311710?file=44625439, including a file, "RR Mediumship.Rmd",
from which I copied this detail:

tsl.JPG


At risk of looking daft to you stats and probability folk, that is saying that the researchers are using 50% chance to determine if results are significant, isn't it? Genuine question.
I'm guessing that the researchers did actually use p=0.05 as claimed, and didn't accidentally set p=0.5 in some software package used to do the donkey work.

The open access peer-reviewed journal PLOS One caused something of a stir when it started off some years ago.
Maybe we should start a journal, p = One, with the tagline, "Where all results are significant."
I'm sure there would be no shortage of submissions...
 
Re the thread Mediumship - triple blind study, discussion has recently moved on to a different "triple-blinded" study conducted in 2022 (from @Merle's post #67 onwards).
Merle provided a useful link to some supporting information from that paper's authors, https://figshare.com/articles/dataset/Mediumship/13311710?file=44625439, including a file, "RR Mediumship.Rmd",
from which I copied this detail:

View attachment 82558

At risk of looking daft to you stats and probability folk, that is saying that the researchers are using 50% chance to determine if results are significant, isn't it? Genuine question.
I'm guessing that the researchers did actually use p=0.05 as claimed, and didn't accidentally set p=0.5 in some software package used to do the donkey work.

The open access peer-reviewed journal PLOS One caused something of a stir when it started off some years ago.
Maybe we should start a journal, p = One, with the tagline, "Where all results are significant."
I'm sure there would be no shortage of submissions...
Not daft at all. This is an easy thing to confuse when you're not familiar with R (the statistical programming language) and the finer details of the field of statistics. I have seen this misunderstanding a lot. The p you have circled there is not the "p-value". The argument p in this case is the probability of success in a binomial random variable. That may sound unfamiliar, but it's not! The number of heads after flipping a coin multiple times is a binomial random variable with a 0.5 probability of heads (assuming fair coin).

Using the coin flip example again. They are testing to see if the number of heads they observed is reasonable under the null hypothesis of 0.5 probability of heads. Here's a sample of what the output looks like, for n = 100 (the number of coin flips), x = 25 (25 heads), p = .5 (null hypothesis, i.e. a fair coin), and conf.level = .95 (95% confidence interval).

Code:
binom.test(25, 100, p = 0.5, conf.level = .95, alternative = "less")

    Exact binomial test

data:  25 and 100
number of successes = 25, number of trials = 100, p-value = 2.818e-07
alternative hypothesis: true probability of success is less than 0.5
95 percent confidence interval:
 0.0000000 0.3313216
sample estimates:
probability of success
                  0.25

With 25 heads, your p̂ (said p-hat) is .25, where p̂ is the estimate of the true probability of heads that is calculated from the data. p̂ is the value you compare against p from your null hypothesis (0.5).

The p-value is 2.818e-07 or 0.00000028, which is, of course, < .05, so we would call this statistically significant. That p-value is the probability you would get 25 or fewer heads, assuming the coin is fair with probability of heads = .5.

Similarly, the 95% confidence interval (where .95 = 1 minus the significance threshold .05) does not contain 0.5, so it is another way to conclude that your coin's probability of heads is less than 0.5.

Simple summary:
pNull hypothesis. The assumed probability of getting heads from a single coin flip. This is the p = 0.5 from the code you circled.
"p-hat". The probability of getting heads calculated from the data
p-valueThe probability of getting the number of heads or fewer in your experiment, calculated using the value of p in the null hypothesis. This is the value typically compared against .05 to determine "statistical significance"

Note: my example used alternative = "less" whereas they used alternative = "greater". Sorry for any confusion that may cause! It doesn't change anything substantial luckily. It is the symmetric equivalent of getting 75 heads out of 100 flips.
 
Last edited:
Note: my example used alternative = "less" whereas they used alternative = "greater". Sorry for any confusion that may cause! It doesn't change anything substantial luckily. It is the symmetric equivalent of getting 75 heads out of 100 flips.
Great post, thank you. I would have created a dog's dinner compared to that.

With a null hypothesis of there being nothing weird, say a coin flip being just a 50/50 outcome (don't put a huge neodymium magnet under the table), if you think you've got a weird outcome, it's common to consider both extremes together - a so-called "two-tailed" test. Failing to so do (doing a "one-tailed test") lets you pretend your p-value is half what it might be had you done things the more traditional way.
 
With a null hypothesis of there being nothing weird, say a coin flip being just a 50/50 outcome (don't put a huge neodymium magnet under the table), if you think you've got a weird outcome, it's common to consider both extremes together - a so-called "two-tailed" test. Failing to so do (doing a "one-tailed test") lets you pretend your p-value is half what it might be had you done things the more traditional way.
In the parallel thread on mediums contacting the dead we seem to be faced with exactly that situation, where it is not a yes/no coin toss type of answer, but a series of responses heavily weighted toward "yes".

https://www.metabunk.org/threads/mediumship-—-triple-blind-study.13149/post-301305
 
With a null hypothesis of there being nothing weird, say a coin flip being just a 50/50 outcome (don't put a huge neodymium magnet under the table), if you think you've got a weird outcome, it's common to consider both extremes together - a so-called "two-tailed" test. Failing to so do (doing a "one-tailed test") lets you pretend your p-value is half what it might be had you done things the more traditional way.
That's a good detail to point out! I figured my comment was getting long enough so I omitted any discussion of one-tailed vs two-tailed tests. It's also one of those things that I don't think doing a one-tailed test is necessarily "right" or "wrong".

(Though it's a good example for the supremacy of Bayesian methodology!!! Hehe partly hyperbolic, but the topic of Bayesian vs Frequentist stats is something I'll post about eventually)
 
In the parallel thread on mediums contacting the dead we seem to be faced with exactly that situation, where it is not a yes/no coin toss type of answer
or so they would have us believe
p may be greater than 0.5, and then the confidence changes
 
I originally started writing this as a response to another MB post, Mediumship — triple blind study, but I decided to make a separate thread since the information is so widely applicable across the woo and pseudoscience world (as well as the real science world sadly - though to a lesser degree. See Replication crisis).

What is "p-hacking"? Wikipedia has a good overview of course:
External Quote:
Data dredging, also known as data snooping or p-hacking[1][a] is the misuse of data analysis to find patterns in data that can be presented as statistically significant, thus dramatically increasing and understating the risk of false positives. This is done by performing many statistical tests on the data and only reporting those that come back with significant results.[2] Thus data dredging is also often a misused or misapplied form of data mining.

I agree with your analysis on p-hacking, but I don't think it applies to the study you mention that was discussed recently at Mediumship — triple blind study,

That study seems to be a specific study testing a particular hypothesis with a statistically significant result in the specific variable they were testing. Yes, if there had been hundreds of variations of the variables in the data they were looking at, and if most variations supported the null hypothesis except the one they reported, then that would be p-hacking.

But, I don't think we can say that, since p-hacking can be a problem, therefore, these people were probably p-hacking, and therefore there is nothing to see here.

As we discussed on that thread, that study appears to be a biased study by people supporting a particular business activity (mediums). That is an obvious conflict of interest. They published a paper that not even they believed represents something that was really happening. For if they thought the spirits of dead people were really seeing what was going on in this world, and able to report those things to us, that has immense application in spying, gambling, and other enterprises. If they really thought this scheme worked, they would be pushing the huge applications available to them. But they stick to the dubious business of cold-reading psychics with gullible people and never try to use this remarkable technology--if it worked--to the huge benefit it would have for humanity.

So, biased, outlier studies like this can be wrong for many reasons, including possible p-hacking. But why was this particular study wrong? We can say many possible reasons this study is not trustworthy. But possible does not equal probable. In that thread, we are discussing what is probable.
 
I agree with your analysis on p-hacking, but I don't think it applies to the study you mention that was discussed recently at Mediumship — triple blind study,
Just a heads up that I haven't read that study in detail, so some of my response could be off base due to that.
That study seems to be a specific study testing a particular hypothesis with a statistically significant result in the specific variable they were testing. Yes, if there had been hundreds of variations of the variables in the data they were looking at, and if most variations supported the null hypothesis except the one they reported, then that would be p-hacking.
P-hacking doesn't require having hundreds of tests and picking the one statistically significant one. It could also be something like: you have a plan for an experiment with statistical test A => you do A, get non-significant results => you look for and find something else (test B) that's significant => report test B with no mention of test A at all. For the specific study you're investigating, do you know if they preregistered it anywhere? "Preregistration" refers to publicly announcing the entire methodology for a study, start to finish, including the statistical tests that will be run. That's one way to know they didn't pull a switcheroo of their statistical tests. There's now a bunch of websites for preregistering studies. (Just google "preregistration website" to see.)
But, I don't think we can say that, since p-hacking can be a problem, therefore, these people were probably p-hacking, and therefore there is nothing to see here.
Sorry, I didn't mean to imply p-hacking was the explanation for sure. That would be a logical fallacy of course!
So, biased, outlier studies like this can be wrong for many reasons, including possible p-hacking. But why was this particular study wrong? We can say many possible reasons this study is not trustworthy. But possible does not equal probable. In that thread, we are discussing what is probable.
I originally meant to take a deep dive into that particular study, but I pivoted to writing a general post on p-hacking. I do intend to take a close look at the study (eventually lol).
 

Avoiding P-Hacking

This comment will focus on just one possible prevention tactic (of many):

Preregistration

Wikipedia has a nice description of Preregistration.
External Quote:
Preregistration is the practice of registering the hypotheses, methods, or analyses of a scientific study before it is conducted.[1][2] Clinical trial registration is similar, although it may not require the registration of a study's analysis protocol. Finally, registered reports include the peer review and in principle acceptance of a study protocol prior to data collection.[3]

Preregistration has the goal to transparently evaluate the severity of hypothesis tests,[4] and can have a number of secondary goals (which can also be achieved without preregistering [5]), including (a) facilitating and documenting research plans, (b) identifying and reducing questionable research practices and researcher biases,[6] (c) distinguishing between confirmatory and exploratory analyses,[7] and, in the case of Registered Reports, (d) facilitating results-blind peer review, and (e) reducing publication bias.[8]
Put really simply: preregistration is just telling everybody publicly what you plan to do in a study. In the context of p-hacking, you can't pretend that test B was your plan all along when you have publicly preregistered test A. As I mentioned above, there's a whole bunch of websites where you can do this. Here's a few just off the top of Google: http://cos.io, aspredicted, osf.io, phdontrack.net has its own list. Actual example from clinicaltrials.gov for a study about dietary cholesterol, eggs, and LDL. This screenshot shows just a portion of what's included in this preregistration. It gets really in depth!
Screenshot 2025-07-27 at 1.47.13 PM.png


A funny example of preregistration working

I learned of this example from fitness YouTuber Layne Norton in the video titled New Keto Sponsored Study Shows High LDL Causes RAPID Plaque Progression! | Educational Video
The topic is a paper Longitudinal Data From the KETO-CTA Study: Plaque Predicts Plaque, ApoB Does Not by a team of scientists who are known for going against the current scholarly consensus of diet research. (KETO, carnivore diet, that whole subculture where its adherents are often very aggressive towards the "establishment".) For example, Nick Norwitz, the second author, is quite zealous for spreading his personal theories and has his own youtube channel where he previously attacked Layne Norton and other PhDs/Drs who represent the consensus positions.

Anyway, back to Longitudinal Data From the KETO-CTA Study, once again, the study found evidence against the current consensus (and evidence for Norwitz's model Lipid Energy Model that he is constantly pushing). Or did it really? Let's check the preregistrated study plan.
This is the entirety of the section

What is the study measuring?

Outcome MeasureMeasure DescriptionTime Frame
Percent change in total non-calcified coronary plaque volumePercent change in total non-calcified coronary plaque volume from baseline (start of the study) till the final visit will be measured using Coronary Computed Tomography Angiography (CCTA).12 months

Now the funny part: The study DOES NOT MENTION Percent change in total non-calcified coronary plaque volume. I.e., the researchers completely OMIT THE PRIMARY OUTCOME, the single piece of data they were supposed to have! It is simply embarrassing for the journal to have let this through peer review.

Some commentary on the situation: A Textbook Case for Unethical Reporting of Research. Note that the article is free, but you have to sign up and claim it as your monthly free article.
External Quote:

So the primary outcome, the measurement upon which the entire study was constructed to test, is omitted in the final results. Thanks to pressure from the scientific community, the lead author, Adrian Soto-Mota, finally admitted in a post on X that the "… numerical pooled NCPVchange value, it is: p50=18.8 mm³."

What does that value mean?

For comparison, in another study evaluating a healthy population of people who did not have any coronary events or risk factors, the change in NCPV was 4.9 mm3. So the LMHR participants within this study developed plaque in their arteries at nearly four times the rate of a healthy population.

...

Why were these findings relegated to the shadows of this paper? It's not difficult to devise why. Each of the study authors, along with the Citizen Science Foundation that funded the paper have a vested interest in promoting the keto-carnivore dietary pattern. Each believes that all of humanity has gotten it wrong, that elevations in a person's LDL-C and ApoB are not risk factors for the development of heart disease. The findings of this paper report exactly the opposite, as these LMHRs with exceedingly high LDL-C are developing arterial plaques at rates which are magnitudes higher than the general population.

But the lack of ethics becomes even more glaring as the authors try to conclude how:

… a 1-year prospective study of 100 persons exhibiting extreme carbohydrate restriction-induced elevations in LDL-C and ApoB, changes in and baseline levels of ApoB were not associated with changes in noncalcified plaque volume or total plaque score as measured by [coronary computed tomography angiography]. — Brackets added.
This is an asinine conclusion for a simple reason — there was no variability of exposure in LDL-C and ApoB with regards to plaque development. The average reported LDL-C among these LMHRs was 254 mg/dL, more than 2.5 times the recommended cutoff for CVD risk.

...

It is difficult to conclude that these omissions and presentations of data were made in good faith or by oversight. Such is confirmed by a physician who walked away from the research project over ethics violations. Dr. Nadolsky, commenting on the conduct of the study team in a reel produced by Dr Idrees Mughal declared:
Yes they are some of the most dishonest "researchers" out there. As the guy who created the original protocol and left the study team, I can tell you this team is about as bad as it comes. Thank you for spreading awareness that they are spinning this to fit their narrative.
[emphasis mine]

Addendum

Common Question: What if I think of something really important to add after I've already preregistered?

Good question! Not a problem. You can add analysis if you really think it's necessary. Preregistration just makes the researchers communicate openly that the analysis was added ad hoc. This makes it more clear when a multiple testing correction should be applied.

Fun aside:
Norwitz is also notable for his self study Oreo Cookie Treatment Lowers LDL Cholesterol More Than High-Intensity Statin therapy in a Lean Mass Hyper-Responder on a Ketogenic Diet: A Curious Crossover Experiment where he claimed
External Quote:

short-term Oreo supplementation lowered LDL-C more than 6 weeks of high-intensity statin therapy in an LMHR subject on a ketogenic diet.
 
Put really simply: preregistration is just telling everybody publicly what you plan to do in a study. In the context of p-hacking, you can't pretend that test B was your plan all along when you have publicly preregistered test A.
By pure coincidence this week's weekend reads from /Retraction Watch/ has this story:
External Quote:

Microbiome company CEO who linked COVID vaccine to bacterial decline now has four retractions


A gastroenterologist and microbiome researcher who has promoted hydroxychloroquine and ivermectin as COVID treatments has lost a paper after a sleuth reported differences between the article and the registered protocol of the clinical trial it purported to describe.
https://retractionwatch.com/2025/07...n-covid-sarscov2-fecal-samples-gut-pathogens/
 
By pure coincidence this week's weekend reads from /Retraction Watch/ has this story:
External Quote:

Microbiome company CEO who linked COVID vaccine to bacterial decline now has four retractions


A gastroenterologist and microbiome researcher who has promoted hydroxychloroquine and ivermectin as COVID treatments has lost a paper after a sleuth reported differences between the article and the registered protocol of the clinical trial it purported to describe.
https://retractionwatch.com/2025/07...n-covid-sarscov2-fecal-samples-gut-pathogens/
Lmao good catch! This kind of fraud blows my mind since it is so easy to catch! You literally just have to look at the study and their preregistration. Do they think they'll get away with it?!
 
Last edited:
Now the funny part: The study DOES NOT MENTION Percent change in total non-calcified coronary plaque volume. I.e., the researchers completely OMIT THE PRIMARY OUTCOME, the single piece of data they were supposed to have!

Yes, your point is valid. Simply firing a shotgun and reporting only the one pellet that hit the target does not make someone a good marksman.

There is a possibility that something like this happened in the study we've been discussing (http://www.patriziotressoldi.it/cmssimpled/uploads/images/IsThereSomeoneInTheHereafter_22.pdf).

However, in this particular case, I think the issue may have more to do with potential social interactions between the sitters and the mediums outside of the study. It appears the sitters were drawn from the same social pool as the mediums. No controls are mentioned for preventing outside interaction. The authors emphasize the various levels of blinding within the study, but if the mediums had already learned through social interactions that some likely sitters would be asking about a particular Teresa, they could have easily worked that information into their cold reading. This would increase the chances that the sitters would identify that reading as being from the Teresa they had in mind.

20250731_0650_Bullseye Achievement Celebration_simple_compose_01k1g1n16afk69p3fjzh6v895t.PNG

By author using ChatGPT SORA.
 
Last edited:
Yes, your point is valid. Simply firing a shotgun and reporting only the one pellet that hit the target does not make someone a good marksman.

There is a possibility that something like this happened in the study we've been discussing (http://www.patriziotressoldi.it/cmssimpled/uploads/images/IsThereSomeoneInTheHereafter_22.pdf).

However, in this particular case, I think the issue may have more to do with potential social interactions between the sitters and the mediums outside of the study. It appears the sitters were drawn from the same social pool as the mediums. No controls are mentioned for preventing outside interaction. The authors emphasize the various levels of blinding within the study, but if the mediums had already learned through social interactions that some likely sitters would be asking about a particular Teresa, they could have easily worked that information into their cold reading. This would increase the chances that the sitters would identify that reading as being from the Teresa they had in mind.

View attachment 82729
By author using ChatGPT SORA.
This is a good point too. We haven't gone down that road yet but I do think we all somewhat know about that sort of dynamic. It's something we saw with "Remote Viewing" too where a lot of the RVers did in fact actually have access too and viewed intelligence products related to the things they were RVing. Eg plenty of cases where they had access too aerial imagery of the very areas they claim to have remote viewed things (tid bit different but very similar issue IMO)
 
Back
Top