Brief Outpatient Rehabilitation Program for Post-COVID-19 Condition, Nerli et al, 2024 - with comment from T. Chalder

A group difference of 10 point on the SF-36-PPS was stated as a “clinically significant difference” in the specific context of sample size determination. However, this statement does not imply that a slightly smaller difference is clinically unsignificant.
What else does it imply? Clinically significant is a minimum threshold value. Why bother even defining such a value if you are going to then ignore it?

Sorry, I forgot. Psycho-behavioural studies don't have to follow standards, not even their own.
 
What else does it imply? Clinically significant is a minimum threshold value.
I know it’s a rhetoric question, but I tried to look at what they’ve said in the protocol.

I believe that they are trying to argue that they only said that 10 points would be clinically significant, as in «10 points is over the threshold for minimally clinical significance», but the threshold might be lower.

The problem with that line of reasoning, is that they stated that «similarly» to their value of 10 from another study, 10 was the threshold for ME/CFS.
The SF-36 subscale Physical Functioning (SF-36-PFS) will serve as the primary endpoint in the present study. A difference of 10 points is considered clinically significant. 43

Similarly, in a study of CFS/ME which shares similarities with post-Covid syndrome, the minimally clinically important difference of SF-36-PFS was reported to be 10.
So they are trying to say that they didn’t say that 10 was the minimal value for CID.

However, other papers defined CID as the threshold, so the minimal part is always implied.
Whether investigators assess the group- or individual-level CID, results demonstrating the effects of an intervention should be presented in relation to the CID.
10.1177_21925682221092721-fig1.jpg

https://pmc.ncbi.nlm.nih.gov/articles/PMC9210237/

So their argument shows that they don’t understand what CID means, or that they chose to ignore the meaning because it suited them. Neither are very flattering..
 
So their argument shows that they don’t understand what CID means, or that they chose to ignore the meaning because it suited them. Neither are very flattering..
It mostly shows that they understand that they can provide a picture of poop on a platter in response and no one but us will be bothered by it.

And they are sadly correct. Which is a terrible indictment of academia, but they are simply correct in that nothing they say or do matters, they can still pretend like it's whatever they market it as.
 
Opinion piece by Richard White - he finally got access to some of the data:
How many actually recovered?
It was only after I complained to the journal JAMA Network Open that the authors published the age- and sex-adjusted recovery rates. These showed that only 8 percent of participants in the control group recovered after twelve months—a startlingly low rate that should raise serious concerns among clinicians and health authorities.

In the intervention group, the proportion was 29 percentage points higher. However, since participants self-assessed the outcomes (for example, “Is your health such that it limits you from moderate activities such as vacuuming?”), and knew whether they received the intervention or not, it is unclear how much of the intervention’s modest effect is due to the placebo effect and response bias.

This means that participants may have overestimated their improvement because they expected an effect, or wanted to confirm the researchers' assumptions. Objective measures, such as the number of steps per day, would likely have shown even weaker results.

An unblinded psychosocial treatment with a recovery rate of only 29 percent on a subjective measure under ideal experimental conditions should not be presented as a solution for a complex, multisystemic condition like long covid.
Nerli answers and ignores most of the criticism or just says that it’s normal with subjective outcomes.
All studies have both strengths and weaknesses. The effect of a method must therefore be assessed across studies: A Dutch study of cognitive behavioral therapy could show similar results to ours, and results from at least two more studiesare expected this year. Together, they will form the basis for further recommendations and knowledge summaries .
The fact that other studies make the same mistakes and get the same results obviously does not matter for the validity of the results. I think this is a textbook example of someone that’s «methodologically challenged».

And he wants to give this treatment to everyone - shocker!
The results from our study are so promising that it should be offered to the patient group that currently has no other documented effective treatment available.
 
There is a huge problem in medicine in how they assess the validity of self-rated improvement, which doesn't come close to understanding just how difficult it is to assess.

I had a really hard lesson over this in the last 2 years. I have improved considerably. 2 years ago, I was bedbound. Mainly because of out-of-control POTS, which has improved a lot since. But even with this, I was much better on several aspects of ME/CFS than I had been the year before. Mainly in odd neurological symptoms, sound intolerance, and so on. I was able to hold basic conversations for the first time in probably 4-5 years.

Since then I have slowly improved, but it only made it so obvious just how hard it is to estimate those things. At some point I used to think I was maybe at a 20-30% range of overall capacity, but seeing how much I improved since then, and how much more there is to improve, made it very clear how this 20-30% range was ridiculously over-inflated, how it was more of a 5-6% range. There's too much to consider, it's like trying to guess the number of items in a jar when there are hundreds of thousands of them, and you can only view them from a single (misleading) angle. There's just no possible accuracy here.

It's basically unrealistic to guesstimate something as complex like this on a limited, and not even linear, single digit scale. It just flat out out doesn't work. There has to be metrics, standards and objective measures to compare to. Even the simple joke of using a banana for scale makes most size guesstimates more reliable than any standard health questionnaire I have seen applied to us.

What health care professionals don't understand is that this isn't a problem for us, it's a problem, period. It's irreconcilable. There is no statistical analysis that can turn this mess of wildly inaccurate guesstimates into real practical numbers. They end up being so inaccurate in many cases it's actually worse than not trying to do that. Because they always do the thing where if the numbers look good, they are important and accurate, and if not, well they just have to figure out how to get the numbers just right. Whatever the numbers mean, which is usually both a number of unrelated things, and nothing at all. It just doesn't work like that at all.

But what it means more than anything is that none of those studies are reliable. Straight up not at all. They are not asking valid questions, making anything that comes after, analysis, interpretation, and so on, entirely arbitrary and useless for the most part.

What they are trying to do is the equivalent of doing chemistry without a reliable thermometer. It's just not possible. They could instead try to train a bunch of assistants at guessing the temperature of things by touch, and it would be exactly as useful as how questionnaires are misused in health care. They could show how most of the assistants usually have the same rough guesstimate ranges, and it still doesn't matter. They can show how multiple guesstimates made by the same person also tend to cluster near the same value, and it still doesn't matter: not accurate.

The root cause for much of this is something they will have to let go eventually: they confuse that the best they can do must be good enough, just because it's the best they can do. No, it's not. Especially not this. They even do the same thing with treatments that they poorly assess, here and usually, thinking that since this rehabilitation stuff is the best they can do, then it must be good enough. I don't know what logical fallacy this is, but it's probably the mother of all logical fallacies: I think X therefore I am right about X because me thinking X is the best I can do.
 
who is he?
Richard is an applied infectious diseases statistician based in Oslo, Norway. He is currently employed as the project manager for the Norwegian Syndromic Surveillance System (NorSySS), a surveillance system of infectious diseases based on consultations with general practitioners and out-of-hours primary care facilities, based in the Norwegian Institute of Public Health (Folkehelseinstituttet).
Really nice and competent guy that has been vocal about Norway’s failure to address the pandemic for many years.
https://www.rwhite.no/publications.html
 
who is he?
I think he is Australian. He has a PhD in biostatistics from Harvard and is now working at the Norwegian Institute for Public Health (yes, that institute! With Flottorp and co..). He has written several superb opinion pieces and given several interviews on Long Covid and has published research into it as well. Great guy with scientific principles! You can connect with him on Bluesky.
 
"Trial By Error: Pushback on “Brief Outpatient Rehab” Trial for Long Covid from Norwegian Ideological Brigades

I often find myself responding to crap studies–such as a Norwegian study called “Brief Outpatient Rehabilitation Program for Post–COVID-19 Condition: A Randomized Clinical Trial,” from Nerli et al., published last December by JAMA Network Open. The senior author was Professor Vegard Wyller, the dean of the Norwegian wing of the CBT/GET/Lightning Process ideological brigades. The trial was as bad as one would expect, given its provenance. (Here’s my post about it.)

It is refreshing when other academics issue their own challenges to the kinds of self-evident methodological lapses that mar this study and so much of this research. Richard Aubrey White, a Harvard-trained biostatistician and a researcher at the Norwegian Institute of Public Health (NIPH), has just published a tough take-down of Nerli et al.—to which the lead author has responded with more of the usual blah blah.

The article appeared in Forskning.no, an online publication focused on news about Norwegian and international research, and includes a note that White is not writing on behalf of his employer. (NIPH has been a hotbed of support for the so-called “biopsychosocial” paradigm, so White’s opinion might be unpopular in some quarters.)

White’s critique—called “Promising treatment for long covid without evidence”–highlights similar issues to those I raised. (I am reliant on Google Translate, so don’t hold me responsible for any mistranslations.)"
 
Finally, one of his suggestions is a real whopper of delusional thinking: “The results from our study are so promising that it should be offered to the patient group that currently has no other documented effective treatment available.”
It's stuff like this that makes it all so... insane. No one in their right mind would ever think such a thing, let alone say it. Results from a single excessively biased study may, in some isolated cases, seem promising, but not for decades, not literally after decades of already offering this exact type of treatment. And, still, no one in their right mind should actually think that it's a good idea to offer some treatment based on being promising on the basis of a single highly biased pilot study, unless that study is truly groundbreaking.

And this is about as groundbreaking as a feather in the wind. It will not break any ground. It will barely sweep a few dust motes before it gets blown by the wind.

The behavior of the medical professionals and its institutions is so out of touch with what they normally do when they take an issue seriously. Most of them would recoil in horror at such suggestions when it isn't about something that has been exempted from the most minimal standards.

It's genuinely insane behavior. And yet it's happening quite rationally, involving people who do know better, and institutions who are supposed to guard against this. And yet they literally all fail miserably at the most basic level. Which can only mean that:
there is no point in claiming that the purpose of a system is to do what it constantly fails to do
https://en.wikipedia.org/wiki/The_purpose_of_a_system_is_what_it_does

No one seems to think about what it means for such exemptions to become commonplace, how degrading standards in how place erodes them everywhere, because then it becomes a single mental flip to simply turn off all their normal training. It's the same slippery slope that starts with dehumanizing a few people, then some more, then a lot more, then literally everyone can be deprived of basic rights and protections.

Or you get stuff like RFK, who is genuinely insane but there's still a significant overlap between what he says and believes in and the medical profession. And they don't even get that, blinded by exceptions.
 
Back
Top