Trial By Error: Trudie Chalder Is Co-Author on Another Bad Exercise Paper

By David Tuller, DrPH

It is a truth universally acknowledged (or at least universally acknowledged by smart researchers), that if the list of authors on an article includes Trudie Chalder, King’s College London’s mathematically and factually challenged professor of cognitive behavior therapy, then the article in question should most assuredly be expected to be short on, or utterly devoid of, intelligence and logical reasoning.

This is certainly the case with a recent publication in European Respiratory Journal titled “Post-Hospitalisation COVID-19 Rehabilitation (PHOSP-R): A randomised controlled trial of exercise-based rehabilitation.” Professor Chalder is one of more than three dozen co-authors, so it is unclear how much she can be held responsible for the article’s poor quality and unwarranted claims. Nonetheless, this new trial continues what appears to be her impressive streak of being involved with scholarship that can accurately be described as rubbish.

The new trial, conducted at the University of Leicester and Northumbria University, was funded by the Medical Research Council and the National Institute for Health Research. The design was not terrible. Specifically, unlike some other Long Covid rehabilitation studies, this one excluded participants with post-exertional malaise (PEM). That’s a good move, given that the presence of PEM is a contra-indication for an exercise-based rehab program.

So this was a study of Long Covid patients who do not meet criteria for an ME/CFS diagnosis—a key point.

The real problem here is that the reporting of the results stinks.

In the study, 181 participants who were experiencing prolonged symptoms after a Covid-related hospital stay were randomized to either an eight-week face-to-face exercise rehabilitation program, an eight-week remote exercise rehabilitation program, or care as usual. The primary outcome was the change in the Incremental Shuttle Walking Test (ISWT). Among the many secondary outcomes were questionnaires measuring health quality of life and symptom burden.

The conclusion: “Exercise-based rehabilitation improved short-term exercise capacity in Post-COVID syndrome following an acute hospitalization.”

First, let’s note that the conclusion does not explicitly state that the study is not about “post-COVID syndrome” overall but only about non-ME/CFS “post-COVID syndrome.” That’s a major limitation of the findings that should have been emphasized prominently throughout, given the significant numbers of Long Covid patients who do experience PEM and qualify for ME/CFS diagnoses. Health care providers will read the conclusions and assume they can be extrapolated to all Long Covid patients. That is clearly not the case.

Beyond that unacceptable oversight, let’s review the data in a bit more detail and see if that claim holds up.

The face-to-face intervention group had a drop-out rate of 29% and the remote intervention group had a drop-out rate of 39%. These drop-out rates are quite high. Remarkably, the article includes no substantive discussion of this. It is hard to argue persuasively that an intervention is successful or effective when so many participants apparently decided not to continue with it, for whatever reasons. Did they find it unhelpful? Too difficult? Harmful? We don’t actually know.

The article also overlooks the fact that the participants were, on average, almost as unhealthy after the eight-week intervention as they were beforehand. After the intervention, the average increase in the IWST for the face-to-face and remote groups was, respectively, 52 meters and 34 meters more than the results* in the care as usual group. But the average meters walked remained way, way below the levels of healthy people in the same age range. [*In this sentence, I originally used the word “increase in the care as usual group” rather than “results.” However, this version of the study does not seem to provide the data on the results in the usual care group, so it is unclear if the average increased or decreased. Perhaps the final published version will include this information. I apolotize for the error.]

The average age of participants in the trial was 59. A 2013 study called “Age-specific normal values for the incremental shuttle walk test in a healthy British population” found that the average distance walked during the ISWT by those in their 40s, 50s, 60s, and over 70 were, respectively, 824 meters, 788 meters, 699 meters, and 633 meters. By comparison, those in the face-to-face group increased from 285 to 312 meters, and those in the remote group from 353 to 388 meters.

It should have been obvious to any intelligent or even minimally competent researcher with that an exploration of both of these issues–the high drop-out rate and the continued poor health of the participants–was essential to put the purportedly “positive” findings in context. There is simply not that much positive to report about interventions that left participants severely disabled and that substantial numbers were unable or unwilling to complete.

Oh, and on top of that, there were null results for the trial’s seven quality-of-life and symptom burden questionnaires: EuroQol five-dimension five-level questionnaire (EQ5D), Patient Health Questionnaire (PHQ9), the Generalised Anxiety Disorder (GAD7) 7-item scale, Dyspnoea-12, the Functional Assessment of Chronic Illness Therapy Fatigue Scale (FACIT), the DePaul Symptom Questionnaire, and the Montreal Cognitive Assessment (MoCA). In other words, whatever incremental improvements might have occurred, participants did not report any overall benefits in the trial’s many subjective measures of well-being.

And to mention one other odd point…According to a 2019 study cited by the authors, the “minimal clinically important difference” (MCID) for the ISWT is 35 meters. In other words, while the results for the face-to-face intervention surpassed that threshold, the results for the remote intervention did not quite reach it. Yet here’s the opening of the discussion section:

“In this fully powered randomised controlled trial, we demonstrated that both face-to-face and remote exercise-based rehabilitation significantly improve exercise capacity compared to usual care alone in those previously hospitalised with COVID-19. These between group improvements exceed the established MCID (35m), highlighting improvements of clinical relevance in those with post-COVID syndrome.”

This last statement is simply not true in relation to the remote intervention, at least when it comes to the final, adjusted, intention-to-treat analysis. It is either a mistake or a deliberate effort to fudge the facts. I assume the latter, because it is very obvious that 35 is a bigger number than 34. With more than three dozen people on the manuscript, it is hard to believe that no one noticed this discrepancy. Either way, this indisputable error requires a correction. (To be clear, a correction won’t make the rest of the paper any better.)

In fact, the authors might have pointed out that there are multiple studies of the MCID of the ISWT, such as a 2008 analysis that found it to be 47.5 meters, and one from 2015 concluding that it was 70 meters. Even thought they chose to cite the MCID most favorable to their argument and ignore the others, they still presented false information to bolster their case.

So here’s the bottom line: Despite some marginal improvements among those who actually were able to or decided to complete the interventions, the trial documented that exercise-based rehabilitation failed dramatically to restore participants’ health. Moreover, participants felt no better subjectively on any measures than beforehand. Given those telling details, along with the fact that significant numbers of participants abandoned the trial’s intervention arms, the boast that these rehabilitation programs “improved short-term exercise capacity” is hard to take seriously.

**********

Disclosure: My academic position at the University of California, Berkeley, is largely supported by donations to the university via the campus crowdfunding platform from people with ME/CFS, Long Covid, and related disorders.

5 thoughts on “Trial By Error: Trudie Chalder Is Co-Author on Another Bad Exercise Paper”

  1. Eleanor Fielding

    Great stuff again – thank you. I can no longer write lengthy articles dissecting arguments but I’ve posted a cartoon about this on BlueSky @eleanorsews #FanningTheFlames

  2. Zachary Grin, PT, DPT

    David Tuller’s critique of the PHOSP-R study has several issues that misrepresent its findings.

    He suggests healthcare providers may misapply the results to all Long Covid patients, despite the study’s clear exclusion criteria. The authors never claimed their findings applied universally, and clinicians understand the need to apply research appropriately.

    Tuller also falsely claims the study doesn’t mention dropout reasons. Figure 1 lists them, including other commitments, unrelated illnesses, and rehabilitation offers. His speculation that patients may have found the intervention ineffective is unnecessary. High dropout rates are common in rehabilitation trials, and the study accounted for them through intention-to-treat and per-protocol analyses

    He dismisses the statistically significant improvements in the Incremental Shuttle Walking Test by comparing participants to healthy populations instead of considering pre-intervention levels. The study never claimed participants would return to full health. The study aimed to measure improvements in exercise capacity, and those were evident. A study like this would not be expected to lead to a full recovery as cardiopulmonary adaptations take time – full recovery would likely take several months.

    He dismisses that the study found potential clinical improvements in several secondary physical outcomes that surpassed the MCIDs. Though not statistically significant compared to usual care, this could be clinically relevant for rehabilitation.

    He focuses on the remote intervention’s 34m ISWT improvement, claiming it doesn’t meet the 35m MCID and therefore the authors cannot claim clinical relevance. However, he fails to acknowledge a 1-meter difference is within the standard error of measurement. The per-protocol analysis showed a 42m improvement, reinforcing the intervention’s clinical relevance when participants stuck with the intervention. He cherry-picks alternative MCID estimates to cast doubt, despite different MCIDs applying to different populations. The 35m threshold is commonly used for this test. Considering this, the author’s conclusion is not false or an effort to “fudge” the facts. Remote care participants who adhered to the intervention experienced clinically relevant improvements in exercise capacity.

    The PHOSP-R study provides evidence that both face-to-face and remote exercise-based rehabilitation lead to statistically significant and clinically meaningful improvements in post-hospitalization COVID-19 patients, within a well-defined patient population. Tuller’s review is not fair or rigorous which has resulted in a distorted evaluation that does a disservice to his readers, who fund his work and deserve better.

    David, please consider making corrections to your post.

  3. As usual, Zachary’s comments are easy to rebut. And given that he has defened the discredited and debunked PACE trial, I’m not surprised by his defense of this poorly reported trial.

    1) Yes, the criteria excluded patients with PEM. The conclusion, which is what many providers will rely on, does not and extrapolates the findings to all Long Covid patients. The idea that all health care providers will just know that these findings should only be extrapolated to patients without PEM when conclusion make no such exclusion is not realistic.

    2) Contrary to Zachary’s claim, I did not say that the article “doesn’t mention drop-out reasons.” I wrote that there is “no substantive discussion of the drop-out rate.” So he can’t even get his critique right. The table with pro forma reasons for stopping does not tell me if they found it unhelpful or a waste or even harmful, nor is it a “substantive discussion” of what it means when a drop-out rate of a purportedly helpful intervention is almost 40%. The authors have made no effort to explain why such a high drop-out rate does not indicate a problem with acceptability of the interventions.If they wanted to argue that this kind of drop-rate is common, they could have done that and cited stats. They didn’t.

    3) Whatever the study claimed or didn’t claim, the notion that patients remained almost as severely disabled after the intervention as before is nowhere mentioned in the study. This is very misleading. It is a critical point.

    4) Statistically insignificant secondary findings, in a study with many, many secondary findings and no evidence of statistical correction for multiple tests, are of more interest to Zachary than to me.

    5) Zachary thinks it’s fine to claim a result met the MCID when it indisputably did not (34 is lower than 35), when the MCID is within the standard error, or when a per-protocol analysis surpasses the MCID. I disagree. The authors are free to explain those details. They’re not free to claim their results met the MCID for the remote intervention when the key outomce–the adjusted analysis of the primary outcome–clearly did not, even if the difference was only a meter.

  4. Also, frankly, if it’s ‘normal’ to have dropout rates of 40% in rehabilitation studies, that makes me wonder to what extent the effectiveness of rehabilitation has been overblown in other medical conditions…

Comments are closed.