When Studies Can't Answer an Important Question (but are still published)
Study of the Week readers may be shocked to read this. I was.
Let’s do a thought experiment about the tricuspid valve. The TCV controls blood flow from the right atrium to the right ventricle.
Background: A common TCV problem occurs when the leaflets don’t close properly during systole and there is too much regurgitation of blood back to the right atrium. We call this TR or tricuspid regurgitation.
TR creates issues for patients because it can lead to terrible swelling in the abdomen or legs; it causes dilation of the RV that can progress to RV pump failure.
Surgeons can operate on TR to either repair or replace the valve. But it’s a tough operation with high risk to the patient.
Design of the Experiment: An important question in cardiology is if and when to operate on patients with TR. You are tasked with designing such an experiment.
First you would find patients with severe TR (with or without symptoms) and then randomize one group to surgery and one group to medical therapy. Since it’s impossible to blind such an experiment, you would definitely need an unbiased endpoint—say death. Randomization would be critical because the only way to judge the effect of the treatment would be if the patients were matched in known and unknown factors.
Perhaps a better way to study this would be to have a sham operative procedure wherein one group got the real repair, one group got an operation but no surgery, and one group got standard medical care. Sham operations are tough to do because of the ethical challenges. (Discussion of sham operations would be a great topic for future posts.)
The Actual Study published in an Actual Journal: A group of surgeons from the Cleveland Clinic reported their study in The Journal of Thoracic and Cardiovascular Surgery, which has an impact factor of 6.
The title gives away the results:
Here is what they did: They took 159 patients who had severe TR and were operated on from 2004-2018. These were operations solely for the tricuspid valve. Most tricuspid surgeries are done as add-on to other surgeries.
After the fact (retrospectively) they made two groups.
One group ( n = 115) had surgery based on the guideline recommendations, which require severe symptoms. The other group (N = 44), called the early group, were patients who had severe TR and some RV dilation by echo, but no symptoms.
They then looked at a primary outcome of death.
The early group had much better results. Patients in the class I group had a higher composite morbidity than early surgery (35.7% vs 18.2% ; P = .036).
Here is the graph: Notice that most of the difference occurs in the first few months. See my later comments.
The surgeons thusly concluded (emphasis mine).
Patients with class I indication for isolated TV surgery had worse survival compared with those undergoing earlier surgery before reaching class I indication. Earlier surgery may improve outcomes in these high-risk patients.
They then spent, and I kid you not, 1400 words in the discussion section promoting the finding that operating earlier on these patients is better. The message was that we wait too long to operate on these patients.
But then, after these 1400 mostly glowing words, they wrote one sentence that exposes the fatal flaw of this exercise.
Differences in baseline characteristics between the 2 groups likely largely explain the differences in their outcomes as discussed above, and propensity matching was not performed.
Comments:
I show you this study because it shocks me. I am shocked that academic researchers would publish such an exercise.
There was no randomization. So the baseline characteristics of the two groups had major differences. I counted 9 factors in which the early surgery group were healthier—for instance, they were younger, leaner, had less heart failure, and less atrial fibrillation, etc.
The comparison, therefore, was between a healthier group vs a sicker group. That is surely why the survival curves separate in the first few months.
Not randomizing groups and comparing healthier to sicker patients are bad, but it gets worse.
A normal practice when comparing two non-randomized groups is to attempt some sort of statistical matching. The idea is to take from the two groups a set of patients who at least look like each other on some baseline factors. The authors did not do this.
What gets me is that their disclaimer statement says that the differences in patients explain the results. The translation is that our experiment can’t answer the question.
But they still publish the study and conclude that we should be operating earlier on patients with tricuspid valve disease.
These are prominent surgeons from a famous center. This is an academic paper with strong conclusions. It may have influence over other surgeons. It may change what surgeons do; or when cardiologists refer patients for consideration of surgery.
But. But.
The study’s methods are utterly unable to answer the question. And the authors know this; they wrote such.
But there the study sits—in an academic journal.
What should we think about this? Why does this happen? What does it say about medical science? How does it not induce cynicism? I am asking.
As always, we at Sensible Medicine remain grateful and surprised at your support. Thank you. The Study of the Week has remained open to all subscribers. JMM
How does it not induce cynicism? It does...............I have gone from an allopathic physician who largely trusted "the science" in medicine and our respected journals and institutions to a total skeptic and a cynic. This was largely due to the mismanagement of the covid pandemic. As if that were not enough, my organization, the American College of Surgeons, has gone off the rails. The Board of Regents in 2020 declared that the ACS was a structurally racist institution in need of radical reform, that (white) surgeons were racists, and that surgery itself is discriminatory. The Board of Regents, without input from the membership, installed DEI/CRT into the College. I, and others objected. I must have been the loudest because the Board of Regents banned me for life from any engagement in the online discussion forums, from access to the members directory, and from my own private messages. This was done without due process and I was denied a hearing because my lifetime ban was not considered a disciplinary action by the Board. Is any wonder I find myself at the end of my career completely disenchanted with the state of medicine today? I weep for the younger generation of physicians and surgeons. R. Bosshardt, MD, FACS
https://www.google.com/search?q=city+journal+isolated&rlz=1C1OKWM_enUS782US782&oq=City+J&aqs=chrome.1.69i57j69i59j35i39i650j0i433i457i512j0i402i512j69i60l3.6818j0j7&sourceid=chrome&ie=UTF-8
Not only is it in a top journal, but its authors are at a hospital consistently ranked in the top 5, and often the #1 hospital, for cardiology. It was where I planned to go if I ever needed cardiology care, but not now! So beyond your Substack and Twitter/X, have you contacted the authors to hear their rationale for not doing propensity matching, and writing such a misleading abstract, which is all many (most?) doctors have time to read? Have you considered writing a Letter to the Editor of the journal with your concerns? I have always known that surgeons tend to be biased towards doing surgery, and it's good to get a second opinion from a non-surgeon. I hope all the non-surgeon cardiologists read your Substack! :) BTW, I agree that writing an article about sham surgeries is a great idea, as well as the best alternative research designs when sham surgeries are not practical/ethical.
https://www.cleveland.com/news/2023/08/cleveland-clinic-again-wins-top-spot-for-heart-care-from-us-news-see-the-changes-in-2023-24-rankings.html#:~:text=CLEVELAND, Ohio — For the 29th,2023-24 Best Hospitals rankings.