Below is the narrated video of my powerpoint presentation on Epistemic Problems in Critical Care Medicine, which provides a framework for understanding why we have both false positives and false negatives in clinical trials in critical care medicine and why we should be circumspect about our "evidence base" and our "knowledge". This is not trivial stuff, and is worth the 35 minutes required to watch the narration of the slideshow. It is a provocative presentation which gives compelling reasons to challenge our "evidence base" in critical care and medicine in general, in ways that are not widely recognized but perhaps should be, with several suggestions about assumptions that need to be challenged and revised to make our models of reality more reliable. Please contact me if you would like me to give an iteration of this presentation at your institution.
This is discussion forum for physicians, researchers, and other healthcare professionals interested in the epistemology of medical knowledge, the limitations of the evidence, how clinical trials evidence is generated, disseminated, and incorporated into clinical practice, how the evidence should optimally be incorporated into practice, and what the value of the evidence is to science, individual patients, and society.
Wednesday, December 23, 2015
Narrated and Abridged: There is (No) Evidence for That: Epistemic Problems in Critical Care Medicine
Tuesday, November 10, 2015
Peersnickety Review: Rant on My Recent Battle With Peer Reviewers
I'd like to relate a tale of exasperation with the peer review process that I recently experienced and that is probably all too familiar - but one that most folks are too timid to complain publicly about.
Nevermind that laypersons think that peer review means that your peers are reviewing your actual data for accuracy and fidelity (they are not, they are reviewing only your manuscript, final analyses, and conclusions), which causes them to be perplexed when revelations of fraudulent data published in top journals are reported. Nevermind that the website Retraction Watch, which began as a small side project now has daily and twice daily postings of retracted papers. Nevermind that some scientists have built entire careers on faked data. Nevermind that the fact that something has been peer reviewed provides very little in the way of assurance that the report contains anything other than rubbish. Nevermind that leading investigators publish the same reviews over and over in different journals with the same figures and sometimes the same text.
The entire process is cumbersome, time consuming, frustrating, and of dubious value as currently practiced.
Last year I was invited by the editors of Chest to write a "contemporary review of ionized calcium in the ICU - should it be measured? should it be treated?" I am not aware of why I was selected for this, but I infer that someone suggested me as the author because of my prior research in medical decision making and because of the monograph we wrote several years back called Laboratory Testing in the ICU which applied principles of rational decision making such as Bayesian methods and back-of-the-envelope cost benefit analyses to make a framework of rational laboratory testing in the ICU. I accepted the invitation, even knowing it would entail a good deal of work for me that would be entirely uncompensated, save for buttressing my fragile ego, he said allegorically.
Now, consider for an instant the extra barriers that I, as a non-academic physician faced in agreeing to do this. As a non-academic physician, I do not have access to a medical library, and of course the Chest editors do not have a way to grant me access. That is, non-academic physicians doing scholarly work such as this are effectively disenfranchised from the infrastructure that they need to do scholarly work. Fortunately for me, my wife was a student at the University of Utah during this time so I was able to access the University library with her help. Whether academic centers and peer-reviewed journals ought to have a monopoly on this information is a matter for debate elsewhere, and not a trivial one.
Nevermind that laypersons think that peer review means that your peers are reviewing your actual data for accuracy and fidelity (they are not, they are reviewing only your manuscript, final analyses, and conclusions), which causes them to be perplexed when revelations of fraudulent data published in top journals are reported. Nevermind that the website Retraction Watch, which began as a small side project now has daily and twice daily postings of retracted papers. Nevermind that some scientists have built entire careers on faked data. Nevermind that the fact that something has been peer reviewed provides very little in the way of assurance that the report contains anything other than rubbish. Nevermind that leading investigators publish the same reviews over and over in different journals with the same figures and sometimes the same text.
The entire process is cumbersome, time consuming, frustrating, and of dubious value as currently practiced.
Last year I was invited by the editors of Chest to write a "contemporary review of ionized calcium in the ICU - should it be measured? should it be treated?" I am not aware of why I was selected for this, but I infer that someone suggested me as the author because of my prior research in medical decision making and because of the monograph we wrote several years back called Laboratory Testing in the ICU which applied principles of rational decision making such as Bayesian methods and back-of-the-envelope cost benefit analyses to make a framework of rational laboratory testing in the ICU. I accepted the invitation, even knowing it would entail a good deal of work for me that would be entirely uncompensated, save for buttressing my fragile ego, he said allegorically.
Now, consider for an instant the extra barriers that I, as a non-academic physician faced in agreeing to do this. As a non-academic physician, I do not have access to a medical library, and of course the Chest editors do not have a way to grant me access. That is, non-academic physicians doing scholarly work such as this are effectively disenfranchised from the infrastructure that they need to do scholarly work. Fortunately for me, my wife was a student at the University of Utah during this time so I was able to access the University library with her help. Whether academic centers and peer-reviewed journals ought to have a monopoly on this information is a matter for debate elsewhere, and not a trivial one.
Sunday, October 11, 2015
When Hell Freezes Over: Trials of Temperature Manipulation in Critical Illness
![]() |
The bed is on fire |
The Eurotherm3235 trial was stopped early because of concerns of harm or futility. This trial enrolled patients with traumatic brain injury (TBI) and elevated intracranial pressure (ICP) and randomized them to induced hypothermia (which reduces ICP) versus standard care. There was a suggestion of worse outcomes in the hypothermia group. I know that the idea that we can help the brain with the simple maneuver of lowering body temperature has great appeal and what some would call "biological plausibility" a term that I henceforth forsake and strike from my vocabulary. You can rationalize the effect of an intervention any way you want using theoretical biological reasoning. So from now on I'm not going to speak of biological plausibility, I will call it biological rationalizing. A more robust principle, as I have claimed before, is biological precedent - that is, this or that pathway has been successfully manipulated in a similar way in the past. It is reasonable to believe that interfering with LDL metabolism will improve cardiovascular outcomes because of decades of trials of statins (though agents used to manipulate this pathway are not all created equal). It is reasonable to believe that intervening with platelet aggregation will improve outcomes from cardiovascular disease because of decades of trials of aspirin and plavix and others. It is reasonable to doubt that manipulation of body temperature will improve any outcome because there is no unequivocal precedent for this, save for warming people with hypothermia from exposure - which basically amounts to treating the known cause of their ailment. This is one causal pathway that we understand beyond a reasonable doubt. If you get exposure, you freeze to death. If we find you still alive and warm you, you may well survive.
Wednesday, October 7, 2015
Early Mobility in the ICU: The Trial That Should Not Be
I learned via twitter yesterday that momentum is building to conduct a trial of early mobility in critically ill patients. While I greatly respect many of the investigators headed down this path, forthwith I will tell you why this trial should not be done, based on principles of rational decision making.
A trial is a diagnostic test of a hypothesis, a complicated and costly test of a hypothesis, and one that entails risk. Diagnostic tests should not be used indiscriminately. That the RCT is a "Gold Standard" in the hierarchy of testing hypotheses does not mean that we should hold it sacrosanct, nor does it follow that we need a gold standard in all cases. Just like in clinical medicine, we should be judicious in our ordering of diagnostic tests.
The first reason that we should not do a trial of early mobility (or any mobility) in the ICU is because in the opinion of this author, experts in critical care, and many others, early mobility works. We have a strong prior probability that this is a beneficial thing to be doing (which is why prominent centers have been doing it for years, sans RCT evidence). When the prior probability is high enough, additional testing has decreasing yield and risks false negative results if people are not attuned to the prior. Here's my analogy - a 35 year old woman with polycystic kidney disease who is taking birth control presents to the ED after collapsing with syncope. She had shortness of breath and chest pain for 12 hours prior to syncope. Her chest x-ray is clear and bedside ultrasound shows a dilated right ventricle. The prior probability of pulmonary embolism is high enough that we don't really need further testing, we give anticoagulants right away. Even if a V/Q scan (creatnine precludes CT) is "low probability" for pulmonary embolism, we still think she has it because the prior probability is so high. Indeed, the prior probability is so high that we're willing to make decisions without further testing, hence we gave heparin. This process follows the very rational Threshold Approach to Decision Making approach proposed by Pauker and Kasirrer in the NEJM in 1980, which is basically a reformulation of VonNeumann and Morganstern's Expected Utility Theory to adapt it to medical decisions. Distilled it states in essence, "when you get to a threshold probability of disease where the benefits of treatment exceed the risks, you treat." And so let it be with early mobility. We already think the benefits exceed the risks, which is why we're doing it. We don't need a RCT. As I used to ask the housestaff over and over until I was cyanotic: "How will the results of that test influence what you're going to do?"
Notice that this logical approach to clinical decision making shines a blinding light upon "evidence based medicine" and the entire enterprise of testing hypotheses with frequentist methods that are deaf to prior probabilities. Can you imagine using V/Q scanning to test for PE without prior probabilities? Can you imagine what a mess you would find yourself in with regard to false negatives and false positives? You would be the neophyte medical student who thinks "test positive, disease present; test negative, disease absent." So why do we continue ad nauseum in critical care medicine to dismiss prior probabilities and decision thresholds and blindly test hypotheses in a purist vacuum?
The next reasons this trial should not be conducted flow from the first. The trial will not have a high enough likelihood ratio to sway the high prior below the decision threshold; if the trial is "positive" we will have spent millions of dollars to "prove" something we already knew at a threshold above our treatment threshold; if the trial is positive, some will squawk "It wasn't blinded" yada yada yada in an attempt to dismiss the results as false positives; if the trial is negative, some will, like the tyro medical student, declare that "there is no evidence for early mobility" and similar hoopla and poppycock; or the worst case: the trial shows harm from early mobility, which will get the naysayers of early mobility very agitated. But of course, our prior probability that early mobility is harmful is hopelessly low, making such a result highly likely to be spurious. When we clamor about "evidence" we are in essence clamoring about "testing hypotheses with RCTs" and eschewing our responsibility to use clinical judgment, recognize the limits of testing, and practice in the face of uncertainty using our "untested" prior probabilities.
Consider a trial of exercise on cardiovascular outcomes in community dwelling adults - what good can possibly come of such a trial? Don't we already know that exercise is good for you? If so, a positive trial reinforces what we already know (but does little to convince sedentary folks to exercise, as they too already know they should exercise), but a negative trial risks sending the message to people that exercise is of no use to you, or that the number needed to treat is too small for you to worry about.
Or consider the recent trials of EGDT which "refuted" the Rivers trial from 14 years ago. Now, everybody is saying, "Well, we know it works, maybe not the catheters and the ScVO2 and all those minutaie , but in general, rapid early resuscitation works. And the trials show that we've already incorporated what works into general practice!"
I don't know the solutions to these difficult quandries that we repeatedly find ourselves in trial after trial in critical care medicine. I'm confused too. That's why I'm thinking very hard and very critically about the limits of our methods and our models and our routines. But if we can anticipate not only the results of the trials, but also the community reaction to them, then we have guidance about how to proceed in the future. Because what value does a mega-trial have, if not to guide care after its completion? And even if that is not its goal, (maybe its goal is just to inform the science), can we turn a blind eye to the fact that it will guide practice after its completion, even if that guidance is premature?
It is my worry that, given the high prior probability that a trial in critical care medicine will be "negative", the most likely result is a negative trial which will embolden those who wish to dismiss the probable benefits of early mobility and give them an excuse to not do it.
Diagnostic tests have risks. A false negative test is one such risk.
A trial is a diagnostic test of a hypothesis, a complicated and costly test of a hypothesis, and one that entails risk. Diagnostic tests should not be used indiscriminately. That the RCT is a "Gold Standard" in the hierarchy of testing hypotheses does not mean that we should hold it sacrosanct, nor does it follow that we need a gold standard in all cases. Just like in clinical medicine, we should be judicious in our ordering of diagnostic tests.
The first reason that we should not do a trial of early mobility (or any mobility) in the ICU is because in the opinion of this author, experts in critical care, and many others, early mobility works. We have a strong prior probability that this is a beneficial thing to be doing (which is why prominent centers have been doing it for years, sans RCT evidence). When the prior probability is high enough, additional testing has decreasing yield and risks false negative results if people are not attuned to the prior. Here's my analogy - a 35 year old woman with polycystic kidney disease who is taking birth control presents to the ED after collapsing with syncope. She had shortness of breath and chest pain for 12 hours prior to syncope. Her chest x-ray is clear and bedside ultrasound shows a dilated right ventricle. The prior probability of pulmonary embolism is high enough that we don't really need further testing, we give anticoagulants right away. Even if a V/Q scan (creatnine precludes CT) is "low probability" for pulmonary embolism, we still think she has it because the prior probability is so high. Indeed, the prior probability is so high that we're willing to make decisions without further testing, hence we gave heparin. This process follows the very rational Threshold Approach to Decision Making approach proposed by Pauker and Kasirrer in the NEJM in 1980, which is basically a reformulation of VonNeumann and Morganstern's Expected Utility Theory to adapt it to medical decisions. Distilled it states in essence, "when you get to a threshold probability of disease where the benefits of treatment exceed the risks, you treat." And so let it be with early mobility. We already think the benefits exceed the risks, which is why we're doing it. We don't need a RCT. As I used to ask the housestaff over and over until I was cyanotic: "How will the results of that test influence what you're going to do?"
Notice that this logical approach to clinical decision making shines a blinding light upon "evidence based medicine" and the entire enterprise of testing hypotheses with frequentist methods that are deaf to prior probabilities. Can you imagine using V/Q scanning to test for PE without prior probabilities? Can you imagine what a mess you would find yourself in with regard to false negatives and false positives? You would be the neophyte medical student who thinks "test positive, disease present; test negative, disease absent." So why do we continue ad nauseum in critical care medicine to dismiss prior probabilities and decision thresholds and blindly test hypotheses in a purist vacuum?
The next reasons this trial should not be conducted flow from the first. The trial will not have a high enough likelihood ratio to sway the high prior below the decision threshold; if the trial is "positive" we will have spent millions of dollars to "prove" something we already knew at a threshold above our treatment threshold; if the trial is positive, some will squawk "It wasn't blinded" yada yada yada in an attempt to dismiss the results as false positives; if the trial is negative, some will, like the tyro medical student, declare that "there is no evidence for early mobility" and similar hoopla and poppycock; or the worst case: the trial shows harm from early mobility, which will get the naysayers of early mobility very agitated. But of course, our prior probability that early mobility is harmful is hopelessly low, making such a result highly likely to be spurious. When we clamor about "evidence" we are in essence clamoring about "testing hypotheses with RCTs" and eschewing our responsibility to use clinical judgment, recognize the limits of testing, and practice in the face of uncertainty using our "untested" prior probabilities.
Consider a trial of exercise on cardiovascular outcomes in community dwelling adults - what good can possibly come of such a trial? Don't we already know that exercise is good for you? If so, a positive trial reinforces what we already know (but does little to convince sedentary folks to exercise, as they too already know they should exercise), but a negative trial risks sending the message to people that exercise is of no use to you, or that the number needed to treat is too small for you to worry about.
Or consider the recent trials of EGDT which "refuted" the Rivers trial from 14 years ago. Now, everybody is saying, "Well, we know it works, maybe not the catheters and the ScVO2 and all those minutaie , but in general, rapid early resuscitation works. And the trials show that we've already incorporated what works into general practice!"
I don't know the solutions to these difficult quandries that we repeatedly find ourselves in trial after trial in critical care medicine. I'm confused too. That's why I'm thinking very hard and very critically about the limits of our methods and our models and our routines. But if we can anticipate not only the results of the trials, but also the community reaction to them, then we have guidance about how to proceed in the future. Because what value does a mega-trial have, if not to guide care after its completion? And even if that is not its goal, (maybe its goal is just to inform the science), can we turn a blind eye to the fact that it will guide practice after its completion, even if that guidance is premature?
It is my worry that, given the high prior probability that a trial in critical care medicine will be "negative", the most likely result is a negative trial which will embolden those who wish to dismiss the probable benefits of early mobility and give them an excuse to not do it.
Diagnostic tests have risks. A false negative test is one such risk.
Wednesday, July 22, 2015
There is (No) Evidence For That: Epistemic Problems in Evidence Based Medicine
Below is a Power Point Presentation that I have delivered several times recently including one iteration at the SMACC conference in Chicago. It addresses epistemic problems in our therapeutic knowledge, and calls into question all claims of "there is evidence for ABC" and "there is no evidence for ABC." Such claims cannot be taken at face value and need deeper consideration and evaluation considering all possible states of reality - gone is the cookbook or algorithmic approach to evidence appraisal as promulgated by the User's Guides. Considered in the presentation are therapies for which we have no evidence, but they undoubtedly work (Category 1 - Parachutes) and therapies for which we have evidence of efficacy or lack thereof (Category 2) but that evidence is subject to false positives and false negatives, for numerous reasons including: the Ludic Fallacy, study bias (See: Why Most Published Research Findings Are False), type 1 and 2 errors, the "alpha bet" (the arbitrary and lax standard used for alpha, namely 0.05), Bayesian interpretations, stochastic dominance of the null hypothesis, inadequate study power in general and that due to delta inflation and subversion of double significance hypothesis testing. These are all topics that have been previously addressed to some degree on this blog, but this presentation presents them together as a framework for understanding the epistemic problems that arise within our "evidence base." It also provides insights into why we have a generation of trials in critical care the results of which converge on the null and why positive studies in this field cannot be replicated.
Tuesday, June 2, 2015
Evolution Based Medicine: A Philosophical Framework for Understanding Why Things Don't Work
An afternoon session at the ATS meeting this year about "de-adoption" of therapies which have been shown to be ineffective was very thought provoking and the contrasts between it and the morning session on ARDS are nothing less than ironic. As I described in the prior post about the baby in the bathwater, physicians seem to have a hard time de-adopting therapies. Ask your colleagues at the next division conference if you should abandon hypothermia after cardiac arrest and rather just treat fever based on the TTM trial and the recent pediatric trial, and see what the response is. Or, suggest that hyperglycemia (at any level in non-diabetic patients) in the ICU be observed rather than treated. Or float the idea to your surgical colleagues that antibiotics be curtailed after four days in complicated intraabdominal infection, and see how quickly you are ushered out of the SICU. Tell your dietition that you're going to begin intentionally underfeeding patients, or not feeding them at all and see what s/he say(s). Propose that you discard sepsis resuscitation bundles, etc. We have a hard time de-adopting. We want to take what we have learned about physiology and pharmacology and apply it, to usurp control of and modify biological processes that we think we understand. We (especially in critical care) are interventionists at heart.
The irony occurred at ATS because in the morning session, we were told that there is incontrovertible (uncontroverted may have been a better word) evidence for the efficacy of prone positioning in ARDS (interestingly, one of the only putative therapies for ARDS that the ARDSnet investigators never trialed), and it was strongly suggested that we begin using esophageal manometry to titrate PEEP in ARDS. So, in the morning, we are admonished to adopt, and in the afternoon we are chided to de-adopt a host of therapies. Is this the inevitable cycle in critical care and medical therapeutics? A headlong rush to adopt, then an uphill battle to de-adopt?
The irony occurred at ATS because in the morning session, we were told that there is incontrovertible (uncontroverted may have been a better word) evidence for the efficacy of prone positioning in ARDS (interestingly, one of the only putative therapies for ARDS that the ARDSnet investigators never trialed), and it was strongly suggested that we begin using esophageal manometry to titrate PEEP in ARDS. So, in the morning, we are admonished to adopt, and in the afternoon we are chided to de-adopt a host of therapies. Is this the inevitable cycle in critical care and medical therapeutics? A headlong rush to adopt, then an uphill battle to de-adopt?
Friday, May 1, 2015
Is There a Baby in That Bathwater? Status Quo Bias in Evidence Appraisal in Critical Care

Status quo bias is a cognitive decision making bias that leads to decision makers' preference for the choice represented by the current status quo, even when the status quo is arbitrary or irrelevant. Decision makers tend to perceive a change from the status quo as a loss and therefore their decisions are biased toward the status quo. This can lead to preference reversals when the status quo reference frame is changed. The status quo can be debiased using a reversal test, i.e., manipulating the status quo either experimentally or via thought experiment to consider a change in the opposite direction. If reluctance to change from the status quo exists in both directions, status quo bias is likely to exist.
My collaborators Peter Terry, Hal Arkes and I reported in a study published in 2006 that physicians were far more likely to abandon a therapy that was status quo or standard therapy based on new evidence of harm than they were to adopt an identical therapy based on the same evidence of benefit from a fictitious RCT (randomized controlled trial) presented in the vignette. These results suggested that there was an asymmetric status quo bias - physicians showed a strong preference for the status quo in the adoption of new therapies, but a strong preference for abandoning the status quo when a standard of care was shown to be harmful. Two characteristics of the vignettes used in this intersubject study deserve attention. First, the vignettes described a standard or status quo therapy that had no support from RCTs prior to the fictitious one described in the vignette. Second, this study was driven in part by what I perceived at the time was a curious lack of adoption of drotrecogin-alfa (Xigris), with its then purported mortality benefit and associated bleeding risk. Thus, our vignettes had very significant trade-offs in terms of side effects in both the adopt and abandon reference frames. Our results seemed to explain s/low uptake of Xigris, and were also consistent with the relatively rapid abandonment of hormone replacement therapy (HRT) after publication of the WHI, the first RCT of HRT.
Thursday, January 29, 2015
The Therapeutic Paradox: What's Right for the Population May Not Be Right for the Patient
![]() |
Bad for the population, good for me |
The authors, bloggers at The Incidental Economist, begin the article with a sobering look at the number needed to treat (NNT). For the primary prevention of myocardial infarction (MI), if 2000 people with a 10% or higher risk of MI in the next 10 years take aspirin for 2 years, one MI will be prevented. 1999 people will have gotten no benefit from aspirin, and four will have an MI in spite of taking aspirin. Aspirin, a very good drug on all accounts, is far from a panacea, and this from a man (me) who takes it in spite of falling far below the risk threshold at which it is recommended.
One problem with NNT is that for patients it is a gratuitous numerical transformation of a simple number that anybody could understand (the absolute risk reduction - "your risk of stroke is reduced 3% by taking coumadin"), into a more abstract one (the NNT - "if we treat 33 people with coumadin, we prevent one stroke among them") that requires retransformation into examples that people can understand, as shown in pictograms in the NYT article. A person trying to understand stroke prevention with coumadin could care less about the other 32 people his doctor is treating with coumadin, he is interested in himself. And his risk is reduced 3%. So why do we even use the NNT, why not just use ARR?
Saturday, January 17, 2015
Clinical Trialists Should Use Economies of Scale to Maximize Profits of Large RCTs
![]() |
The lever is a powerful tool |
Thinking about calcium levels and causation and repletion, one cannot help but think about all sorts of other levels we check in the ICU - potassium, magnesium, phosphate - and may other things we routinely do but about which we have no real inkling of an idea as to whether we're doing any patients any good. (Arterial lines are another example.) Are we just wasting our time with many of the things we do? This question becomes more urgent as evidence mounts that much of what we do (in the ICU and elsewhere) is useless, wasteful, or downright harmful. But who or what agency is going to fund a trial of potassium or calcium replacement in the ICU? It certainly seems unglamorous. Don't we have other disease-specific priorities that are paramount in importance to such a trial?
I then realized that a good businessman, wanting to maximize the "profit" from a large, randomized controlled trial (and the dollars "invested" in it), would take advantage of economies of scale. For those who are not business savvy (I do not imply that I am), business costs can be roughly divided into fixed costs and variable costs. If you have a factory making widgets you have certain costs such as the rent, advertising, widget making machines. These costs are "fixed" meaning that they are invariable whether you make 100 widgets or 10,000 widgets. Variable costs are the costs of materials, electricity, and human resources which must be scaled up as you make more widgets. In general, the cost of making each widget goes down as the fixed costs are spread out over more widget units. Additionally, if you can leverage your infrastructure to make wadgets, a product similar to a widget, you likewise increase profits by lowering costs per unit.
Saturday, October 11, 2014
Enrolling Bad Patients After Good: Sunk Cost Bias and the Meta-Analytic Futility Stopping Rule
Four (relatively) large critical care randomized controlled trials were published early in the NEJM in the last week. I was excited to blog on them, but then I realized they're all four old news, so there's nothing to blog about. But alas, the fact that there is no news is the news.
In the last week, we "learned" that more transfusion is not helpful in septic shock, that EGDT (the ARISE trial) is not beneficial in sepsis, that simvastatin (HARP-2 trial) is not beneficial in ARDS, and that parental administration of nutrition is not superior to enteral administration in critical illness. Any of that sound familiar?
I read the first two articles, then discovered the last two and I said to myself "I'm not reading these." At first I felt bad about this decision, but then that I realized it is a rational one. Here's why.
In the last week, we "learned" that more transfusion is not helpful in septic shock, that EGDT (the ARISE trial) is not beneficial in sepsis, that simvastatin (HARP-2 trial) is not beneficial in ARDS, and that parental administration of nutrition is not superior to enteral administration in critical illness. Any of that sound familiar?
I read the first two articles, then discovered the last two and I said to myself "I'm not reading these." At first I felt bad about this decision, but then that I realized it is a rational one. Here's why.
Labels:
ARDS,
ARISE,
Bayes theorem,
EGDT,
HARP-2,
MAFSR,
meta-analysis,
meta-analytic futility stopping rule,
parenteral nutrition,
prior probability,
PROCESS,
SAILS,
Sepsis,
statins,
transfusion
Saturday, July 12, 2014
Better the Devil You Know: Thrombolysis for Pulmonary Embolism
In my view, the task of the expert is to render the complex simple. And the expert does do this, except when reality goes against his bets and complexity becomes a tool for obfuscating an unwanted result.
In 2002, Konstantanidis compared alteplase plus heparin versus heparin alone for submassive pulmonary embolism (PE). The simple message from this study was "alteplase now saves you from alteplase later" and the simple strategy is to wait until there is hemodynamic deterioration (shock) and then give alteplase. Would that it were actually viewed so simply - I would not then get calls from stressed providers hemming and hawing about the septum bowing on the echo and the sinus tachycardia and the....
If you're a true believer, you think alteplase works - you want it to work. So, you do another study, hoping that biomarkers better identify a subset of patients that will benefit from an up front strategy of thrombolysis. Thus, the PEITHO study appeared in the April 10th, 2014 issue of the NEJM. It too showed that fibrinolysis (with tenecteplase) now simply saved you from tenecteplase later. But fibrinolysis now also causes stroke later with an increase from 0.2% in the control group versus 2.4% in the fibrinolysis group - and most of them were hemorrhagic. Again, the strategic path is in stark relief - if your patient is dying of shock from PE, give fibrinolysis. If not, wait - because less than 5% of them are going to deteriorate.
So we have vivid clarity provided by large modern randomized controlled trials guiding us on what to do with that subset of patients with PE that is not in shock. For those that are in shock, most agree that we should give thrombolysis.
To muddy that clarity, Chatterjee et al report the results of a meta-analysis in the June 18th issue of JAMA in which they combine all trials they could find over the past 45 years (back to 1970!) of all patients with PE, regardless of hemodynamic status. The result: fewer patients died but more had bleeding. We have now made one full revolution, from trying to identify subsets likely to benefit, to combining them all back together - I think I'm getting dizzy.
If the editorialist would look at his numbers as his patients likely would (and dispense with relative risk reductions), he would see that:
For almost every life that is saved, there is almost one (0.74) case of bleeding in the brain and there are 3.4 more cases of major bleeding. And bear in mind that these are the aggregate meta-analysis numbers that include patients in shock and those not in shock - the picture is worse if you exclude those in shock.
Better the devil you know.
In 2002, Konstantanidis compared alteplase plus heparin versus heparin alone for submassive pulmonary embolism (PE). The simple message from this study was "alteplase now saves you from alteplase later" and the simple strategy is to wait until there is hemodynamic deterioration (shock) and then give alteplase. Would that it were actually viewed so simply - I would not then get calls from stressed providers hemming and hawing about the septum bowing on the echo and the sinus tachycardia and the....
If you're a true believer, you think alteplase works - you want it to work. So, you do another study, hoping that biomarkers better identify a subset of patients that will benefit from an up front strategy of thrombolysis. Thus, the PEITHO study appeared in the April 10th, 2014 issue of the NEJM. It too showed that fibrinolysis (with tenecteplase) now simply saved you from tenecteplase later. But fibrinolysis now also causes stroke later with an increase from 0.2% in the control group versus 2.4% in the fibrinolysis group - and most of them were hemorrhagic. Again, the strategic path is in stark relief - if your patient is dying of shock from PE, give fibrinolysis. If not, wait - because less than 5% of them are going to deteriorate.
So we have vivid clarity provided by large modern randomized controlled trials guiding us on what to do with that subset of patients with PE that is not in shock. For those that are in shock, most agree that we should give thrombolysis.
To muddy that clarity, Chatterjee et al report the results of a meta-analysis in the June 18th issue of JAMA in which they combine all trials they could find over the past 45 years (back to 1970!) of all patients with PE, regardless of hemodynamic status. The result: fewer patients died but more had bleeding. We have now made one full revolution, from trying to identify subsets likely to benefit, to combining them all back together - I think I'm getting dizzy.
If the editorialist would look at his numbers as his patients likely would (and dispense with relative risk reductions), he would see that:
Death | Bleeding in the brain | Other Major Bleeding | |
Blood Thinner | 3.89% | 0.19% | 3.42 |
Clot Buster | 2.17% | 1.46% | 9.24 |
Difference | 1.72% | -1.27% | -5.82 |
For almost every life that is saved, there is almost one (0.74) case of bleeding in the brain and there are 3.4 more cases of major bleeding. And bear in mind that these are the aggregate meta-analysis numbers that include patients in shock and those not in shock - the picture is worse if you exclude those in shock.
Better the devil you know.
Monday, May 19, 2014
Sell Side Bias and Scientific Stockholm Syndrome: A Report from the Annual Meeting of the American Thoracic Society
What secrets lie inside? |
Well, I'm at the American Thoracic Society (ATS) meeting in San Diego right now, and it certainly does feel like people - everyone - is trying to sell me something. From the giant industry sponsored banners, to the emblazoned tote bags, to the bags of propaganda left at my hotel room door every morning, to the exhibitor hall filled with every manner of new and fancy gadgets (but closed to cameras), to the investigators themselves, everybody is trying to convince me to buy (or prescribe) something. Especially ideas. Investigators have a promotional interest in their ideas. And they want you and me to buy into their ideas. I have become convinced that investigators without industry ties (that dying breed) are just about as susceptible to sell side bias as those with industry ties. Indeed, I have also noted that the potential consumer of many of the ideas himself seems biased - he wants things to work, too, and he has a ready explanation for why some ideas didn't pan out in the data (see below). It's like an epidemic of scientific Stockholm Syndrome.
The first session I attended was a synopsis of the SAILS trial by the ARDSnet investigators, testing whether use of a statin, rosuvastatin, in patients with sepsis-incited lung injury would influence 60 day mortality. The basis of this trial was formed by observational associations that patients on statins had better outcomes in this, that, and the other thing, including sepsis. If you are not already aware of the results, guess whether rosuvastatin was beneficial in this study.
Saturday, April 26, 2014
Dear SIRS: Your Septic System Stinks
I perused with interest the April 2nd JAMA article on the temporal improvement in severe sepsis outcomes in Australia and New Zealand (ANZ) by Kaukonen et al this week. Epidemiological studies like this remind me why I'm so fond of reading reports of RCTs: because they're so much easier to think about. Epidemiological studies have so many variables, measured and unmeasured, and so much confounding you have to consider. I spent at least five hours poring over the ANZ report, and then comparing it to the recent NEJM article about improved diabetes complications between 1990 and 2010, which is similar, but a bit more convincing (perhaps the reason it's in the NEJM).
I was delighted that the authors of the ANZ study twice referenced our delta inflation article and that the editorialists agree with the letter I wrote to AJRCCM last year advocating composite morbidity outcomes in trials of critical illness. These issues dovetail - we have a consistent track record of failure to demonstrate mortality improvements in critical care, while we turn a blind eye to other outcomes which may be more tractable and which are often of paramount concern to patients.
I was delighted that the authors of the ANZ study twice referenced our delta inflation article and that the editorialists agree with the letter I wrote to AJRCCM last year advocating composite morbidity outcomes in trials of critical illness. These issues dovetail - we have a consistent track record of failure to demonstrate mortality improvements in critical care, while we turn a blind eye to other outcomes which may be more tractable and which are often of paramount concern to patients.
Monday, April 21, 2014
Stowaway and Accidental Empiricist Humbles Physiological Theorists: The Boy in the Wheel Well
![]() |
Kessler Peak in the Wasatch: 10,400 feet |
Sunday, April 6, 2014
Underperforming the Market: Why Researchers are Worse than Professional Stock Pickers and A Way Out
I was reading in the NYT yesterday a story about Warren Buffet and how the Oracle of Omaha has trailed the S&P 500 for four of the last five years. It was based on an analysis done by a statistician who runs a blog called Statistical Ideas, which has a post on p-values that links to this Nature article a couple of months back that describes how we can be misled by P-values. And all of this got me thinking.
We have a dual problem in medical research: a.) of conceiving alternative hypotheses which cannot be confirmed in large trials free of bias; and b.) not being able to replicate the findings of positive trials. What are the reasons for this?
We have a dual problem in medical research: a.) of conceiving alternative hypotheses which cannot be confirmed in large trials free of bias; and b.) not being able to replicate the findings of positive trials. What are the reasons for this?
Tuesday, April 1, 2014
Absolute Confusion: How Researchers Mislead the Public with Relative Risk
This article in Sunday's New York Times about gauging the risk of autism highlights an important confusion in the appraisal of evidence from clinical trials and epidemiological studies that appears to be shared by laypersons, researchers, and practitioners alike: we focus on relative risks when we should be concerned with absolute risks.
The rational decision maker, when evaluating a risk or a benefit, is concerned with the absolute magnitude of that risk or benefit. A proportional change from an arbitrary baseline (a relative risk) is irrelevant. Here's an example that should bring this into keen focus.
If you are shopping and you find a 50% off sale, that's a great sale. Unless you're shopping for socks. At $0.99 a pair, you save $0.50 with that massive discount. Alternatively, if you come across a 3% sale, but it's at the Audi dealership, that paltry discount can save you $900 on a $30,000 Audi A4. Which discount should you spend the day pursuing? The discount rate mathematically obscures the value of the savings. If we framed the problem in terms of absolute savings, we would be better consumers. But retailers know that saying "50% OFF!" attracts more attention than "$0.50 OFF!" in the sock department. Likewise, car salesmen know that writing "$1000 BELOW INVOICE!" on the windshield looks a lot more attractive than "3% BELOW INVOICE!"
The rational decision maker, when evaluating a risk or a benefit, is concerned with the absolute magnitude of that risk or benefit. A proportional change from an arbitrary baseline (a relative risk) is irrelevant. Here's an example that should bring this into keen focus.
If you are shopping and you find a 50% off sale, that's a great sale. Unless you're shopping for socks. At $0.99 a pair, you save $0.50 with that massive discount. Alternatively, if you come across a 3% sale, but it's at the Audi dealership, that paltry discount can save you $900 on a $30,000 Audi A4. Which discount should you spend the day pursuing? The discount rate mathematically obscures the value of the savings. If we framed the problem in terms of absolute savings, we would be better consumers. But retailers know that saying "50% OFF!" attracts more attention than "$0.50 OFF!" in the sock department. Likewise, car salesmen know that writing "$1000 BELOW INVOICE!" on the windshield looks a lot more attractive than "3% BELOW INVOICE!"
Sunday, March 23, 2014
Lost Without a MAP: Blood Pressure Targets in Septic Shock
Saturday, March 22, 2014
Thursday, March 20, 2014
Sepsis Bungles: The Lessons of Early Goal Directed Therapy
On March 18th, the NEJM published early online three original trials of therapies for the critically ill that will serve as fodder for several posts. Here, I focus on the ProCESS trial of protocol guided therapy for early septic shock. This trial is in essence a multicenter version of the landmark 2001 trial of Early Goal Directed Therapy (EGDT) for severe sepsis by Rivers et al. That trial showed a stunning 16% absolute reduction in mortality in sepsis attributed to the use of a protocol based on physiological goals for hemodynamic management. That absolute reduction in mortality is perhaps the largest for any therapy in critical care medicine. If such a reduction were confirmed, it would make EGDT the single most important therapy in the field. If such reduction cannot be confirmed, there are several reasons why the Rivers results may have been misleading:
- As I have blogged in the case of intensive insulin therapy, Single center studies inflate treatment effects when compared to multicenter studies for reasons that are unclear, but which may be related to bias especially in unblinded studies. (The revelation that Rivers was an investor in one of the devices used in the trial raised additional concerns about bias.)
- Regression to the mean may lead to reduced effect sizes when trials are repeated, especially when the index trial has a very large effect size. In a similar vein, since large absolute mortality reductions are statistically unlikely in critical care medicine, Bayesian inference means that trials reporting large reductions are likely to represent type I statistical errors.
There were other concerns about the Rivers study and how it was later incorporated into practice, but I won't belabor them here. The ProCESS trial randomized about 1350 patients among three groups, one simulating the original Rivers protocol, one to a modified Rivers protocol, and one representing "standard care" that is, care directed by the treating physician without a protocol. The study had 80% power to demonstrate a mortality reduction of 6-7%. Before you read further, please wager, will the trial show any statistically significant differences in outcome that favor EGDT or protocolized care?
Friday, February 28, 2014
Overediagnosis and Mitigated Overdiagnosis: Ongoing problems with Breast and Lung Cancer Screening
I got to thinking about cancer screening (again) in the last week after reading in BMJ about the Canadian National Breast Screening Study (CNBSS). That article piqued my interest because I immediately recalled the brouhaha that ensued after the U.S. Preventative Services Task Force (USPSTF) recommended that women not get mammograms until age 50 rather than age 40. That uproar was similar to the outcry by urologists when the USPSTF recommended against screening for prostate cancer with PSA testing. Meanwhile, changes in the cholesterol guidelines have incited intellectual swashbuckling among experts in that field. Without getting into the details, observers of these events might generate the following hypotheses:
- People are comfortable with the status quo and uncomfortable with change
- People get emotionally connected to good causes and this makes the truth blurry, or invisible. After you've participated in the Race for the Cure, it's hard to swallow the possibility that the linchpin of the Race might not be as useful as we thought; and is no longer recommended for a whole swath of women.
- People are terrified of cancer
- Screening costs money. Somebody pockets that money. Urologists and radiologists and gastroenterologists LOVE screening programs. So do Porche dealers.
Monday, February 10, 2014
Brief Updates on Hypothermia, Hyperglycemia, Cholesterol, Blood Pressure Lowering in Stroke and Testosterone
I've read a lot of interesting articles recently, but none that are sufficient fodder for a dedicated post. So here I will update some themes from previous blog posts with recent articles from NEJM and JAMA that relate to them.
Prehospital Induction of Hypothermia After Cardiac Arrest
In this article in the January 1st issue of JAMA, investigators from King County Washington report the results of a trial which tested the hypothesis that earlier (prehospital) induction of hypothermia, by infusing cold saline, would augment the assumed benefit of hypothermia that is usually initiated in the hospital for patients with ventricular fibrillation. Please guess what was the effect of this intervention on survival to hospital discharge and neurological outcomes.
You were right. There was not even a signal, not a trend towards benefit, even though body temperature was lower by 1 degree Celcius and time to target hypothermia temperature in the hospital was one hour shorter. However, the intervention group experienced re-arrest in the field significantly more often than the control group and had more pulmonary edema and diuretic use. Readers interested in exploring this topic further are referred to this post on Homeopathic Hypothermia.
Hyperglycemic Control in Pediatric Intensive Care
In this article in the January 9th issue of NEJM, we are visited yet again by the zombie topic that refuses to die. We keep looking for subgroups or populations that will benefit, and if we find one that appears to, it will be a Type I error, thinks the blogger with Bayesian inclinations. In this trial, 1369 pediatric patients at 13 centers in England were randomized to tight versus conventional glycemic control. Consistent with other trials in other populations, there was no benefit in the primary outcome, but tightly "controlled" children had much more and severe hypoglycemia. The "cost effectiveness" analysis they report is irrelevant. You can't have "cost effectiveness" of an ineffective therapy. My, my, how we continue to grope.
Prehospital Induction of Hypothermia After Cardiac Arrest
In this article in the January 1st issue of JAMA, investigators from King County Washington report the results of a trial which tested the hypothesis that earlier (prehospital) induction of hypothermia, by infusing cold saline, would augment the assumed benefit of hypothermia that is usually initiated in the hospital for patients with ventricular fibrillation. Please guess what was the effect of this intervention on survival to hospital discharge and neurological outcomes.
You were right. There was not even a signal, not a trend towards benefit, even though body temperature was lower by 1 degree Celcius and time to target hypothermia temperature in the hospital was one hour shorter. However, the intervention group experienced re-arrest in the field significantly more often than the control group and had more pulmonary edema and diuretic use. Readers interested in exploring this topic further are referred to this post on Homeopathic Hypothermia.
Hyperglycemic Control in Pediatric Intensive Care
In this article in the January 9th issue of NEJM, we are visited yet again by the zombie topic that refuses to die. We keep looking for subgroups or populations that will benefit, and if we find one that appears to, it will be a Type I error, thinks the blogger with Bayesian inclinations. In this trial, 1369 pediatric patients at 13 centers in England were randomized to tight versus conventional glycemic control. Consistent with other trials in other populations, there was no benefit in the primary outcome, but tightly "controlled" children had much more and severe hypoglycemia. The "cost effectiveness" analysis they report is irrelevant. You can't have "cost effectiveness" of an ineffective therapy. My, my, how we continue to grope.
Wednesday, January 29, 2014
Does Investigating Delirium Make You Delirious? A Sober Look at Sedation and Analgesia in the ICU
![]() |
Michael's Milk |
As chronicled in the accompanying perspective article by D.S. Jones, delirium has been around as long as ICUs have, and the longer you're there, the more likely you will become delirious. It's an exposure thing. Thus, until somebody reports the results of a trial of delirium treatment or prevention that has important and undeniable effects on clinically relevant outcomes, I will continue to approach delirium as I always have - by going to great lengths to get patients out of bed, off the vent, and out of the ICU as fast as I possibly can - because these things benefit all patients regardless of whether they have an impact on delirium.
Labels:
CAM-ICU,
delirium,
demand elasticity,
dexmedetomidine,
diprivan,
endotracheal tube,
ETT,
ICU,
intensive care,
mechanical ventilation,
nocebo,
pain,
propofol,
sedation
Thursday, January 23, 2014
White Noise: Trials of Pharmaceuticals for Alzheimer's Disease
In yesterday's NEJM, the results of two trials of antiamyloid monoclonal antibodies (sonalezumab and bapeneuzumab) for Alzheimer's Disease (AD) are published. I became interested in the evidence for AD treatments after the recent trial of Vitamin E and Mematine for AD (the TEAM-AD VA Cooperative Trial) was published in JAMA earlier this month. Regular readers know that I think that the prior probability that vitamins, minerals, and antioxidants are beneficial for any disease outside of deficiency states is very low. The vitamin E trial was the impetus for some background investigation which I will summarize below.
Tuesday, January 7, 2014
The Girl is Brain Dead But the Emperor Has No Clothes
On my other blog, Status Iatrogenicus, the story of Jahi McMath, the brain dead girl from Oakland Children's Hospital is being chronicled and dissected, for those who may be interested.....
Friday, December 27, 2013
Billions and Billions of People on Statins? Damn the Torpedos and Full Speed Ahead
![]() |
Absolutely Relative Risk is in the Mind of the Taker |
Are the premises of the guidelines flawed leading to flawed extrapolations, or are the premises correct and we just don't like the implications? Let's look at the premises - because if they're flawed, we may find that other premises we have accepted are flawed.
Wednesday, November 20, 2013
Chill Out: Homeopathic Hypothermia after Cardiac Arrest

Readers of this blog may know that I harbor measured skepticism for HACA even though I recognize that it may be beneficial. From a pragmatic perspective, it makes sense to use it, since the outcome of hypoxic-ischemic encephalopathy (HIE) and ABI (Anoxic Brain Injury) is so dismal. But what did the original two studies actually show?
- The HACA group multicenter trial randomized 273 patients to hypothermia versus control and found that the hypothermia group had higher rates of "favorable neurological outcome" (a cerebral performance category of 1 or 2 - the primary endpoint) with RR of 1.40 and 95% CI 1.08-1.81; moreover, mortality was lower in the hypothermia group, with RR 0.74 and 95% CI 0.58-0.95
- The Bernard et al study randomized 77 patients to hypothermia versus control and found that survival (the primary outcome) was 49% and 26% in the hypothermia and control groups, respectively, with P=0.046
Monday, November 18, 2013
Dead in the Water: Colloids versus Crystalloids for Fluid Resuscitation in the ICU

- Because the prior probability of success is so low (based on extant trials) that a subsequent trial is unlikely to influence the posterior probability that any success represents the truth. (This is a Bayesian or meta-analytic worldview.)
- Because the low probability of success does not justify the expense of additional trials
- Because the low probability of success violates bioethical precepts mandating that trials must have added value for patients and society
And so we have, in the November 6th edition of JAMA, the CRISTAL trial of colloids versus crystalloids for resuscitation in the ICU. As is customary, I will leave it to interested readers to peruse the manuscript for details. My task here is to provide some background and nuance.
Saturday, November 16, 2013
The Cardiologist Giveth, then the Cardiologist Taketh Away: Revision of the Cholesterol Guidelines

- 3 Things to Know About the New Cholesterol Guidelines by Harlan Krumholz of Yale
- Don’t Give More Patients Statins by authors from Harvard and UCSF
- Experts Reshape Treatment Guide for Cholesterol
- New Cholesterol Advice Startles Even Some Doctors
- Questions For A New Class of Cholesterol Drugs
As the old Spanish proverb states, "rio revuelto, ganancia de pescadores" - when the river is stirred up, the fishermen benefit. I will admit that I'm gloating a bit since I consider the new guidelines to be a tacit affirmative nod to several posts on the topic of the cholesterol hypothesis (CH). (More posts here and here and here, among several others - search for "cholesterol" or "causal pathways" on the Medical Evidence Blog search bar.)
Sunday, November 3, 2013
The Intensivist Giveth Then the Intensivist Taketh Away: Esmolol in Septic Patients Receiving High Dose Norepinephrine
Two studies in the October 23/30 issue of JAMA serve as fodder for reflection on the history and direction of critical care research and the hypotheses that drive it. Morelli et all report the results of a study of Esmolol in septic shock. To quickly summarize, this was a single center dose ranging study the primary aim of which was to determine if esmolol could be titrated to a heart rate goal (primary outcome), presumably with the later goal of performing a phase 3 clinical trial to see if esmolol, titrated in such a fashion, could favorably influence clinical outcomes of interest. 154 patients with septic shock on high dose norepinephrine with a heart rate greater than 95 were enrolled, and heart rate was indeed lower in the esmolol group (P less than 0.001). Perhaps surprisingly, hemodynamic parameters, lactate clearance, and pressor and fluid requirements were (statistically significantly) improved in the esmolol group. Most surprising (and probably the reason why we find this published in JAMA rather than Critical Care Medicine - consider that outlier results such as this may get disproportionate attention), mortality in the esmolol group was 50% compared to 80% in the control group (P less than 0.001). The usual caveats apply here: a small study, a single center, lack of blinding. And regular readers will guess that I won't swallow the mortality difference. I'm a Bayesian (click here for a nice easy-to-use Bayesian calcluator), there's no biological precedent for such a finding and it's too big a bite for me to swallow. So I will go on the record here as stating that I'm betting against similar results in a larger trial.
I'm more interested in how we formulate the hypothesis that esmolol will provide benefit in septic shock. I was a second year medical student in 1995 when Gattinoni et al published the results of a trial of "goal-oriented hemodynamic therapy" in critically ill patients in the NEJM. I realize that critical care research as we now recognize it was in its adolescence then, as a quick look at the methods section of that article demonstrates. I also recognize that they enrolled a heterogenous patient population. But it is worth reviewing the wording of the introduction to their article:
I'm more interested in how we formulate the hypothesis that esmolol will provide benefit in septic shock. I was a second year medical student in 1995 when Gattinoni et al published the results of a trial of "goal-oriented hemodynamic therapy" in critically ill patients in the NEJM. I realize that critical care research as we now recognize it was in its adolescence then, as a quick look at the methods section of that article demonstrates. I also recognize that they enrolled a heterogenous patient population. But it is worth reviewing the wording of the introduction to their article:
Recently, increasing attention has been directed to the hemodynamic treatment of critically ill patients, because it has been observed in several studies that patients who survived had values for the cardiac index and oxygen delivery that were higher than those of patients who died and, more important, higher than standard physiologic values.1-3 Cardiac-index values greater than 4.5 liters per minute per square meter of body-surface area and oxygen-delivery values greater than 650 ml per minute per square meter — derived empirically on the basis of the median values for patients who previously survived critical surgical illness — are commonly referred to as supranormal hemodynamic values.4
Saturday, October 12, 2013
Goldilocks Meets Walter White in the ICU: Finding the Temperature (for Sepsis and Meningitis) that's Just Right

Reading the Point and Counterpoint piece (in addition to an online first article in JAMA describing a trial of induced hypothermia in severe bacterial meningitis - more on that later) allowed me to synthesize some ideas about the epistemology (and psychology) of medical evidence and its evaluation that I have been tossing about in my head for a while. Both the proponent pair and the opponent pair of authors give some background physiological reasoning as to why fever may be, by turns, beneficial and detrimental in sepsis. The difference, and I think this is typical, is that the proponents of fever reduction: a.) seem much more smitten by their presumed understanding of the underlying physiology of sepsis and the febrile response; b.) focus more on minutiae of that physiology; c.) fail to temper their faith in application of physiological principles with the empirical data; and d.) grope for subtle signals in the empirical data that appear to rescue the sinking hypothesis.
Friday, August 2, 2013
Sause for the Goose, Sauce for the Gander: Low Tidal Volume Ventilation in the Operating Theatre
![]() |
PIBW is based on height, not weight. |
Following my usual procedure, I read the title and abstract of the methods of this article on Intraoperative Low Tidal Volume Ventilation in this week's NEJM, and I made a wager with myself on what the outcome would be. Because there are both biological plausibility and biological precedent for low tidal volume, and because it is one of the few interventions in critical care in which I have supreme confidence (yes, you can conclude that I'm biased), my prior probability for this intervention is high and I wagered that the study would be positive. If you have not already done so, read the methods in the abstract and make your own wager before you read on.
This trial is solid but not bombproof. Outcomes assessors were blinded and so were post-operative care providers, but anesthesiologists administering tidal volumes were not. Outcomes themselves, while mostly based on consensus definitions (sometimes a consensus of collective ignorance), are susceptible to ascertainment and misclassification biases. The outcome was a composite, something that I like, as will be elaborated in a now published letter in AJRCCM. A composite outcome allows an additive effect between component outcomes and effectively increases study power. This is essential in a study such as this, where only 400 patients were enrolled and the study had "only" 80% power to detect a reduction in the primary outcome from 20% to 10%. As we have shown, detecting a difference of this magnitude in mortality is a difficult task indeed, and most critical care studies seeking such a difference are effectively underpowered. How many effective (in some aspect other than mortality) therapies have been dismissed because of this systemic underpowering in critical care research is anybody's guess.
Thursday, June 20, 2013
More is Not Less, It Just Costs More: Early Tracheostomy, Early Parenteral Nutrition, and Rapid Blood Pressure Lowering in ICH

Firstly, Young et al (May 22/29, 2013 issue of JAMA) report the results of the TracMan multicenter trial of early tracheostomy in ICUs in the UK. These data seal the deal on an already evolving shift in my views on early tracheostomy that were based on anecdotal experience and earlier data from Rumbak and Terragni. Briefly, the authors enrolled 899 patients expected to receive at least 7 more days of mechanical ventilation (that prediction was no more reliable in the current trial than it had been in previous trials) and randomized them to receive a trach on day 4 (early) versus on day 10 (late). The early patients did end up receiving less sedatives and had a trend toward shorter duration of respiratory support. But their KM curves are basically superimposable and the mortality rates virtually identical at 30 days. These data, combined with other available studies, leave little room for subjective interpretation. Early tracheostomy, it is very likely, does not favorably affect outcomes enough to justify its costs and risks.
Friday, May 31, 2013
Over Easy? Trials of Prone Positioning in ARDS

First, a general principle:
regression to the mean. Few, if
any, therapies in critical care (or in medicine in general) confer a mortality
benefit this large. I refer the reader (again) to our study of delta inflation which tabulated over 30 critical care trials in the top 5
medical journals over 10 years and showed that few critical care trials show mortality
deltas (absolute mortality differences) greater than 10%. Almost
all those that do are later refuted.
Indeed it was our conclusion that searching for deltas greater than or
equal to 10% is akin to a fool's errand, so unlikely is the probability of
finding such a difference. Jimmy T.
Sylvester, my attending at JHH in late 2001 had already recognized this. When the now infamous sentinel trail of intensive insulin therapy (IIT) was published, we discussed it at our ICU
pre-rounds lecture and he said something like "Either these data are
faked, or this is revolutionary." We
now know that there was no revolution (although many ICUs continue to practice as if there had been one). He
could have just as easily said that this is an anomaly that will regress to the
mean, that there is inherent bias in this study, or that "trials stopped early for benefit...."
Monday, May 20, 2013
It All Hinges on the Premises: Prophylactic Platelet Transfusion in Hematologic Malignancy
A quick update before I proceed with the current post: The Institute of Medicine has met and they agree with me that sodium restriction is for the birds. (Click here for a New York Times summary article.) In other news, the oh-so-natural Omega-3 fatty acid panacea did not improve cardiovascular outcomes as reported in the NEJM on May 9th, 2013.
An article by the TOPPS investigators in the May 9th NEJM is very useful to remind us not to believe everything we read, to always check our premises, and that some data are so dependent on the perspective from which they're interpreted or the method or stipulations of analysis that they can be used to support just about any viewpoint.
An article by the TOPPS investigators in the May 9th NEJM is very useful to remind us not to believe everything we read, to always check our premises, and that some data are so dependent on the perspective from which they're interpreted or the method or stipulations of analysis that they can be used to support just about any viewpoint.
The authors sought to determine if a strategy of withholding
prophylactic platelet transfusions for platelet counts below 10,000 in patients
with hematologic malignancy was non-inferior to giving prophylactic platelet
transfusions. I like this idea, because
I like "less is more" and I think the body is basically
antifragile. But non-inferior how? And what do we mean by non-inferior in this
trial?
Saturday, April 27, 2013
Tell Them to Go Pound Salt: Ideology and the Campaign to Legislate Dietary Sodium Intake
In the March 28th, 2013 issue of the NEJM, a review of sorts
entitled "Salt in Health and Disease - A Delicate Balance" by Kotchen et al can be
found. My interest in this topic stems
from my interest in the question of association versus causation, my personal predilection for salt, my observation that I lose a good deal of sodium in outdoor activities
in the American Southwest, and my concern for bias in the generation of and especially
the implementation of evidence in medicine as public policy.
This is an important topic, especially because sweeping
policy changes regarding the sodium content of food are proposed, but it is a
nettlesome topic to study, rife with hobgoblins. First we need
a well-defined research question: does reduction
in dietary sodium intake: a.) reduce
blood pressure in hypertensive people? in
all people? b.) does this reduction in
hypertension lead to improved outcomes (hypertension is in some ways a
surrogate marker)? In a utopian world,
we would randomize thousands of participants to diets low in sodium and "normal"
in sodium, we would measure sodium intake carefully, and we would follow the
participants for changes in blood pressure and clinical outcomes for a
protracted period. But alas, this has
not been done, and it will not likely be done because of cost and logistics,
among other obstacles (including ideology).
Friday, April 19, 2013
David versus Goliath on the Battlefield of Non-inferiority: Strangeness is in the Eye of the Beholder
In this week's JAMA is my letter to the editor about the CONSORT statement revision for the reporting of non-inferiority trials, and the authors' responses. I'll leave it to interested readers to view for themselves the revised CONSORT statement, and the letter and response.
In sum, my main argument is that Figure 1 in the article is asymmetric, such that inferiority is stochastically less likely than superiority and an advantage is therefore conferred to the "new" [preferred; proprietary; profitable; promulgated] treatment in a non-inferiority trial. Thus the standards for interpretation of non-inferiority trials are inherently biased. There is no way around this, save for revising the standards.
The authors of CONSORT say that my proposed solution is "strange" because it would require revision of the standards of interpretation for superiority trials as well. For me it is "strange" that we would endorse asymmetric and biased standards of interpretation in any trial. The compromised solution, as I suggested in my letter, is that we force different standards for superiority only in the context of a non-inferiority trial. Thus, superiority trial interpretation standards remain untouched. It is only if you start with a non-inferiority trial that you have a higher hurdle to claiming superiority that is contingent on evidence of non-inferiority in the trial that you designed. This would disincentivise the conduct of non-inferiority trials for a treatment that you hope/think/want to be superior. In the current interpretation scheme, it's a no-brainer - conduct a non-inferiority trial and pass the low hurdle for non-inferiority, and then if you happen to be superior too, BONUS!
In my proposed scheme, there is no bonus superiority that comes with a lower hurdle than inferiority. As I said in the last sentence, "investigators seeking to demonstrate superiority should design a superiority trial." Then, there is no minimal clinically important difference (MCID) hurdle that must be cleared, and a statistical difference favoring new therapy by any margin lets you declare superiority. But if you fail to clear that low(er) hurdle, you can't go back and declare non-inferiority.
Which leads me to something that the word limit of the letter did not allow me to express: we don't let unsuccessful superiority trials test for non-inferiority contingently, so why do we let successful non-inferiority trials test for superiority contingently?
Symmetry is beautiful; Strangeness is in the eye of the beholder.
(See also: Dabigatran and Gefitinib especially the figures, analogs of Figure 1 of Piaggio et al, on this blog.)
In sum, my main argument is that Figure 1 in the article is asymmetric, such that inferiority is stochastically less likely than superiority and an advantage is therefore conferred to the "new" [preferred; proprietary; profitable; promulgated] treatment in a non-inferiority trial. Thus the standards for interpretation of non-inferiority trials are inherently biased. There is no way around this, save for revising the standards.
The authors of CONSORT say that my proposed solution is "strange" because it would require revision of the standards of interpretation for superiority trials as well. For me it is "strange" that we would endorse asymmetric and biased standards of interpretation in any trial. The compromised solution, as I suggested in my letter, is that we force different standards for superiority only in the context of a non-inferiority trial. Thus, superiority trial interpretation standards remain untouched. It is only if you start with a non-inferiority trial that you have a higher hurdle to claiming superiority that is contingent on evidence of non-inferiority in the trial that you designed. This would disincentivise the conduct of non-inferiority trials for a treatment that you hope/think/want to be superior. In the current interpretation scheme, it's a no-brainer - conduct a non-inferiority trial and pass the low hurdle for non-inferiority, and then if you happen to be superior too, BONUS!
In my proposed scheme, there is no bonus superiority that comes with a lower hurdle than inferiority. As I said in the last sentence, "investigators seeking to demonstrate superiority should design a superiority trial." Then, there is no minimal clinically important difference (MCID) hurdle that must be cleared, and a statistical difference favoring new therapy by any margin lets you declare superiority. But if you fail to clear that low(er) hurdle, you can't go back and declare non-inferiority.
Which leads me to something that the word limit of the letter did not allow me to express: we don't let unsuccessful superiority trials test for non-inferiority contingently, so why do we let successful non-inferiority trials test for superiority contingently?
Symmetry is beautiful; Strangeness is in the eye of the beholder.
(See also: Dabigatran and Gefitinib especially the figures, analogs of Figure 1 of Piaggio et al, on this blog.)
Subscribe to:
Posts (Atom)