Amid a Pandemic, a Health Care Algorithm Shows Promise and Peril

Posted on by

Yves here. Aside from my general allergy to AI and algos (among other things, they are only as good as their training sets, which raises questions of accuracy and consistency of inputs), another reason to be concerned with health care algos is they require collection of patient data to work, which means yet another source of data vulnerability. Our reader IM Doc pointed out:

In the USA – we have multiple large tertiary referral centers that have quite the national reputation – I would include in that list MD Anderson, MAYO, Johns Hopkins, and …… The Scripps Clinic in La Jolla, California. I have innumerable patients that are seen there – they cater to that type of clientele. I first heard about this impending disaster over the weekend – and today things appeared to get immeasurably worse there…… see the following article……

https://www.nbcsandiego.com/news/local/what-we-know-about-scripps-health-cyberattack/2598969/

I know we have a major pipeline down from ransomware now – but this is just as scary if not more so. This is a major medical system in this country – and it has been hobbled. All of my patient’s appointments there have been cancelled until June – they are admitting no one – and no one seems to know if it will be back or not anytime soon. It has already been going on for a week.

ANNNNDDDDD – they use Epic – which has repeatedly touted itself ( I have been in the meetings multiple times in my life) as completely impervious to hacking.

Again – I knew this day was coming at some point. These EMR systems are a complete disaster waiting to happen. The hackers have managed up to this point to take down non-EPIC systems at Bugtussle Memorial Hospital across the country – but nothing like Scripps.

The patient portal was among the systems taken out by hackers..and it was “Epic powered” or some such corporate jargon. While Scripps has been remarkably close-mouthed, available evidence says it’s not a reach to think Epic is implicated.

By Vishal Khetpal, M.D., MSc is an internal medicine resident physician training in the Brown University Internal Medicine Program and Nishant Shah, M.D., MPH is an assistant professor of medicine at the Alpert Medical School of Brown University and an assistant professor of health services, practice, and policy at the Brown University School of Public Health. Originally published at Undark

Last spring, physicians like us were confused. Covid-19 was just starting its deadly journey around the world, afflicting our patients with severe lung infections, strokes, skin rashes, debilitating fatigue, and numerous other acute and chronic symptoms. Armed with outdated clinical intuitions, we were left disoriented by a disease shrouded in ambiguity.

In the midst of the uncertainty, Epic, a private electronic health record giant and a key purveyor of American health data, accelerated the deployment of a clinical prediction tool called the Deterioration Index. Built with a type of artificial intelligence called machine learning and in use at some hospitals prior to the pandemic, the index is designed to help physicians decide when to move a patient into or out of intensive care, and is influenced by factors like breathing rate and blood potassium level. Epic had been tinkering with the index for years but expanded its use during the pandemic. At hundreds of hospitals, including those in which we both work, a Deterioration Index score is prominently displayed on the chart of every patient admitted to the hospital.

The Deterioration Index is poised to upend a key cultural practice in medicine: triage. Loosely speaking, triage is an act of determining how sick a patient is at any given moment to prioritize treatment and limited resources. In the past, physicians have performed this task by rapidly interpreting a patient’s vital signs, physical exam findings, test results, and other data points, using heuristics learned through years of on-the-job medical training.

Ostensibly, the core assumption of the Deterioration Index is that traditional triage can be augmented, or perhaps replaced entirely, by machine learning and big data. Indeed, a study of 392 Covid-19 patients admitted to Michigan Medicine that the index was moderately successful at discriminating between low-risk patients and those who were at high-risk of being transferred to an ICU, getting placed on a ventilator, or dying while admitted to the hospital. But last year’s hurried rollout of the Deterioration Index also sets a worrisome precedent, and it illustrates the potential for such decision-support tools to propagate biases in medicine and change the ways in which doctors think about their patients.

The use of algorithms to support clinical decision making isn’t new. But historically, these tools have been put into use only after a rigorous peer review of the raw data and statistical analyses used to develop them. Epic’s Deterioration Index, on the other hand, remains proprietary despite its widespread deployment. Although physicians are provided with a list of the variables used to calculate the index and a rough estimate of each variable’s impact on the score, we aren’t allowed under the hood to evaluate the raw data and calculations.

Furthermore, the Deterioration Index was not independently validated or peer-reviewed before the tool was rapidly deployed to America’s largest health care systems. Even now, there have been, to our knowledge, only two peer-reviewed published studies of the index. The deployment of a largely untested proprietary algorithm into clinical practice — with minimal understanding of the potential unintended consequences for patients or clinicians — raises a host of issues.

It remains unclear, for instance, what biases may be encoded into the index. Medicine already has a fraught history with race and gender disparities and biases. Studies have shown that, among other injustices, physicians underestimate the pain of minority patients and are less likely to refer women to total knee replacement surgery when it is warranted. Some clinical scores, including calculations commonly used to assess kidney and lung function, have traditionally been adjusted based on a patient’s race — a practice that many in the medical community now oppose. Without direct access to the equations underlying Epic’s Deterioration Index, or further external inquiry, it is impossible to know whether the index incorporates such race-adjusted scores in its own algorithm, potentially propagating biases.

Introducing machine learning into the triage process could fundamentally alter the way we teach medicine. It has the potential to improve inpatient care by highlighting new links between clinical data and outcomes — links that might otherwise have gone unnoticed. But it could also over-sensitize young physicians to the specific tests and health factors that the algorithm deems important; it could compromise trainees’ ability to hone their own clinical intuition. In essence, physicians in training would be learning medicine on Epic’s terms.

Thankfully, there are safeguards that can be relatively painlessly put in place. In 2015, the international Equator Network created a 22-point Tripod checklist to guide the responsible development, validation, and improvement of clinical prediction tools like the Deterioration Index. For example, it asks tool developers to provide details on how risk groups were created, report performance measures with confidence intervals, and discuss limitations of validation studies. Private health data brokers like Epic should always be held to this standard.

Now that its Deterioration Index is already being used in clinical settings, Epic should immediately release for peer review the underlying equations and the anonymized datasets it used for its internal validation, so that doctors and health services researchers can better understand any potential implications they may have for health equity. There need to be clear communication channels to raise, discuss, and resolve any issues that emerge in peer review, including concerns about the score’s validity, prognostic value, bias, or unintended consequences. Companies like Epic should also engage more deliberately and openly with the physicians who use their algorithms; they should share information about the populations on which the algorithms were trained, the questions the algorithms are best equipped to answer, and the flaws the algorithms may carry. Caveats and warnings should be communicated clearly and quickly to all clinicians who use the indices.

The Covid-19 pandemic, having accelerated the widespread deployment of clinical prediction tools like the Deterioration Index, may herald a new coexistence between physicians and machines in the art of medicine. Now is the time to set the ground rules to ensure that this partnership helps us change medicine for the better, and not the worse.

Print Friendly, PDF & Email

18 comments

  1. Andreas Agas

    Underlying the problem, the technology industry (and the media) continues to represent that systems, such as described, use “artificial intelligence”.

    Nothing remotely like artificial intelligence exists. Intelligence requires consciousness. No one has yet clearly defined either consciousness or intelligence.

    Even the idea of “machine learning” overstates what these machines and their algorithms can do.

    They can only do search and pattern matching. Two things that don’t sound any where close to as impressive as “artificial intelligence” or “machine learning”.

    The hype and overstatement of these tools sets the public up to trust them more than the tools (and their operators) deserve.

    1. Synoia

      Artificial Intelligence: Where is, or where are, the calibration and test results?

      Are these results repeatable and consistent across multiple AI systems?

      How are disagreements between AIs adjudicated?

  2. TimH

    I suspect that AI will prove popular as a method to downgrade point-of-health-provision costs because the provider will be able to deny culpability. “The expert [system] told me so!”. There will be finger pointing between AI provider and health provider when SHTF and disclaim of responsibility from both sides.

    Need mandatory algorithm transparency and legal clarification that the health provider is responsible regardless of tool used. Which means also that they need experts to look inside the machine before deployment, just as you’d check that scalpels are sharp and sterilizers work properly.

    1. JTMcPhee

      And the provider, if conscientious like imdoc, will be mandated by “standard of care” and insurance company and Medicare protocols, to utilize these “tools,” at risk of liability or non-reimbursement….

      Another way station on the race to the crapified bottom.

  3. upstater

    On my provider’s EPIC portal, each and every login comes with an acceptance of their T&C. It is the ONLY business portal requiring consent every time. It is MANY pages of legalese. One can only imagine that agreeing to AI derived metrics will be buried in that monstrosity.

  4. Jeremy Grimm

    I am feeling especially grim and cynical at the moment.

    ” The deployment of a largely untested proprietary algorithm into clinical practice — with minimal understanding of the potential unintended consequences for patients or clinicians — raises a host of issues.” Indeed, questions similar to those regarding the algorithm discussed in this post, might be raised regarding the EUA approved ‘Warp Speed’ vaccines.

    Physicians perform triage through “rapidly interpreting a patient’s vital signs, physical exam findings, test results, and other data points, using heuristics learned through years of on-the-job medical training.” The AI program can add income, credit score, and personal assets to its triage scoring algorithm to better fit into medical care as practiced in the US.

    “Without direct access to the equations underlying Epic’s Deterioration Index…” — really? the equations? Does the author of this post want the equations or a list of the variables used to calculate the index and their relative weightings in calculating the triage index? While it is possible to infer these variables and weightings, or with a series of test entries to ‘black box’ extract many of them out of the algorithm, I have little doubt the equations proffered, if any might be, could hide information almost as well as the program’s black box.

    1. Skip Intro

      ‘No matter how cynical I get, I can’t keep up’ – Lily Tomlin

      I think it is worse than you indicate, and suggests a fundamental misunderstanding of machine learning algorithms on the part of the author, Dr. Khetpal. The equations the good doctor imagines are not related to the input variables in any human comprehensible way, so revealing them or more importantly, the internal weights used by the algorithm would be fundamentally useless. No single decision can be traced back to the inputs, it is a truly black box, even to the engineers working on it, as the equations and weights emerge from training in fundamentally inscrutable ways. This will be part of the fun for self-crashing cars, for fans of litigation.

      1. JTMcPhee

        The corps rolling out these ‘tools’ will, like the nuclear power industry and the autonomous vehicle vendors, do their damndest to change “the law” to insulate themselves from liability and push the resulting externalities off onto the mopery.

  5. SamT

    I sent this to a friend of mine who works at Epic. They said the determination index formula is available on Epic’s forum for customers which at least every hospitals IT team and administrators have access to. The IT team at each hospital had to take time to set it up, and it is not meant to be the main source of decision making on triage. They also said it’s just a simple regression. There is nothing complicated like machine learning happening.

    1. Yves Smith Post author

      While that sounds nice, you know that in a world where ERs are staffed by outsourced companies run by private equity, this Epic cute position “Oh, this isn’t to be the guidebook” is likely to prove to be laughable in practice. The big trend is corporatized medicine and fixed protocols as opposed to practitioner judgment.

      And IT people do not have clinical or statistical chops, nor are they players in decision-making. So acting as if having an hospital IT person involved = quality control or independent validation is also laughable

      Their PR calls it a “predictive model” which is a way stronger claim than a regression.

      https://www.epic.com/epic/post/epic-ai-helps-clinicians-predict-covid-19-patients-might-need-intensive-care

      And non-subscribers can’t read their damned white paper, which does not inspire confidence. So I can’t verify your contact’s beliefs.

      1. Equitable > Equal

        The same dynamic can be seen in Tesla’s implicit description of their driver aids as ‘self-driving’, with implicit links to the concept that is it artificially intelligent. Unfortunately, the death toll shows both that is it far from artificially intelligent, and also reliably misused in practice despite the disclaimers.

  6. lambert strether

    I love the term “Deterioration Index,” but why restrict it to the ER?

  7. Old Mainer

    As folks with long term diseases are often reminded, “You are an individual, not a statistic.” Statistics apply to population, not individuals. An individual may, may slightly, may somewhat, may mostly or may not at all match the population on which a statistic is based.

    AI as presently constituted is more statistics. A hospital may find the Deterioration Index improves total outcomes for a population of patients while at the same time dramatically worsening outcomes for some individual patients. This is is direct opposition to the first principle of medicine, “First, do no harm.”

  8. Gregory Etchason

    AI in healthcare could truly be transformative. If your current job in healthcare can be reduced to zeros and ones( Derm, Pathology Radiology) you’ll likely getdownsized in the next decade. AI is already superior to most humans ability to visually make a Dx. Unfortunately more likely Corp Medicine will use it to continue the discounting of expertise and get rid of workers. Same for EVs it’s not about the environment it’s about a chance for manufacturers to get rib of 30-50% of the workers.

  9. Gregory Etchason

    AI in healthcare could truly be transformative. If your current job in healthcare can be reduced to zeros and ones( Derm, Pathology Radiology) you’ll likely get downsized in the next decade. AI is already superior to most humans ability to visually make a Dx. Unfortunately more likely Corp Medicine will use it to continue the discounting of expertise and get rid of workers. Same for EVs it’s not about the environment it’s about a chance for manufacturers to get rib of 30-50% of the workers.

Comments are closed.