Heuristics, language and medical errors

Cover Page


Cite item

Full Text

Open Access Open Access
Restricted Access Access granted
Restricted Access Subscription or Fee Access

Abstract

The use of heuristics, cognitive “rules of thumb” or “shortcuts”, are a common part of medical decision making. While using heuristics lessens the cognitive burden of decision making and often comes up with the correct answer, reliance on heuristics can also lead to medical errors. In this introduction we will define heuristics and discuss some of the more common heuristics/biases that may lead to biased decision making and patient harm. We will also look at the use of language in medicine and how that can lead us to make diagnostic errors.

Full Text

Introduction

Heuristics are cognitive shortcuts that allow us to make decisions rapidly and often with incomplete information [1–7].

Using a heuristic requires little cognitive effort. For example, we may be in the midst of an influenza outbreak and see a patient with symptoms suggestive of influenza. We briefly interview and examine the patient and decide that it is influenza. Of course, it could be SARS or MERS. However, even though we used an intuitive approach, a cognitive shortcut, and didn’t consciously direct our workup to rule out everything else, it would be ludicrous to do a full evaluation for MERS on every patient with influenza symptoms. Our mental shortcut (heuristic) is probably right. We are using the “availability heuristic” to make our decision; influenza is in the news and seemingly every patient we see has influenza. It is in the top of our mind and is the most “available” (and in this case the most likely) diagnosis [8].

Most of the time heuristics work fine, and in some situations, may be more accurate than conscious deliberation [9]. In fact, about 95% of our decisions are based on intuition/heuristics [10]. However, the use of heuristics can also lead us to make wrong decisions. Take, for example, the “availability heuristic.” We are more likely to think of a pulmonary embolism in our next patient with shortness of breath if we recently misdiagnosed a patient with a pulmonary embolism. Or, when we see a patient with an uncommon illness, we suddenly “see” that same illness in patients with similar symptoms even though the disease itself may be uncommon. The illness is top in our mind (most available) and so we “see it” everywhere.

Heuristics are an evolutionary adaptation and are hard wired in our brains. For example, a heuristic may suggest that we run away from a charging lion without consciously balancing the pros and cons of various options for 10 minutes (at which time we likely will be dinner with a side of fries). If we are wrong and it is a Russian Blue Cat, we haven’t wasted a lot of mental effort making our decision and we have survived [12]. Functional MRI (fMRI) has identified separate neural pathways for heuristic-based decisions and those made by contemplation – a distinction now referred to as “the dual process model” [13–16].

Being general rules, heuristics can cause us to make mistakes in an individual case if that case deviates from the general rule. The problem is not just theoretical. In one study, up to 74% of medical errors were determined to be the result of cognitive errors. 30% of emergency physicians self-identified cognitive errors as a source of mistakes while cognitive errors were operant by self-report in 42% internal medicine errors [17–19]. There are many cognitive and affective errors that are well recognized as sources of errors in medical decision making. A partial compilation of 50 of these by Pat Croskerry MD, PhD, FRCP from Dalhousie University in Nova Scotia can be found here: http://sjrhem.ca/wp-content/uploads/2015/11/CriticaThinking-Listof50-biases.pdf [5]. It would be impossible to address all, or even a significant portion of these, in a short paper. However, with this background, let’s consider some examples.

Case 1

NHK is a 64-year-old male with a history of hypertension who presents to the emergency department with abdominal pain, described as a burning in the midepigastric region after he eats a large meal. The pain usually resolves with an H2 blocker or an antacid. The patient turns out to be a retired physician as well as a major benefactor of the hospital. The Chief-of-Staff and head of surgery show up. The clinical evaluation reveals diffuse mild, tenderness, especially midepigastric. An appropriate workup, including an EKG, is normal.

This is not a patient that you would evaluate further given the history and exam. It sounds like simple esophageal reflux. However, the Chief-of-Staff wonders out loud (and just within hearing range of the patient) if a computerized tomography scan of the abdomen should be done. Since the patient is a friend yours, you really don’t want to miss anything and order a CT scan. The CT scan shows an adrenal incidentaloma which prompts the surgeon to recommend an open biopsy “just to be sure.” Usually these are followed every 6 months, but the surgeon doesn’t want to miss anything either. The patient dies on the table from uncontrolled bleeding, the result of an unneeded test. (This process of following up one unneeded test with others is called “The Ulysses Syndrome”. It often results in an adverse patient event [20].)

Affect Heuristic/Visceral Bias

Having a friend (or disagreeable individual or anyone, really) as a patient may lead to decisions invoked by the “affect heuristic” also known as “visceral bias” [11]. The affect heuristic is ope rative when we have positive or negative feelings  about our patient. This is at least part of the reason behind the admonition to not care for close family members and why many of don’t like taking care of colleagues and other “important” people such as major donors to the hospital.

These feelings can skew our decision-making in two ways. The first is “omission bias,” which might lead us to avoid an indicated but painful procedure on a friend or family member. For example , we might be more hesitant to perform a lumbar puncture on our own child or significant other than we would be on a patient who is other wise a stranger. This omission could lead to a missed or delayed diagnosis of meningitis or sub-arachnoid hemorrhage. A second potential outcome of the affect heuristic is “commission bias” whereby we are more likely to perform tests on a patient in an attempt to not miss anything. Commission bias is what led to a bad outcome in the case above. It isn’t unusual to find an “incidentaloma”, an unexpected finding on a CT scan [21]. There are guidelines to manage these findings, but we feel pressured by our friendship with our colleague in the case above to “do more” to rule out any significant illness. This may lead to a bad outcome (bleeding from a procedure, etc.).

Case 2

SM is a 41-year-old complaining of the sudden onset of a severe headache. He states he never gets headaches and that this is the worst headache of his life. The nurse tells you he is a “drug seeker” and had an addiction to narcotics in the past (although she hasn’t seen him for 12 years). His physical examination is intact except the clothing he is wearing are old and a bit shabby. The patient is discharged with a diagnosis of a severe migraine and returns by ambulance to the hospital comatose later that day. The CT scan shows a large subarachnoid hemorrhage.

Attribution Error

If the individual was a businessperson dressed in a suit and tie, and the nurse hadn’t told us that he had a history of drug seeking in the past, we may have done the CT scan at the first visit. The symptoms are suggestive of a subarachnoid hemorrhage; sudden onset of a severe headache in patient without a history of headache. This is an example of “attribution error”. An “attribution error” occurs when we make decisions based on irrelevant information based on a patient’s race, age, or other “attribute” that doesn’t change the probability of disease. For example, both a 41-year-old well-dressed banker and a 41-year-old homeless individual can have a subarachnoid. Yet we treat the banker diffe rently than the homeless individual. In our case, we have no idea if this patient currently has a drug problem. The last history we have of this is from 12 years ago. Even if he does, it is still irrelevant to the chief complaint given that the mortality from each subarachnoid hemorrhage approaches 50%: It isn’t something we can afford to miss.

Think about how you feel if the nurse tells you that your next patient “is looking for narco tics” or is a “drug seeker” or “is crazy”? Do you look forward to seeing this patient? When we are told a patient is a “drug seeker”, we subconsciously assign other characteristics to the patient. Perhaps you are expecting an angry, aggressive and demanding patient. If we are told that a patient is depressed, we are less likely to believe their symptoms (for example a subarachnoid headache symptoms) are real [22]. The way people dress, their race and even their gender, effect how we react to the patient. We may assume that a homeless person is uneducated, and an alcoholic or addict. The truth may be that the patient has a PhD in physics but has been unable to find a job and therefore can’t afford a flat. But we “attribute” certain characteristics to the patient solely based on the fact that he is homeless, poorly dressed, or has a past history of drug addiction. The same is true if we are  seeing a minority or a wo man. We tend to minimize the symptoms if our patient is female, perhaps  thinking “she is just emotional”. We are less likely to do an invasive cardiac workup if our patient is female [23,  24]. This is the same reason that female doctors get called “nurse” or a recent medical school graduate is told “you are too young to be a doctor”. The patient is associating (attributing) female = nurse and a young-looking = can’t possibility be a doctor.

Once you hear something negative about a patient, it is hard to put it out of your mind. So, if you find for some reason you are angry or frustrated or don’t like the patient, take a step back before you make a medical decision and ask yourself why you are having this reaction to the patient and whether or not this is clouding your decision making process. Is this affective bias? Could it be attribution bias? Simply recognizing the role these are playing in your decision-making process can help you avoid errors.

The “one word one meaning” fallacy

The next case is one that we have all experienced (and many of us on frequent basis). Recall when you were a junior student or resident and went into a patient’s room. You walk out of the room and go talk to the senior physician to present the case. The senior doctor goes into the patient’s room and gets an entirely different history. Raise your hand if this has happened to you…more than once! So why does this happen? Part of the reason is the “one word one meaning” fallacy [25]. For example, the patient may deny chest pain when asked. But if you ask, “any pain, pressure, sharp feeling, heaviness, tightness, or discomfort in your chest?” the patient may answer yes. To the patient, “chest pain” may mean one thing (for example sharp). To us, “chest pain” means any cardiac discomfort including pressure, heaviness, a tearing sensation, pain with brea thing, etc. So, start with an open-ended questions such as, “What kind of symptoms are you having today?” If you do need to prompt the patient, give them several options to help them describe their symptoms (pressure, tightness, sharp, stabbing, etc.).

Another example of the “one-word-one-mea ning-fallacy”: JC is a 32-year-old patient who was in the hospital for “shortness of breath”. He had an extensive workup including a left heart ca theterization, a right heart catheterization, pulmonary function tests, a bronchoscopy, a chest CT, etc. He was discharged and promptly walked from his inpatient room to the emergency department. It was very puzzling. Why did he get discharged and come right to emergency department? His response, “I am still short of breath”. Here is this guy who looks like he could be the striker on the Russian National Football team complaining he is short of breath. So, I asked him, “what do you mean by shortness of breath?” As doctors and other care providers we expect him to say something like, “I can’t walk 20 feet without having to stop to catch my breath.” His response? “I can’t breathe through my nose”. This happens a lot: Don’t assume you and the patient are speaking the same language even if you are speaking the same language. To those of us in medicine, “shortness of breath” usually means dyspnea. But, as in this case, the patient may use words differently. So always ask the patient to describe their symptoms. For example, what does the patient mean when she complains of a “migraine headache” or “the flu?” To the patient, a migraine is any bad headache. To us it is a specific kind of headache (unilateral, pulsating, nausea, photophobia). To us, “flu” means “influenza”. To the patient it may be nausea and vomiting. In the case of JC we could have avoided 2 weeks in the hospital and tens of thousands of dollars of tests simply by asking, “What do you mean by shortness of breath?” So any time the patient tells you their diagnosis (“I am dizzy’, “I have a migraine”, “I have a slipped disc”), ask them to describe their symptoms. It is our job to figure out what is going on and not just to put down in the chart what the patient believes that they have. We may need to take the time to educate the patient, especially in the age of the World Wide Web, where patients often think they already know the answer. But that is part of what it means to be a doctor in the 2020s.

Case 3

The last case is also case that everyone has seen. BFL is a 22-year-old female who presents complaining of “twelve out of ten” pain; this is the worst pain she can imagine. Of course, she is smiling and eating a bag of potato chips. You are scratching your head wondering if she is making this up. How can she really be in such bad pain while eating potato chips? She looks fine. Is she just looking for narcotics?

Frame of Reference

It is possible she is looking for narcotics. But it also may be that her frame of reference is different than ours. What is the worst pain a patient can imagine or has seen? Maybe an ingrown toenail, a sprained ankle or falling and hitting her head. What is the worst pain you can ima gine? Probably 30% second degree burns, a kidney stone or maybe an open femur fracture. So, for the patient, this may be the worst pain imaginable. The point is that we should not dismiss this complaint and assume it is fake or embellished. Her idea of what constitutes “worst pain” is going to be different than ours.

Discussion and Conclusion

Language and heuristics both have a roll in diagnostic errors. Teaching about the way we make decisions may help to reduce medical errors [26, 27]. In this paper we discussed the affective heuristic (AKA “visceral bias”), attribution error, availability bias and “the one word on meaning fallacy”. In the follow-up paper we will discuss dealing with uncertainty in medicine, “representiveness”, diagnosis momentum as well as other common sources of medical  error.

×

About the authors

Mark A. Graber

University of Iowa Carver College of Medicine

Author for correspondence.
Email: mark-graber@uiowa.edu

MD MSHCE FACEP, Emeritus Professor of Emergency and Family Medicine

United States, Iowa City

References

  1. Blumenthal-Barby JS, Krieger H. Cognitive biases and heuristics in medical decision making: A critical review using a systematic search strategy. Med Decis Making. 2015;35(4):539-557. https://doi.org/
  2. 1177/0272989X14547740.
  3. Bobadilla-Suarez S, Love BC. Fast or frugal, but not both: Decision heuristics under time pressure. J Exp Psychol Learn Mem Cogn. 2018;44(1):24-33. https://doi.org/10.1037/xlm0000419.
  4. Reilly JB, Ogdie AR, Von Feldt JM, et al. Teaching about how doctors think: A longitudinal curriculum in cognitive bias and diagnostic error for residents. BMJ Qual Saf. 2013;22:1044-1050. https://doi.org/10.1136/bmjqs-2013-001987.
  5. Gilovic J, Griffin D, Kahneman D. Heuristics and biases: The psychology of intuitive judgment. Cambridge, UK: Cambridge University Press; 2002.
  6. Croskerry P. 50 Cognitive and affective biases. Available from: http://sjrhem.ca/wp-content/uploads/2015/11/CriticaThinking-Listof50-biases.pdf. [cited 2020 Oct 22]
  7. Coroskerry P. Bias: A normal operating characteristic of the diagnosing brain. Diagnosis. 2014;1:23-27. https://doi.org/10.1515/dx-2013-0028.
  8. Reilly JB, Ogdie AR, Von Feldt JM, et al. Teaching about how doctors think: A longitudinal curriculum in cognitive bias and diagnostic error for residents. BMJ Qual Saf. 2013;22:1044-1050. https://doi.org/10.1136/bmjqs-2013-001987.
  9. Redelmeier DA, Ng K. Approach to making the availability heuristic less available. BMJ Qual Saf. 2020;29(7):528-530. https://doi.org/10.1136/bmjqs-2020-010831.
  10. Croskerry P. Clinical cognition and diagnostic error: Applications of a dual process model of reasoning. Adv Health Sci Educ Theory Pract. 2009;14(Suppl 1):27-35. https://doi.org/10.1007/s10459-009-9182-2.
  11. Lakoff G, Johnson M. Philosophy in the flesh: The embodied mind and its challenge to Western thought. New York: Basic Books; 1999.
  12. Kralik JD, Xu ER, Knight EJ, et al. When less is more: Evolutionary origins of the affect heuristic. PLOS One. 2012;7(10):e46240. https://doi.org/10.1371/journal.pone.0046240.
  13. Santos LR, Rosati AG. The evolutionary roots of human decision making. Annu Rev Psychol. 2015;66:321-347. https://doi.org/10.1146/annurev-psych-010814-015310.
  14. Baumeister RF, Masicampo EJ, Vohs KD. Do conscious thoughts cause behavior? Annual Review of Psychology. 2011;62:331-361. https://doi.org/10.1146/annurev.psych.093008.131126.
  15. Evans JS, Frankish K. In two minds: Dual processes and beyond. Oxford, England: Oxford University Press; 2009.
  16. Andersson L, Eriksson J, Stillesjö S, et al. Neurocognitive processes underlying heuristic and normative probability judgments. Cognition. 2020;196:104153. https://doi.org/10.1016/j.cognition.2019.104153.
  17. Goel V, Dolan RJ. Explaining modulation of reasoning by belief. Cognition. 2003;87:B11-22. https://doi.org/10.1016/s0010-0277(02)00185-3.
  18. Graber ML, Franklin N, Gordon R. Diagnostic error in internal medicine. Arch Intern Med. 2005;165:1493-1499. https://doi.org/10.1001/archinte.165.13.1493.
  19. Kachalia A, Gandhi TK, Puopolo AL, et al. Missed and delayed diagnoses in the emergency department: A study of closed malpractice claims from 4 liability insurers. Ann Emerg Med. 2007;49:196-205. https://doi.org/10.1016/j.annemergmed.2006.06.035.
  20. O’Sullivan ED, Schofield SJ. Bias in Clinical Medicine. J R Coll Physicians Edinb. 2018;48:225-232. https://doi.org/10.4997/JRCPE.2018.306.
  21. Essex C. Ulysses syndrome. BMJ. 2005;330:1268. https://doi.org/10.1136/bmj.330.7502.1268.
  22. Hitzeman N, Cotton E. Incidentalomas: Initial management. Am Fam Physician. 2014;90(11):784-789.
  23. Graber MA, Bergus G, Dawson JD, et al. The effect of a patient’s psychiatric history on physicians’ estimation of the probability of disease. J Gen Intern Med. 2000;15:204-206. https://doi.org/10.1046/j.1525-1497.2000.04399.x.
  24. Blum M, Slade M, Boden D, Cabin H, Caulin-Glaser T. Examination of gender bias in the evaluation and treatment of angina pectoris by cardiologists. Am J Cardiol. 2004;93(6):765-767. https://doi.org/10.1016/
  25. j.amjcard.2003.12.007.
  26. Samulowitz A, Gremyr I, Eriksson E, Hensing G. “Brave Men” and “Emotional Women”: A theory-guided literature review on gender bias in health care and gendered norms towards patients with chronic pain. Pain Res Manag. 2018;2018:6358624. https://doi.org/10.1155/2018/6358624.
  27. Hayakawa SI. Language in thought and action. 4th ed. San Diego: Harcourt Brace Jovanovich; 1978.
  28. Lambe KA, O’Reilly G, Kelly BD, et al. Dual-process cognitive interventions to enhance diagnostic reasoning: a systematic review. BMJ Qual Saf. 2016;25:808-820. https://doi.org/10.1136/bmjqs-2015-004417.
  29. Feufel MA, Flach JM. Medical education should teach heuristics rather than train them away. Med Educ. 2019;53(4):334-344. https://doi.org/10.1111/medu.13789

Supplementary files

Supplementary Files
Action
1. JATS XML

Copyright (c) 2021 Graber M.A.

Creative Commons License
This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License.

СМИ зарегистрировано Федеральной службой по надзору в сфере связи, информационных технологий и массовых коммуникаций (Роскомнадзор).
Регистрационный номер и дата принятия решения о регистрации СМИ: серия ПИ № ФС 77 - 70763 от 21.08.2017 г.


This website uses cookies

You consent to our cookies if you continue to use our website.

About Cookies