Back to Journals » Advances in Medical Education and Practice » Volume 15

Overconfidence, Time-on-Task, and Medical Errors: Is There a Relationship?

Authors Al-Maghrabi M, Mamede S, Schmidt HG, Omair A , Al-Nasser S, Alharbi NS, Magzoub MEMA

Received 23 October 2023

Accepted for publication 18 February 2024

Published 22 February 2024 Volume 2024:15 Pages 133—140

DOI https://doi.org/10.2147/AMEP.S442689

Checked for plagiarism Yes

Review by Single anonymous peer review

Peer reviewer comments 2

Editor who approved publication: Dr Md Anwarul Azim Majumder



Mohsin Al-Maghrabi,1 Silvia Mamede,2 Henk G Schmidt,3 Aamir Omair,4– 6 Sami Al-Nasser,4– 6 Nouf Sulaiman Alharbi,4– 6 Mohi Eldin Mohammed Ali Magzoub7

1Department of Pediatrics, Imam Abdulrahman Alfaisal Hospital, Riyadh, Saudi Arabia; 2Institute of Medical Education Research Rotterdam, Erasmus University Rotterdam, Rotterdam, the Netherlands; 3Department of Psychology, Erasmus University Rotterdam, Rotterdam, the Netherlands; 4Department of Medical Education, College of Medicine, King Saud bin Abdulaziz University for Health Sciences, Riyadh, Saudi Arabia; 5King Abdullah International Medical Research Center, Riyadh, Saudi Arabia; 6Ministry of the National Guard - Health Affairs, Riyadh, Saudi Arabia; 7Department of Medical Education, United Arab Emirates University, Al Ain, United Arab Emirates

Correspondence: Nouf Sulaiman Alharbi, Department of Medical Education, College of Medicine, King Saud bin Abdulaziz University for Health Sciences, Riyadh, Saudi Arabia, Email [email protected]

Background: Literature suggest that physicians’ high level of confidence has a negative impact on medical decisions, and this may lead to medical errors. Experimental research is lacking; however, this study investigated the effects of high confidence on diagnostic accuracy.
Methods: Forty internal medicine residents from different hospitals in Saudi Arabia were divided randomly into two groups: A high-confidence group as an experimental and a low-confidence group acting as a control. Both groups solved each of eight written complex clinical vignettes. Before diagnosing these cases, the high-confidence group was led to believe that the task was easy, while the low-confidence group was presented with information from which it could deduce that the diagnostic task was difficult. Level of confidence, response time, and diagnostic accuracy were recorded.
Results: The participants in the high-confidence group had a significantly higher confidence level than those in the control group: 0.75 compared to 0.61 (maximum 1.00). However, neither time on task nor diagnostic accuracy significantly differed between the two groups.
Conclusion: In the literature, high confidence as one of common cognitive biases has a strong association with medical error. Even though the high-confidence group spent somewhat less time on the cases, suggesting potential premature decision-making, we failed to find differences in diagnostic accuracy. It is suggested that overconfidence should be studied as a personality trait rather than as a malleable characteristic.

Keywords: overconfidence, diagnostic accuracy, medical errors, premature closure

Background

In the past decades, medical errors have become a center of discussion as all health care systems attempted to achieve a high quality of care. This has been supported by a well-established system that has fostered the incident investigations of every adverse event and the formulation of policies to plug the staff or operation failures.1 The goal for quality improvement is building a safer health care environment in which attention is given to provide proper care to patients and families affected by medical errors.2,3

How is the health of the patients affected by medical errors? The Institute of Medicine (IOM) answered this question in 1999 and estimated that the annual incidence of deaths from medical errors in the United States was between 44,000 and 98,000 from 2403351 deaths.4 There was strong controversy whether these numbers underestimated or exaggerated the problem. However, this report highlighted the importance and consequences of medical errors. As a result, the US government supported the IOM recommendations as this problem was considered a challenge to the US health care systems. Many follow-up studies have been conducted periodically to observe and assess the improvement and progression of patient safety.5,6 Despite measures taken, the incidence of medical errors has been increasing. For instance, in 2013, the estimated incidence jumped to 210,000–400,000 deaths a year associated with medical errors among hospitalized patients in the US,7 making medical errors the third leading cause of death in the US after heart disease (635,000/year) and cancer (598,000/year). Beyond the loss of lives and the catastrophic effects of patients’ morbidity, these medical errors add a great burden on the economy of approximately $735 to $980 billion, according to the IOM.8

To comprehend the impact of medical errors in patient care, the whole process of clinical reasoning and problem solving is better understood.9 For instance, one important area where errors have occurred frequently is the area of “diagnostic uncertainty”. Physicians’ work is partly based on personal interpretations and estimating disease probabilities.10 In studies conducted in the US and the UK, the incidence of diagnostic failure was highest in family medicine (FM), emergency medicine (EM), and internal medicine (IM), disciplines in which diagnosis is indeed subject to higher degrees of diagnostic uncertainty.11

A dual-systems theory was developed by Kahneman and Tversky to explain the decision-making process under uncertainty.12–15 System 1 is a fast and automated approach. It is unconscious and effortless. This is done through “pattern recognition”, in which quick diagnosis is possible because a physician has encountered similar cases in the past. On the other hand, System 2 is a non-programmed, conscious approach. In contrast to System 1, it is slow, analytical, and effortful. This system is activated when pattern recognition fails. Accordingly, an appropriate amount of data is collected, and a systematic case analysis is performed. Kahneman12 strongly suggests that System 1 thinking is a source of error, particularly through falling prey to cognitive biases. Over forty cognitive biases have been described in clinical medicine.12.It is estimated that 75% of medical errors in internal medicine practice are attributed to cognitive biases.16

Kahneman or Croskerry reported that error is often the result of neglect, time pressure, or lack of appropriate knowledge,17 it is certainly likely that cognitive biases play an important part in the emergence of errors. Among the most clearly recognized biases is overconfidence.18 It is considered a major factor that leads to diagnostic error.19 Premature diagnostic closure is defined as “narrowing the choice of diagnostic hypothesis too early”.18 As a result, the most likely diagnosis is never considered.20,21 Premature closure occurs when a physician either gets attached to data collected early from miscellaneous sources of information or does not gather enough data to reach the correct diagnosis.11 As a result, the proposed diagnosis is not verified. Overconfidence is a form of excess and unfounded confidence in one’s abilities. It is a risky personal trait in medicine, especially in fields where patient life is in a critical situation, and prognosis depends on the right medical decisions such as emergency medicine. It occurs when physicians think they perform better than they do. Without relying on enough patient data, the physician’s opinion becomes the base for the diagnosis. Generally, overconfident people have a strong belief that their contribution usually has a positive outcome (regardless of whether it does). Rather than relying on reasonable and relevant information, overconfident people trust their hunches.22

Although studies suggest that physicians’ high confidence can lead to medical errors through premature closure,23,24 experimental research is lacking to support this claim. Therefore, in the experiment described below we influenced the amount of confidence in their decisions that physicians have and studied its effect on diagnostic accuracy and time on task in a subsequent diagnostic assignment. We expected physicians who displayed a higher level of confidence to spend less time on the task and be less diagnostically accurate.

Methods

Overview

For the purpose of investigating the effects of high confidence on diagnostic accuracy, the current study was designed as a randomized controlled experiment conducted over a four-week period during October–November 2019. The independent variable was high confidence, and the dependent variables were response time and diagnostic accuracy.

Participants

Forty third-year internal medicine residents participated in this study. All internal medicine residents at different main hospitals in the Riyadh region were invited to participate in this study. Residents were informed that their participation is not mandatory and that their refusal to participate will not affect their learning experience in any way. Those who accepted to be part of the study were randomly assigned to either the experimental (high confidence) group or the control (low-confidence) group. The participants included 27 men and 13 women. Their mean (SD) age was 31.8 (4.1) years, and their mean clinical experience was 7.9 (3.9) years. There were 20 participants in each group. After randomization, no significant difference was found in terms of sex, age, or years of training. Participation was voluntary, and informed consent was obtained from residents who agreed to participate. The specific purpose of the study was not announced to the residents because this would probably affect the validity of the results. After the collection of data, participants were informed about the main aim of the study.

The internal medicine residency program in Saudi Arabia is a four-year program that covers general internal medicine and subspecialties. It has two levels: junior level (first and second year) and senior-level (third and fourth year) residents. Choosing internal medicine residents for the current study is justified given their exposure to diverse and complicated chronic cases. These residents often deal with overlapping chronic medical conditions; thus, they are more subject to the demands of achieving a high level of diagnostic accuracy. Moreover, they are not consistently required to function under the immediate pressure of handling emergencies, a factor that can significantly influence one’s confidence and diagnostic accuracy. This study is particularly pertinent as it provides insights into how confidence levels may impact their diagnostic decision-making in less urgent but equally critical situations.

Materials

In total, eight written clinical vignettes were used for this study with the following diagnoses: stomach cancer, vitamin B12 deficiency, pulmonary thromboembolism, celiac disease, acute viral pericarditis, acute myeloid leukemia, acute bacterial endocarditis, and sarcoidosis. The cases were written by experts in internal medicine. Each case was presented in English and included a patient’s medical history, relevant physical examination findings, and laboratory test results. See Box 1 for an example case.

Box 1 Example of a Medical Case

Procedure

Both groups were informed that the main purpose of the study was to identify factors affecting problem solving. A booklet included cases and short-answer questions given to the two groups. The high-confidence group was informed that the cases in the booklet were given to fourth-year medical students to assess the difficulty of the cases, and we found that the students correctly solved 6 to 7 out of 8 cases. This was done to ensure that the residents feel confident solving the cases. The low-confidence group was informed that those cases were alternatively given to two professors of internal medicine with more than 15 years of experience, and the participants got only 3 out of the 8 cases correct. It was done to ensure that the low-confidence group would experience a lower level of confidence. Inducing low confidence in this group does not imply that we expect the group to be less confident or high confident; rather, it is a planned experiment designed to investigate the consequences of low and high confidence on diagnosis accuracy. After the information of the purported difficulty level of the cases was given, the participants were asked to indicate how many of the cases they expected to diagnose correctly.

Subsequently, the participants were given an example case to practice before solving the eight actual cases. Then, the participants typed in their diagnosis for each individual case. Response time was recorded in minutes and seconds for each case. After completing the task, the candidates were asked four short-answer questions: 1. How suitable do you think these cases are for helping medical students to learn to diagnose cases? 2. How many cases out of 8 you think you have diagnosed correctly? 3. How complex were the cases on average in your opinion? 4. How often have you seen in practice on average cases such as the ones presented? The questions were answered using a scale from 0 to 10, with exception of the question referring to the number of cases diagnosed correctly.

Analysis

Cases were scored for diagnostic accuracy using the following rule: 0 = incorrect answer, 1 = partially correct, and 2 = correct. A diagnosis was considered correct if it included the main component of the true diagnosis (eg, “bacterial endocarditis” in the case of acute bacterial endocarditis). A diagnosis was considered partially correct when part of the diagnosis was mentioned, but the main correct diagnosis was missing (eg, leukemia in the case of acute myeloid leukemia). A diagnosis was considered incorrect when it did not match the main diagnosis or even a part of it (eg, “myocardial infarction” in the case of acute viral pericarditis).

For each participant, the mean score of confidence level, diagnostic accuracy, and mean response time was calculated for all the eight cases. The group’s mean (average) confidence level was obtained by adding the sum of participants’ ratings on different scenarios and dividing it by the total number of participants. The diagnostic accuracy was estimated as a % by calculating the correct diagnoses and dividing them by the total number of diagnoses done. This percentage shows the group’s mean diagnostic accuracy. We measured the amount of time each participant took to react to each diagnosis. The total of all individual reply times was then divided by the total number of participants to obtain the group’s average or mean response time. All differences between groups were assessed through t-tests. The p-value <0.05 was considered significant.

Results

There were a total of 40 participants in the study who were divided equally into the two groups of High confidence (n = 20) and Low confidence (n = 20) as the independent variables. The results of the independent samples t-test revealed that the high confidence group had a significantly higher confidence level compared with the control group: 0.75 compared to 0.61 with a p-value of 0.03. This implies that our experimental manipulation had succeeded. The dependent variables for the study included the Response Time and Diagnostic. See Table 1.

Table 1 Mean Differences with Standard Deviations (SD) Between the High-Confidence and Low-Confidence Groups on Confidence Level, Diagnostic Accuracy, and Response Time

Accuracy

Response Time

There was no significant difference in response time, although the high-confidence group needed on average 12% less time to arrive at a diagnosis.

Diagnostic Accuracy

The results showed that the diagnostic accuracy was almost similar in both groups. The mean score in diagnostic accuracy was equal to 0.65 indicating that the residents were able to respond correctly to 5 out of 8 cases. This shows that the accurate diagnosis can be established by initiative from the less confident group, whereas the overconfident group will end the case quickly, resulting in a diagnostic error. Interestingly, the confidence instilled by telling the participants that the cases to be diagnosed would be either rather easy or quite difficult seems to have spilled over to self-assessment of performance and the frequency by which the cases are said to be seen by both groups. Both groups misjudge the number of cases successfully diagnosed, but it is the high-confidence group that significantly overestimates its own performance. Surprisingly, this group also indicates to have encountered the cases in professional practice significantly more frequently, a statement that is unlikely true because the groups were randomly assigned to the conditions of the experiment. A high confidence seems to alter the perception of reality, even in the light of actual performance. See Table 2.

Table 2 Mean Differences with Standard Deviations (SD) Between the High-Confidence and Low-Confidence Groups on the Post Experiment Variables

Discussion

This study represents a first attempt to experimentally study the effect of overconfidence on premature closure, as expressed by shorter processing times of clinical vignettes, and diagnostic accuracy. To that end, we influenced the amount of confidence that third-year residents in internal medicine had in future diagnostic decisions by presenting cases to be diagnosed as either rather easy or particularly difficult. We studied its effect on confidence, diagnostic accuracy, and time on task in a subsequent diagnostic task. We expected physicians who displayed a higher level of confidence to spend less time on the task and be less diagnostically accurate.

We were successful in inducing appropriate levels of confidence in our randomly assigned groups of third-year residents. The high-confidence group even after the diagnostic task rated its own performance significantly higher than its actual performance warranted. Although high-confidence participants spent somewhat less time on the diagnostic task, however, we failed to find effects of this experimentally induced effect on time on task and diagnostic accuracy. This indicates that the overconfident group is more likely to make a quick, inaccurate diagnosis, and prematurely close the contact with patients, whereas the control group may spend more time ruling out rare diagnoses, making them less susceptible to premature closure. This would necessitate a clinical picture that corresponds to both a simpler, incorrect diagnosis and a more complex accurate diagnosis. The results are consistent to the findings of an experiment that failed to display relationships between overconfidence, premature closure, and diagnostic accuracy, as suggested by the extant literature.23

The question is why this is so. One possible answer is that such relationship does not exist. Indeed, there is yet no convincing empirical evidence supporting the idea of a causal relationship between overconfidence, premature closure (as expressed by lesser processing time), and diagnostic error, but theoretical analyses and common sense make it likely that overconfident physicians do not adequately appreciate the full findings presented by a patient; they are unable to “swap horses midstream”,25 because they think that their initial interpretations must be correct. This inability would result in premature closure followed by incorrect diagnoses.23 On the other hand, one to the few studies in which the effect of premature closure was investigated directly failed to find effects on diagnostic accuracy.26 Even a comprehensive review of the relationship between overconfidence and diagnostic error was unable to unearth direct evidence of the causal relationship between overconfidence, premature closure, and diagnostic accuracy.23

A second answer to the question why we failed to find a causal relationship between overconfidence on the one hand and time on task and diagnostic accuracy on the other may be that overconfidence is a personality trait of some physicians rather than an attribute that can be induced. Cassam27 argues that overconfidence can be considered an “epistemic vice”. The author is of the opinion that bettering physician self-knowledge can reduce diagnostic errors brought on by overconfidence is proven to be dubious. Overconfidence is a cognitive bias that is “stealthy” in the sense that it prevents its own detection. If this is true, if overconfidence is an enduring aspect of one’s personality, an experiment in which confidence in one’s own performance is manipulated by the experimenter may not really be appropriate.

Although the study results are favorable, there are certain limitations that limit the generalizability of the findings that must be addressed in future investigations. First, interns who volunteered to participate were randomly allocated to one of two groups. As a result, the participants may have been more motivated than those who did not participate, however this does not ensure a precise representation of the population. Second, there is the possibility of a lack of ecological validity. The external validity including ecological validity of studies such as the one discussed here is a challenge. A case in point is the use of vignettes rather than “real” patients. Although vignettes in general have been demonstrated to have sufficient external validity, they may be too easy for participants in an overconfidence study, because they present all findings necessary to arrive at an appropriate diagnosis. To ensure internal validity, an experimental approach compels researchers to simplify the event under study. The internal validity ascribed to an experimental approach ensures the accuracy of any causal inferences made from the findings. Premature closure may express itself particularly in situations where the physician decides how much information he or she needs. In such cases, experiments are not the way to go. The third is the failure to account for individual differences in cognitive biases, which may lead to the mistaken belief that all experts are equally susceptible to biases. To overcome this issue, we propose improving or developing reliable, specific measurement methods for cognitive biases. In addition, examine retrospectively the amount of diagnostic error produced as a result of premature closure in each of their own practices. For identifying errors, criteria would need to be defined. This might reduce a relative risk between overconfidence and diagnostic error.

Conclusion

This study shows that cognitive bias can be a source of medical errors. Current research has demonstrated that when health care professionals display high levels of confidence combined with insufficient time spent on performing tasks during patient diagnosis, can increase the risk of medical errors in many ways including premature closure and ignoring the complexities of a patient’s condition. Moreover, the results revealed that the high-confidence group spent less time diagnosing the presented cases, signifying diagnostic errors made due to premature closure. Yet, no differences appeared in diagnostic accuracy between them and the low-confidence (control) group. Further studies should consider confidence as a personal trait instead of an adaptable one as well as to teach doctors how to overcome error-related tendencies.

Data Sharing Statement

The datasets utilized to reach the findings of this study are not publicly available because the ethical approval was obtained from KAIMRC on the basis that only the research team who are involved in the study can access it. The retention period is five years from publication. Supporting documents are available upon request to the corresponding author.

Ethical Approval and Consent to Participate

This study involved human participants; thus, it was conducted in compliance with relevant guidelines and regulations in the Helsinki Declaration. Ethical clearance was attained from King Abdullah International Research Center (KAIMRC), King Saud Bin Abdulaziz University for Health Sciences (RYD-19-H19812-191375). Moreover, a written informed consent was obtained from all the participants prior to their involvement with clear explanation of their rights to not to participate or withdraw their participation at any time. Confidentiality of the information was considered by hiding personal identifiers during analysis and publications of results.

Acknowledgments

The authors thank the residents who dedicated their free limited time to take part in the research. In addition, the authors would like to thank Prof. Sajida Agha for her comments and efforts towards improving the final draft of this manuscript.

Author Contributions

All authors made a significant contribution to the work reported, whether that is in the conception, study design, execution, acquisition of data, analysis and interpretation, or in all these areas; took part in drafting, revising or critically reviewing the article; gave final approval of the version to be published; have agreed on the journal to which the article has been submitted; and agree to be accountable for all aspects of the work.

Funding

This research did not receive any specific grant from any agency or sector.

Disclosure

The authors declare that they have no competing interests in this work.

References

1. Wu AW, Lipshutz AK, Pronovost PJ. Effectiveness and efficiency of root cause analysis in medicine. JAMA. 2008;299(6):685–687. doi:10.1001/jama.299.6.685

2. Gallagher TH, Studdert D, Levinson W. Disclosing harmful medical errors to patients. N Engl J Med. 2007;356(26):2713–2719. doi:10.1056/NEJMra070568

3. Conway J, Federico F, Stewart K, Campbell M. Respectful Management of Serious Clinical Adverse Events. In: IHI Innovation Series White Paper. 2nd ed. Cambridge, MA: Institute for Healthcare Improvement; 2011.

4. Donaldson MS, Corrigan JM, Kohn LT. To Err is Human: Building a Safer Health System. USA: National Academies Press; 2000.

5. Leape LL, Berwick DM. Five years after to err is human: what have we learned? JAMA. 2005;293(19):2384–2390. doi:10.1001/jama.293.19.2384

6. Bates DW, Singh H. Two decades since to err is human: an assessment of progress and emerging priorities in patient safety. Health Aff (Millwood). 2018;37(11):1736–1743. doi:10.1377/hlthaff.2018.0738

7. Makary MA, Daniel M. Medical error—the third leading cause of death in the US. BMJ. 2016;353. doi:10.1136/bmj.i2139

8. Andel C, Davidow SL, Hollander M, Moreno DA. The economics of health care quality and medical errors. J Health Care Finance. 2012;39(1):39.

9. Neuspiel DR, Schuman AJ. Prevent medical errors in your practice. Contemp Pediatr. 2018;35(7):31–34.

10. Aitken C, Mavridis D. Reasoning under uncertainty. Evid Based Ment Health. 2019;22(1):44–48. doi:10.1136/ebmental-2018-300074

11. Graber M. Diagnostic errors in medicine: a case of neglect. Jt Comm J Qual Patient Saf. 2005;31(2):106–113. doi:10.1016/s1553-7250(05)31015-4

12. Kahneman D. Thinking, Fast and Slow. USA: Macmillan; 2011.

13. Tversky A, Kahneman D. Judgment under uncertainty: heuristics and biases. Judgment and decision making: an interdisciplinary reader; 1986:38–55.

14. Croskerry P. A universal model of diagnostic reasoning. Acad Med. 2009;84(8):1022–1028. doi:10.1097/ACM.0b013e3181ace703

15. Tay SW, Ryan P, Ryan CA. Systems 1 and 2 thinking processes and cognitive reflection testing in medical students. Can Med Educ J. 2016;7(2):e97. doi:10.36834/cmej.36777

16. Croskerry P. Achieving quality in clinical decision making: cognitive strategies and detection of bias. Acad Emerg Med. 2002;9(11):1184–1204. doi:10.1197/aemj.9.11.1184

17. Norman GR, Monteiro SD, Sherbino J, Ilgen JS, Schmidt HG, Mamede S. The causes of errors in clinical reasoning: cognitive biases, knowledge deficits, and dual process thinking. Acad Med. 2017;92(1):23–30. doi:10.1097/ACM.0000000000001421

18. Eva KW, Cunnington JP. The difficulty with experience: does practice increase susceptibility to premature closure? J Contin Educ Health Prof. 2006;26(3):192–198. doi:10.1002/chp.69

19. Croskerry P, Norman G. Overconfidence in clinical decision making. Am J Med. 2008;121(5):S24–S29. doi:10.1016/j.amjmed.2008.02.001

20. Kruglanski AW, Webster DM. Motivated closing of the mind: ”Seizing” and ”freezing”. Psychol Rev. 1996;103(2):263. doi:10.1037/0033-295X.103.2.263

21. Blissett S, Morrison D, McCarty D, Sibbald M. Should learners reason one step at a time? A randomized trial of two diagnostic scheme designs. Medical Education. 2017;51(4):432–441. doi:10.1111/medu.13221

22. West RF, Stanovich KE. The domain specificity and generality of overconfidence: individual differences in performance estimation bias. Psychonomic Bull Rev. 1997;4(3):387–392. doi:10.3758/BF03210798

23. Berner ES, Graber ML. Overconfidence as a cause of diagnostic error in medicine. Am J Med. 2008;121(5):S2–S23. doi:10.1016/j.amjmed.2008.01.001

24. Ely JW, Levinson W, Elder NC, Mainous AG, Vinson DC. Perceived causes of family physicians’ errors. J Fam Pract. 1995;40(4):337–444.

25. Eva KW, Link CL, Lutfey KE, McKinlay JB. Swapping horses midstream: factors related to physicians’ changing their minds about a diagnosis. Acad Med. 2010;85(7):1112–1117. doi:10.1097/ACM.0b013e3181e16103

26. Krupat E, Wormwood J, Schwartzstein RM, Richards JB. Avoiding premature closure and reaching diagnostic accuracy: some key predictive factors. Medical Education. 2017;51(11):1127–1137. doi:10.1111/medu.13382

27. Cassam Q. Diagnostic error, overconfidence and self-knowledge. Palgrave Commun. 2017;3(1):17025. doi:10.1057/palcomms.2017.25

Creative Commons License © 2024 The Author(s). This work is published and licensed by Dove Medical Press Limited. The full terms of this license are available at https://www.dovepress.com/terms.php and incorporate the Creative Commons Attribution - Non Commercial (unported, v3.0) License. By accessing the work you hereby accept the Terms. Non-commercial uses of the work are permitted without any further permission from Dove Medical Press Limited, provided the work is properly attributed. For permission for commercial use of this work, please see paragraphs 4.2 and 5 of our Terms.