Jump to ContentJump to Main Navigation
Show Summary Details
More options …


Official Journal of the Society to Improve Diagnosis in Medicine (SIDM)

Editor-in-Chief: Graber, Mark L. / Plebani, Mario

Ed. by Argy, Nicolas / Epner, Paul L. / Lippi, Giuseppe / Singhal, Geeta / McDonald, Kathryn / Singh, Hardeep / Newman-Toker, David

Editorial Board: Basso , Daniela / Crock, Carmel / Croskerry, Pat / Dhaliwal, Gurpreet / Ely, John / Giannitsis, Evangelos / Katus, Hugo A. / Laposata, Michael / Lyratzopoulos, Yoryos / Maude, Jason / Sittig, Dean F. / Sonntag, Oswald / Zwaan, Laura

CiteScore 2018: 0.69

SCImago Journal Rank (SJR) 2018: 0.359
Source Normalized Impact per Paper (SNIP) 2018: 0.424

See all formats and pricing
More options …

Pediatric faculty knowledge and comfort discussing diagnostic errors: a pilot survey to understand barriers to an educational program

Joseph A. Grubenhoff
  • Corresponding author
  • University of Colorado Denver School of Medicine, Aurora, CO, USA
  • Children’s Hospital Colorado, Aurora, CO, USA
  • Email
  • Other articles by this author:
  • De Gruyter OnlineGoogle Scholar
/ Sonja I. Ziniel / Lalit Bajaj / Daniel Hyman
Published Online: 2019-02-13 | DOI: https://doi.org/10.1515/dx-2018-0056



Improving Diagnosis in Healthcare calls for improved training in diagnostic reasoning and establishing non-judgmental forums to learn from diagnostic errors arising from heuristic-driven reasoning. Little is known about pediatric providers’ familiarity with heuristics or the culture surrounding forums where diagnostic errors are discussed. This study aimed to describe pediatric providers’ familiarity with common heuristics and perceptions surrounding public discussions of diagnostic errors.


We surveyed pediatric providers at a university-affiliated children’s hospital. The survey asked participants to identify common heuristics used during clinical reasoning (five definitions; four exemplar clinical vignettes). Participants answered questions regarding comfort publicly discussing their own diagnostic errors and barriers to sharing them.


Seventy (30.6% response rate) faculty completed the survey. The mean number of correctly selected heuristics was 1.60/5 [standard deviation (SD)=1.13] and 1.01/4 (SD=1.06) for the definitions and vignettes, respectively. A low but significant correlation existed between correctly identifying a definition and selecting the correct heuristic in vignettes (Spearman’s ρ=0.27, p=0.02). Clinicians were significantly less likely to be “pretty” or “very” comfortable discussing diagnostic errors in public vs. private conversations (28.3% vs. 74.3%, p<0.01). The most frequently cited barriers to discussing errors were loss of reputation (62.9%) and fear of knowledge-base (58.6%) or decision-making (57.1%) being judged.


Pediatric providers demonstrated limited familiarity with common heuristics leading to diagnostic error. Greater years in practice is associated with more comfort discussing diagnostic errors, but negative peer and personal perceptions of diagnostic performance are common barriers to discussing errors publicly.

Keywords: cognitive bias; communication; diagnostic errors; heuristics; pediatrics


The 2015 National Academy of Medicine’s report, Improving Diagnosis in Healthcare [1], called attention to diagnostic errors and their contribution to patient harm. Among the major themes highlighted in the report is the need to improve training in clinical reasoning among diagnosticians. Theoretically, diagnosticians who are aware of potential flaws in human decision-making might be better prepared to recognize and avoid pitfalls in their diagnostic reasoning, thereby reducing patient harm [2].

Diagnostic reasoning can be understood using the dual process theory of decision-making, a general model describing how humans make decisions [3]. Dual process theory suggests two modes of reasoning – System 1 and System 2 – are used to make decisions. Deliberate, conscious, time-consuming and effortful judgments characterize System 2 reasoning. System 1 reasoning, in contrast, is primarily intuitive, subconscious, fast and relatively effortless [4]. System 1 reasoning relies on heuristics (often called cognitive biases). These mental shortcuts allow experienced clinicians to arrive at decisions quickly with a reasonable degree of certainty in familiar situations. Using heuristics to make decisions is neither avoidable nor an inherently inappropriate method for clinical reasoning [5]. While experienced clinicians may recognize their reliance on System 1 reasoning in clinical practice, the theories underlying decision-making are not formally taught in most medical school or residency curricula [6].

In clinical vignettes, cognitive errors contribute to overall diagnostic errors in as many as 37–77% of cases [7] leading some authors to suggest that understanding the etiology of these errors is paramount to reducing them [2]. Others have argued that knowing about heuristics may not necessarily allow clinicians to avoid diagnostic errors [8]. Consequently, some authors have suggested novel interventions (e.g. diagnostic checklists, differential diagnosis generators) to mitigate the risks posed by heuristic-driven reasoning [9], [10]. However, the deliberate use of such tools requires awareness of situations at high risk for cognitive errors. It is not clear that diagnosticians currently possess this awareness, but some understanding of the role of heuristics in diagnostic reasoning may be helpful in reducing harm from diagnostic errors.

Discussion of diagnostic errors during morbidity and mortality (M&M) conferences represents an approach to reducing harm through structured analysis of the errors including identification of cognitive factors [11]. Katz and Detsky suggest that M&M is an ideal venue to introduce discussion of cognitive errors [12]. Fruitful discussions of diagnostic error require an environment that is open to honest discussion focused on patient safety rather than harmful criticism of individuals [13]. Yet, discussions of harm arising from flawed diagnostic reasoning are uncomfortable for clinicians [14], [15]. For example, loss of reputation and fear of incurring blame are two significant components that influence clinicians’ comfort discussing their own errors with colleagues [16]. Recently, the promotion of both psychological safety (including a non-blaming atmosphere) and transparency when reporting and reviewing medical errors in public venues has received significant attention [17]. While some institutions have deliberately revamped their M&M conferences to promote psychological safety [18], there remains an undercurrent of blame culture in medicine [19].

We aimed to better understand two obstacles to improving training in clinical reasoning for clinicians. First, little is known about how familiar clinicians are with heuristics operating during the diagnostic process. Second, it is not clear whether current culture promotes psychological safety when discussing diagnostic errors in public venues such as M&M. In anticipation of developing a curriculum on diagnostic errors at our institution, we wanted to better characterize these obstacles among a subset of pediatric faculty. We hypothesized that participants’ awareness of heuristic definitions would positively correlate with an ability to recognize heuristics operating in clinical vignettes. We also sought to describe participants’ comfort discussing their own diagnostic errors in public venues and the perceived barriers to such discussions.

Materials and methods

Study design and participants

An anonymous web-based survey was distributed by e-mail invitation to all faculty members of the Divisions of Critical Care, Emergency Medicine and Hospital Medicine within the Department of Pediatrics at our institution. These divisions were selected because they have mature tiered case review programs amenable to discussing cognitive errors alongside systems issues. Physicians and advanced-practice providers (APPs; i.e. nurse practitioners and physician assistants) were eligible. The survey was available from January 31 to March 31, 2017. E-mail invitations were distributed at the beginning of the enrollment period with e-mail reminders sent at 3 and 6 weeks after the survey opened. Full- and part-time clinical faculty were eligible to participate; we did not enroll trainees. The Local Institutional Review Board deemed the study exempt from review.

Survey instrument

The survey consisted of three sections. The first two sections asked participants to select definitions of common heuristics (base rate neglect, outcome bias, search satisfying, anchoring, gambler’s fallacy) and to name heuristics operating in four general pediatric clinical vignettes (premature closure, posterior probability error, visceral bias, framing effect). For each definition and vignette, the correct answer, three foils, and an “I don’t know” option were provided. No correct answer in the definition questions was used as a correct answer for the vignettes. We used definitions for heuristics as described by Croskerry [20]. We selected heuristics that commonly contribute to diagnostic errors discussed in M&M in our own clinical experience; data on the true frequency of diagnostic errors attributable to specific heuristics are lacking [7]. To ensure that the vignettes accurately demonstrated the heuristic of interest, two investigators (JAG, SIZ) held two focus groups with representatives from each division prior to finalizing the survey instrument; focus group members did not participate in the survey.

The third section of the survey assessed participants’ comfort discussing diagnostic errors and diagnostic uncertainty with trusted colleagues in private and in public venues using a five-point Likert scale. Additionally, we asked participants to identify the three most significant barriers to discussing their own diagnostic errors in public.

We pilot tested the survey with general pediatric residents, chief residents and fellows in the specialties to whom the survey was ultimately distributed. The survey was administered using REDCap [21] and took less than an average of 10 min to complete.


Fisher’s exact and chi-squared (χ2) tests were used to compare survey completion by age, gender, years in practice and provider type. Descriptive statistics were used to evaluate demographic characteristics. Mean number of correct responses with standard deviation (SD) were calculated for the definition and clinical vignette items. “I don’t know” responses were categorized as incorrect. Tetrachoric and Spearman rank correlation coefficients were used to explore associations between the total number of correctly answered heuristic definitions relative to the total number of correctly identified heuristics in clinical vignettes. Fisher’s exact and χ2 tests were used to assess differences in the attitudinal items by years in practice (dichotomized to <6 years and ≥6 years), gender and provider type. Differences related to age were compared using the Kruskal-Wallis test. For items regarding comfort discussing diagnostic errors, we dichotomized the responses into “comfortable” (including ‘pretty’ and ‘very’ comfortable) and “uncomfortable” (‘not at all’, ‘a little’ and ‘somewhat’ comfortable) as efforts to improve comfort surrounding such discussions would likely target those who are less comfortable or at least ambivalent about such discussions.


There were 229 eligible faculty from the three divisions surveyed. Overall, 88 (38.4%) faculty members accessed the survey and provided demographic information. Eight surveys were excluded from all further analyses because participants only provided demographic information; an additional 10 were excluded because they were only partially complete. There were no significant differences with respect to age, gender, years in practice or provider type between those that provided complete information compared to those with truncated responses. The overall response rate was 30.6% [22]. The response rate was significantly lower for APPs (9.0%) compared to physicians (44.3%) (p<0.001). Table 1 shows the demographics of participants. Only 15.9% endorsed receiving formal training in clinical reasoning beyond training received in medical school or, for APPs, during graduate training. All participants endorsing formal training were less than 50 years old. Most participants (84.3%) reported that they had been involved in a diagnostic error resulting in patient harm during their career.

Table 1:

Demographic characteristics of respondents with completed surveys.

Knowledge of heuristic definitions

No participant answered all five heuristic definition questions correctly; the mean number of correctly selected heuristic definitions was 1.60/5 (SD=1.13) (Figure 1). Reporting formal training in clinical reasoning did not significantly increase the likelihood of correctly selecting the heuristic term for any definition (p-values range from p=0.05 to 1.00). Age, gender, years in practice and provider type were not significantly associated with answer accuracy.

Knowledge of heuristic definitions. Solid bars indicate the proportion of participants answering correctly. Stippled bars indicate the proportion of participants answering incorrectly. Hashed bars indicate the proportion of participants responding “I don’t know”.
Figure 1:

Knowledge of heuristic definitions.

Solid bars indicate the proportion of participants answering correctly. Stippled bars indicate the proportion of participants answering incorrectly. Hashed bars indicate the proportion of participants responding “I don’t know”.

Identifying heuristics in vignettes

Only two participants selected the correct heuristic for all four vignettes; the mean number of correctly selected heuristics among the vignettes was 1.01/4 (SD=1.06) (Figure 2). Reporting formal training in clinical reasoning did not significantly increase the likelihood of correctly identifying the heuristic for any vignette (p=0.72–1.00). Age, gender, years in practice and provider type were not significantly associated with answer accuracy for any of the vignettes.

Identifying heuristics in clinical vignettes. Solid bars indicate the proportion of participants answering correctly. Stippled bars indicate the proportion of participants answering incorrectly. Hashed bars indicate the proportion of participants responding “I don’t know”.
Figure 2:

Identifying heuristics in clinical vignettes.

Solid bars indicate the proportion of participants answering correctly. Stippled bars indicate the proportion of participants answering incorrectly. Hashed bars indicate the proportion of participants responding “I don’t know”.

There was a low but significant correlation between scores on the definition and vignette portions of the survey (Spearman’s ρ=0.27, p=0.02).

Comfort surrounding discussion of diagnostic errors

The proportion of participants reporting being “pretty” or “very” comfortable discussing diagnostic errors was significantly lower for public compared to private discussions: 28.6% vs. 74.3%, respectively (p<0.01). The proportion of participants reporting being “pretty” or “very” comfortable acknowledging diagnostic uncertainty to themselves was not significantly different than acknowledging diagnostic uncertainty to colleagues: 82.9% vs. 75.7%, respectively (p=0.06). Age, gender, formal training and provider type were not significantly associated with comfort levels. Years in practice was significantly associated with comfort discussing errors in public and acknowledging uncertainty to colleagues (Table 2).

Table 2:

Provider comfort discussing diagnostic errors in private and public settings by years in practice.

Participants identified the three most significant barriers to discussing their own diagnostic errors in public venues (Table 3). Over 50% of the participants indicated that the most significant barriers were their reputation as a provider being at stake (62.9%), and their knowledge-base (58.6%) and decision-making being judged (57.1%).

Table 3:

Barriers to discussing one’s own diagnostic errors in public.


Pediatric faculty in this study demonstrated limited ability to identify heuristic definitions. Similarly, they were frequently unable to correctly name heuristics employed in clinical vignettes. We did find evidence to support our hypothesis that knowledge of heuristic definitions correlates with being able to identify heuristics in clinical vignettes, although the correlation is low. Pediatric faculty expressed significant discomfort discussing their own errors publicly in venues such as M&M despite being generally comfortable discussing diagnostic errors in private and acknowledging diagnostic uncertainty to themselves and their peers. This discomfort appears to arise from concerns about negative peer perceptions regarding clinical reputation and the diagnosticians’ clinical reasoning skills and knowledge.

Clinicians are central to efforts to reduce patient harm resulting from diagnostic errors. Success in these efforts may be dependent on two related conditions: (1) the diagnostician’s ability to recognize how diagnostic reasoning might be subconsciously influenced through reliance on heuristics; and (2) the willingness to discuss cognitive errors in public in order to identify solutions to avoid them in the future.

Pediatricians can appreciate their own diagnostic errors in hindsight. Singh and colleagues demonstrated that 54% of surveyed pediatricians admit to making 1–2 diagnostic errors per month, and 45% reported making 1–2 diagnostic errors per year that result in patient harm [23]. However, reducing patient harm due to diagnostic error requires that diagnosticians recognize and avoid heuristic failures prior to committing them.

There are few studies investigating strategies to avoid, or mitigate the effects of, heuristics and results are mixed [24]. Most studies have included trainees and have focused on clinical vignettes rather than real patient encounters. While some studies have demonstrated small but significant improvements in diagnostic accuracy when residents were instructed to rely on slow deliberate System 2 reasoning (through structured reflection) [25], [26], one study did not demonstrate a benefit with this approach [27]. Sherbino and colleagues provided instruction on two common heuristics (search satisfying and availability) to senior medical students followed by training on cognitive forcing strategies to avoid heuristic errors. Despite immediate testing after the educational intervention, fewer than 50% of students successfully “debiased” themselves [28]. Little work has been done to evaluate practicing clinicians’ familiarity with and ability to recognize heuristics. We conducted this study to begin to address that gap.

Participants in this study were unable to correctly name several defined heuristics or identify heuristics operating in several clinical vignettes. We found that formal training in diagnostic reasoning did not improve participants’ ability to identify heuristics. We did not collect information regarding the timing or content of this training so it is difficult to know whether more robust or more proximate education would have yielded improved accuracy. However, our results align with another study in which experts in diagnostic error were asked to identify heuristics (cognitive biases) perceived to be operating in eight clinical vignettes [29]. Participants demonstrated no agreement when naming heuristics represented by the vignettes (κ=0.0–0.6).

The goal of introducing diagnosticians to concepts including heuristics and dual process theory is not to teach new vocabulary. Rather, improving diagnosticians’ awareness of these concepts aims to improve recognition of the risk of diagnostic errors arising from System 1 reasoning. Nemes-Walsh and Lee evaluated the effect of an educational intervention introducing topics including dual process theory and cognitive bias pertaining to diagnostic errors [30]. This study demonstrated a slight improvement in responses on the National Patient Safety Foundation Reducing Diagnostic Errors: Strategies for Solutions Quiz and increased willingness to take a “diagnostic timeout” following the intervention. There was no indication, however, that fewer diagnostic errors occurred after the training. In a more rigorous study, residents completing a yearlong curriculum on cognitive biases and diagnostic errors demonstrated improvement in correctly naming heuristics [31]. Additionally, they were frequently able to identify heuristics in standardized clinical vignettes and describe potential debiasing strategies. The modest yet significant association between correctly naming defined heuristics and identifying heuristics in clinical vignettes in the present study aligns with those findings. While mastery of vocabulary alone is unlikely to reduce diagnostic errors arising from use of heuristics, conceptual familiarity with this type of reasoning may improve recognition of high-risk clinical scenarios and warrants further study to determine whether these interventions reduce diagnostic errors in clinical encounters.

Diagnosticians in this study were generally comfortable speaking privately to peers about their diagnostic errors. That comfort did not persist when participants considered discussing diagnostic errors in public settings. This discomfort appears to stem from a general concern over how diagnosticians’ knowledge, decision-making and reputation are judged rather than the error simply being made public. Diagnosticians also reported a poorer self-perception of themselves as clinicians as a barrier. These findings suggest significant cultural changes are necessary to advance patient safety. We did not inquire about the fear of malpractice litigation which may also be a substantial barrier to openly discussing diagnostic errors.

We were encouraged by the finding that more experienced clinicians were significantly more comfortable with public discussions of their errors than their less experienced peers. These data imply that senior clinicians may serve as influential role models in the cultural shift required to share experiences publicly in efforts to reduce harm from diagnostic errors.

Interpretation of our study must be considered in light of some limitations. We experienced a low response rate from APPs and, as a pilot study, sampled a relatively narrow segment of our hospital’s pediatric faculty. Our findings may not be representative of faculty at other hospitals or providers in non-academic and outpatient settings. We studied a small number of heuristics. Croskerry lists over 30 heuristics encountered in clinical reasoning but data describing which heuristics are most common or significant in diagnostic reasoning are limited [20]. Thus, we chose heuristics that our own clinical experience indicates are employed relatively frequently but acknowledge that they are not proven to be the most common culprits in diagnostic errors. We also acknowledge that our survey attempted to assess familiarity with heuristic concepts using definitions as a surrogate. It would be more useful to assess familiarity using more qualitative methods discussing vignettes with diagnosticians; this is an opportunity for future study.


Clinicians in our study were unable to match heuristic definitions or clinical vignettes exemplifying heuristics to the proper heuristic term. Attempts to foster a culture supportive of discussing diagnostic errors will require both improved awareness of the role heuristics play in the diagnostic process and better methods to promote psychological safety when sharing errors.


  • 1.

    Balogh E, Miller BT, Ball J, Institute of Medicine (U.S.). Committee on Diagnostic Error in Health Care. Improving diagnosis in health care. Washington, DC: The National Academies Press, 2015. Google Scholar

  • 2.

    Bordini BJ, Stephany A, Kliegman R. Overcoming diagnostic errors in medical practice. J Pediatr 2017;185:19–25.e1. CrossrefPubMedGoogle Scholar

  • 3.

    Croskerry P, Singhal G, Mamede S. Cognitive debiasing 1: origins of bias and theory of debiasing. BMJ Qual Saf 2013;22:ii58–ii64. Web of ScienceGoogle Scholar

  • 4.

    Evans JS. Dual-processing accounts of reasoning, judgment, and social cognition. Annu Rev Psychol 2008;59:255–78. Web of ScienceCrossrefPubMedGoogle Scholar

  • 5.

    McLaughlin K, Eva KW, Norman GR. Reexamining our bias against heuristics. Adv Health Sci Educ Theor Pract 2014;19:457–64. CrossrefGoogle Scholar

  • 6.

    Rencic J, Trowbridge RL, Jr, Fagan M, Szauter K, Durning S. Clinical reasoning education at US medical schools: results from a national survey of internal medicine clerkship directors. J Gen Intern Med 2017;32:1242–6. Web of ScienceCrossrefPubMedGoogle Scholar

  • 7.

    Saposnik G, Redelmeier D, Ruff CC, Tobler PN. Cognitive biases associated with medical decisions: a systematic review. BMC Med Inform Decis Mak 2016;16:138. PubMedCrossrefWeb of ScienceGoogle Scholar

  • 8.

    Gordon R, Franklin N. Cognitive underpinnings of diagnostic error. Acad Med 2003;78:782. CrossrefPubMedGoogle Scholar

  • 9.

    Graber ML, Sorensen AV, Biswas J, Modi V, Wackett A, Johnson S, et al. Developing checklists to prevent diagnostic error in Emergency Room settings. Diagnosis 2014;1:223–31. CrossrefPubMedGoogle Scholar

  • 10.

    Riches N, Panagioti M, Alam R, Cheraghi-Sohi S, Campbell S, Esmail A, et al. The effectiveness of electronic differential diagnoses (DDX) generators: a systematic review and meta-analysis. PLoS One 2016;11:e0148991. CrossrefWeb of SciencePubMedGoogle Scholar

  • 11.

    Cifra CL, Jones KL, Ascenzi JA, Bhalala US, Bembea MM, Newman-Toker DE, et al. Diagnostic errors in a PICU: insights from the morbidity and mortality conference. Pediatr Crit Care Med 2015;16:468–76. Web of ScienceCrossrefGoogle Scholar

  • 12.

    Katz D, Detsky AS. Incorporating metacognition into morbidity and mortality rounds: the next frontier in quality improvement. J Hosp Med 2016;11:120–2. Web of ScienceCrossrefPubMedGoogle Scholar

  • 13.

    Woodward HI, Lemer C, Wu AW. An end to the witch hunts: responding to the defenders of blame and shame. A commentary on Collins, Block, Arnold and Christakis. Soc Sci Med 2009;69:1291–3. CrossrefPubMedWeb of ScienceGoogle Scholar

  • 14.

    Hilfiker D. Facing our mistakes. NEJM 1984;310:118–22. PubMedCrossrefGoogle Scholar

  • 15.

    Luu S, Leung SO, Moulton CA. When bad things happen to good surgeons: reactions to adverse events. Surg Clin North Am 2012;92:153–61. PubMedWeb of ScienceCrossrefGoogle Scholar

  • 16.

    Kaldjian LC, Forman-Hoffman VL, Jones EW, Wu BJ, Levi BH, Rosenthal GE. Do faculty and resident physicians discuss their medical errors? J Med Ethics 2008;34:717–22. CrossrefWeb of SciencePubMedGoogle Scholar

  • 17.

    Khatri N, Brown GD, Hicks LL. From a blame culture to a just culture in health care. Health Care Manage Rev 2009;34:312–22. Web of ScienceCrossrefPubMedGoogle Scholar

  • 18.

    Tad YD, Pierce RG, Pell JM, Stephan L, Kneeland PP, Wald HL. Leveraging a redesigned morbidity and mortality conference that incorporates the clinical and educational missions of improving quality and patient safety. Acad Med 2016;91:1239–43. PubMedCrossrefWeb of ScienceGoogle Scholar

  • 19.

    Collins ME, Block SD, Arnold RM, Christakis NA. On the prospects for a blame-free medical culture. Soc Sci Med 2009;69:1287–90. PubMedCrossrefWeb of ScienceGoogle Scholar

  • 20.

    Croskerry P. The importance of cognitive errors in diagnosis and strategies to minimize them. Acad Med 2003;78:775–80. PubMedCrossrefGoogle Scholar

  • 21.

    Harris PA, Taylor R, Thielke R, Payne J, Gonzalez N, Conde JG. Research electronic data capture (REDCap) – a metadata-driven methodology and workflow process for providing translational research informatics support. J Biomed Sci 2009;42:377–81. Google Scholar

  • 22.

    American Association of Public Opinion Research. Standard definitions: final dispositions of case codes and outcome rates for surveys, 9th edition; 2016. Google Scholar

  • 23.

    Singh H, Thomas EJ, Wilson L, Kelly PA, Pietz K, Elkeeb D, et al. Errors of diagnosis in pediatric practice: a multisite survey. Pediatrics 2010;126:70–9. CrossrefWeb of ScienceGoogle Scholar

  • 24.

    McDonald KM, Matesic B, Contopoulos-Ioannidis DG, Lonhart J, Schmidt E, Pineda N, et al. Patient safety strategies targeted at diagnostic errors: a systematic review. Ann Intern Med 2013;158:381–9. PubMedWeb of ScienceCrossrefGoogle Scholar

  • 25.

    Monteiro SD, Sherbino J, Patel A, Mazzetti I, Norman GR, Howey E. Reflecting on diagnostic errors: taking a second look is not enough. J Gen Intern Med 2015;30:1270–4. CrossrefWeb of SciencePubMedGoogle Scholar

  • 26.

    Mamede S, van Gog T, van den Berge K, Rikers RM, van Saase JL, van Guldener C, et al. Effect of availability bias and reflective reasoning on diagnostic accuracy among internal medicine residents. J Am Med Assoc 2010;304:1198–203. Web of ScienceCrossrefGoogle Scholar

  • 27.

    Norman G, Sherbino J, Dore K, Wood T, Young M, Gaissmaier W, et al. The etiology of diagnostic errors: a controlled trial of system 1 versus system 2 reasoning. Acad Med 2014;89:277–84. Web of SciencePubMedCrossrefGoogle Scholar

  • 28.

    Sherbino J, Yip S, Dore KL, Siu E, Norman GR. The effectiveness of cognitive forcing strategies to decrease diagnostic error: an exploratory study. Teach Learn Med 2011;23:78–84. Web of ScienceCrossrefPubMedGoogle Scholar

  • 29.

    Zwaan L, Monteiro S, Sherbino J, Ilgen J, Howey B, Norman G. Is bias in the eye of the beholder? A vignette study to assess recognition of cognitive biases in clinical case workups. BMJ Qual Saf 2017;26:104–10. PubMedCrossrefWeb of ScienceGoogle Scholar

  • 30.

    Nemes-Walsh JK, Lee AJ. Diagnostic errors: impact of an educational intervention on pediatric primary care. J Pediatr Healthcare 2017;32:10. Google Scholar

  • 31.

    Reilly JB, Ogdie AR, Von Feldt JM, Myers JS. Teaching about how doctors think: a longitudinal curriculum in cognitive bias and diagnostic error for residents. BMJ Qual Saf 2013;22:1044–50. Web of SciencePubMedCrossrefGoogle Scholar

About the article

Corresponding author: Joseph A. Grubenhoff, MD, MSCS, Associate Professor of Pediatrics, Associate Medical Director of Clinical Effectiveness, University of Colorado Denver School of Medicine, 13123 E. 16th Avenue, B-251, Aurora, CO 80045, USA; and Children’s Hospital Colorado, Aurora, CO, USA, Phone: 3037242581

Received: 2018-07-19

Accepted: 2019-01-21

Published Online: 2019-02-13

Published in Print: 2019-06-26

Author contributions: All the authors have accepted responsibility for the entire content of this submitted manuscript and approved submission.

Research funding: None declared.

Employment or leadership: None declared.

Honorarium: None declared.

Competing interests: The funding organization(s) played no role in the study design; in the collection, analysis, and interpretation of data; in the writing of the report; or in the decision to submit the report for publication.

Citation Information: Diagnosis, Volume 6, Issue 2, Pages 101–107, ISSN (Online) 2194-802X, ISSN (Print) 2194-8011, DOI: https://doi.org/10.1515/dx-2018-0056.

Export Citation

©2019 Walter de Gruyter GmbH, Berlin/Boston.Get Permission

Citing Articles

Here you can find all Crossref-listed publications in which this article is cited. If you would like to receive automatic email messages as soon as this article is cited in other publications, simply activate the “Citation Alert” on the top of this page.

Andrew P.J. Olson, Geeta Singhal, and Gurpreet Dhaliwal
Diagnosis, 2019, Volume 6, Number 2, Page 75

Comments (0)

Please log in or register to comment.
Log in