D iagnostic errors result in an estimated 40,000 to 80,000 US hospital deaths annually 1 ; they are the most common type of medical error cited in malpractice claims and result in the largest payouts 2 ; they lead to more patient readmissions than other types of errors 3 ; and autopsy studies confirm the ubiquity of missed diagnoses. 4 Although not all diagnostic errors result in harm, many do. 5 Typically these mistakes represent either cognitive errors, such as those resulting from a lack of knowledge, or communication errors, such as those resulting from poor handoffs and follow-up. 6,7 Differential diagnosis (DDX) generators have the potential to assist in the former, while other types of clinical decision support can address the latter.In this issue of JGIM, Bond and colleagues examine the features and performance of four DDX generators. 8 They identified these tools through a systematic search, compared each tool against a set of consensus criteria, and tested each tool using diagnostic conundrums from the New England Journal of Medicine (NEJM) and the Medical Knowledge Self Assessment Program (MKSAP) of the American College of Physicians. The best tools demonstrated sensitivities as high as 65% and a number of features which advanced their usability, including direct links to evidence resources, the ability to input multiple variables including negative findings, the potential to be integrated with electronic health records (EHRs) to obviate the need for manual data entry, and mobile access.Unfortunately, the performance results were not much better than programs examined in a similar study published in the NEJM nearly 20 years ago by Berner and colleagues. 9 In that study, four DDX generators were evaluated, two of which are no longer on the market, and only one of which met the inclusion criteria of the current study. Berner and colleagues concluded that on average the tools had a relatively mediocre sensitivity of approximately 50-70% for the correct diagnosis, and a companion editorial by Kassirer gave the tools a grade of C. 10 The current study by Bond and colleagues has many strengths. First, their search for DDX generators was systematic and comprehensive, including searches of the internet and medical literature databases, as well as expert consultation. The appendix includes a list of programs which were identified but did not meet inclusion criteria. The only major tool type that was not considered for inclusion in their study was the internet search engine (such as the generic Google or the medically targeted WebMD). Evidence suggests that these search tools can lead to the correct diagnosis in over 50% of cases, similar to the sensitivities of the DDX generators in this study. 11 Thus, head-to-head comparisons of DDX generators with internet search engines will be critical to demonstrating the incremental value of the proprietary, specialized software. Of note, at least one of the tools examined in Bond's study offered structured Google searching of pre-selected medical websites.A second stre...