Several potential advantages for students' learning are claimed for problem-based learning (PBL). Students in PBL curricula may be more highly motivated; they may be better problem solvers and self-directed learners; they may be better able to learn and recall information; and they may be better able to integrate basic science knowledge into the solutions of clinical problems. Although some of these claims find theoretical support from the literature on the psychology of learning, to date there has been no review of the experimental evidence supporting the possible differences in students' learning that can be attributed to PBL. In this review article, the authors examine each claim critically in light of that evidence. They conclude that (1) there is no evidence that PBL curricula result in any improvement in general, content-free problem-solving skills; (2) learning in a PBL format may initially reduce levels of learning but may foster, over periods up to several years, increased retention of knowledge; (3) some preliminary evidence suggests that PBL curricula may enhance both transfer of concepts to new problems and integration of basic science concepts into clinical problems; (4) PBL enhances intrinsic interest in the subject matter; and (5) PBL appears to enhance self-directed learning skills, and this enhancement may be maintained.
A new theory of the development of expertise in medicine is outlined. Contrary to existing views, this theory assumes that expertise is not so much a matter of superior reasoning skills or in-depth knowledge of pathophysiological states as it is based on cognitive structures that describe the features of prototypical or even actual patients. These cognitive structures, referred to as "illness scripts," contain relatively little knowledge about pathophysiological causes of symptoms and complaints but a wealth of clinically relevant information about disease, its consequences, and the context under which illness develops. By contrast, intermediate-level students without clinical experience typically use pathophysiological, causal models of disease when solving problems. The authors review evidence supporting the theory and discuss its implications for the understanding of five phenomena extensively documented in the clinical-reasoning literature: (1) content specificity in diagnostic performance; (2) typical differences in data-gathering techniques between medical students and physicians; (3) difficulties involved in setting standards; (4) a decline in performance on certain measures of clinical reasoning with increasing expertise; and (5) a paradoxical association between errors and longer response times in visual diagnosis.
CONTEXT Although health sciences programmes continue to value non-cognitive variables such as interpersonal skills and professionalism, it is not clear that current admissions tools like the personal interview are capable of assessing ability in these domains. Hypothesising that many of the problems with the personal interview might be explained, at least in part, by it being yet another measurement tool that is plagued by context specificity, we have attempted to develop a multiple sample approach to the personal interview.METHODS A group of 117 applicants to the undergraduate MD programme at McMaster University participated in a multiple mini-interview (MMI), consisting of 10 short objective structured clinical examination (OSCE)-style stations, in which they were presented with scenarios that required them to discuss a health-related issue (e.g. the use of placebos) with an interviewer, interact with a standardised confederate while an examiner observed the interpersonal skills displayed, or answer traditional interview questions.RESULTS The reliability of the MMI was observed to be 0.65. Furthermore, the hypothesis that context specificity might reduce the validity of traditional interviews was supported by the finding that the variance component attributable to candidate-station interaction was greater than that attributable to candidate. Both applicants and examiners were positive about the experience and the potential for this protocol.DISCUSSION The principles used in developing this new admissions instrument, the flexibility inherent in the multiple mini-interview, and its feasibility and cost-effectiveness are discussed.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
customersupport@researchsolutions.com
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.
Copyright © 2024 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.