Authors

  1. Sears, Kim RN, PhD
  2. M, Christina Godfrey, RN, PhD
  3. Luctkar, Marian Flude, RN, MScN (PhD candidate)
  4. Ginsburg, Liane RN, PhD
  5. Tregunno, Deborah RN, PhD
  6. Ross-White, Amanda MLIS, AHIP

Abstract

Review question/objective: The objective of this systematic review is to compare the use of self-assessment instruments versus an objective structured clinical examination (OSCE) to measure the competence of healthcare learners and healthcare professionals. The specific question that will guide this review is: When measuring the competence of healthcare learners and healthcare professionals is the evaluation obtained by self-assessment instruments comparable to performance on an OSCE?

 

Background: Establishing the effectiveness of the health professional education process is complex and requires a multifaceted approach to assess the outcomes.1 Typically, outcomes are assessed in terms of the competence of the professional, level of confidence, performance and/or skills. Throughout the literature on this topic, these terms are used interchangeably, but there is overlap and some terms may encompass others. It is important to begin by providing descriptions/definitions of these terms.

 

Competence

 

In their paper that discusses the definition and assessment of professional competence, Epstein and Hundert define professional competence as: "the habitual and judicious use of communication, knowledge, technical skills, clinical reasoning, emotions, values and reflection in daily practice for the benefit of the individual and community being served."2(p226)

 

Confidence

 

Holland's concept analysis of professional confidence describes four components, namely affect (feelings associated with action), reflection (thoughtfully examine one's actions and intentions), higher cognitive functioning (which includes aspects such as learning and integration of concepts, decision making, attention, motivation and memory) and action.3(p219)

 

Performance

 

The on-line Merriam-Webster dictionary defines performance as "the execution of an action or something accomplished - a deed or feat".4

 

Skill

 

Skill is defined as "proficiency, facility, or dexterity that is acquired or developed through training or experience."5

 

Self-efficacy

 

Self-efficacy is defined as "people's beliefs about their capabilities to produce designated levels of performance that exercise influence over events that affect their lives."6

 

Knowledge

 

Knowledge is defined as "(1) the fact or condition of knowing something with familiarity gained through experience or association; (2) acquaintance with or understanding of a science, art, or technique."7

 

Empathy

 

Empathy is defined as the "ability to understand and share the feelings of another."8

 

Looking at the above definitions of competence, confidence and performance it is clear that there is considerable overlap. The term competence was found to be the most inclusive. However, because there are nuances involved in each term, in this review the following concepts will be referred to: that of competence (including knowledge and performance) and confidence (including self-efficacy).

 

There are a variety of ways to measure health professionals' competence and self-assessment is a cost effective source of information. Self-assessment has been defined as: the evaluation or judgment of 'the worth' of one's performance and the identification of one's strengths and weaknesses with a view to improving one's learning outcomes.9(p146) For example, self-reported patient safety competence may provide data about learners' insights into and likely safety of, their own practice,10 and about their perceived strengths or limitations.11,12 The value of using more objective methods to assess competence is unclear. Recent studies examining self versus expert assessment of technical and non-technical skills have produced mixed results. Surgeons seem to be able to accurately assess their own technical skills but not their non-technical skills;13 however, an earlier study of junior medical officers found no correlation between their self-assessments of confidence and their measured competencies on routine procedural skills.14

 

In a systematic review published in 2006, Davis and colleagues describe how accurately health professionals, primarily physicians, subjectively evaluate their own competence compared with external observations of their competence.15 They concluded that physicians have a limited ability to accurately self-assess and this may be particularly true among those rated as the least skilled and those who were also the most confident. These results were found to be consistent with other professions.16

 

The objective structured clinical examination (OSCE) is another method that has shown to be a useful means to assess the competence of a learner. Typically an OSCE consists of a specific scenario established by the examiners that requires the learner to demonstrate their proficiency in that area. The evaluator can control the environment and standardize the patient and in this manner use the OSCE to objectively assess competencies (i.e. knowledge, attitudes and behaviors). There is growing recognition that OSCEs are appropriate for evaluating the interpersonal skills associated with breaking bad news or cross-cultural interviewing.17 The use of the OSCE to assess physician communication skills is also becoming more common (e.g. Huntley, Salmon, Fisher, Fletcher, & Young;18 Iramaneerat, Myford, Yudkowsky, & Lowenstein;19 Ponton-Carss, Hutchison, & Violato;20 Van Nuland, Van den Noortgate, van der Vleuten, & Jo.21)

 

In the realm of patient safety there is a small but emerging body of literature encouraging the use of OSCEs to assess aspects of patient safety competence among medical trainees.22-28 In this area most OSCEs assess the technical aspects of patient safety or quality improvement competence,25,27-29 or clinical aspects of patient safety such as hand hygiene compliance and medication labeling.25 Few studies describe the use of OSCEs to assess socio-cultural aspects of patient safety,24,26 and those that do tend to focus on communicating or disclosing an error and are discipline-specific in nature.30,31

 

In nursing, a recent integrative review by Walsh and colleagues located 41 papers and identified major gaps regarding the psychometrics of nursing OSCEs.32 In concluding their review, the researchers highlighted the need for additional research on using the OSCE as an evaluative tool in nursing.

 

The OSCE is thought to be a more objective measure than self-assessment. However, while limited, examinations of the extent to which OSCE performance predicts outcomes on other performance metrics are somewhat equivocal. Some studies have failed to detect a significant positive relationship between OSCE performance and other forms of summative evaluations of health professionals and learners.23 A study by Tamblyn found that scores achieved in a patient-physician communication and clinical decision-making OSCE, that was part of a national licensing examination, predicted complaints to medical regulatory authorities up to 10 years later.33

 

In an environment where providing optimal student learning and quality patient care is a goal, there is a need to explore whether a link exists between self-assessment scores and OSCEs in light of providing the best learning for the most affordable means. It has been noted that some studies comparing self and external assessments of competence (such as the OSCE) have had several methodological problems. Davis and colleagues report that fewer than half of the studies they included in their systematic review (1) used pretested or validated OSCEs or standardized patients or assessment instruments, or (2) described objective criteria for performance assessment.15 Others have noted there is insufficient methodological detail in most published research involving standardized patients (SP), in particular details pertaining to SP characteristics and their training.

 

An examination of the Cochrane Library of Systematic Reviews, the Joanna Briggs Institute Database of Systematic Reviews and Implementation Reports and the PROSPERO database indicates that no systematic reviews have been completed (or proposed) on this topic since the Davis review in 2006. Building on the Davis review, this systematic review will explore research that examines the relationship between self-assessed competence and objective assessments of competence using the OSCE. The proposed synthesis is part of a broader program of research which builds on recommendations from numerous international bodies regarding the need to restructure health professional education to ensure it equips learners with the knowledge, skills and attitudes they need to function safely.1,34-37 Notably, there is also recognition that what is evaluated drives what is taught and learnt.38,39 Accordingly, development of an OSCE for adoption by various health professional education programs may be crucial for truly integrating patient safety into health professional education. Just as written examinations and OSCEs assess different things,40,41 so do subjective and objective assessments; however, both are understood to yield important data.10

 

Article Content

Inclusion criteria

Types of participants

In this review all healthcare learners and healthcare professionals will be considered, including but not limited to physicians, nurses, dentists, occupational therapists, physiotherapists, social workers and respiratory therapists.

 

Types of intervention(s)/phenomena of interest

In this review studies in which participants are first administered a self-assessment (related to performance), followed by an OSCE will be considered; the results of which will then be compared. The review will consider studies that explore the evaluation of competence by self-assessment, which is then compared with the evaluation of competence using an OSCE exam.

 

Types of outcomes

In this review studies that include the following outcome measures will be considered: competence, confidence, performance, self-efficacy, knowledge and empathy as defined in the background section.

 

Types of studies

In this review both experimental and epidemiological study designs will be considered, including randomized controlled trials, non-randomized controlled trials, quasi-experimental, before and after studies, prospective and retrospective cohort studies, case control studies and analytical cross-sectional studies for inclusion.

 

Descriptive epidemiological study designs will also be considered, including case series, individual case reports and descriptive cross-sectional studies for inclusion.

 

Search strategy

The search strategy aims to find both published and unpublished studies. A three-step search strategy will be utilized in this review. An initial limited search of MEDLINE and CINAHL will be undertaken, followed by an analysis of the text words contained in the title and abstract and of the index terms used to describe the article. A second search using all identified keywords and index terms will then be undertaken across all included databases. Thirdly, the reference lists of all identified reports and articles will be searched for additional studies. In this review only studies published in English will be included. In order to provide the broader picture of all available literature on this topic the non-English literature will be tallied, but not translated. Although this review is building on a review done in 2006, in order to be thorough, no date limit will be placed on the search strategy.

 

The databases to be searched include:

 

Medline, CINAHL, Embase, ERIC, Education Research Complete, Education Full Text, CBCA Education, GlobalHealth, Sociological Abstracts, Cochrane, Mosby's Nursing Consult and PsycInfo.

 

The search for unpublished studies will include: Dissertation Abstracts and Google Scholar.

 

Initial keywords to be used will be: OSCE; objectiv$ structur$ clinic$ exam$; self-assessment; self-report; competence; confidence; self-efficacy.

 

Assessment of methodological quality

Papers selected for retrieval will be assessed by two independent reviewers for methodological validity prior to inclusion in the review using standardized critical appraisal instruments from the Joanna Briggs Institute Meta Analysis of Statistics Assessment and Review Instrument (JBI-MAStARI) (Appendix V). Any disagreements that arise between the reviewers will be resolved through discussion, or with a third reviewer.

 

Data collection

Data will be extracted from papers included in the review using the standardized data extraction tool from JBI-MAStARI (Appendix VI). The data extracted will include specific details about the interventions, populations, study methods and outcomes of significance to the review question and specific objectives.

 

Data synthesis

Quantitative data will, where possible, be pooled in statistical meta-analysis using JBI-MAStARI. All results will be subject to double data entry. Effect sizes expressed as odds ratios (for categorical data) and weighted mean differences (for continuous data) and their 95% confidence intervals will be calculated for analysis. Heterogeneity will be assessed statistically using the standard Chi-square and also explored using subgroup analyses based on the different study designs included in this review. Where statistical pooling is not possible, the findings will be presented in narrative form including tables and figures to aid in data presentation where appropriate.

 

The data will be analyzed using the "Four Stages of Learning" theory developed by Noel Burch (Glordon Training International) in the early 1970s.42 The model is comprised of four stages including: Stage 1 Unconscious Incompetence; Stage 2 Conscious Incompetence; Stage 3 Conscious Competence and Stage 4 Unconscious Competence. According to Burch, everyone progresses through the same four stages regardless of the skill that needs to be acquired. From the beginning stage of acquiring a new skill, the learner progresses from Stage 1 towards Stage 4 as they gain experience and their level of competence increases. In this systematic review the focus will be on the assessment of competence; thus this model provides a useful framework to interpret the levels of competence demonstrated by the participants in the included studies.

 

Conflicts of interest

The reviewers declare no conflict of interest.

 

Acknowledgements

Nil.

 

References

 

1 Lucian Leap Institute. Unmet needs: Teaching physicians to provide safe patient care. Boston: National Patient Safety Foundation; 2010. [Context Link]

 

2 Epstein RM, Hundert EM. Defining and assessing professional competence. [Review] [171 refs]. JAMA 2002 Jan 9;287(2):226-35. [Context Link]

 

3 Holland K, Middleton L, Uys L. Professional confidence: a concept analysis. Scandinavian Journal of Occupational Therapy 2012 Mar;19(2):214-24. [Context Link]

 

4 Merriam-Webster I. Performance. Merriam-Webster,Incorporated 2013 [cited 2013 Jun 9];Available from: URL: http://www.merriam-webster.com/dictionary/performance[Context Link]

 

5 Houghton Mifflin Company. Skill. The American Heritage Dictionary of the English Language 2009 [cited 2013 Jun 9];(Fourth Edition)Available from: URL: http://www.thefreedictionary.com/skill[Context Link]

 

6 Bandura A. Self-efficacy. In: Ramachaudran VS, editor. Encyclopedia of human behavior Volume 4.New York: Academic Press; 1994. p. 71-81. [Context Link]

 

7 Merriam-Webster I. Knowledge. Merriam-Webster,Incorporated 2013 [cited 2013 Jun 9];Available from: URL: http://www.merriam-webster.com/dictionary/knowledge[Context Link]

 

8 Oxford Dictionary. Empathy. Google 2013Available from: URL: http://www.oxforddictionaries.com/definition/english/empathy[Context Link]

 

9 Klenowski V. Student Self_evaluation Processes in Student-centred Teaching and Learning Contexts of Australia and England. Assessment in Education: Principles, Policy & Practice 1995;2(2):145-63. [Context Link]

 

10 Carlisle C. Reflecting on levels of confidence and competence in skills acquisition. Med Educ 2000;34(11):886-7. [Context Link]

 

11 Ginsburg LR, Tregunno D, orton PG. Self-reported patient safety competencies among new graduates in medicine, nursing and pharmacy. BMJ quality and safety 2013;22(2):147-54. [Context Link]

 

12 Regehr G, Hodges B, Tiberius R, Lofchy J. Measuring self-assessment skills: an innovative relative ranking model. Acad Med 1996;71(10 Suppl):S52-S54. [Context Link]

 

13 Arora S, Miskovic D, Hull L, Moorthy K, Aggarwal R, Johannsson H, et al. Self vs expert assessment of technical and non-technical skills in high fidelity simulation. Am J Surg 2011;202(4):500-6. [Context Link]

 

14 Barnsley L, Lyon PM, Ralston SJ, Hibbert EJ, Cunningham I, Gordon FC, et al. Clinical skills in junior medical officers: a comparison of self-reported confidence and observed competence. Med Educ 2004;38(4):358-67. [Context Link]

 

15 Davis DA, Mazmanian PE, Fordis M, Van Harrison R, Thorpe KE, Perrier L. Accuracy of physician self-assessment compared with observed measures of competence: a systematic review. JAMA 2006;296(9):1094-102. [Context Link]

 

16 Hodges B, Regehr G, Martin D. Difficulties in recognizing one's own incompetence: novice physicians who are unskilled and unaware of it. Academic Medicine 2001 Oct;76(10:Suppl):Suppl-9. [Context Link]

 

17 Casey PM, Goepfert AR, Espey EL, Hammoud MM, Kaczmarczyk JM, Katz NT, et al. To the point: reviews in medical education-the Objective Structured Clinical Examination. Am J Obstet Gynecol 2009;200(1):25-34. [Context Link]

 

18 Huntley CD, Salmon P, Fisher PL, Fletcher I, Young B. LUCAS: a theoretically informed instrument to assess clinical communication in objective structured clinical examinations. Med Educ 2012;46(3):267-76. [Context Link]

 

19 Iramaneerat C, Myford CM, Yudkowsky R, Lowenstein T. Evaluating the effectiveness of rating instruments for a communication skills assessment of medical residents. Adv Health Sci Educ Theory Pract 2009;14(4):575-94. [Context Link]

 

20 Ponton-Carss A, Hutchison C, Violato C. Assessment of communication, professionalism, and surgical skills in an objective structured performance-related examination (OSPRE): a psychometric study. Am J Surg 2011;202(4):433-40. [Context Link]

 

21 Van Nuland M, Van den Noortgate W, van der Vleuten C, Jo G. Optimizing the utility of communication OSCEs: omit station-specific checklists and provide students with narrative feedback. Patient Educ Couns 2012;88(1):106-12. [Context Link]

 

22 Battles JB, Wilkinson SL, Lee SJ. Using standardised patients in an objective structured clinical examination as a patient safety tool. Qual Saf Health Care 2004;13 Suppl 1:i46-i50. [Context Link]

 

23 Varkey P, Natt N. The Objective Structured Clinical Examination as an educational tool in patient safety. Jt Comm J Qual Patient Saf 2007;33(1):48-53. [Context Link]

 

24 Daud-Gallotti RM, Morinaga CV, Arlindo-Rodrigues M, Velasco IT, Martins MA, Tiberio IC. A new method for the assessment of patient safety competencies during a medical school clerkship using an objective structured clinical examination. Clinics (Sao Paulo) 2011;66(7):1209-15. [Context Link]

 

25 Pernar LI, Shaw TJ, Pozner CN, Vogelgesang KR, Lacroix SE, Gandhi TK, et al. Using an Objective Structured Clinical Examination to test adherence to Joint Commission National Patient Safety Goal-associated behaviors. Jt Comm J Qual Patient Saf 2012;38(9):414-8. [Context Link]

 

26 Singh R, Singh A, Fish R, McLean D, Anderson DR, Singh G. A patient safety objective structured clinical examination. J Patient Saf 2009;5(2):55-60. [Context Link]

 

27 Varkey P, Natt N, Lesnick T, Downing S, Yudkowsky R. Validity evidence for an OSCE to assess competency in systems-based practice and practice-based learning and improvement: a preliminary investigation. Acad Med 2008;83(8):775-80. [Context Link]

 

28 Wagner DP, Hoppe RB, Lee CP. The patient safety OSCE for PGY-1 residents: a centralized response to the challenge of culture change. Teach Learn Med 2009;21(1):8-14. [Context Link]

 

29 Gupta P, Varkey P. Developing a tool for assessing competency in root cause analysis. Joint Commission Journal on Quality & Patient Safety 2009 Jan;35(1):36-42. [Context Link]

 

30 Posner G, Naik V, Bidlake E, Nakajima A, Sohmer B, Arab A, et al. Assessing residents' disclosure of adverse events: traditional objective structured clinical examinations versus mixed reality. J Obstet Gynaecol Can 2012;34(4):367-73. [Context Link]

 

31 Stroud L, McIlroy J, Levinson W. Skills of internal medicine residents in disclosing medical errors: a study using standardized patients. Acad Med 2009;84(12):1803-8. [Context Link]

 

32 Walsh M, Bailey PH, Koren I. Objective structured clinical evaluation of clinical competence: an integrative review. J Adv Nurs 2009;65(8):1584-95. [Context Link]

 

33 Tamblyn R, Abrahamowicz M, Dauphinee D, Wenghofer E, Jacques A, Klass D, et al. Physician scores on a national clinical skills examination as predictors of complaints to medical regulatory authorities. JAMA 2007;298(9):993-1001. [Context Link]

 

34 American Association of Colleges of Nursing. Hallmarks of quality and patient safety: Recommended baccalaureate competencies and curricular guidelines to ensure high-quality and safe patient care. Journal of professional nursing: official journal of the American Association of Colleges of Nursing 2006;22(6):329-30. [Context Link]

 

35 Cronenwett L, Sherwood G, Barnsteiner J, Disch J, Johnson J, Mitchell P, et al. Quality and Safety Education for Nurses. Nursing Outlook 2007 May;55(3):122-31. [Context Link]

 

36 Department of Health. Modernising medical careers: The new curriculum for the foundation years in postgraduate education and training. London United Kingdom: Department of Health; 2007. [Context Link]

 

37 Frank JR, Brien S, on behalf of The Safety Competencies Steering Committee. The safety competencies: Enhancing patient safety across the health professions. Ottawa, Ontario: Canadian Patient Safety Institute; 2013. [Context Link]

 

38 Seiden SC, Galvan C, Lamm R. Role of medical students in preventing patient harm and enhancing patient safety. Qual Saf Health Care 2006 Aug;15(4):272-6. [Context Link]

 

39 van der Vleuten C. Validity of final examinations in undergraduate medical training. BMJ 2000;321(7270):1217-9. [Context Link]

 

40 Dong T, Saguil A, Artino AR, Jr., Gilliland WR, Waechter DM, Lopreaito J, et al. Relationship between OSCE scores and other typical medical school performance indicators: a 5-year cohort study. Mil Med 2012;177(9 Suppl):44-6. [Context Link]

 

41 Probert CS, Cahill DJ, McCann GL, Ben-Shlomo Y. Traditional finals and OSCEs in predicting consultant and self-reported clinical skills of PRHOs: a pilot study. Med Educ 2003;37(7):597-602. [Context Link]

 

42 Burch N. Four stages of learning. Gordon Training International 1970 Available from: URL: http://www.gordontraining.com/free-workplace-articles/learning-a-new-skill-is-ea[Context Link]

Appendix I: Appraisal instruments

 

MAStARI appraisal instrument

Appendix II: Data extraction instruments

 

MAStARI data extraction instrument

 

Keywords: OSCE; objectiv$ structur$ clinic$ exam$; self-assessment; self-report; competence; confidence; self-efficacy