Objective structured clinical examination

An objective structured clinical examination (OSCE) is an approach to the assessment of clinical competence in which the components are assessed in a planned or structured way with attention being paid to the objectivity of the examination which is basically an organization framework consisting of multiple stations around which students rotate and at which students perform and are assessed on specific tasks.[1] OSCE is a modern[2] type of examination often used for assessment in health care disciplines.

History and purpose

edit

The development of OSCE is credited to Ronald Harden. Since the publication of the first paper in the British Medical Journal in 1975, OSCE has been widely adopted in many medical schools and professional bodies. The format of OSCE is continuously evolving and may include real or simulated patients, clinical specimens, and other clinical materials. OSCE is primarily used to assess focused clinical skills such as history taking, physical examination, diagnosis, communication, and counseling.[3][page needed]

In the last three decades the OSCE has seen a steady exponential growth and usage in both undergraduate and postgraduate examinations around the globe. OSCEs are also used for licensure examinations and as a feedback tool in formative settings. Common uses of the OSCE include:[1]

  • as a performance based assessment tool for testing the minimum accepted standards of students or trainees as barrier (exit) examinations during the undergraduate years in most of the medical schools
  • as a postgraduate high stakes assessment tool in Royal College examinations
  • as a formative assessment tool in undergraduate medical education
  • as a tool for the assessment of graduates seeking highstakes licensure and certification to practise medicine
  • as an educational tool to provide immediate feedback

Objective structured clinical examinations evaluate learners "showing how" to perform complex clinical tasks including those infrequently observed and those core to practice.[4]

Design

edit

An OSCE usually consists of a circuit of short stations, usually 5–10 minutes, though some use up to 15 minutes. In each station, the candidate is examined on a one-to-one basis with one or two examiner(s) and either real or simulated (actors or electronic patient simulators) patients. Each station has a different examiner, as opposed to the traditional method of clinical examinations where a candidate would be assigned to one examiner for the entire examination. Candidates rotate through the stations, completing all the stations on their circuit. In this way, all candidates take the same stations. It is considered to be an improvement over traditional examination methods[citation needed] because the stations can be standardised, enabling fairer peer comparison and complex procedures can be assessed without endangering patients health.[clarification needed]

OSCEs are designed to assess candidates' clinical skills more objectively. This is done by giving all candidates the same stations (though signs of real patients may vary slightly), where they are assessed with the same marking scheme and awarded marks for each step performed correctly. If theoretical knowledge is examined, such as with the examiner asking questions at the end of the station, the questions will also be standardised. The candidate will only be asked questions on the marking scheme, and will not be awarded marks for any other questions asked.

OSCEs are also designed to be structured, with instructions carefully written to ensure that the candidate has a very specific task to complete in each station. Where simulated patients are used, detailed scripts are provided to ensure that the information provided is the same for all candidates, even including the emotions displayed by the patient. The examination is carefully structured to include parts from all elements of the curriculum as well as a wide range of skills.

Variation

edit

There are several variations of OSCE, those are:[1]

  • Objective Structured Practical Examination (OSPE), which assess practical skills, knowledge and/or interpretation of data in non clinical settings.
  • Objective Structured Assessment of Technical Skills (OSATS), which designed for objective skills assessment, consisting of a global rating scale and a procedure specific checklist. It is primarily used for feedback or measuring progress of training in surgical specialities.
  • Objective Structured Video Examinations (OSVE). The variation consists of videotaped recordings of patient-doctor encounters are shown to students simultaneously and questions related to the video clip are asked. Written answers are marked in a standardised manner.
  • Team Objective Structured Clinical Examination (TOSCE). Formative assessment covering common consultations in general practice. A team of students visits each station in a group, performing one task each in a sequence. The candidates are marked for their performance and feedback is provided. The team approach improves efficiency and encourages learning from peers.

Advantages

edit

The advantages of OSCE are:[3][page needed]

  • Broader content coverage: Unlike the conventional short or long case examination format, multi-station OSCE allows broader content and domain coverage.
  • Decreased bias: Student performance in each station is marked by independent examiners on a predetermined marking template that is customized to each patient scenario, thus reduces the variability of examination.
  • Practicability: OSCE allows the use of simulated patients and clinical materials, thereby decreasing the need for real patients during the examination.

Disadvantages

edit

The disadvantages of OSCE are:[3][page needed]

  • Fragmentation of tasks: OSCE is often criticized for fragmenting a physician’s task as the candidates are asked to focus on a particular task, thus potentially undermines what a physician is expected to do in real life.
  • Construct invalidity: OSCE is based on a predefined list of activities that a candidate is required to perform during the examination. However, this may not be a true reflection of how a competent and experienced physician works in real life.
  • Preparation time and budget: The preparation time for simulated patient (SP)-based OSCE is longer, as it includes script writing, training, and pilot testing. It may take several sessions for an SP to become familiar with the case and realistically portray the findings in a consistent manner. The increased budget needed to run a successful SP program is also a valid concern.

Marking

edit

Marking in OSCEs is done by the examiner. Occasionally written stations, for example, writing a prescription chart, are used and these are marked like written examinations, again usually using a standardized mark sheet. One of the ways an OSCE is made objective is by having a detailed mark scheme and standard set of questions. For example, a station concerning the demonstration to a simulated patient on how to use a metered dose inhaler (MDI) would award points for specific actions which are performed safely and accurately. The examiner can often vary the marks depending on how well the candidate performed the step. At the end of the mark sheet, the examiner often has a small number of marks that they can use to weight the station depending on performance and if a simulated patient is used, then they are often asked to add marks depending on the candidates approach. At the end, the examiner is often asked to give a "global score". This is usually used as a subjective score based on the candidates overall performance, not taking into account how many marks the candidate scored. The examiner is usually asked to rate the candidate as pass/borderline/fail or sometimes as excellent/good/pass/borderline/fail. This is then used to determine the individual pass mark for the station.

Many centres allocate each station an individual pass mark. The sum of the pass marks of all the stations determines the overall pass mark for the OSCE. Many centres also impose a minimum number of stations required to pass which ensures that a consistently poor performance is not compensated by a good performance on a small number of stations.

There are, however, criticisms that the OSCE stations can never be truly standardized and objective in the same way as a written exam. It has been known for different patients / actors to afford more assistance, and for different marking criteria to be applied. Finally, it is not uncommon at certain institutions for members of teaching staff be known to students (and vice versa) as the examiner. This familiarity does not necessarily affect the integrity of the examination process, although there is a deviation from anonymous marking. However, in OSCEs that use several circuits of the same stations the marking is repeatedly shown to be very consistent which supports the validity that the OSCE is a fair clinical examination. There are arguments for and against quarantining OSCE examinees to prevent sharing of exam information.[5] Although the data tend to show no improvement in the overall scores in a later OSCE session, the research methodology is flawed and validity of the claim is questionable.[6] A study suggested that marks do not give a sound inference of student collusion in an OSCE.[7]

Preparation

edit

Preparing for OSCEs is very different from preparing for an examination on theory. In an OSCE, clinical skills are tested rather than pure theoretical knowledge.[importance?] It is essential to learn correct clinical methods, and then practice repeatedly until one perfects the methods whilst simultaneously developing an understanding of the underlying theory behind the methods used.[importance?] Marks are awarded for each step in the method; hence, it is essential to dissect the method into its individual steps, learn the steps, and then learn to perform the steps in a sequence.

Most hospitals and universities have clinical skills labs where students have the opportunity to practice clinical skills such as taking blood or mobilizing patients in a safe and controlled environment. It is often very helpful to practise in small groups with colleagues, setting a typical OSCE scenario and timing it with one person role playing a patient, one person doing the task and if possible, one person either observing and commenting on technique or even role playing the examiner using a sample mark sheet. Many OSCE textbooks have sample OSCE stations and mark sheets that can be helpful when studying in the manner. In doing this the candidate is able to get a feel of running to time and working under pressure.

In many OSCEs the stations are extended using data interpretation. For example, the candidate may have to take a brief history of chest pain and then interpret an electrocardiogram. It is also common to be asked for a differential diagnosis, to suggest which medical investigations the candidate would like to do or to suggest a management plan for the patient.

The peer-assisted mock OSCE improved tutee confidence, reduced the anxieties associated with OSCEs, and improved candidate confidence for OSCE.[8][9]

See also

edit

References

edit
  1. ^ a b c Khan, Kamran Z.; Ramachandran, Sankaranarayanan; Gaunt, Kathryn; Pushkar, Piyush (2013). "The Objective Structured Clinical Examination (OSCE): AMEE Guide No. 81. Part I: An historical and theoretical perspective". Medical Teacher. 35 (9): e1437 – e1446. doi:10.3109/0142159X.2013.818634. ISSN 0142-159X. PMID 23968323. S2CID 28150274.
  2. ^ Harden, R M; Stevenson, M; Downie, W W; Wilson, G M (1975). "Assessment of clinical competence using objective structured examination". BMJ. 1 (5955): 447–451. doi:10.1136/bmj.1.5955.447. PMC 1672423. PMID 1115966.
  3. ^ a b c Amin, Zubair; Eng, Khoo Hoon (2009). Basics in Medical Education (2 ed.). WORLD SCIENTIFIC. doi:10.1142/7044. ISBN 978-981-283-538-3.
  4. ^ Criscione-Schreiber, Lisa (2020). "Turning Objective Structured Clinical Examinations into Reality". Rheumatic Disease Clinics of North America. 46 (1): 21–35. doi:10.1016/j.rdc.2019.09.010. PMID 31757285. S2CID 208234190.
  5. ^ "Calls to ensure end-of-year exam fairness". BMA - Connecting doctors. Retrieved 2017-04-13.
  6. ^ Yeung, E (April 2, 2017). "Sharing of information in objective structured clinical examinations -- does no evidence of difference mean evidence of no difference?". BMJ Open. 6 (10): e012541. doi:10.1136/bmjopen-2016-012541. PMC 5073653. PMID 27729351.
  7. ^ Parks, R; Warren, P M; Boyd, K M; Cameron, H; Cumming, A; Lloyd-Jones, G (2006). "The Objective Structured Clinical Examination and student collusion: Marks do not tell the whole truth". Journal of Medical Ethics. 32 (12): 734–738. doi:10.1136/jme.2005.015446. PMC 2563344. PMID 17145916.
  8. ^ Young, Ian; Montgomery, Kieran; Kearns, Patrick; Hayward, Samantha; Mellanby, Ed (7 May 2014). "The benefits of a peer-assisted mock OSCE". The Clinical Teacher. 11 (3): 214–218. doi:10.1111/tct.12112. PMID 24802924. S2CID 12122898.
  9. ^ Braier-Lorimer, Darryl Alexander; Warren-Miell, Hannah (2021-11-25). "A peer-led mock OSCE improves student confidence for summative OSCE assessments in a traditional medical course". Medical Teacher. 44 (5): 535–540. doi:10.1080/0142159X.2021.2004306. ISSN 0142-159X. PMID 34822315. S2CID 244660969.
edit