Performance and workplace assessment

Chapter 38


Performance and workplace assessment




Introduction


The apprenticeship model of medical training has existed for thousands of years: the apprentice learns from watching the master, and the master, in turn, observes the apprentice’s performance and helps them improve. Therefore, performance assessment is not a new concept. However, in the modern healthcare environment, with its discourse of accountability, performance assessment has an increasing role in ensuring that professionals maintain the knowledge and skills required for practice. A number of international academic and professional bodies have incorporated performance assessment into their overall assessment frameworks for licensing, training and continued professional development. In the United States, the Medical Licensing Examination (USMLE) makes use of a structured test of clinical skills in the second stage of the licensing assessment. In the UK and Australasia, the main Royal Colleges have assessment frameworks for trainees that include a portfolio of workplace-based assessment (WPBA) tools.


The terms performance and competence are often used interchangeably. However, competence should describe what an individual is able to do, while performance should describe what an individual actually does in clinical practice. Clinical competence is the term being used most frequently by many of the professional regulatory bodies and in the educational literature. There are several dimensions to competence, and a wide range of well-validated assessments have been developed examining these. Traditional methods focus on the assessment of competence in artificial settings built to resemble the clinical environment. However, more novel methods of performance assessment concentrate on building up a structured picture of how the individual practitioner acts in his or her everyday working life, in interactions with patients and other practitioners, using technical, professional and interpersonal skills. Miller’s model, shown in Fig. 38.1, provides a framework for understanding the different facets of clinical competence and the assessment methods that test these.




In this chapter we will look at the different methods used to assess clinical performance, which we will define as the assessment of clinical skills and behaviours in both academic and workplace settings. These methods fall in the top two layers of Miller’s pyramid. We will discuss how and why they are used and some of the practical aspects to be considered for educators wishing to make use of them. We will also explore the strengths and weaknesses of the various tools and consider some of the outstanding issues concerning their use.



Choosing the right assessment


When planning assessments, it is important to be aware of the purpose of the assessment and where it fits into the wider educational programme: a tool is only as useful as the system within which it resides. Considerations would include how to make pass/fail decisions, how to give feedback to candidates, effects on the learning of candidates and whether the assessment is ‘high-stakes’. For example, assessments for the purpose of certification may require different criteria than some medical school assessments, where the primary purpose is to encourage and direct the learning of students (Downing 2003).


In considering the best tools for the purpose of a particular assessment system, educators need to evaluate their validity, reliability, educational impact and acceptability (Schurwith & van der Vleuten 2009).


Validity has been classically regarded as the answer to the question ‘Does this assessment tool measure what it is intended to measure?’ Evidence to support the validity of an assessment has to be gathered from various sources to support the use of the assessment in any particular context. Evidence includes:



Reliability is a measure of the reproducibility of the scores of an assessment, so that the outcome is the same if the assessment is repeated over time. There are various mathematical models for calculating a reliability coefficient, the commonest being Cronbach’s alpha.


Generalizability is a form of reliability measure which takes into account the conditions of a typical objective structured clinical examination (OSCE), where not all the candidates see the same patients, and are not all examined by the same examiners.


Educational impact is important to consider, as inevitably assessments will influence students’ learning strategies. Aligning the content of any assessment to the desired learning objectives would be a useful way to encourage students to attend to the most important outcomes. Other factors such as timing, outcomes (e.g. pass/fail) and format of the assessment will also influence student behaviours.


Acceptability and cost-effectiveness are important features to consider when setting up an assessment system. The administrative setup required to manage a robust and defensible system must be planned and supported, with due regard for costs. The issue of having sound educationally based systems and high quality has to be balanced against such costs. Acceptability extends to not only whether the candidates and examiners believe in the system of assessment, but also other stakeholders such as regulatory bodies, employers and members of the public.



Assessments of clinical competence


Assessments of what a student or doctor is able to do can take place in artificial settings and have the advantage of being able to examine a number of individuals at the same time. The best known of these types of assessment is the OSCE.



Objective structured clinical examination (OSCE)


What is it? An OSCE consists of a series of structured stations around which a candidate moves in sequence. At each station specific tasks have to be performed, usually involving clinical skills such as history taking, clinical examination or practical skills. Different degrees of simulation may be used to test a wide range of psychomotor and communication skills, using simulated patients, part-task trainers, charts and results, resuscitation manikins or computer-based simulations. There is a time limit for each station, and the marking scheme is structured and determined in advance.



How is it used? The OSCE is typically used in high-stakes summative assessments at both the undergraduate and postgraduate level. The main advantages are that large numbers of candidates can be assessed in the same way across a range of clinical skills. High levels of reliability and validity can be achieved in the OSCE due to four main features (Newble 2004):



Overall, candidate scores are less dependent on who is examining and which patient is selected than in traditional long case or viva voce examinations. The key to reliability is the number of stations; the more stations marked by different examiners there are, the more reliable the OSCE will be. However, this has to be balanced with practicality, as clearly the longer an OSCE, the more onerous it will be for all involved.


Organization: OSCEs can be complex to organize. Planning should begin well in advance of the examination date, and it is essential to ensure that there are enough patients, simulated patients, examiners, staff, refreshments and equipment to run the complete circuit for all candidates on the day of the exam. Careful calculations of the numbers of candidates, the length of each complete circuit and how many circuits have to be run need to be made. The mix of stations is chosen in advance and depends on the curriculum and the purpose of the assessment. A blueprint should be drawn up which outlines how the assessment is going to meet its goals. An example of a simple blueprint, detailing the different areas of a curriculum and how the stations chosen will cover these, can be seen in Fig. 38.2. Development of a blueprint ensures adequate content validity, the level to which the sampling of skills in the OSCE matches the learning objectives of the whole curriculum. It is also necessary to think about the timing of the stations. The length of the station should fit the task requested as closely as possible. Ideally, stations should be practised in advance to clarify this and anticipate potential problems with the setup or the mark sheet. Thought also has to be given to the training of assessors and standardized or simulated patients.


Dec 9, 2016 | Posted by in GENERAL & FAMILY MEDICINE | Comments Off on Performance and workplace assessment

Full access? Get Clinical Tree

Get Clinical Tree app for offline access