Improving assessment and selection practices within the Health Care professions and internationally
Submitting Institution
City University, LondonUnit of Assessment
Psychology, Psychiatry and NeuroscienceSummary Impact Type
SocietalResearch Subject Area(s)
Medical and Health Sciences: Public Health and Health Services
Summary of the impact
Applied psychology research undertaken at City University London has had
a major impact on improving high stakes medical selection. Previously
doctors were selected by curriculum vitae and unstructured
interviews which resulted in bias and discrimination in selection. The
research has provided evidence for using new selection methodologies:
machine-marked tests (job knowledge and situational judgment tests) and
assessment centres, informed by best practice in Occupational Psychology.
These apply to all medical specialties (e.g., surgeons and anaesthetists),
several healthcare professions (e.g., doctors, nurses, dentists) and
selection points of entry (medical school through to specialty roles).
Impact includes:
- UK Medical Royal Colleges and Medical Schools Council adopted the
selection methods
- Major UK policy impact and significant cost savings for the NHS, the
second largest organisation in the world
- Internationally, new methods were adopted due to the research outputs.
Underpinning research
The underpinning research for this case study was carried out at City
University London by Fiona Patterson (Professor of Organisational
Psychology 2003 to 2010, Visiting Professor 2010 to date), Dr Lara
Zibarras (Research Assistant/PhD 2005 to 2011, Lecturer since 2008) and
the Work Psychology Group, a research-led occupational psychology firm
based in the UK. The underpinning research builds on earlier work
conducted while Patterson was at the University of Nottingham, but focuses
on significant new research undertaken at City from 2003. The research was
conducted in partnership with senior medical figures such as directors of
Postgraduate General Practice (GP) Education (e.g., Professors Simon Plint
and Bill Irish, Former Chair and Chair, GP National Recruitment Office)
and Professor Paul O'Neill (Medical Schools Council); and key NHS
stakeholders (Royal Colleges and Deaneries) to design and evaluate
improvements to medical healthcare selection processes. This partnership
developed over several years by focusing on the needs of the stakeholders.
Previously there had been little validated research investigating
selection issues in medical education and training. Most selection
entailed a panel interview focusing on technical and clinical
competencies, with limited focus on non-clinical attributes such as
communication, empathy, problem solving and team-working.
Research has been carried out in three areas: (1) Identifying selection
criteria for success in job roles (via job analysis); (2) Designing and
validating selection methods to measure these criteria; (3) Evaluation and
validation of selection processes (fairness, reliability, validity).
Identifying selection criteria
Job analysis techniques identified theoretical competency-based models of
the knowledge, skills, abilities and other attributes required for
successful performance (e.g., Patterson, Ferguson & Thomas, 2008).
Uniquely, these models included important non-cognitive attributes
(empathy, integrity) required by healthcare professionals. The models have
been tested through validity studies exploring the short- and long-term
performance of doctors (Koczwara et al, 2012) and showed that
non-cognitive attributes contribute to successful performance as a doctor.
These new competencies were assessed during the selection of all doctors
during annual selection processes.
Design and validation of selection methods
Patterson and her team have led the field, with extensive research to
design and implement best practice selection methodologies for selecting
healthcare professionals. Examples include:
- Design, implementation and validation of selection methodologies,
including the first application of Situational Judgment Tests (SJT) and
assessment centres in GP medical training (Patterson et
al. 2009b)
- Design, implementation and validation of an assessment centre to
select GPs (with the GP National Recruitment Office, see Patterson et
al 2005, 2010, 2011).
- Developing the Royal College of Physicians' selection test
methodology; including machine-marked job knowledge and SJTs for
Clinical Medicine(Patterson et al. 2009a)
Research findings provided evidence to introduce innovative methodologies
in medical selection such as SJTs and assessment centres. This has changed
the way doctors are selected via all medical Royal Colleges in the UK
(through the Academy of Medical Royal Colleges) and the Medical Schools
Council. For example, the "foundation programme" (the two years of
training that doctors are required to undertake following medical school)
now uses SJTs as part of its selection process.
Validation and evaluation of selection processes
Validation activities have included (1) predictive validation studies; (2)
a focus on candidate reactions and perceptions (i.e., face validity); and
(3) the development of the new concept political validity. Due to the high
stakes nature of medical selection, selection processes must be seen to be
fair by candidates and by other stakeholders who hold potentially diverse
views (political validity). Research outputs from this case study have
shown the process to be fair from the point of view of applicants and to
predict future job performance (Patterson et al. 2009
a, b).
References to the research
Patterson's research group has published more than 20 articles
underpinning this impact. The underpinning research is supported by an
exceptionally strong evidence base widely published in the medical (e.g.,
BMJ, Medical Education) and psychology literature (Journal of Applied
Psychology, International Journal of Selection & Assessment), all of
which are highly regarded peer-reviewed journals. In particular, Medical
Education is a leading international journal for health professionals; its
readership includes medical educators, teachers and researchers globally
so it reaches key stakeholders.
Development of theoretical models of important non-cognitive
attributes and skills required by doctors (e.g. empathy, communication and
integrity)
a. Patterson F., Ferguson, E., & Thomas, S. (2008) Using job
analyses to identify core and specific competencies for three secondary
care specialties: Implications for selection and recruitment. Medical
Education, 42 1195-204 10.1111/j.1365-2923.2008.03174.x
Design and validation of new selection methodologies used in
various contexts (e.g., GP, Core Medical Training, Obstetrics and
Gynaecology).
b. Lievens F. & Patterson F. (2011) The validity and
incremental validity of knowledge tests, low- fidelity simulations, and
high-fidelity simulations for predicting job performance in advanced level
high-stakes selection. Journal of Applied Psychology, 96 927-940 10.1037/a0023496
c. Patterson F., Carr V., Zibarras L., Burr B., Berkin L.,
Plint S., Irish B. & Gregory S. (2009). New machine-marked tests for
selection into core medical training: Evidence from two validation
studies. Clinical Medicine, 9(5) 417-420 http://www.ncbi.nlm.nih.gov/pubmed/19886098
d. Patterson F., Baron H., Carr V., Lane P. & Plint S. (2009)
Evaluation of three short-listing methodologies for selection into
postgraduate training: the case of General Practice in the UK. Medical
Education, 43 50-57 10.1111/j.1365-2923.2008.03238.x
This paper won the "Silver Quill Award" in 2013 for the most downloaded
research article in Medical Education in 2012.
Evaluation of selection processes — candidate perceptions and
political validity
e. Prideaux D., Roberts C., Eva K., Centeno A., McCrorie P.,
McManus C., Patterson F., Powis D., Tekian A. & Wilkinson D.
(2011) Assessment for selection for the health care professions and
specialty training: International consensus statement &
recommendations. Medical Teacher, 33(3) 215-23 10.3109/0142159X.2011.551560
f. Patterson F., Zibarras L., Carr V., Irish B. & Gregory S.
(2011) Evaluating candidate reactions to selection practices using
organisational justice theory. Medical Education, 45 289-297 10.1111/j.1365-2923.2010.03808.x
This research was initially funded by several grants:
Two funded PhD studentships (2003, Ruth Price, ESRC funded; 2005, Lara
Zibarras, City University London Psychology Department studentship).
(2008 to 2009) Fiona Patterson: South York and Humberside workforce
development, £150,000, New Competency-based Selection Systems for
Secondary Care.
(2011 to 12) Lara Zibarras was awarded £4,000 by the Work Psychology Group
Ltd for consultancy.
Details of the impact
Prior to this research, an unstandardised and unstructured approach to
selection and assessment of healthcare professionals existed and resulted
in an unfair and biased way of selecting individuals. The underpinning
research has identified clear criteria with which to identify and select
healthcare professionals and the evidence to implement better and
innovative selection and assessment methodologies (e.g., SJTs, Assessment
Centres) that are proven through validation and evaluation techniques to
identify the best individuals for the job roles.
The major impact of this research has been:
- Improved selection criteria: Patterson and her colleagues have
conducted several job analyses across a range of specialties, initially
with GPs and Surgeons, but extending to Obstetrics & Gynaecology;
Psychiatry; Public Health; Paediatrics; and Histopathology.
- Substantial improvement in the assessment and selection practices for
doctors through the design and widespread implementation of new
selection methodologies. The SJTs designed and validated by Patterson
and her colleagues have been used since 2006 for selection into
postgraduate training, including General Practice and Public Health.
SJTs have been piloted for other specialties including Surgery,
Radiology, Histopathology, Core Medical Training, Anaesthesia, Dentistry
and Acute Specialties. In addition, Patterson and her colleagues have
designed assessment centres which are used across different specialties:
Neurosurgery; Paediatrics; Public Health; and Obstetrics &
Gynaecologyg.
Patterson's profile has garnered invitations to advisory roles and
appointments. Initially, she was invited to become assessment and
psychometric advisor for the Royal College of General Practitioners, UK
(2005 to 2010) and for the Royal College of Surgeons, England (2006 to
2011). During this time she was also advisor to the UK Department of
Health and the Academy of Medical Royal Colleges on assessment and
selection (2006 to date). Patterson and her research team therefore worked
in close collaboration with Royal Colleges and the Department of Health to
develop and implement the selection methodologies. This meant that they
were not only relevant to the target audience but gained significant
buy-in from these key stakeholders.
As a result of Patterson et al.'s work in developing selection
and assessment methodologies, following publication of the Government
White Paper, Liberating the NHS, in 2010, she was invited by the Royal
College of GPs to direct research and evaluation to define knowledge,
skills, abilities and attributes required for future UK GPs. In May 2012
she was awarded an Honorary Fellowship by the Royal College of GPs in
recognition of her contribution to selection for GP specialty training. In
2010, Patterson was invited by Professor Sir John Tooke and the Medical
Schools Council to chair an International Expert Panel to review selection
for entry into foundation training in the UK (the point after medical
school when doctors do two years of generalist training). The work was
conducted in collaboration with the UK Medical Schools' Council and
resulted in the use of SJTs for selection into foundation training.h,i,j,k
The extensive work conducted in the UK led to Patterson being an invited
expert in the development of an international policy statement on
Assessment for Selection in Healthcare Professions; Ottawa Conference on
the Assessment of Competence in Medicine and Healthcare Professions,
Miami, USA, 2010. In addition, she was invited to be Selection/Assessment
Advisor to GP Education Training (GPET), Canberra, Australia (2009 to
date) and was invited to advise on selection methodology being adopted
internationally (e.g., Royal Colleges of different medical specialties,
Australasia; Australian GP Education & Training; Netherlands; Denmark;
and Asian-Pacific network including Malaysia and Indonesia).
Methodologies implemented as a direct result of the research include:
The beneficiaries of the work include the candidates going through the
selection process, the administrators of the process and key stakeholders
such as the Medical Royal Colleges and patients.
Candidates: Since better selection methods now exist, tens of
thousands of candidates (medical school students, doctors, nurses,
dentists etc.) now benefit from improved fairness and evidence- based
transparent selection processes which also have a positive influence on
the diversity of health-related professionalsl. Improvements to
the selection process also mitigate the longer-term costs of appointing
doctors who do not complete training or follow a medical career.
Recruiters/professional staff: Currently tens of thousands of
applicants go through selection processes each year. Taking General
Practitioners alone, 8,000 candidates sit the new assessments annually
(initially adopted in 2006). Selection using the new procedures represents
a significant annual cost saving for the Department of Health/NHS and time
saving for recruiters and professional services staff. Plint and Patterson
(2010) estimate the cost of the new machine- marked tests used in the GP
selection process as £20 per applicant, where the previous hand- marked
application form questions cost £50 per applicant. The changed approach
reduces costs by £240,000 annually. This saving is even more significant
when extended beyond General Practice to Surgery, Radiology,
Histopathology, Core Medical Training, Anaesthesia, Dentistry and Acute
Specialties.
Key Stakeholders: The underpinning research has provided evidence
to implement various selection criteria and assessment methodologies. This
has led to a major change in UK policy where healthcare professionals are
increasingly being assessed and selected using SJTs and assessment centresg,h,i,j,k,l
with beneficiaries including the Royal Colleges with whom the methods have
been developed and validated, including the Royal College of General
Practice, Royal College of Physicians, Deaneries (e.g., Kent, Surrey &
Sussex, Yorkshire and the Humber, Oxford) and the National Recruitment
Office for General Practice Training.
Patients and the general public: the quality of care for patients
has improved through the reform and improvement of medical education and
training. Identifying the right person for the job ensures better patient
outcomes and experiences.
The impact of this work is therefore far reaching and highly significant
in ensuring that appropriate staff are chosen for critical roles.
Sources to corroborate the impact
g. March 2010: Selection in Speciality Training. Report supporting the
move towards the use of assessment centres (refers to Patterson and
colleagues' work): www.gmc-uk.org/Final_Selection_report_PMETB_Board_March_2010__2_.pdf_34128269.pdf.
h. August 2011: Final Report of Improving Selection to the Foundation
Programme Project Group. Outlines the work conducted by Patterson and
colleagues with regards to: Job Analysis; and development and
piloting of situational judgement tests for foundation doctors:
www.medschools.ac.uk/SiteCollectionDocuments/Final%20Report%20of%20ISFP%20Project%20Group.pdf.
i. May 2012: Final report of Foundation recruitment exercise — Medical
Schools' Council report explaining introduction and use of SJTs for
foundation doctors, live implementation in 2013. www.medschools.ac.uk/AboutUs/Projects/studentfitnesstopractise/resources/Documents/Final_report_of_PRE_Full_Appendices.pdf and details of the SJT
development and validation: www.isfp.org.uk/AboutISFP/Pages/default1.aspx.
j. All applicants to Foundation Programme from August 2013 will complete
an SJT, based on Patterson et al's work: www.foundationprogramme.nhs.uk/pages/home/how-to-apply/SJT/EPM.
k. February 2013: First SJT for Foundation Programme applicants a
success: hailed an "effective and innovative way of testing applicants'
attributes". Professor Derek Gallen says, "I am delighted that
the SJT has been such a success in its first year as part of the
application..."
www.mddus.com/mddus/news-and-media/news/february-2013/first-sjt-test-%E2%80%9Ca-success%E2%80%9D.aspx.
l. February 2013 — General Medical Councils report argues that using best
practice selection methodologies will widen access and participation into
medical school. www.gmc-uk.org/Identifying_best_practice_in_the_selection_of_medical_students.pdf_51119804.pdf
— References Patterson and colleagues' work as best practice selection
methods.