Bristol research leads to better ways of evaluating schools and promoting learning, achievement and improvement in the UK and Internationally
Submitting Institution
University of BristolUnit of Assessment
EducationSummary Impact Type
SocietalResearch Subject Area(s)
Education: Curriculum and Pedagogy, Specialist Studies In Education
Summary of the impact
Since 2008, UK and overseas policies, practices and tools aimed at
evaluating and promoting
quality in schools and supporting student learning, attainment and
progress have been
profoundly influenced by research conducted at the University of Bristol.
The work began in
2001 in the Graduate School of Education; from 2005, the School's efforts
were complemented
by those of the Centre for Multilevel Modelling. The research has
generated original knowledge
about school performance measures and school, teacher and context factors
which promote
student learning. This knowledge has transformed government and
institutional policies and
practices. New improved methods of evaluating schools and interventions in
education (and
other sectors) have been demonstrated and widely disseminated, thereby
enhancing public
understanding of institutional league tables and facilitating the
scaling-up of new approaches
nationally. The development of statistical methodology and MLwiN software
and training has
enabled more rigorous and sensitive quantitative analysis of educational
datasets around the
world, as well as wider take-up of this methodology by non academics.
Underpinning research
The research comprises studies of educational quality, effectiveness and
improvement
pioneering innovative "value added" measures of school performance to
report original
knowledge on the nature and extent of school effectiveness in a range of
contexts.
Sophisticated (multilevel) methodology and software for statistical
modelling (MLwiN), have also
been developed, extended and utilised to provide new evaluation tools and
substantive findings.
The research has involved the creation of new and detailed longitudinal
datasets in the UK and
overseas (including measures of student academic and attitude outcomes) to
analyse, measure
and evaluate school performance as well as the influence of other levels
within education
systems (eg regions, within school departments). Thomas and Peng
(University of Bristol staff
since 2001), Goldstein and Steele (Bristol staff since 2005), Rasbash
(Bristol staff 2005-10) and
Leckie (Bristol staff since 2009) have built upon previous studies by
extending earlier datasets,
analyses and research. For example, the China 2009-2012 and Lancashire
1992-2006 large-scale
datasets were unique in collecting new regional student attainment
and related information
over 4-14 consecutive cohorts for the first time. The latter enabled time
trends in value-added
school performance to be examined over a longer period (14 years) than any
other studies
worldwide. The UK Department for Children, Schools and Families (DCSF)
national pupil
database (NPD) introduced in 2002 has also been extensively employed, as
well as other
national surveys.
Educational effectiveness and improvement underpinning research since
2001 includes:
- UK and/or overseas evidence of internal variations in school
effectiveness (eg
subject/departmental effectiveness; differential effectiveness for
different pupil groups or
curriculum stages), the impact of pupils moving schools and time trends
of school effects, as
well as evidence that national and regional differences exist in terms
of school effectiveness.
This demonstrates that effectiveness is best seen as a feature that is
outcome-, context- and
time-specific and indicates that school league tables have little to
offer as guides to school
choice. (Thomas, Goldstein, Leckie, Peng) [3][4][5][6].
- Evidence of the extent to which school input, process and context
factors link to school
effectiveness in China (Thomas, Peng) [6] and the impact of school
resources and parental
divorce on pupil attainment (Steele), thereby highlighting relevant
factors that need to be
considered when evaluating schools.
Methodological and software development underpinning research since
2005 includes:
- New equations have been created for predicting the group effects in
repeated cross-section
multilevel models in order to predict accurately how schools are likely
to perform in future. In
addition, new (simulation-based) graphical approaches have been
developed for
communicating the statistical uncertainty of predicted group effects in
multilevel models.
These have helped in conveying to parents whether the academic
performances of several
local schools can be statistically separated from one another (Leckie,
Goldstein) [3].
- Multilevel models for complex, non-hierarchical data structures have
been developed to
model correctly the effects of schools and neighbourhoods on pupils'
academic progress
when students are changing schools and moving neighbourhoods during
their schooling
(Leckie). Multilevel models for segregation and inequality have also
been developed, for
example to measure the extent to which social or ethnic segregation of
students across
schools has significantly changed over time (Leckie, Goldstein).
- Modelling of multivariate data with different response types at
several levels and procedures
for handling correlated measurement and misclassification errors.
(Goldstein) [1][2].
References to the research
[1] Goldstein, H. (2010) Multilevel Statistical models. 4th
Edition. Whiley. [Citations 2008-13:
2,480] Listed in REF2
[3] Leckie, G. and Goldstein, H. (2009) The limitations of using school
league tables to inform
school choice, Journal of the Royal Statistical Society: Series A,
172, 835-851. Listed in
REF2
[5] Thomas, S. (2001) Dimensions of Secondary School Effectiveness:
Comparative Analyses
Across Regions, School Effectiveness & School Improvement Journal.
Vol 12(3): 285-322.
http://dx.doi.org/10.1076/sesi.12.3.285.3448
[6] Thomas, Sally et al (2012) 学校效能增值评量研究. [Research on Value Added
Evaluation
of School Effectiveness].Jiaoyu yanjiu. [Educational research]. 33 (7),
pp. 29-35 Beijing:
Zhongyang Jiaoyu Kexue Yanjiusuo. Listed in REF2
Related research grants supporting and evidencing quality of
publications
These were awarded, for example, by the ESRC and the Department for
International
Development following a rigorous process of review by the respective
agencies.
• Steele, F., Goldstein, H. and Leckie, G. (2011-2013) Longitudinal
Effects, Multilevel
Modelling and Applications (LEMMA III), ESRC: 750K
• Steele, F., Goldstein, H. and Leckie, G. (2008-2011) ESRC:
STRUCTURES for Building,
Learning, Applying and Computing Statistical Models (LEMMA II),
ESRC: 700K
• Rasbash, J., Steele, F. and Thomas, S. (2005-2008) Learning
Environment for Multilevel
Modelling Applications (LEMMA), ESRC: 650K
• Goldstein, H. (2003-2005) Developing Multilevel Models for
Realistically Complex Social
Science Data, ESRC: 300K
• Thomas, S. and Peng, W.-J. (2010-2013) Improving Teacher
Development and Educational
Quality in China [ITDEQC], ESRC/DfID: 500K
• Thomas, S. and Peng, W.-J. (2008-2011) Improving Educational
Evaluation and Quality in
China [IEEQC], ESRC/DfID: 250K
• Thomas, S. and Peng, W.-J. (2001-2007) Lancashire LEA: Value Added
Project (this was a
continuation of a project that began in 1992 and moved to the University
of Bristol): 200K
• Thomas, S. et al (2002-2004) Effective professional learning
communities, DfES: 600K
Details of the impact
The impact of this research on a wide range of beneficiaries (policy,
practitioner, NGO, public)
worldwide has reach and is significant in two ways since 2008. First, it
informed and
underpinned new policy and, practice. Second, it has generated
methodological developments in
key areas. Each of these is outlined below.
Impact on UK and international policy and government thinking relating
to measuring
educational effectiveness and school performance
In the UK, Goldstein and Thomas' research [1][3][4][5] has contributed
evidence to inform and
influence key national policies such as the utility of school
self-evaluation, national pupil
databases (eg the Pupil Level Annual Schools Census (PLASC)),
contextualised value-added
measures of school performance (introduced by the DCSF in 2006 and almost
identical to
measures used in Lancashire LEA, 1993-2006) and separate value-added
measures for different
student groups (introduced by the DfE in 2011). The research has also
promoted the use of a
wider range of outcomes and measures by the DfE/DCSF/OFSTED/LSC [a][b].
Goldstein's
research was referenced as underpinning evidence in a 2012 Northern
Ireland Assembly
Research and Information Service Research Paper, "Providing information on
pupil and school
performance". He was also a member of the UK government select committee
invited seminar
to advise on Accountability and League Tables (2013). In addition,
Goldstein co-directs the
PLASC Users Group, set up in 2006 with support from the DfE. Since then
regular meetings
have been held with 40-plus participants, involving researchers who have
used, or are interested
in using the PLASC/NPD datasets. Civil servants from the DfE also attended
and frequently
returned to report on current developments and participate in discussions
[c].
Advice on evaluating educational quality has also been frequently sought
by policymakers
internationally, which demonstrates the reach and significance of this
work. This has resulted in
citations in OECD publications that provide guidance to member states [a],
as well as invitations
to speak in many international contexts, often introducing new ideas on
school evaluation to non-academics
for the first time and elucidating the input, process and
context factors associated
with school effectiveness (eg Goldstein (2011) Queensland University of
Technology [attended
by Australian Government officials]; Thomas (2010) Chilean Ministry of
Education; Thomas
(2008) EU education conference for the French Presidency).
Impact on UK and international educational and school practices and
public
understanding relating to evaluating educational quality, improving
school effectiveness
and best practice in school self-evaluation and use of data
Leckie and Goldstein's research [3] demonstrates the limitations of using
the government's
school league tables to inform school choice. Since 2008 this has promoted
stakeholders' and
the public's understanding of the problems with league tables through
widespread national and
international communication to non-academics via popular articles and
other media, including
interviews for the BBC Radio 4 programmes "Analysis" and "The Learning
Curve", and articles in
the Financial Times, the Daily Telegraph and the Times Education
Supplement. Goldstein,
Leckie and Thomas' work on critiquing school performance measures
demonstrates impact in
terms of both reach and significance. It has been cited by numerous UK
NGOs (eg NUT, RSA,
RSS, the Institute for Government) and overseas NGOs and governments
seeking to evidence
the complexity, dangers and limitations of school performance measures;
thereby influencing
public thinking and new policy development on educational accountability
and improvement
initiatives [e][f][g].
Pilot school evaluation studies using value-added techniques have been
conducted in UK
and several countries worldwide (eg China, Africa) [3][4][5][6] and this
has raised the awareness
of policymakers and teachers and resulted in new evaluation practices by
schools [d]. Professor
Xiaoman Zhu, President of the National Institute of Education Sciences
(NIES) until 2010,
Ministry of Education, Beijing, has emphasised the contribution of the
IEEQC research [6] and
collaboration between the University of Bristol and NIES to better
understanding the concept of
educational quality and evaluation methods in the Chinese context, as well
as to capacity-building
for NIES researchers [h]. Mr Xiaoqiang Ma, NIES and IEEQC project
researcher has
gone on to publish a 2012 book, "Value added evaluation: a new
perspective on school
evaluation". The head teacher of a Chinese senior secondary school
participating in the IEEQC
project stated that "This approach [value added method] is particularly
good. From next year
[2009], we will use this approach to evaluate our key schools. That is,
taking account of the
intake of senior high school year 1 when comparing schools with college
entrance examination
results. We [as a school] particularly welcome the method" [i].
Thomas has also applied methods
to evaluate schools' performance alongside other key elements in creating
and sustaining
English schools as professional learning communities (PLCs), resulting in
new tools used by
school leaders to develop their schools as PLCs.
Expansion of use of quantitative methods in educational research and
social sciences
more broadly, which in turn shapes research that influences policy and
practice
The impact of new statistical methodology [1][2] has been achieved through
further development
of the user-friendly MLwiN software, the REALCOM-Impute software for
multiple imputation and
through dissemination and training events. Since 2008 the MLwiN software
(available free to UK
academics), together with extensive user guides, has been downloaded by
3,846 new users and
it has been purchased by 5,518 overseas academics and 613 non-academic
users. Moreover,
67 organisations have purchased MLwiN site licences (50 users) since 2008;
of these 8
organisations hold extension licences (250 users).
The CMM website is widely acknowledged as the premier resource for
research and
training in multilevel modelling. There are around 1,100 page-loads and
360 unique visitors per
day (65% from outside the UK). The LEMMA Virtual Learning Environment
(launched in April
2008) has around 10,000 registered users, of whom 70% are international
and 14% are non-academic,
thereby demonstrating the reach and significance of the research
impact. Some
training events are targeted at non-academics, eg a session on multilevel
modelling given at
Ofsted in 2008 (Steele). UK non-academic beneficiaries and users include
the Departments of
Education, Health, and Work and Pensions, the Scottish Executive and the
Office for National
Statistics [b]. For example, Trevor Knight (consultant statistician to
DfE) reported in July 2010
that MLwiN was used by DfE statisticians to calculate Contextual Value
Added (CVA) and other
value-added school performance measures, employed as an integral part of
the OFSTED school
inspection process and used to construct the Learning Achievement Tracker
— a tool for schools
and FE colleges to appreciate progress made by students since the end of
compulsory
schooling. MLwiN was also used in the National Evaluation of the Sure
Start Local Programmes
for the DfE [j], the NatCen (2009) report for the Department for
Environment, Food and Rural
Affairs on educational attainment in rural areas and by Higher Education
Funding Council for
England and others in conducting new analyses to support higher education
institutions in
developing "contextualised" admissions policies and equality and diversity
policy for REF2014
submissions. Overseas non-academic MLwiN users include Statistics Canada,
Statistics
Norway, the Netherlands Bureau of Statistics, UNESCO and the World Health
Organisation.
Sources to corroborate the impact
[a] Evans, H. (2008) Value-Added
in English Schools. A DFE paper updated from the OECD
Project on the Development of Value-Added Models in Education Systems and
OECD (2008)
Measuring
Improvements in Learning Outcomes: Best Practices to Assess the
Value-added
of Schools. Organisation for Economic Co-operation and Development and Spanishtranslation
cites work by Goldstein and Thomas regarding use/methodology of VA
measures.
[b] Director General, Monitoring and Assessment, UK Statistics Authority
provided information
(June 2013) about influence of University of Bristol research on
government and public
understanding and UK policy on school evaluation.
[c] PLUG
website lists 9 seminars presented by non-academic/government
researchers hosted
by PLUG since 2008.
[d] Executive Headteacher, Bradley Stoke Community School and Abbeywood
Community
School provided information (September 2013) about influence of University
of Bristol
research on teachers understanding and good practice in student and school
evaluation and
links to improved student outcomes.
[e] Wildman, R (2011) Beware
of the Misleading Means and Measures. Chapter 3 in
Transformation Audit (The Inclusive Economies Project). Policy and
Analysis Unit of the
Institute for Justice and Reconciliation (IJR), South Africa cites work by
Leckie & Goldstein
on dangers of school league tables. IJC is dedicated to researching and
influencing policy
debates around the issue of socio-economic justice in South African and
elsewhere on the
continent.
[f] Mulgan, R. (2012) Transparency
and Public Sector Performance. Report prepared for the
Australia and New Zealand School of Government cites work by Leckie &
Goldstein on
dangers of school league tables.
[g] Cipollone, P. et al. (2010) Value-Added
Measures in Italian High Schools: Problems and
Findings. Bank of Italy Temi di Discussione (Working Paper) No. 754. cites
work of Thomas
regarding use and methodology of value added measures.
[h] 2011 Celebration Book on 70th anniversary of National Institute of
Educational Sciences,
Ministry of Education, Beijing 70周年所庆纪念文集. Text emphasises the influence
of
University of Bristol research on understanding school evaluation in
Chinese context (pg 42).
[i] ESRC Research Impact Evaluation Report of Improving Educational
Evaluation and Quality
in China (IEEQC) project: overall rating "Outstanding", January 2013.
[j] NESS Team (2010). The impact of
Sure Start Local Programmes on five years olds and their families.
DfE Research Report RR067. states use of multilevel modeling analysis in
conducting the evaluation (page 23)