Log in
Stroke and other forms of brain injury often result in debilitating communication impairments. For example, patients with acquired apraxia of speech (AOS) experience difficulties that affect their capacity to verbally express thoughts and needs. Such individuals have benefitted from the development of a novel computerised treatment — "Sheffield Word" (SWORD). Patients who took part in clinical trials showed improvements in aspects of speech that were impaired after stroke. SWORD is now used by healthcare teams worldwide, providing benefits to a large patient population. The SWORD computerised treatment is convenient to use at home, fosters users' autonomy, and delivers higher treatment doses than possible through traditional clinical sessions. Clinicians who treat AOS have also benefitted through education, training and access to online materials about SWORD which were provided by the research team.
Literacy and language work reciprocally to support children's learning and attainment. Professors Morag Stuart and Julie Dockrell, in complementary and sometimes co-authored work, have had a substantial influence on the way children of all abilities and with a range of needs are taught to read and develop their broader oral language and literacy skills. Their work influenced national programmes in England on early reading, dyslexia and spoken language (National Literacy Strategy and National Curriculum) and speech, language and communication needs (the Better Communication Research Programme). Their findings provided educational professionals and speech and language therapists with the conceptual understanding that underpins good practice, and helped equip teachers to identify difficulties and intervene.
Research in robust speech enhancement and audio-visual processing has led to impact on a range of different fronts:
(i) Collaboration with CSR, a leading $1 billion consumer electronics company, has shaped its R&D research agenda in speech enhancement, has inspired ideas for new product improvements, and has helped establish Belfast as an audio research centre of excellence within the company.
(ii) Our technology has changed the strategic R&D direction of a company delivering healthcare monitoring systems, with potential for multi-million pound savings in NHS budgets.
(iii) Audio-visual speech processing research has led to a proof-of-concept biometric system, Liopa: a novel, robust and convenient person authentication and verification technology exploiting lip and facial movements (www.liopa.co.uk). A start-up company is in an advanced stage of being established to commercialise this product. The product and commercialisation strategy was awarded First Prize in the 2013 NISP Connect £25K entrepreneurship competition in the Digital Media and Software category. The first commercial partner for Liopa has been engaged.
(iv) A system-on-chip implementation of a version of our speech recognition engine, which was developed through an EPSRC project, was awarded first prize in the High Technology Award in the 2010 NISP £25K Awards competition, and contributed to the founding of a spin-out company, Analytics Engines (www.analyticsengines.com).
Research at Newcastle on speech language and communication needs (SLCN) has significantly extended the range and quality of evidence in this field. The research is significantly informing high level policy makers, and has been used to set up an accessible database which is having impact on the practice of a range of different professionals. It has contributed to a new programme of work which has produced positive outcomes for schools, children and young people and has been used as an evidence base to secure charity funding for a third sector organisation.
The impact is primarily in Public Health. It mainly concerns the adoption of and demand for a speech research technology, Electropalatography (EPG), for clinical diagnosis and treatment of speech disorders. Our continuing long-term and interdisciplinary research into EPG has increased our impact in this census period from the previous RAE2008, during which time the UOA had already been awarded a Queen's Anniversary Prize (2002) for working towards the clinical application of speech science.
Financial Support from the charitable sector and the NHS for the training of classroom assistants and SLTs in EPG therapy is highlighted, along with user testimonials, unmet demand, and small-scale provision of the therapy.
Speech Graphics Ltd is a spinout company from the University of Edinburgh, building on research into the animation of talking heads during 2006-2011. Speech Graphics' technology is the first high fidelity lip-sync solution driven by audio. Speech Graphics market a multi-lingual, scalable solution to audio-driven animation that uses acoustic analysis and muscle dynamics to drive the faces of computer game characters accurately matching the words and emotion in the audio. The industry-leading technology developed by Speech Graphics has been used to animate characters in computer games developed by Supermassive games in 2012 and in music videos for artists such as Kanye West in 2013.
This impact case study provides evidence of economic impacts of our research because:
i) a spin-out company, Speech Graphics Ltd, has been created, established its viability, and gained international recognition;
ii) the computer games industry and the music video industry have adopted a new technology founded on University of Edinburgh research into a novel technique to synthesize lip motion trajectories using Trajectory Hidden Markov Models; and
iii) this led to the improvement of the process of cost-effective creation of computer games which can be sold worldwide because their dialogue can be more easily specialised into different human languages with rapid creation of high-quality facial animation replacing a combination of motion capture and manual animation.
Professor Usha Goswami's research on reading development and dyslexia and in the relatively new field of educational neuroscience has gained her international renown as an expert in this field that brings together research in neurobiology and education. Her literacy research, which she and her group have undertaken in the Departments of Education and Psychology in the University of Cambridge, has focused on cross-linguistic factors underpinning reading development and developmental dyslexia, producing innovative data. She has also been an influential critic of the Government's focus on `synthetic phonics'. During this REF impact period, Professor Goswami's work has had significant impact on UK Government educational and other public policy, on public debate and understanding about reading and dyslexia, and on practitioners and services concerned with written material in every language.
Professor Sue Roulstone's research has contributed significantly and widely to a growing political appreciation of the crucial role that language plays in children's development, socially, emotionally and educationally. The research has contributed evidence to underpin policy initiatives to monitor language as an indicator of successful child development, to support the importance of children's early language development and to ensure services for those with impairments are evidence-based. These initiatives have in turn impacted upon the development of services and the use of parent and child perspectives in service design and development.
Edinburgh's research in multilingual speech synthesis has had clinical and commercial impact, and has resulted in a large and diverse community of users.
Clinical applications: Our research has enabled the construction of natural-sounding, personalised synthetic voices from recordings of speech from people with disordered speech due to conditions such as Parkinson's disease or Motor Neurone Disease. These synthetic voices are used in assistive technology devices that allow sufferers of these conditions to communicate more easily and effectively.
Commercial take-up: Our research has achieved commercial impact through the licensing of technology components, and through the activities of start-up companies.
Community of users: The Festival Speech Synthesis System (v2.1 released in November 2010) is a complete open-source text-to-speech system released under an unrestrictive X11-type license, and is distributed as part of many major Linux distributions.
One of the world-leading systems for large-vocabulary Automatic Speech Recognition (ASR) has been developed by a team led from the University of Sheffield. This system, which won the international evaluation campaigns for rich speech transcription organised by the US National Institute for Standards and Technology (NIST) in 2007 and 2009, has led directly to the creation of one spin-out, been largely instrumental in the launch of a second, has had significant impact on the development and growth of three existing companies, and has made highly advanced technology available free for the first time to a broad range of individual and organisational users, with applications including language learning, speech-to-speech translation and access to education for those with reading and writing difficulties.