Log in
Within this case study we present the TrOWL technology developed at the University of Aberdeen that enables more efficient and scalable exploitation of semantic data. TrOWL and its component algorithms — REL, Quill and the Aberdeen Profile Checker — have had non-academic impact in two key areas. With respect to practitioners and professional services, the technology has enabled the introduction of two important World Wide Web Consortium (W3C) standards: OWL2 and SPARQL 1.1. This has led to impact in the way that many companies work, across a range of sectors. Further, through partnership with specific companies, the use of TrOWL has changed the way they operate and the technical solutions they provide to clients. These collaborations have led to economic impacts in companies such as Oracle in "mitigat[ing] the losses of potential customers", and IBM in "using the TrOWL reasoning infrastructure in [their] Smarter Cities solutions".
Ontologies are used to describe the meaning of terms in a domain. Manchester has had a leading role in the design of ontology languages, algorithms and tools. Through standardization, algorithm development and tool creation, we have significantly influenced the uptake of the Ontology Web Language (OWL) and Semantic Web Technologies by public service providers and industry. For example, the NCI thesaurus and SNOMED CT are medical terminologies in OWL; specialised semantic web companies such as Clark & Parsia, Racer Systems and TopQuadrant provide semantic technologies and services that build on OWL; and companies such as Oracle and B2i Healthcare include tool support for OWL.
Extracting information and meaning from natural language text is central to a wide variety of computer applications, ranging from social media opinion mining to the processing of patient health-care records. Sentic Computing, pioneered at the University of Stirling, underpins a unique set of related tools for incorporating emotion and sentiment analysis in natural language processing. These tools are being employed in commercial products, with performance improvements of up to 20% being reported in accuracy of textual analysis, matching or even exceeding human performance (Zoral Labs). Current applications include social media monitoring as part of a web content management system (Sitekit Solutions Ltd), personal photo management systems (HP Labs India) and patient opinion mining (Patient Opinion Ltd). Impact has also been achieved through direct collaboration with other commercial partners such as Microsoft Research Asia, TrustPilot and Abies Ltd. Moreover, international organisations such as the Brain Sciences Foundation and the A*Star Institute for High Performance Computing have realised major impact by drawing upon our research.
Aston University researchers developed and maintain the Uncertainty Markup Language (UncertML) for quantitative specification and interoperable communication of uncertainty measures in the Web. It is the only complete mechanism for representation of uncertainty in a web context. UncertML has been:
- Used in policy and decision making by UK (Food and Environment Research Agency) and international (European Commission) government agencies, and many research / industrial institutes;
- Presented at industrial /technical workshops, leading to ongoing international collaborations with bodies such as national space agencies (ESA and NASA) and government data providers;
- Accepted as a discussion paper for formal standardisation by the Open Geospatial Consortium;
- Chosen by independent data providers for efficient sharing of complex information and rigorous risk analysis across scientific domains such as pharmacy, global soil mapping and air quality.
Our research has enabled archaeological professional and commercial organisations to integrate diverse archaeology excavation datasets and significantly develop working practices. Commercial archaeological datasets are usually created on a per-site basis structured via differing schema and vocabularies. These isolated information silos hinder meaningful cross search and comparison. As the only record of unrepeatable fieldwork, it is essential that these data are made available for re-use and re-interpretation. As a result of the research, the Archaeology Data Service, English Heritage, the Royal Commissions on the Ancient and Historical Monuments of Scotland and Wales have published as Linked Data important excavation datasets and national vocabularies that can act as hubs in the web of archaeological data.
Based in the School of English, the Research and Development Unit for English Studies (RDUES) conducts research in the field of corpus linguistics and develops innovative software tools to allow a wide range of external audiences to locate, annotate and use electronic data more effectively. This case study details work carried out by the RDUES team (Matt Gee, Andrew Kehoe, Antoinette Renouf) in building large-scale corpora of web texts, from which examples of language use have been extracted, analysed, and presented in a form suitable for teaching and research across and beyond HE, including collaboration with commercial partners.
COnnecting REpositories (CORE) is a system for aggregating, harvesting and semantically enriching documents. As at July 2013, CORE contains 15m+ open access research papers from worldwide repositories and journals, on any topic and in more than 40 languages. In July 2013, CORE recorded 500k+ visits from 90k+ unique visitors. By processing both full-text and metadata, CORE serves four communities: researchers searching research materials; repository managers needing analytical information about their repositories; funders wanting to evaluate the impact of funded projects; and developers of new knowledge-mining technologies. The CORE semantic recommender has been integrated with digital libraries and repositories of cultural institutions, including the European Library and UNESCO. CORE has been selected to be the metadata aggregator of the UK's national open access services.
The impact of this work stems from the provision of better quality information models, and is manifest via: (a) reduced cost through improved reuse and less rework; (b) improved system interoperability; and (c) enhanced assurance and checking that information requirements are supported by the resultant systems. The approach has been applied in commercial environments, such as Shell (UK), where it has reduced development costs by up to 50% ($1m in one case). It has also been applied in the defence environment, forming a part of underpinning standards currently being implemented by the UK and Swedish Armed Forces.
The Airbus company has used OntoREM, a semi-automated methodology developed at UWE Bristol, for developing systems' requirements specifications and improving the quality of such specifications. This has saved Airbus [text removed for publication] cost and time to develop aircraft operability requirements for wing design and industrialisation in two different aircraft programmes — with a significant increase in requirements reusability. It has enabled improved assessment of risk in advance of a project's start through prior estimation of the cost and time of developing requirements. This has allowed reliable forecasts and scheduling, and better management of the expectations of a project's key stakeholders.
GATE (a General Architecture for Text Engineering—see http://gate.ac.uk/) is an experimental apparatus, R&D platform and software suite with very wide impact in society and industry. There are many examples of applications: the UK National Archive uses it to provide sophisticated search mechanisms over its .gov.uk holdings; Oracle includes it in its semantics offering; Garlik Ltd. uses it to mine the web for data that might lead to identity theft; Innovantage uses it in intelligent recruiting products; Fizzback uses it for customer feedback analysis; the British Library uses it for environmental science literature indexing; the Stationery Office for value-added services on top of their legal databases. It has been adopted as a fundamental piece of web infrastructure by major organisations like the BBC, Euromoney and the Press Association, enabling them to integrate huge volumes of data with up-to-the-minute currency at an affordable cost, delivering cost savings and new products.