REF 2014: Discipline Mattered in How Impact Calculated

This is how they calculated in 1877. Things changed in 2014.
In our latest Digital Research Report on impact evidence we showed that corroborating documents included in impact case studies and submitted to REF 2014 varied in type by subject panel. The chart reproduced below shows how media items (like television programmes, YouTube videos, news articles …) were used more in panel D (Arts and Humanities) than in any other panel, whilst reports appeared in panel A (Biological Sciences and Medicine) the most.

LSE Impact logo
This article by Tamar Loach and Martin Szomszor originally appeared on the LSE Impact of Social Sciences blog as “What impact evidence was used in REF 2014? Disciplinary differences in how researchers demonstrate and assess impact” and is reposted under the Creative Commons license (CC BY 3.0).
This kind of analysis is broad-brush and not particularly surprising – we have counted clinical guidelines as reports and media pieces are surely closer to the research carried out in panel D than in any other panel. What it does do is show that by allowing for flexibility in the way impact is reported (case studies could cite any evidence that corroborated impact so long as it was auditable) diversity in approach and content is encouraged and realized. Here we show this at the level of main panels, but the variety and multiplicity of impact stories in the full set of case studies is backed up by a huge array of impact evidence – evidence includes  technical documentation on commercial websites, social media and audience responses to public activities and a huge number of URLs pointing to unknown resources.

The percentage of REF 2014 Impact Case Studies, by main panel, that contain categorized types of evidence. (Source: Loach, Tamar; Adams, Jonathan; Szomszor, Martin (2016): Digital Research Report: The Societal and Economic Impacts of Academic Research – International perspectives on good practice and managing evidence. Digital Science. figshare.
Not all of these evidence types are represented in the aggregate numbers in the bar chart – we classified strings of text representing pieces of evidence using a collection of keywords and patterns to group evidence pieces into commonly occurring categories, but this inevitably misses out the less common evidence types. This obscures finer grained diversity, but with the evidence types we are able to categorize, we can look at the association between use of evidence in the various categories and the scores received by submissions in peer review. We are also limited by the public availability of scores, which are available at the Unit of Assessment/Institution level only. This means that we know the percentage of 4*,3*,2*,1* case studies that a university submitted to a particular panel, but we can’t say what any particular case study scored.

At the high level we are able to access the association between scores and evidence types can be tested. The correlations between the amount of evidence of a given type, and the grade point average (GPA) score for a set of case studies is shown in the colored table below (again reproduced from our recent report). This looks at average behavior and obscures outliers (which may themselves be more interesting than group trends) but we can see differences across panels, and use these differences to point to the benefit of flexibility in assessment systems. We saw that arts and humanities case studies were those most likely to include media as evidence of impact. But within Arts and Humanities, use of media does not affect the GPA associated with a case study set in a significant way.

In contrast, the reports used in nearly 40 percent of Medical and Health case studies have a statistically significant positive correlation with GPA. Although not a causal link, it does seem that use of reports as impact evidence is associated with highly scoring impact case studies across the panels. On the other hand testimonials seem to be associated with higher scores in Arts and Humanities but are negatively related to score in the Medical Sciences.

Spearman correlation between the indicative score and the amount of various types of evidence. There is a column for each subject panel. (Source: Loach, Tamar; Adams, Jonathan; Szomszor, Martin (2016): Digital Research Report: The Societal and Economic Impacts of Academic Research – International perspectives on good practice and managing evidence. Digital Science. figshare.
A value of 1 implies maximal positive correlation, 0 no correlation, and -1 a maximal inverse correlation. The values in bold are significant (p value < 0.05, where the null hypothesis is that the indicative score and the amount of a given evidence type are uncorrelated.)

In our report we conclude that the case study format allowed a variety of evidence to be displayed, and that these types appeared – and were assessed – differently across research communities. The evidence ontology we use neglects less common forms of evidence, and the limited availability of scores means we’re really looking at averages and associating them with individual case studies. Nevertheless, we see different practices across broad subject categories. We also know that there is great variety in the impact case study database, and that there are multifarious differences  across individual case studies. We use this variety to argue that the surprises and outliers in the impact case studies provide the most value and interest, and to encourage diversity in content, certainly not ruled by a formulaic impression of what good impact or good impact evidence looks like.

0 0 votes
Article Rating

Tamar Loach and Martin Szomszor

Tamar Loach is research metrics analyst at Digital Science. Her background is in theoretical physics, and she has undertaken research projects both in this area and on alternative research metrics with the Complexity and Networks group, all at Imperial College London. She currently works with all kinds of research data, and helped develop the impact case study website and a synthetic analysis of the impact case studies for HEFCE after REF 2014. Martin Szomszor is head of data science at Digital Science. He was previously Deputy Head of Centre at the City eHealth Research Centre (2009-2011) where he led research on the use of social media in epidemic intelligence.

Notify of

This site uses Akismet to reduce spam. Learn how your comment data is processed.

Inline Feedbacks
View all comments
Would love your thoughts, please comment.x