Skip to content

gdanezis/refreerank

Repository files navigation

refreerank

A project using the UK REF data sets to collaboratively re-rank the publishing venues and UK Computer Science departments.

Objective

The UK Research Excellence Framework (REF) that took place in 2014 was a peer-reviewed process that attempted to objectively rank UK Higher Education academic departments based on the quality of their research. Academic staff from each department were asked to submit their top four outputs from between 2009 and 2014 for peer review. The submitted pieces were reviewed -- at great expense (of time and money) -- and the departments were ranked based on the opinions of the reviewers. These submissions, as well as the outcomes, can be accessed from the REF 2014 website.

Despite being peer-reviewed, the REF process was not uncontroversial: A number of top UK scientists devoted a very large amount of time to evaluating the thousands of submitted works; the final judgments of quality depend on what is still a relatively small college of experts and may still contain biases; the task of selecting the experts to sit on the expert panels implies a pre-existing judgment of research quality. The final rankings produced were widely discussed.

Was all of this expense necessary?

We note that the in mere act of selecting papers for review, authors and institutions are inherently making a value judgment about the quality of their recent work. Thus, it can be assumed that, for any particular author, the works selected tend to be of a higher quality than works that they did not choose to include.

This selection enables us to perform an independent evaluation of publication venues, and hence academic institutions. By aggregating over the subjective quality decisions of tens of institutions and hundreds of authors, we create a participatory ranking system. We show that, when applied to the field of Computer Science, this peer-to-peer ranking reproduces some of the findings of the REF, whilst also creating some surprises.

What we did

We matched the works submitted for evaluation with publications recorded in the dblp database, an open-source record of computer science publications, using custom fuzzy matching techniques. From this database, we were also able to identify other published work that authors chose not to submit to the REF process between 2008 and 2014. Based on the combination of these selected and unselected works, we evaluated various publication venues (conferences and journals) to see whether research presented at some venues was systematically selected by authors for inclusion in the REF submission over research that they presented in other (presumably less prestigious) venues.

To estimate the venue quality, we made flow graphs describing researchers' quality judgments about different publication venues: given a set of selected and unselected publications at specific venues, we build a directed graph from all the venues of unselected papers to all the venues of selected ones. We then compute the stationary distribution of this directed graph, representing the probability of reaching a venue after a long random walk through this graph. Heuristically, the steps follow the subjective quality judgments and higher quality venues are more likely to be reached.

Once we estimate a quality score for venues we use them as proxies for judging the quality of research per department: for each author put forward we chose their best 12 papers evaluated by venue, and aggregate their score into the score of the institution. We experimented with selecting the top-4 or even all papers, without any major effect on most rankings. We call this the Peer Score, and the resulting ranking the Peer Rank. We publish the full procedure for computing them.

Results

The following table summarizes the Peer Rank and Peer Score of the top-75 UK Computer Science departments. We also compare the Peer Rank with the REF Output rank, and provide the difference in ranks. We observe some departments are not majorly re-ranked, while others see their position change significantly. As a reminder, Peer Rank and Score are computed using the submissions alone (see above) -- we do not use the judgments and outcomes of the REF peer-review.

Peer Rank (Score) REF Rank (Diff.) Computer Science Department
1 (0.26) 2 (+1) University College London
2 (0.24) 5 (+3) University of Oxford
3 (0.22) 13 (+10) University of Edinburgh
4 (0.16) 22 (+18) University of Nottingham
5 (0.15) 4 (-1) Imperial College London
6 (0.12) 6 (+0) King's College London
7 (0.12) 34 (+27) University of Southampton
8 (0.11) 30 (+22) University of Glasgow
9 (0.10) 8 (-1) University of Cambridge
10 (0.10) 3 (-7) University of Liverpool
11 (0.10) 19 (+8) Newcastle University
12 (0.09) 12 (+0) Lancaster University
13 (0.09) 9 (-4) University of Manchester
14 (0.09) 20 (+6) University of Birmingham
15 (0.08) 23 (+8) University of Bristol
16 (0.07) 49 (+33) City University London
17 (0.07) 17 (+0) University of York
18 (0.07) 1 (-17) University of Warwick
19 (0.06) 15 (-4) Swansea University
20 (0.06) 10 (-10) Queen Mary University of London
21 (0.06) 29 (+8) Aberystwyth University
22 (0.06) 11 (-11) Royal Holloway, University of London
23 (0.05) 38 (+15) University of Bath
24 (0.05) 40 (+16) Brunel University London
25 (0.05) 14 (-11) University of St Andrews
26 (0.05) 24 (-2) Cardiff University
27 (0.05) 37 (+10) Open University
28 (0.05) 7 (-21) University of Sheffield
29 (0.05) 25 (-4) University of Leicester
30 (0.05) 66 (+36) Middlesex University
31 (0.05) 26 (-5) University of Durham
32 (0.05) 52 (+20) University of Ulster
33 (0.04) 21 (-12) Birkbeck College
34 (0.04) 31 (-3) University of Leeds
35 (0.04) 39 (+4) University of Kent
36 (0.04) 45 (+9) Heriot-Watt University
37 (0.04) 35 (-2) University of Essex
38 (0.04) 42 (+4) University of Lincoln
39 (0.03) 57 (+18) De Montfort University
40 (0.03) 41 (+1) Queen's University Belfast
41 (0.03) 36 (-5) University of Dundee
42 (0.03) 56 (+14) University of Strathclyde
43 (0.03) 33 (-10) University of Sussex
44 (0.03) 32 (-12) University of Aberdeen
45 (0.03) 68 (+23) University of Bedfordshire
46 (0.03) 53 (+7) University of Surrey
47 (0.02) 28 (-19) Oxford Brookes University
48 (0.02) 47 (-1) University of Stirling
49 (0.02) 43 (-6) Teesside University
50 (0.02) 75 (+25) Coventry University
51 (0.02) 70 (+19) University of the West of Scotland
52 (0.02) 67 (+15) Loughborough University
53 (0.02) 18 (-35) University of East Anglia
54 (0.02) 59 (+5) Goldsmiths' College
55 (0.02) 64 (+9) University of Hertfordshire
56 (0.02) 50 (-6) Aston University
57 (0.02) 44 (-13) University of Portsmouth
58 (0.02) 16 (-42) University of Exeter
59 (0.01) 62 (+3) University of Huddersfield
60 (0.01) 51 (-9) University of Northumbria at Newcastle
61 (0.01) 63 (+2) Glasgow Caledonian University
62 (0.01) 58 (-4) Kingston University
63 (0.01) 73 (+10) Liverpool John Moores University
64 (0.01) 46 (-18) University of the West of England, Bristol
65 (0.01) 69 (+4) Edinburgh Napier University
66 (0.01) 55 (-11) Manchester Metropolitan University
67 (0.01) 61 (-6) University of Hull
68 (0.01) 27 (-41) University of Plymouth
69 (0.01) 76 (+7) University of Westminster
70 (0.01) 71 (+1) University of Brighton
71 (0.01) 48 (-23) Bangor University
72 (0.01) 72 (+0) University of Derby
73 (0.01) 77 (+4) Robert Gordon University
74 (0.01) 74 (+0) Glynd?r University
75 (0.01) 79 (+4) University of Greenwich

More Results & Summary

As a means to ranking departments we also produced a ranking of computer science academic venues. Note that larger communities and larger venues have a disproportionately high ranking score. Just for fun we also made a table summarizing how many papers from authors of each institutions were used by other institutions as part of their REF submission.

So we conclude, that it is possible to rank both publishing venues and institutions from individual self-rankings from each researcher. It requires no appointment of expert peer-reviewers (solving the chicken-and-egg problem of assessing expertise), and incurs little expense. However, the subjective judgments of authors had to be accurate because of the expectation of high-quality peer review: thus, the submissions should come under some degree of scrutiny to incentivize accurate selection of higher quality works. Note that we were only able to perform this analysis on the Computer Science corpus thanks to the existence of the DBLP open data records of publications.

In terms of the results, we were surprised by how widely the Peer Ranking differed from the REF Output scores from some institutions. This could be due to venues being only a weak proxy for research quality. However, we would not have expected ranking of large departments to be affected by this to the extent observed. Our analysis only considers papers in conferences and journals, introducing possibly a bias. The fraction of REF authors and papers automatically matched to DBLP records is high for all institutions but sometimes far from perfect.

In terms of public policy, the fact that self-rankings can be aggregated to rank institutions is very interesting: they can be used to rank the quality of the peer-review process of established scientific venues, and this in turn to rank departments according to which venues they publish in. Furthermore, since self-ranking is not as socially awkward as ranking other peoples' works, an open and transparent process may be used to rank venues and departments. One may even foresee a process by which a continuous assesement of research quality is put in place, by requiring UK academics to highlight their top-4 works alongside depositing them into open access repositories, as they are now required. However, the expectation of peer-review is what provides incentives to do so diligently -- and preserving this expectation is a key to the success of this method.

Contributors

Code & Datasets

The matching and ranking was implemented and tested in about 1000 lines of Python code. All our findings are publicly reproducible and only depend on public datasets: you need to download and uncompress the DBLP xml database and dtd file into the data folder. Then, execute extractdblppapers.py (builds a list of paper records), extractrefdata.py (builds databases of UK ref data), matchdblp.py (matches DBLP records likely to be in the REF), matchauthors.py (matches REF authors to DBLP authors and papers) and finally extractUKdblp.py (performs the ranking).

About

A project using the UK REF data sets to collaborativelly re-rank the publishing venues in Computer Science

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages